@mulmochat-plugin/generate-image 0.1.0 → 0.1.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +20 -38
- package/dist/index.cjs +1 -1
- package/dist/index.d.ts +1 -0
- package/dist/index.js +145 -43
- package/dist/plugin/ImageGenerationConfig.vue.d.ts +11 -0
- package/dist/plugin/index.d.ts +2 -1
- package/dist/plugin/types.d.ts +7 -0
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -1,6 +1,10 @@
|
|
|
1
|
-
#
|
|
1
|
+
# MulmoChat Plugin
|
|
2
2
|
|
|
3
|
-
|
|
3
|
+
A plugin for [MulmoChat](https://github.com/receptron/MulmoChat) - a multi-modal voice chat application with OpenAI's GPT-4 Realtime API.
|
|
4
|
+
|
|
5
|
+
## What this plugin does
|
|
6
|
+
|
|
7
|
+
Generate images using AI based on text prompts. This plugin provides the UI components and tool definition, while the actual image generation is handled by the host application via `context.app.generateImage()`.
|
|
4
8
|
|
|
5
9
|
## Installation
|
|
6
10
|
|
|
@@ -11,40 +15,14 @@ yarn add @mulmochat-plugin/generate-image @mulmochat-plugin/ui-image
|
|
|
11
15
|
## Usage
|
|
12
16
|
|
|
13
17
|
```typescript
|
|
14
|
-
import
|
|
18
|
+
import Plugin from "@mulmochat-plugin/generate-image";
|
|
15
19
|
import "@mulmochat-plugin/generate-image/style.css";
|
|
16
20
|
|
|
17
|
-
//
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
This plugin requires `@mulmochat-plugin/ui-image` as a peer dependency for the view components.
|
|
23
|
-
|
|
24
|
-
## Host App Integration
|
|
25
|
-
|
|
26
|
-
This plugin calls `context.app.generateImage(toolName, prompt)` to generate images. The host application (MulmoChat) must provide this function via the `ToolContext.app` object.
|
|
27
|
-
|
|
28
|
-
The host app is responsible for:
|
|
29
|
-
- Selecting the image generation backend (Google Gemini, OpenAI, ComfyUI, etc.)
|
|
30
|
-
- Configuring API keys and model settings
|
|
31
|
-
- Handling the actual image generation API calls
|
|
32
|
-
|
|
33
|
-
Example host app implementation:
|
|
34
|
-
|
|
35
|
-
```typescript
|
|
36
|
-
const context: ToolContext = {
|
|
37
|
-
app: {
|
|
38
|
-
generateImage: async (toolName: string, prompt: string) => {
|
|
39
|
-
// Call your image generation API here
|
|
40
|
-
const imageData = await yourImageGenerationAPI(prompt);
|
|
41
|
-
return {
|
|
42
|
-
message: "Image generated",
|
|
43
|
-
data: { imageData, prompt },
|
|
44
|
-
};
|
|
45
|
-
},
|
|
46
|
-
},
|
|
47
|
-
};
|
|
21
|
+
// Add to pluginList
|
|
22
|
+
const pluginList = [
|
|
23
|
+
// ... other plugins
|
|
24
|
+
Plugin,
|
|
25
|
+
];
|
|
48
26
|
```
|
|
49
27
|
|
|
50
28
|
## Development
|
|
@@ -53,15 +31,19 @@ const context: ToolContext = {
|
|
|
53
31
|
# Install dependencies
|
|
54
32
|
yarn install
|
|
55
33
|
|
|
56
|
-
# Start dev server
|
|
34
|
+
# Start dev server (http://localhost:5173/)
|
|
57
35
|
yarn dev
|
|
58
36
|
|
|
37
|
+
# Build
|
|
38
|
+
yarn build
|
|
39
|
+
|
|
59
40
|
# Type check
|
|
60
41
|
yarn typecheck
|
|
61
42
|
|
|
62
43
|
# Lint
|
|
63
44
|
yarn lint
|
|
64
|
-
|
|
65
|
-
# Build
|
|
66
|
-
yarn build
|
|
67
45
|
```
|
|
46
|
+
|
|
47
|
+
## License
|
|
48
|
+
|
|
49
|
+
MIT
|
package/dist/index.cjs
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
"use strict";Object.defineProperties(exports,{__esModule:{value:!0},[Symbol.toStringTag]:{value:"Module"}});const
|
|
1
|
+
"use strict";Object.defineProperties(exports,{__esModule:{value:!0},[Symbol.toStringTag]:{value:"Module"}});const e=require("vue"),m=require("@mulmochat-plugin/ui-image"),s="generateImage",c={type:"function",name:s,description:"Generate an image based on the prompt and display it on the screen. Be descriptive and specify the concrete details of the images in the prompt. Each call generates one image.",parameters:{type:"object",properties:{prompt:{type:"string",description:"A detailed prompt describing the image to generate"}},required:["prompt"]}},f=[{name:"Sunset Beach",args:{imageData:"https://picsum.photos/id/28/800/600",prompt:"A beautiful sunset over a calm ocean beach with palm trees"}},{name:"Mountain Lake",args:{imageData:"https://picsum.photos/id/29/800/600",prompt:"A serene mountain lake surrounded by pine trees and snow-capped peaks"}},{name:"City Skyline",args:{imageData:"https://picsum.photos/id/43/800/600",prompt:"A modern city skyline at night with glowing skyscrapers"}},{name:"Forest Path",args:{imageData:"https://picsum.photos/id/15/800/600",prompt:"A winding path through an enchanted forest with sunlight filtering through the leaves"}}],v=e.defineComponent({__name:"View",props:{selectedResult:{},sendTextMessage:{type:Function}},emits:["updateResult"],setup(a){const l=a,n=e.ref(null);return e.watch(()=>l.selectedResult,i=>{i?.toolName===s&&i.data&&(n.value=i)},{immediate:!0,deep:!0}),(i,o)=>n.value?(e.openBlock(),e.createBlock(e.unref(m.ImageView),{key:0,selectedResult:n.value},null,8,["selectedResult"])):e.createCommentVNode("",!0)}}),y=e.defineComponent({__name:"Preview",props:{result:{}},setup(a){return(l,n)=>(e.openBlock(),e.createBlock(e.unref(m.ImagePreview),{result:a.result},null,8,["result"]))}}),h={class:"space-y-4"},b=["value"],k={key:0},M=["value"],E={key:1},N=["value"],x=["value"],V=e.defineComponent({__name:"ImageGenerationConfig",props:{value:{}},emits:["update:value"],setup(a,{emit:l}){const n=a,i=l,o=e.computed(()=>typeof n.value=="string"?{backend:n.value,styleModifier:"",geminiModel:"gemini-2.5-flash-image",openaiModel:"gpt-image-1"}:{...n.value,geminiModel:n.value.geminiModel||"gemini-2.5-flash-image",openaiModel:n.value.openaiModel||"gpt-image-1"}),u=r=>{const t=r.target.value;i("update:value",{...o.value,backend:t})},d=r=>{const t=r.target.value;i("update:value",{...o.value,styleModifier:t})},p=r=>{const t=r.target.value;i("update:value",{...o.value,geminiModel:t})},g=r=>{const t=r.target.value;i("update:value",{...o.value,openaiModel:t})};return(r,t)=>(e.openBlock(),e.createElementBlock("div",h,[e.createElementVNode("div",null,[t[1]||(t[1]=e.createElementVNode("label",{class:"block text-sm font-medium text-gray-700 mb-2"}," Image Generation Backend ",-1)),e.createElementVNode("select",{value:o.value.backend,onChange:u,class:"w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"},[...t[0]||(t[0]=[e.createElementVNode("option",{value:"gemini"},"Google Gemini",-1),e.createElementVNode("option",{value:"openai"},"OpenAI",-1),e.createElementVNode("option",{value:"comfyui"},"ComfyUI (Local)",-1)])],40,b),t[2]||(t[2]=e.createElementVNode("p",{class:"text-xs text-gray-500 mt-1"}," Choose between cloud-based Gemini or OpenAI, or a local ComfyUI workflow for image generation. ",-1))]),o.value.backend==="gemini"?(e.openBlock(),e.createElementBlock("div",k,[t[4]||(t[4]=e.createElementVNode("label",{class:"block text-sm font-medium text-gray-700 mb-2"}," Gemini Model ",-1)),e.createElementVNode("select",{value:o.value.geminiModel||"gemini-2.5-flash-image",onChange:p,class:"w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"},[...t[3]||(t[3]=[e.createElementVNode("option",{value:"gemini-2.5-flash-image"},"Gemini 2.5 Flash Image",-1),e.createElementVNode("option",{value:"gemini-3-pro-image-preview"}," Gemini 3 Pro Image (Preview) ",-1)])],40,M),t[5]||(t[5]=e.createElementVNode("p",{class:"text-xs text-gray-500 mt-1"}," Select which Gemini model to use for image generation. ",-1))])):e.createCommentVNode("",!0),o.value.backend==="openai"?(e.openBlock(),e.createElementBlock("div",E,[t[7]||(t[7]=e.createElementVNode("label",{class:"block text-sm font-medium text-gray-700 mb-2"}," OpenAI Model ",-1)),e.createElementVNode("select",{value:o.value.openaiModel||"gpt-image-1",onChange:g,class:"w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"},[...t[6]||(t[6]=[e.createElementVNode("option",{value:"gpt-image-1"},"GPT Image 1",-1),e.createElementVNode("option",{value:"gpt-image-1.5"},"GPT Image 1.5",-1),e.createElementVNode("option",{value:"gpt-image-1-mini"},"GPT Image 1 Mini",-1)])],40,N),t[8]||(t[8]=e.createElementVNode("p",{class:"text-xs text-gray-500 mt-1"}," Select which OpenAI image model to use. ",-1))])):e.createCommentVNode("",!0),e.createElementVNode("div",null,[t[9]||(t[9]=e.createElementVNode("label",{class:"block text-sm font-medium text-gray-700 mb-2"}," Additional Style Modifier ",-1)),e.createElementVNode("input",{type:"text",value:o.value.styleModifier||"",onInput:d,placeholder:"e.g., ghibli-style anime, oil painting, cyberpunk",class:"w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"},null,40,x),t[10]||(t[10]=e.createElementVNode("p",{class:"text-xs text-gray-500 mt-1"}," This style will be automatically appended to all image generation prompts. Leave empty for no additional styling. ",-1))])]))}}),I=async(a,l)=>{const{prompt:n}=l;return a.app?.generateImage?a.app.generateImage(a,n):{message:"generateImage function not available"}};function w(a,l,n){return{toolName:s,data:{imageData:a,prompt:n},message:"",title:l}}const C={toolDefinition:c,execute:I,generatingMessage:"Generating image...",isEnabled:()=>!0,viewComponent:v,previewComponent:y,fileUpload:{acceptedTypes:["image/png","image/jpeg"],handleUpload:w},systemPrompt:`When you are talking about places, objects, people, movies, books and other things, you MUST use the ${s} API to draw pictures to make the conversation more engaging.`,config:{key:"imageGenerationBackend",defaultValue:{backend:"gemini",styleModifier:"",geminiModel:"gemini-2.5-flash-image",openaiModel:"gpt-image-1"},component:V},samples:f},G={plugin:C};exports.default=G;
|
package/dist/index.d.ts
CHANGED
package/dist/index.js
CHANGED
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
import { defineComponent as
|
|
2
|
-
import { ImageView as
|
|
3
|
-
const
|
|
1
|
+
import { defineComponent as g, ref as b, watch as k, createBlock as d, createCommentVNode as p, openBlock as r, unref as c, computed as x, createElementBlock as m, createElementVNode as t } from "vue";
|
|
2
|
+
import { ImageView as M, ImagePreview as I } from "@mulmochat-plugin/ui-image";
|
|
3
|
+
const u = "generateImage", w = {
|
|
4
4
|
type: "function",
|
|
5
|
-
name:
|
|
5
|
+
name: u,
|
|
6
6
|
description: "Generate an image based on the prompt and display it on the screen. Be descriptive and specify the concrete details of the images in the prompt. Each call generates one image.",
|
|
7
7
|
parameters: {
|
|
8
8
|
type: "object",
|
|
@@ -14,7 +14,7 @@ const r = "generateImage", d = {
|
|
|
14
14
|
},
|
|
15
15
|
required: ["prompt"]
|
|
16
16
|
}
|
|
17
|
-
},
|
|
17
|
+
}, C = [
|
|
18
18
|
{
|
|
19
19
|
name: "Sunset Beach",
|
|
20
20
|
args: {
|
|
@@ -43,62 +43,164 @@ const r = "generateImage", d = {
|
|
|
43
43
|
prompt: "A winding path through an enchanted forest with sunlight filtering through the leaves"
|
|
44
44
|
}
|
|
45
45
|
}
|
|
46
|
-
],
|
|
46
|
+
], G = /* @__PURE__ */ g({
|
|
47
47
|
__name: "View",
|
|
48
48
|
props: {
|
|
49
49
|
selectedResult: {},
|
|
50
50
|
sendTextMessage: { type: Function }
|
|
51
51
|
},
|
|
52
52
|
emits: ["updateResult"],
|
|
53
|
-
setup(
|
|
54
|
-
const
|
|
55
|
-
return
|
|
56
|
-
() =>
|
|
57
|
-
(
|
|
58
|
-
|
|
53
|
+
setup(n) {
|
|
54
|
+
const l = n, a = b(null);
|
|
55
|
+
return k(
|
|
56
|
+
() => l.selectedResult,
|
|
57
|
+
(i) => {
|
|
58
|
+
i?.toolName === u && i.data && (a.value = i);
|
|
59
59
|
},
|
|
60
60
|
{ immediate: !0, deep: !0 }
|
|
61
|
-
), (
|
|
61
|
+
), (i, o) => a.value ? (r(), d(c(M), {
|
|
62
62
|
key: 0,
|
|
63
|
-
selectedResult:
|
|
64
|
-
}, null, 8, ["selectedResult"])) :
|
|
63
|
+
selectedResult: a.value
|
|
64
|
+
}, null, 8, ["selectedResult"])) : p("", !0);
|
|
65
65
|
}
|
|
66
|
-
}),
|
|
66
|
+
}), A = /* @__PURE__ */ g({
|
|
67
67
|
__name: "Preview",
|
|
68
68
|
props: {
|
|
69
69
|
result: {}
|
|
70
70
|
},
|
|
71
|
-
setup(
|
|
72
|
-
return (
|
|
71
|
+
setup(n) {
|
|
72
|
+
return (l, a) => (r(), d(c(I), { result: n.result }, null, 8, ["result"]));
|
|
73
73
|
}
|
|
74
|
-
}),
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
}
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
74
|
+
}), _ = { class: "space-y-4" }, P = ["value"], O = { key: 0 }, T = ["value"], S = { key: 1 }, B = ["value"], E = ["value"], N = /* @__PURE__ */ g({
|
|
75
|
+
__name: "ImageGenerationConfig",
|
|
76
|
+
props: {
|
|
77
|
+
value: {}
|
|
78
|
+
},
|
|
79
|
+
emits: ["update:value"],
|
|
80
|
+
setup(n, { emit: l }) {
|
|
81
|
+
const a = n, i = l, o = x(() => typeof a.value == "string" ? {
|
|
82
|
+
backend: a.value,
|
|
83
|
+
styleModifier: "",
|
|
84
|
+
geminiModel: "gemini-2.5-flash-image",
|
|
85
|
+
openaiModel: "gpt-image-1"
|
|
86
|
+
} : {
|
|
87
|
+
...a.value,
|
|
88
|
+
geminiModel: a.value.geminiModel || "gemini-2.5-flash-image",
|
|
89
|
+
openaiModel: a.value.openaiModel || "gpt-image-1"
|
|
90
|
+
}), f = (s) => {
|
|
91
|
+
const e = s.target.value;
|
|
92
|
+
i("update:value", {
|
|
93
|
+
...o.value,
|
|
94
|
+
backend: e
|
|
95
|
+
});
|
|
96
|
+
}, v = (s) => {
|
|
97
|
+
const e = s.target.value;
|
|
98
|
+
i("update:value", {
|
|
99
|
+
...o.value,
|
|
100
|
+
styleModifier: e
|
|
101
|
+
});
|
|
102
|
+
}, y = (s) => {
|
|
103
|
+
const e = s.target.value;
|
|
104
|
+
i("update:value", {
|
|
105
|
+
...o.value,
|
|
106
|
+
geminiModel: e
|
|
107
|
+
});
|
|
108
|
+
}, h = (s) => {
|
|
109
|
+
const e = s.target.value;
|
|
110
|
+
i("update:value", {
|
|
111
|
+
...o.value,
|
|
112
|
+
openaiModel: e
|
|
113
|
+
});
|
|
91
114
|
};
|
|
115
|
+
return (s, e) => (r(), m("div", _, [
|
|
116
|
+
t("div", null, [
|
|
117
|
+
e[1] || (e[1] = t("label", { class: "block text-sm font-medium text-gray-700 mb-2" }, " Image Generation Backend ", -1)),
|
|
118
|
+
t("select", {
|
|
119
|
+
value: o.value.backend,
|
|
120
|
+
onChange: f,
|
|
121
|
+
class: "w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"
|
|
122
|
+
}, [...e[0] || (e[0] = [
|
|
123
|
+
t("option", { value: "gemini" }, "Google Gemini", -1),
|
|
124
|
+
t("option", { value: "openai" }, "OpenAI", -1),
|
|
125
|
+
t("option", { value: "comfyui" }, "ComfyUI (Local)", -1)
|
|
126
|
+
])], 40, P),
|
|
127
|
+
e[2] || (e[2] = t("p", { class: "text-xs text-gray-500 mt-1" }, " Choose between cloud-based Gemini or OpenAI, or a local ComfyUI workflow for image generation. ", -1))
|
|
128
|
+
]),
|
|
129
|
+
o.value.backend === "gemini" ? (r(), m("div", O, [
|
|
130
|
+
e[4] || (e[4] = t("label", { class: "block text-sm font-medium text-gray-700 mb-2" }, " Gemini Model ", -1)),
|
|
131
|
+
t("select", {
|
|
132
|
+
value: o.value.geminiModel || "gemini-2.5-flash-image",
|
|
133
|
+
onChange: y,
|
|
134
|
+
class: "w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"
|
|
135
|
+
}, [...e[3] || (e[3] = [
|
|
136
|
+
t("option", { value: "gemini-2.5-flash-image" }, "Gemini 2.5 Flash Image", -1),
|
|
137
|
+
t("option", { value: "gemini-3-pro-image-preview" }, " Gemini 3 Pro Image (Preview) ", -1)
|
|
138
|
+
])], 40, T),
|
|
139
|
+
e[5] || (e[5] = t("p", { class: "text-xs text-gray-500 mt-1" }, " Select which Gemini model to use for image generation. ", -1))
|
|
140
|
+
])) : p("", !0),
|
|
141
|
+
o.value.backend === "openai" ? (r(), m("div", S, [
|
|
142
|
+
e[7] || (e[7] = t("label", { class: "block text-sm font-medium text-gray-700 mb-2" }, " OpenAI Model ", -1)),
|
|
143
|
+
t("select", {
|
|
144
|
+
value: o.value.openaiModel || "gpt-image-1",
|
|
145
|
+
onChange: h,
|
|
146
|
+
class: "w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"
|
|
147
|
+
}, [...e[6] || (e[6] = [
|
|
148
|
+
t("option", { value: "gpt-image-1" }, "GPT Image 1", -1),
|
|
149
|
+
t("option", { value: "gpt-image-1.5" }, "GPT Image 1.5", -1),
|
|
150
|
+
t("option", { value: "gpt-image-1-mini" }, "GPT Image 1 Mini", -1)
|
|
151
|
+
])], 40, B),
|
|
152
|
+
e[8] || (e[8] = t("p", { class: "text-xs text-gray-500 mt-1" }, " Select which OpenAI image model to use. ", -1))
|
|
153
|
+
])) : p("", !0),
|
|
154
|
+
t("div", null, [
|
|
155
|
+
e[9] || (e[9] = t("label", { class: "block text-sm font-medium text-gray-700 mb-2" }, " Additional Style Modifier ", -1)),
|
|
156
|
+
t("input", {
|
|
157
|
+
type: "text",
|
|
158
|
+
value: o.value.styleModifier || "",
|
|
159
|
+
onInput: v,
|
|
160
|
+
placeholder: "e.g., ghibli-style anime, oil painting, cyberpunk",
|
|
161
|
+
class: "w-full border border-gray-300 rounded px-3 py-2 focus:outline-none focus:ring-2 focus:ring-blue-500"
|
|
162
|
+
}, null, 40, E),
|
|
163
|
+
e[10] || (e[10] = t("p", { class: "text-xs text-gray-500 mt-1" }, " This style will be automatically appended to all image generation prompts. Leave empty for no additional styling. ", -1))
|
|
164
|
+
])
|
|
165
|
+
]));
|
|
92
166
|
}
|
|
93
|
-
},
|
|
94
|
-
|
|
95
|
-
|
|
167
|
+
}), D = async (n, l) => {
|
|
168
|
+
const { prompt: a } = l;
|
|
169
|
+
return n.app?.generateImage ? n.app.generateImage(n, a) : { message: "generateImage function not available" };
|
|
170
|
+
};
|
|
171
|
+
function L(n, l, a) {
|
|
172
|
+
return {
|
|
173
|
+
toolName: u,
|
|
174
|
+
data: { imageData: n, prompt: a },
|
|
175
|
+
message: "",
|
|
176
|
+
title: l
|
|
177
|
+
};
|
|
178
|
+
}
|
|
179
|
+
const U = {
|
|
180
|
+
toolDefinition: w,
|
|
181
|
+
execute: D,
|
|
96
182
|
generatingMessage: "Generating image...",
|
|
97
183
|
isEnabled: () => !0,
|
|
98
|
-
viewComponent:
|
|
99
|
-
previewComponent:
|
|
100
|
-
|
|
101
|
-
|
|
184
|
+
viewComponent: G,
|
|
185
|
+
previewComponent: A,
|
|
186
|
+
fileUpload: {
|
|
187
|
+
acceptedTypes: ["image/png", "image/jpeg"],
|
|
188
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
189
|
+
handleUpload: L
|
|
190
|
+
},
|
|
191
|
+
systemPrompt: `When you are talking about places, objects, people, movies, books and other things, you MUST use the ${u} API to draw pictures to make the conversation more engaging.`,
|
|
192
|
+
config: {
|
|
193
|
+
key: "imageGenerationBackend",
|
|
194
|
+
defaultValue: {
|
|
195
|
+
backend: "gemini",
|
|
196
|
+
styleModifier: "",
|
|
197
|
+
geminiModel: "gemini-2.5-flash-image",
|
|
198
|
+
openaiModel: "gpt-image-1"
|
|
199
|
+
},
|
|
200
|
+
component: N
|
|
201
|
+
},
|
|
202
|
+
samples: C
|
|
203
|
+
}, F = { plugin: U };
|
|
102
204
|
export {
|
|
103
|
-
|
|
205
|
+
F as default
|
|
104
206
|
};
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
import type { ImageGenerationConfigValue } from "./types";
|
|
2
|
+
type __VLS_Props = {
|
|
3
|
+
value: "gemini" | "openai" | "comfyui" | ImageGenerationConfigValue;
|
|
4
|
+
};
|
|
5
|
+
declare const __VLS_export: import("vue").DefineComponent<__VLS_Props, {}, {}, {}, {}, import("vue").ComponentOptionsMixin, import("vue").ComponentOptionsMixin, {
|
|
6
|
+
"update:value": (value: ImageGenerationConfigValue) => any;
|
|
7
|
+
}, string, import("vue").PublicProps, Readonly<__VLS_Props> & Readonly<{
|
|
8
|
+
"onUpdate:value"?: ((value: ImageGenerationConfigValue) => any) | undefined;
|
|
9
|
+
}>, {}, {}, {}, {}, string, import("vue").ComponentProvideOptions, false, {}, any>;
|
|
10
|
+
declare const _default: typeof __VLS_export;
|
|
11
|
+
export default _default;
|
package/dist/plugin/index.d.ts
CHANGED
|
@@ -10,8 +10,9 @@
|
|
|
10
10
|
* // Use plugin directly
|
|
11
11
|
* ```
|
|
12
12
|
*/
|
|
13
|
-
import type { ToolPlugin } from "../common";
|
|
13
|
+
import type { ToolPlugin, ToolResult } from "../common";
|
|
14
14
|
import type { ImageToolData, GenerateImageArgs } from "./types";
|
|
15
|
+
export declare function createUploadedImageResult(imageData: string, fileName: string, prompt: string): ToolResult<ImageToolData, never>;
|
|
15
16
|
/**
|
|
16
17
|
* GenerateImage plugin instance
|
|
17
18
|
*/
|
package/dist/plugin/types.d.ts
CHANGED
|
@@ -7,3 +7,10 @@ export type { ImageToolData } from "@mulmochat-plugin/ui-image";
|
|
|
7
7
|
export interface GenerateImageArgs {
|
|
8
8
|
prompt: string;
|
|
9
9
|
}
|
|
10
|
+
/** Configuration value for image generation backend */
|
|
11
|
+
export interface ImageGenerationConfigValue {
|
|
12
|
+
backend: "gemini" | "openai" | "comfyui";
|
|
13
|
+
styleModifier?: string;
|
|
14
|
+
geminiModel?: "gemini-2.5-flash-image" | "gemini-3-pro-image-preview";
|
|
15
|
+
openaiModel?: "gpt-image-1" | "gpt-image-1.5" | "gpt-image-1-mini";
|
|
16
|
+
}
|