@reminix/vercel-ai 0.0.7 → 0.0.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +14 -30
- package/dist/agent-adapter.d.ts +12 -24
- package/dist/agent-adapter.d.ts.map +1 -1
- package/dist/agent-adapter.js +27 -92
- package/dist/agent-adapter.js.map +1 -1
- package/package.json +2 -2
package/README.md
CHANGED
|
@@ -69,8 +69,7 @@ serve({ agents: [reminixAgent], port: 8080 });
|
|
|
69
69
|
```
|
|
70
70
|
|
|
71
71
|
Your agent is now available at:
|
|
72
|
-
- `POST /agents/<name>/
|
|
73
|
-
- `POST /agents/<name>/chat` - Conversational chat
|
|
72
|
+
- `POST /agents/<name>/execute` - Execute the agent
|
|
74
73
|
|
|
75
74
|
## API Reference
|
|
76
75
|
|
|
@@ -123,27 +122,23 @@ serve({ agents: [gpt, claude, gemini], port: 8080 });
|
|
|
123
122
|
|
|
124
123
|
## Endpoint Input/Output Formats
|
|
125
124
|
|
|
126
|
-
### POST /agents/{name}/
|
|
125
|
+
### POST /agents/{name}/execute
|
|
127
126
|
|
|
128
|
-
|
|
127
|
+
Execute the agent with a prompt or messages.
|
|
129
128
|
|
|
130
|
-
**Request:**
|
|
129
|
+
**Request with prompt:**
|
|
131
130
|
```json
|
|
132
131
|
{
|
|
133
|
-
"
|
|
134
|
-
"prompt": "Summarize this text: ..."
|
|
135
|
-
}
|
|
132
|
+
"prompt": "Summarize this text: ..."
|
|
136
133
|
}
|
|
137
134
|
```
|
|
138
135
|
|
|
139
|
-
|
|
136
|
+
**Request with messages:**
|
|
140
137
|
```json
|
|
141
138
|
{
|
|
142
|
-
"
|
|
143
|
-
"
|
|
144
|
-
|
|
145
|
-
]
|
|
146
|
-
}
|
|
139
|
+
"messages": [
|
|
140
|
+
{"role": "user", "content": "Hello!"}
|
|
141
|
+
]
|
|
147
142
|
}
|
|
148
143
|
```
|
|
149
144
|
|
|
@@ -154,29 +149,18 @@ Or with messages:
|
|
|
154
149
|
}
|
|
155
150
|
```
|
|
156
151
|
|
|
157
|
-
###
|
|
152
|
+
### Streaming
|
|
158
153
|
|
|
159
|
-
|
|
154
|
+
For streaming responses, set `stream: true` in the request:
|
|
160
155
|
|
|
161
|
-
**Request:**
|
|
162
156
|
```json
|
|
163
157
|
{
|
|
164
|
-
"
|
|
165
|
-
|
|
166
|
-
]
|
|
158
|
+
"prompt": "Tell me a story",
|
|
159
|
+
"stream": true
|
|
167
160
|
}
|
|
168
161
|
```
|
|
169
162
|
|
|
170
|
-
|
|
171
|
-
```json
|
|
172
|
-
{
|
|
173
|
-
"output": "The capital of France is Paris.",
|
|
174
|
-
"messages": [
|
|
175
|
-
{"role": "user", "content": "What is the capital of France?"},
|
|
176
|
-
{"role": "assistant", "content": "The capital of France is Paris."}
|
|
177
|
-
]
|
|
178
|
-
}
|
|
179
|
-
```
|
|
163
|
+
The response will be sent as Server-Sent Events (SSE).
|
|
180
164
|
|
|
181
165
|
## Runtime Documentation
|
|
182
166
|
|
package/dist/agent-adapter.d.ts
CHANGED
|
@@ -5,7 +5,7 @@
|
|
|
5
5
|
*/
|
|
6
6
|
import type { ToolLoopAgent } from 'ai';
|
|
7
7
|
import { generateText, streamText, type LanguageModel } from 'ai';
|
|
8
|
-
import { AgentAdapter, type ServeOptions, type
|
|
8
|
+
import { AgentAdapter, type ServeOptions, type ExecuteRequest, type ExecuteResponse } from '@reminix/runtime';
|
|
9
9
|
/**
|
|
10
10
|
* Options for wrapping a Vercel AI model or agent.
|
|
11
11
|
*/
|
|
@@ -46,38 +46,26 @@ export declare class VercelAIAgentAdapter extends AgentAdapter {
|
|
|
46
46
|
*/
|
|
47
47
|
private toModelMessages;
|
|
48
48
|
/**
|
|
49
|
-
*
|
|
50
|
-
*
|
|
51
|
-
* For task-oriented operations. Expects input with 'messages' key
|
|
52
|
-
* or a 'prompt' key for simple text generation.
|
|
53
|
-
*
|
|
54
|
-
* @param request - The invoke request with input data.
|
|
55
|
-
* @returns The invoke response with the output.
|
|
49
|
+
* Build prompt or messages from execute request.
|
|
56
50
|
*/
|
|
57
|
-
|
|
51
|
+
private buildInputFromRequest;
|
|
58
52
|
/**
|
|
59
|
-
* Handle
|
|
60
|
-
*
|
|
61
|
-
* For conversational interactions.
|
|
53
|
+
* Handle an execute request.
|
|
62
54
|
*
|
|
63
|
-
*
|
|
64
|
-
*
|
|
65
|
-
*/
|
|
66
|
-
chat(request: ChatRequest): Promise<ChatResponse>;
|
|
67
|
-
/**
|
|
68
|
-
* Handle a streaming invoke request.
|
|
55
|
+
* For both task-oriented and chat-style operations. Expects input with 'messages' key
|
|
56
|
+
* or a 'prompt' key for simple text generation.
|
|
69
57
|
*
|
|
70
|
-
* @param request - The
|
|
71
|
-
* @
|
|
58
|
+
* @param request - The execute request with input data.
|
|
59
|
+
* @returns The execute response with the output.
|
|
72
60
|
*/
|
|
73
|
-
|
|
61
|
+
execute(request: ExecuteRequest): Promise<ExecuteResponse>;
|
|
74
62
|
/**
|
|
75
|
-
* Handle a streaming
|
|
63
|
+
* Handle a streaming execute request.
|
|
76
64
|
*
|
|
77
|
-
* @param request - The
|
|
65
|
+
* @param request - The execute request with input data.
|
|
78
66
|
* @yields JSON-encoded chunks from the stream.
|
|
79
67
|
*/
|
|
80
|
-
|
|
68
|
+
executeStream(request: ExecuteRequest): AsyncGenerator<string, void, unknown>;
|
|
81
69
|
}
|
|
82
70
|
/**
|
|
83
71
|
* Wrap a Vercel AI SDK model or agent for use with Reminix Runtime.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"agent-adapter.d.ts","sourceRoot":"","sources":["../src/agent-adapter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,IAAI,CAAC;AACxC,OAAO,EAAE,YAAY,EAAE,UAAU,EAAE,KAAK,aAAa,EAAqB,MAAM,IAAI,CAAC;AAErF,OAAO,EACL,YAAY,EAEZ,KAAK,YAAY,EACjB,KAAK,
|
|
1
|
+
{"version":3,"file":"agent-adapter.d.ts","sourceRoot":"","sources":["../src/agent-adapter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAEH,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,IAAI,CAAC;AACxC,OAAO,EAAE,YAAY,EAAE,UAAU,EAAE,KAAK,aAAa,EAAqB,MAAM,IAAI,CAAC;AAErF,OAAO,EACL,YAAY,EAEZ,KAAK,YAAY,EACjB,KAAK,cAAc,EACnB,KAAK,eAAe,EAErB,MAAM,kBAAkB,CAAC;AAE1B;;GAEG;AACH,MAAM,WAAW,2BAA2B;IAC1C,IAAI,CAAC,EAAE,MAAM,CAAC;CACf;AAGD,KAAK,gBAAgB,GAAG,aAAa,CAAC,GAAG,EAAE,GAAG,EAAE,GAAG,CAAC,CAAC;AAcrD;;;;;;GAMG;AACH,qBAAa,oBAAqB,SAAQ,YAAY;IACpD,MAAM,CAAC,WAAW,SAAe;IAEjC,OAAO,CAAC,YAAY,CAAmC;IACvD,OAAO,CAAC,OAAO,CAAU;IACzB,OAAO,CAAC,KAAK,CAAS;IAEtB;;OAEG;IACH,SAAS,CAAC,aAAa,sBAAgB;IAEvC;;OAEG;IACH,SAAS,CAAC,WAAW,oBAAc;IAEnC;;;;;OAKG;gBAED,YAAY,EAAE,aAAa,GAAG,gBAAgB,EAC9C,OAAO,GAAE,2BAAgC;IAQ3C,IAAI,IAAI,IAAI,MAAM,CAEjB;IAED;;OAEG;IACH,OAAO,CAAC,eAAe;IAOvB;;OAEG;IACH,OAAO,CAAC,qBAAqB;IAgB7B;;;;;;;;OAQG;IACG,OAAO,CAAC,OAAO,EAAE,cAAc,GAAG,OAAO,CAAC,eAAe,CAAC;IAyBhE;;;;;OAKG;IACI,aAAa,CAAC,OAAO,EAAE,cAAc,GAAG,cAAc,CAAC,MAAM,EAAE,IAAI,EAAE,OAAO,CAAC;CAwBrF;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAsCG;AACH,wBAAgB,SAAS,CACvB,YAAY,EAAE,aAAa,GAAG,gBAAgB,EAC9C,OAAO,GAAE,2BAAgC,GACxC,oBAAoB,CAEtB;AAED;;GAEG;AACH,MAAM,WAAW,mBAAoB,SAAQ,2BAA2B,EAAE,YAAY;CAAG;AAEzF;;;;;;;;;;;;;;;;GAgBG;AACH,wBAAgB,UAAU,CACxB,YAAY,EAAE,aAAa,GAAG,gBAAgB,EAC9C,OAAO,GAAE,mBAAwB,GAChC,IAAI,CAIN"}
|
package/dist/agent-adapter.js
CHANGED
|
@@ -59,61 +59,39 @@ export class VercelAIAgentAdapter extends AgentAdapter {
|
|
|
59
59
|
}));
|
|
60
60
|
}
|
|
61
61
|
/**
|
|
62
|
-
*
|
|
63
|
-
*
|
|
64
|
-
* For task-oriented operations. Expects input with 'messages' key
|
|
65
|
-
* or a 'prompt' key for simple text generation.
|
|
66
|
-
*
|
|
67
|
-
* @param request - The invoke request with input data.
|
|
68
|
-
* @returns The invoke response with the output.
|
|
62
|
+
* Build prompt or messages from execute request.
|
|
69
63
|
*/
|
|
70
|
-
|
|
64
|
+
buildInputFromRequest(request) {
|
|
71
65
|
const input = request.input;
|
|
72
|
-
|
|
73
|
-
let prompt;
|
|
74
|
-
if ('prompt' in input) {
|
|
75
|
-
prompt = String(input.prompt);
|
|
76
|
-
}
|
|
77
|
-
else if ('messages' in input) {
|
|
66
|
+
if ('messages' in input) {
|
|
78
67
|
const messages = input.messages;
|
|
79
|
-
|
|
80
|
-
}
|
|
81
|
-
else {
|
|
82
|
-
prompt = JSON.stringify(input);
|
|
68
|
+
return { messages: this.toModelMessages(messages) };
|
|
83
69
|
}
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
// Use ToolLoopAgent.generate()
|
|
87
|
-
const agent = this.modelOrAgent;
|
|
88
|
-
const result = await agent.generate({ prompt, options: {} });
|
|
89
|
-
output = result.text;
|
|
70
|
+
else if ('prompt' in input) {
|
|
71
|
+
return { prompt: String(input.prompt) };
|
|
90
72
|
}
|
|
91
73
|
else {
|
|
92
|
-
|
|
93
|
-
const model = this.modelOrAgent;
|
|
94
|
-
const result = await this._generateText({
|
|
95
|
-
model,
|
|
96
|
-
prompt,
|
|
97
|
-
});
|
|
98
|
-
output = result.text;
|
|
74
|
+
return { prompt: JSON.stringify(input) };
|
|
99
75
|
}
|
|
100
|
-
return { output };
|
|
101
76
|
}
|
|
102
77
|
/**
|
|
103
|
-
* Handle
|
|
78
|
+
* Handle an execute request.
|
|
104
79
|
*
|
|
105
|
-
* For
|
|
80
|
+
* For both task-oriented and chat-style operations. Expects input with 'messages' key
|
|
81
|
+
* or a 'prompt' key for simple text generation.
|
|
106
82
|
*
|
|
107
|
-
* @param request - The
|
|
108
|
-
* @returns The
|
|
83
|
+
* @param request - The execute request with input data.
|
|
84
|
+
* @returns The execute response with the output.
|
|
109
85
|
*/
|
|
110
|
-
async
|
|
111
|
-
const messages = this.
|
|
86
|
+
async execute(request) {
|
|
87
|
+
const { prompt, messages } = this.buildInputFromRequest(request);
|
|
112
88
|
let output;
|
|
113
89
|
if (this.isAgent) {
|
|
114
90
|
// Use ToolLoopAgent.generate()
|
|
115
91
|
const agent = this.modelOrAgent;
|
|
116
|
-
|
|
92
|
+
// Agent API expects either prompt or messages, not both
|
|
93
|
+
const agentInput = prompt ? { prompt, options: {} } : { messages: messages, options: {} };
|
|
94
|
+
const result = await agent.generate(agentInput);
|
|
117
95
|
output = result.text;
|
|
118
96
|
}
|
|
119
97
|
else {
|
|
@@ -121,69 +99,26 @@ export class VercelAIAgentAdapter extends AgentAdapter {
|
|
|
121
99
|
const model = this.modelOrAgent;
|
|
122
100
|
const result = await this._generateText({
|
|
123
101
|
model,
|
|
124
|
-
messages,
|
|
102
|
+
...(prompt ? { prompt } : { messages: messages }),
|
|
125
103
|
});
|
|
126
104
|
output = result.text;
|
|
127
105
|
}
|
|
128
|
-
|
|
129
|
-
const responseMessages = [
|
|
130
|
-
...request.messages,
|
|
131
|
-
{ role: 'assistant', content: output },
|
|
132
|
-
];
|
|
133
|
-
return { output, messages: responseMessages };
|
|
134
|
-
}
|
|
135
|
-
/**
|
|
136
|
-
* Handle a streaming invoke request.
|
|
137
|
-
*
|
|
138
|
-
* @param request - The invoke request with input data.
|
|
139
|
-
* @yields JSON-encoded chunks from the stream.
|
|
140
|
-
*/
|
|
141
|
-
async *invokeStream(request) {
|
|
142
|
-
const input = request.input;
|
|
143
|
-
// Build prompt from input
|
|
144
|
-
let prompt;
|
|
145
|
-
if ('prompt' in input) {
|
|
146
|
-
prompt = String(input.prompt);
|
|
147
|
-
}
|
|
148
|
-
else if ('messages' in input) {
|
|
149
|
-
const messages = input.messages;
|
|
150
|
-
prompt = messages.map((m) => m.content).join('\n');
|
|
151
|
-
}
|
|
152
|
-
else {
|
|
153
|
-
prompt = JSON.stringify(input);
|
|
154
|
-
}
|
|
155
|
-
if (this.isAgent) {
|
|
156
|
-
// Use ToolLoopAgent.stream()
|
|
157
|
-
const agent = this.modelOrAgent;
|
|
158
|
-
const result = await agent.stream({ prompt, options: {} });
|
|
159
|
-
for await (const chunk of result.textStream) {
|
|
160
|
-
yield JSON.stringify({ chunk });
|
|
161
|
-
}
|
|
162
|
-
}
|
|
163
|
-
else {
|
|
164
|
-
// Use streamText with LanguageModel
|
|
165
|
-
const model = this.modelOrAgent;
|
|
166
|
-
const result = this._streamText({
|
|
167
|
-
model,
|
|
168
|
-
prompt,
|
|
169
|
-
});
|
|
170
|
-
for await (const chunk of result.textStream) {
|
|
171
|
-
yield JSON.stringify({ chunk });
|
|
172
|
-
}
|
|
173
|
-
}
|
|
106
|
+
return { output };
|
|
174
107
|
}
|
|
175
108
|
/**
|
|
176
|
-
* Handle a streaming
|
|
109
|
+
* Handle a streaming execute request.
|
|
177
110
|
*
|
|
178
|
-
* @param request - The
|
|
111
|
+
* @param request - The execute request with input data.
|
|
179
112
|
* @yields JSON-encoded chunks from the stream.
|
|
180
113
|
*/
|
|
181
|
-
async *
|
|
182
|
-
const messages = this.
|
|
114
|
+
async *executeStream(request) {
|
|
115
|
+
const { prompt, messages } = this.buildInputFromRequest(request);
|
|
183
116
|
if (this.isAgent) {
|
|
184
117
|
// Use ToolLoopAgent.stream()
|
|
185
118
|
const agent = this.modelOrAgent;
|
|
186
|
-
|
|
119
|
+
// Agent API expects either prompt or messages, not both
|
|
120
|
+
const agentInput = prompt ? { prompt, options: {} } : { messages: messages, options: {} };
|
|
121
|
+
const result = await agent.stream(agentInput);
|
|
187
122
|
for await (const chunk of result.textStream) {
|
|
188
123
|
yield JSON.stringify({ chunk });
|
|
189
124
|
}
|
|
@@ -193,7 +128,7 @@ export class VercelAIAgentAdapter extends AgentAdapter {
|
|
|
193
128
|
const model = this.modelOrAgent;
|
|
194
129
|
const result = this._streamText({
|
|
195
130
|
model,
|
|
196
|
-
messages,
|
|
131
|
+
...(prompt ? { prompt } : { messages: messages }),
|
|
197
132
|
});
|
|
198
133
|
for await (const chunk of result.textStream) {
|
|
199
134
|
yield JSON.stringify({ chunk });
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"agent-adapter.js","sourceRoot":"","sources":["../src/agent-adapter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAGH,OAAO,EAAE,YAAY,EAAE,UAAU,EAAyC,MAAM,IAAI,CAAC;AAErF,OAAO,EACL,YAAY,EACZ,KAAK,
|
|
1
|
+
{"version":3,"file":"agent-adapter.js","sourceRoot":"","sources":["../src/agent-adapter.ts"],"names":[],"mappings":"AAAA;;;;GAIG;AAGH,OAAO,EAAE,YAAY,EAAE,UAAU,EAAyC,MAAM,IAAI,CAAC;AAErF,OAAO,EACL,YAAY,EACZ,KAAK,GAKN,MAAM,kBAAkB,CAAC;AAY1B;;GAEG;AACH,SAAS,eAAe,CAAC,KAAc;IACrC,OAAO,CACL,KAAK,KAAK,IAAI;QACd,OAAO,KAAK,KAAK,QAAQ;QACzB,UAAU,IAAI,KAAK;QACnB,OAAQ,KAA0B,CAAC,QAAQ,KAAK,UAAU,CAC3D,CAAC;AACJ,CAAC;AAED;;;;;;GAMG;AACH,MAAM,OAAO,oBAAqB,SAAQ,YAAY;IACpD,MAAM,CAAC,WAAW,GAAG,WAAW,CAAC;IAEzB,YAAY,CAAmC;IAC/C,OAAO,CAAU;IACjB,KAAK,CAAS;IAEtB;;OAEG;IACO,aAAa,GAAG,YAAY,CAAC;IAEvC;;OAEG;IACO,WAAW,GAAG,UAAU,CAAC;IAEnC;;;;;OAKG;IACH,YACE,YAA8C,EAC9C,UAAuC,EAAE;QAEzC,KAAK,EAAE,CAAC;QACR,IAAI,CAAC,YAAY,GAAG,YAAY,CAAC;QACjC,IAAI,CAAC,OAAO,GAAG,eAAe,CAAC,YAAY,CAAC,CAAC;QAC7C,IAAI,CAAC,KAAK,GAAG,OAAO,CAAC,IAAI,IAAI,iBAAiB,CAAC;IACjD,CAAC;IAED,IAAI,IAAI;QACN,OAAO,IAAI,CAAC,KAAK,CAAC;IACpB,CAAC;IAED;;OAEG;IACK,eAAe,CAAC,QAAmB;QACzC,OAAO,QAAQ,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC;YAC1B,IAAI,EAAE,CAAC,CAAC,IAAuC;YAC/C,OAAO,EAAE,CAAC,CAAC,OAAO,IAAI,EAAE;SACzB,CAAC,CAAC,CAAC;IACN,CAAC;IAED;;OAEG;IACK,qBAAqB,CAAC,OAAuB;QAInD,MAAM,KAAK,GAAG,OAAO,CAAC,KAAgC,CAAC;QAEvD,IAAI,UAAU,IAAI,KAAK,EAAE,CAAC;YACxB,MAAM,QAAQ,GAAG,KAAK,CAAC,QAAqB,CAAC;YAC7C,OAAO,EAAE,QAAQ,EAAE,IAAI,CAAC,eAAe,CAAC,QAAQ,CAAC,EAAE,CAAC;QACtD,CAAC;aAAM,IAAI,QAAQ,IAAI,KAAK,EAAE,CAAC;YAC7B,OAAO,EAAE,MAAM,EAAE,MAAM,CAAC,KAAK,CAAC,MAAM,CAAC,EAAE,CAAC;QAC1C,CAAC;aAAM,CAAC;YACN,OAAO,EAAE,MAAM,EAAE,IAAI,CAAC,SAAS,CAAC,KAAK,CAAC,EAAE,CAAC;QAC3C,CAAC;IACH,CAAC;IAED;;;;;;;;OAQG;IACH,KAAK,CAAC,OAAO,CAAC,OAAuB;QACnC,MAAM,EAAE,MAAM,EAAE,QAAQ,EAAE,GAAG,IAAI,CAAC,qBAAqB,CAAC,OAAO,CAAC,CAAC;QAEjE,IAAI,MAAc,CAAC;QAEnB,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;YACjB,+BAA+B;YAC/B,MAAM,KAAK,GAAG,IAAI,CAAC,YAAgC,CAAC;YACpD,wDAAwD;YACxD,MAAM,UAAU,GAAG,MAAM,CAAC,CAAC,CAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC,CAAC,EAAE,QAAQ,EAAE,QAAS,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC;YAC3F,MAAM,MAAM,GAAG,MAAM,KAAK,CAAC,QAAQ,CAAC,UAAU,CAAC,CAAC;YAChD,MAAM,GAAG,MAAM,CAAC,IAAI,CAAC;QACvB,CAAC;aAAM,CAAC;YACN,sCAAsC;YACtC,MAAM,KAAK,GAAG,IAAI,CAAC,YAA6B,CAAC;YACjD,MAAM,MAAM,GAAG,MAAM,IAAI,CAAC,aAAa,CAAC;gBACtC,KAAK;gBACL,GAAG,CAAC,MAAM,CAAC,CAAC,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC,CAAC,EAAE,QAAQ,EAAE,QAAS,EAAE,CAAC;aACnD,CAAC,CAAC;YACH,MAAM,GAAG,MAAM,CAAC,IAAI,CAAC;QACvB,CAAC;QAED,OAAO,EAAE,MAAM,EAAE,CAAC;IACpB,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,CAAC,aAAa,CAAC,OAAuB;QAC1C,MAAM,EAAE,MAAM,EAAE,QAAQ,EAAE,GAAG,IAAI,CAAC,qBAAqB,CAAC,OAAO,CAAC,CAAC;QAEjE,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;YACjB,6BAA6B;YAC7B,MAAM,KAAK,GAAG,IAAI,CAAC,YAAgC,CAAC;YACpD,wDAAwD;YACxD,MAAM,UAAU,GAAG,MAAM,CAAC,CAAC,CAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC,CAAC,EAAE,QAAQ,EAAE,QAAS,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC;YAC3F,MAAM,MAAM,GAAG,MAAM,KAAK,CAAC,MAAM,CAAC,UAAU,CAAC,CAAC;YAC9C,IAAI,KAAK,EAAE,MAAM,KAAK,IAAI,MAAM,CAAC,UAAU,EAAE,CAAC;gBAC5C,MAAM,IAAI,CAAC,SAAS,CAAC,EAAE,KAAK,EAAE,CAAC,CAAC;YAClC,CAAC;QACH,CAAC;aAAM,CAAC;YACN,oCAAoC;YACpC,MAAM,KAAK,GAAG,IAAI,CAAC,YAA6B,CAAC;YACjD,MAAM,MAAM,GAAG,IAAI,CAAC,WAAW,CAAC;gBAC9B,KAAK;gBACL,GAAG,CAAC,MAAM,CAAC,CAAC,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC,CAAC,EAAE,QAAQ,EAAE,QAAS,EAAE,CAAC;aACnD,CAAC,CAAC;YACH,IAAI,KAAK,EAAE,MAAM,KAAK,IAAI,MAAM,CAAC,UAAU,EAAE,CAAC;gBAC5C,MAAM,IAAI,CAAC,SAAS,CAAC,EAAE,KAAK,EAAE,CAAC,CAAC;YAClC,CAAC;QACH,CAAC;IACH,CAAC;;AAGH;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;GAsCG;AACH,MAAM,UAAU,SAAS,CACvB,YAA8C,EAC9C,UAAuC,EAAE;IAEzC,OAAO,IAAI,oBAAoB,CAAC,YAAY,EAAE,OAAO,CAAC,CAAC;AACzD,CAAC;AAOD;;;;;;;;;;;;;;;;GAgBG;AACH,MAAM,UAAU,UAAU,CACxB,YAA8C,EAC9C,UAA+B,EAAE;IAEjC,MAAM,EAAE,IAAI,EAAE,QAAQ,EAAE,GAAG,cAAc,EAAE,GAAG,OAAO,CAAC;IACtD,MAAM,KAAK,GAAG,SAAS,CAAC,YAAY,EAAE,cAAc,CAAC,CAAC;IACtD,KAAK,CAAC,EAAE,MAAM,EAAE,CAAC,KAAK,CAAC,EAAE,IAAI,EAAE,QAAQ,EAAE,CAAC,CAAC;AAC7C,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@reminix/vercel-ai",
|
|
3
|
-
"version": "0.0.
|
|
3
|
+
"version": "0.0.9",
|
|
4
4
|
"description": "Reminix adapter for Vercel AI SDK - serve agents as REST APIs",
|
|
5
5
|
"license": "Apache-2.0",
|
|
6
6
|
"author": {
|
|
@@ -47,7 +47,7 @@
|
|
|
47
47
|
"LICENSE"
|
|
48
48
|
],
|
|
49
49
|
"dependencies": {
|
|
50
|
-
"@reminix/runtime": "^0.0.
|
|
50
|
+
"@reminix/runtime": "^0.0.9"
|
|
51
51
|
},
|
|
52
52
|
"peerDependencies": {
|
|
53
53
|
"ai": ">=6.0.0"
|