@rimori/client 2.5.30 → 2.5.31
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -73,7 +73,6 @@ export declare class AIModule {
|
|
|
73
73
|
* @param params.messages The messages to generate text from.
|
|
74
74
|
* @param params.tools Optional tools to use for generation.
|
|
75
75
|
* @param params.cache Whether to cache the result (default: false).
|
|
76
|
-
* @param params.model The model to use for generation.
|
|
77
76
|
* @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
78
77
|
* @param params.variables Variables for the server-side prompt template.
|
|
79
78
|
* @returns The generated text.
|
|
@@ -82,7 +81,6 @@ export declare class AIModule {
|
|
|
82
81
|
messages: Message[];
|
|
83
82
|
tools?: Tool[];
|
|
84
83
|
cache?: boolean;
|
|
85
|
-
model?: string;
|
|
86
84
|
prompt?: string;
|
|
87
85
|
variables?: Record<string, any>;
|
|
88
86
|
}): Promise<string>;
|
|
@@ -92,7 +90,6 @@ export declare class AIModule {
|
|
|
92
90
|
* @param params.onMessage Callback for each message chunk.
|
|
93
91
|
* @param params.tools Optional tools to use for generation.
|
|
94
92
|
* @param params.cache Whether to cache the result (default: false).
|
|
95
|
-
* @param params.model The model to use for generation.
|
|
96
93
|
* @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
97
94
|
* @param params.variables Variables for the server-side prompt template.
|
|
98
95
|
*/
|
|
@@ -101,7 +98,6 @@ export declare class AIModule {
|
|
|
101
98
|
onMessage: OnLLMResponse;
|
|
102
99
|
tools?: Tool[];
|
|
103
100
|
cache?: boolean;
|
|
104
|
-
model?: string;
|
|
105
101
|
prompt?: string;
|
|
106
102
|
variables?: Record<string, any>;
|
|
107
103
|
}): Promise<string>;
|
|
@@ -126,7 +122,6 @@ export declare class AIModule {
|
|
|
126
122
|
* Generate a structured object from a request using AI.
|
|
127
123
|
* @param request.cache Whether to cache the result (default: false).
|
|
128
124
|
* @param request.tools The tools to use for generation.
|
|
129
|
-
* @param request.model The model to use for generation.
|
|
130
125
|
* @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
131
126
|
* @param request.variables Variables for the server-side prompt template.
|
|
132
127
|
* @returns The generated object.
|
|
@@ -134,7 +129,6 @@ export declare class AIModule {
|
|
|
134
129
|
getObject<T = any>(params: {
|
|
135
130
|
cache?: boolean;
|
|
136
131
|
tools?: Tool[];
|
|
137
|
-
model?: string;
|
|
138
132
|
prompt?: string;
|
|
139
133
|
variables?: Record<string, any>;
|
|
140
134
|
}): Promise<T>;
|
|
@@ -143,7 +137,6 @@ export declare class AIModule {
|
|
|
143
137
|
* @param request.onResult Callback for each result chunk.
|
|
144
138
|
* @param request.cache Whether to cache the result (default: false).
|
|
145
139
|
* @param request.tools The tools to use for generation.
|
|
146
|
-
* @param request.model The model to use for generation.
|
|
147
140
|
* @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
148
141
|
* @param request.variables Variables for the server-side prompt template.
|
|
149
142
|
*/
|
|
@@ -151,7 +144,6 @@ export declare class AIModule {
|
|
|
151
144
|
onResult: OnStreamedObjectResult<T>;
|
|
152
145
|
cache?: boolean;
|
|
153
146
|
tools?: Tool[];
|
|
154
|
-
model?: string;
|
|
155
147
|
prompt?: string;
|
|
156
148
|
variables?: Record<string, any>;
|
|
157
149
|
}): Promise<T>;
|
|
@@ -78,18 +78,16 @@ export class AIModule {
|
|
|
78
78
|
* @param params.messages The messages to generate text from.
|
|
79
79
|
* @param params.tools Optional tools to use for generation.
|
|
80
80
|
* @param params.cache Whether to cache the result (default: false).
|
|
81
|
-
* @param params.model The model to use for generation.
|
|
82
81
|
* @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
83
82
|
* @param params.variables Variables for the server-side prompt template.
|
|
84
83
|
* @returns The generated text.
|
|
85
84
|
*/
|
|
86
85
|
getText(params) {
|
|
87
86
|
return __awaiter(this, void 0, void 0, function* () {
|
|
88
|
-
const { messages, tools, cache = false,
|
|
87
|
+
const { messages, tools, cache = false, prompt, variables } = params;
|
|
89
88
|
const { result } = yield this.streamObject({
|
|
90
89
|
cache,
|
|
91
90
|
tools,
|
|
92
|
-
model,
|
|
93
91
|
messages,
|
|
94
92
|
prompt,
|
|
95
93
|
variables,
|
|
@@ -103,18 +101,16 @@ export class AIModule {
|
|
|
103
101
|
* @param params.onMessage Callback for each message chunk.
|
|
104
102
|
* @param params.tools Optional tools to use for generation.
|
|
105
103
|
* @param params.cache Whether to cache the result (default: false).
|
|
106
|
-
* @param params.model The model to use for generation.
|
|
107
104
|
* @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
108
105
|
* @param params.variables Variables for the server-side prompt template.
|
|
109
106
|
*/
|
|
110
107
|
getStreamedText(params) {
|
|
111
108
|
return __awaiter(this, void 0, void 0, function* () {
|
|
112
|
-
const { messages, onMessage, tools, cache = false,
|
|
109
|
+
const { messages, onMessage, tools, cache = false, prompt, variables, } = params;
|
|
113
110
|
const messageId = Math.random().toString(36).substring(3);
|
|
114
111
|
const { result } = yield this.streamObject({
|
|
115
112
|
cache,
|
|
116
113
|
tools,
|
|
117
|
-
model,
|
|
118
114
|
messages,
|
|
119
115
|
prompt,
|
|
120
116
|
variables,
|
|
@@ -188,19 +184,17 @@ export class AIModule {
|
|
|
188
184
|
* Generate a structured object from a request using AI.
|
|
189
185
|
* @param request.cache Whether to cache the result (default: false).
|
|
190
186
|
* @param request.tools The tools to use for generation.
|
|
191
|
-
* @param request.model The model to use for generation.
|
|
192
187
|
* @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
193
188
|
* @param request.variables Variables for the server-side prompt template.
|
|
194
189
|
* @returns The generated object.
|
|
195
190
|
*/
|
|
196
191
|
getObject(params) {
|
|
197
192
|
return __awaiter(this, void 0, void 0, function* () {
|
|
198
|
-
const { cache = false, tools = [],
|
|
193
|
+
const { cache = false, tools = [], prompt, variables } = params;
|
|
199
194
|
return yield this.streamObject({
|
|
200
195
|
messages: [],
|
|
201
196
|
cache,
|
|
202
197
|
tools,
|
|
203
|
-
model,
|
|
204
198
|
prompt,
|
|
205
199
|
variables,
|
|
206
200
|
});
|
|
@@ -211,19 +205,17 @@ export class AIModule {
|
|
|
211
205
|
* @param request.onResult Callback for each result chunk.
|
|
212
206
|
* @param request.cache Whether to cache the result (default: false).
|
|
213
207
|
* @param request.tools The tools to use for generation.
|
|
214
|
-
* @param request.model The model to use for generation.
|
|
215
208
|
* @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
|
|
216
209
|
* @param request.variables Variables for the server-side prompt template.
|
|
217
210
|
*/
|
|
218
211
|
getStreamedObject(params) {
|
|
219
212
|
return __awaiter(this, void 0, void 0, function* () {
|
|
220
|
-
const { onResult, cache = false, tools = [],
|
|
213
|
+
const { onResult, cache = false, tools = [], prompt, variables } = params;
|
|
221
214
|
return yield this.streamObject({
|
|
222
215
|
messages: [],
|
|
223
216
|
onResult,
|
|
224
217
|
cache,
|
|
225
218
|
tools,
|
|
226
|
-
model,
|
|
227
219
|
prompt,
|
|
228
220
|
variables,
|
|
229
221
|
});
|
|
@@ -232,14 +224,13 @@ export class AIModule {
|
|
|
232
224
|
streamObject(params) {
|
|
233
225
|
return __awaiter(this, void 0, void 0, function* () {
|
|
234
226
|
var _a, _b, _c, _d, _e;
|
|
235
|
-
const { messages, onResult = () => null, cache = false, tools = [],
|
|
227
|
+
const { messages, onResult = () => null, cache = false, tools = [], prompt, variables, } = params;
|
|
236
228
|
const chatMessages = messages.map((message, index) => (Object.assign(Object.assign({}, message), { id: `${index + 1}` })));
|
|
237
229
|
const payload = {
|
|
238
230
|
cache,
|
|
239
231
|
tools,
|
|
240
232
|
stream: true,
|
|
241
233
|
messages: chatMessages,
|
|
242
|
-
model,
|
|
243
234
|
session_token_id: (_a = this.sessionTokenId) !== null && _a !== void 0 ? _a : undefined,
|
|
244
235
|
};
|
|
245
236
|
if (prompt) {
|