@rimori/client 2.5.30 → 2.5.31

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -73,7 +73,6 @@ export declare class AIModule {
73
73
  * @param params.messages The messages to generate text from.
74
74
  * @param params.tools Optional tools to use for generation.
75
75
  * @param params.cache Whether to cache the result (default: false).
76
- * @param params.model The model to use for generation.
77
76
  * @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
78
77
  * @param params.variables Variables for the server-side prompt template.
79
78
  * @returns The generated text.
@@ -82,7 +81,6 @@ export declare class AIModule {
82
81
  messages: Message[];
83
82
  tools?: Tool[];
84
83
  cache?: boolean;
85
- model?: string;
86
84
  prompt?: string;
87
85
  variables?: Record<string, any>;
88
86
  }): Promise<string>;
@@ -92,7 +90,6 @@ export declare class AIModule {
92
90
  * @param params.onMessage Callback for each message chunk.
93
91
  * @param params.tools Optional tools to use for generation.
94
92
  * @param params.cache Whether to cache the result (default: false).
95
- * @param params.model The model to use for generation.
96
93
  * @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
97
94
  * @param params.variables Variables for the server-side prompt template.
98
95
  */
@@ -101,7 +98,6 @@ export declare class AIModule {
101
98
  onMessage: OnLLMResponse;
102
99
  tools?: Tool[];
103
100
  cache?: boolean;
104
- model?: string;
105
101
  prompt?: string;
106
102
  variables?: Record<string, any>;
107
103
  }): Promise<string>;
@@ -126,7 +122,6 @@ export declare class AIModule {
126
122
  * Generate a structured object from a request using AI.
127
123
  * @param request.cache Whether to cache the result (default: false).
128
124
  * @param request.tools The tools to use for generation.
129
- * @param request.model The model to use for generation.
130
125
  * @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
131
126
  * @param request.variables Variables for the server-side prompt template.
132
127
  * @returns The generated object.
@@ -134,7 +129,6 @@ export declare class AIModule {
134
129
  getObject<T = any>(params: {
135
130
  cache?: boolean;
136
131
  tools?: Tool[];
137
- model?: string;
138
132
  prompt?: string;
139
133
  variables?: Record<string, any>;
140
134
  }): Promise<T>;
@@ -143,7 +137,6 @@ export declare class AIModule {
143
137
  * @param request.onResult Callback for each result chunk.
144
138
  * @param request.cache Whether to cache the result (default: false).
145
139
  * @param request.tools The tools to use for generation.
146
- * @param request.model The model to use for generation.
147
140
  * @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
148
141
  * @param request.variables Variables for the server-side prompt template.
149
142
  */
@@ -151,7 +144,6 @@ export declare class AIModule {
151
144
  onResult: OnStreamedObjectResult<T>;
152
145
  cache?: boolean;
153
146
  tools?: Tool[];
154
- model?: string;
155
147
  prompt?: string;
156
148
  variables?: Record<string, any>;
157
149
  }): Promise<T>;
@@ -78,18 +78,16 @@ export class AIModule {
78
78
  * @param params.messages The messages to generate text from.
79
79
  * @param params.tools Optional tools to use for generation.
80
80
  * @param params.cache Whether to cache the result (default: false).
81
- * @param params.model The model to use for generation.
82
81
  * @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
83
82
  * @param params.variables Variables for the server-side prompt template.
84
83
  * @returns The generated text.
85
84
  */
86
85
  getText(params) {
87
86
  return __awaiter(this, void 0, void 0, function* () {
88
- const { messages, tools, cache = false, model, prompt, variables } = params;
87
+ const { messages, tools, cache = false, prompt, variables } = params;
89
88
  const { result } = yield this.streamObject({
90
89
  cache,
91
90
  tools,
92
- model,
93
91
  messages,
94
92
  prompt,
95
93
  variables,
@@ -103,18 +101,16 @@ export class AIModule {
103
101
  * @param params.onMessage Callback for each message chunk.
104
102
  * @param params.tools Optional tools to use for generation.
105
103
  * @param params.cache Whether to cache the result (default: false).
106
- * @param params.model The model to use for generation.
107
104
  * @param params.prompt Server-side prompt name (e.g. 'writing.analysis').
108
105
  * @param params.variables Variables for the server-side prompt template.
109
106
  */
110
107
  getStreamedText(params) {
111
108
  return __awaiter(this, void 0, void 0, function* () {
112
- const { messages, onMessage, tools, cache = false, model, prompt, variables, } = params;
109
+ const { messages, onMessage, tools, cache = false, prompt, variables, } = params;
113
110
  const messageId = Math.random().toString(36).substring(3);
114
111
  const { result } = yield this.streamObject({
115
112
  cache,
116
113
  tools,
117
- model,
118
114
  messages,
119
115
  prompt,
120
116
  variables,
@@ -188,19 +184,17 @@ export class AIModule {
188
184
  * Generate a structured object from a request using AI.
189
185
  * @param request.cache Whether to cache the result (default: false).
190
186
  * @param request.tools The tools to use for generation.
191
- * @param request.model The model to use for generation.
192
187
  * @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
193
188
  * @param request.variables Variables for the server-side prompt template.
194
189
  * @returns The generated object.
195
190
  */
196
191
  getObject(params) {
197
192
  return __awaiter(this, void 0, void 0, function* () {
198
- const { cache = false, tools = [], model = undefined, prompt, variables } = params;
193
+ const { cache = false, tools = [], prompt, variables } = params;
199
194
  return yield this.streamObject({
200
195
  messages: [],
201
196
  cache,
202
197
  tools,
203
- model,
204
198
  prompt,
205
199
  variables,
206
200
  });
@@ -211,19 +205,17 @@ export class AIModule {
211
205
  * @param request.onResult Callback for each result chunk.
212
206
  * @param request.cache Whether to cache the result (default: false).
213
207
  * @param request.tools The tools to use for generation.
214
- * @param request.model The model to use for generation.
215
208
  * @param request.prompt Server-side prompt name (e.g. 'writing.analysis').
216
209
  * @param request.variables Variables for the server-side prompt template.
217
210
  */
218
211
  getStreamedObject(params) {
219
212
  return __awaiter(this, void 0, void 0, function* () {
220
- const { onResult, cache = false, tools = [], model = undefined, prompt, variables } = params;
213
+ const { onResult, cache = false, tools = [], prompt, variables } = params;
221
214
  return yield this.streamObject({
222
215
  messages: [],
223
216
  onResult,
224
217
  cache,
225
218
  tools,
226
- model,
227
219
  prompt,
228
220
  variables,
229
221
  });
@@ -232,14 +224,13 @@ export class AIModule {
232
224
  streamObject(params) {
233
225
  return __awaiter(this, void 0, void 0, function* () {
234
226
  var _a, _b, _c, _d, _e;
235
- const { messages, onResult = () => null, cache = false, tools = [], model = undefined, prompt, variables, } = params;
227
+ const { messages, onResult = () => null, cache = false, tools = [], prompt, variables, } = params;
236
228
  const chatMessages = messages.map((message, index) => (Object.assign(Object.assign({}, message), { id: `${index + 1}` })));
237
229
  const payload = {
238
230
  cache,
239
231
  tools,
240
232
  stream: true,
241
233
  messages: chatMessages,
242
- model,
243
234
  session_token_id: (_a = this.sessionTokenId) !== null && _a !== void 0 ? _a : undefined,
244
235
  };
245
236
  if (prompt) {
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@rimori/client",
3
- "version": "2.5.30",
3
+ "version": "2.5.31",
4
4
  "main": "dist/index.js",
5
5
  "types": "dist/index.d.ts",
6
6
  "repository": {