ai 3.2.33 → 3.2.34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "ai",
3
- "version": "3.2.33",
4
- "description": "The AI Toolkit for TypeScript",
3
+ "version": "3.2.34",
4
+ "description": "Vercel AI SDK - The AI Toolkit for TypeScript and JavaScript",
5
5
  "license": "Apache-2.0",
6
6
  "sideEffects": false,
7
7
  "main": "./dist/index.js",
@@ -58,13 +58,13 @@
58
58
  }
59
59
  },
60
60
  "dependencies": {
61
- "@ai-sdk/provider": "0.0.12",
62
- "@ai-sdk/provider-utils": "1.0.2",
63
- "@ai-sdk/react": "0.0.26",
64
- "@ai-sdk/solid": "0.0.19",
65
- "@ai-sdk/svelte": "0.0.20",
66
- "@ai-sdk/ui-utils": "0.0.17",
67
- "@ai-sdk/vue": "0.0.21",
61
+ "@ai-sdk/provider": "0.0.13",
62
+ "@ai-sdk/provider-utils": "1.0.3",
63
+ "@ai-sdk/react": "0.0.27",
64
+ "@ai-sdk/solid": "0.0.20",
65
+ "@ai-sdk/svelte": "0.0.21",
66
+ "@ai-sdk/ui-utils": "0.0.18",
67
+ "@ai-sdk/vue": "0.0.22",
68
68
  "@opentelemetry/api": "1.9.0",
69
69
  "eventsource-parser": "1.1.2",
70
70
  "jsondiffpatch": "0.6.0",
@@ -245,6 +245,13 @@ type CallSettings = {
245
245
  */
246
246
  topP?: number;
247
247
  /**
248
+ Only sample from the top K options for each subsequent token.
249
+
250
+ Used to remove "long tail" low probability responses.
251
+ Recommended for advanced use cases only. You usually only need to use temperature.
252
+ */
253
+ topK?: number;
254
+ /**
248
255
  Presence penalty setting. It affects the likelihood of the model to
249
256
  repeat information that is already in the prompt.
250
257
 
@@ -265,6 +272,12 @@ type CallSettings = {
265
272
  */
266
273
  frequencyPenalty?: number;
267
274
  /**
275
+ Stop sequences.
276
+ If set, the model will stop generating text when one of the stop sequences is generated.
277
+ Providers may have limits on the number of stop sequences.
278
+ */
279
+ stopSequences?: string[];
280
+ /**
268
281
  The seed (integer) to use for random sampling. If set and supported
269
282
  by the model, calls will generate deterministic results.
270
283
  */
@@ -243,6 +243,13 @@ type CallSettings = {
243
243
  */
244
244
  topP?: number;
245
245
  /**
246
+ Only sample from the top K options for each subsequent token.
247
+
248
+ Used to remove "long tail" low probability responses.
249
+ Recommended for advanced use cases only. You usually only need to use temperature.
250
+ */
251
+ topK?: number;
252
+ /**
246
253
  Presence penalty setting. It affects the likelihood of the model to
247
254
  repeat information that is already in the prompt.
248
255
 
@@ -263,6 +270,12 @@ type CallSettings = {
263
270
  */
264
271
  frequencyPenalty?: number;
265
272
  /**
273
+ Stop sequences.
274
+ If set, the model will stop generating text when one of the stop sequences is generated.
275
+ Providers may have limits on the number of stop sequences.
276
+ */
277
+ stopSequences?: string[];
278
+ /**
266
279
  The seed (integer) to use for random sampling. If set and supported
267
280
  by the model, calls will generate deterministic results.
268
281
  */
@@ -489,6 +489,7 @@ function prepareCallSettings({
489
489
  topP,
490
490
  presencePenalty,
491
491
  frequencyPenalty,
492
+ stopSequences,
492
493
  seed,
493
494
  maxRetries
494
495
  }) {
@@ -575,6 +576,7 @@ function prepareCallSettings({
575
576
  topP,
576
577
  presencePenalty,
577
578
  frequencyPenalty,
579
+ stopSequences: stopSequences != null && stopSequences.length > 0 ? stopSequences : void 0,
578
580
  seed,
579
581
  maxRetries: maxRetries != null ? maxRetries : 2
580
582
  };