ai 4.0.0-canary.6 → 4.0.0-canary.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "ai",
3
- "version": "4.0.0-canary.6",
3
+ "version": "4.0.0-canary.7",
4
4
  "description": "AI SDK by Vercel - The AI Toolkit for TypeScript and JavaScript",
5
5
  "license": "Apache-2.0",
6
6
  "sideEffects": false,
@@ -174,24 +174,6 @@ type CallSettings = {
174
174
  headers?: Record<string, string | undefined>;
175
175
  };
176
176
 
177
- /**
178
- Represents the number of tokens used in a prompt and completion.
179
- */
180
- type LanguageModelUsage = {
181
- /**
182
- The number of tokens used in the prompt.
183
- */
184
- promptTokens: number;
185
- /**
186
- The number of tokens used in the completion.
187
- */
188
- completionTokens: number;
189
- /**
190
- The total number of tokens used (promptTokens + completionTokens).
191
- */
192
- totalTokens: number;
193
- };
194
-
195
177
  /**
196
178
  Reason why a language model finished generating a response.
197
179
 
@@ -229,6 +211,24 @@ functionality that can be fully encapsulated in the provider.
229
211
  */
230
212
  type ProviderMetadata = LanguageModelV1ProviderMetadata;
231
213
 
214
+ /**
215
+ Represents the number of tokens used in a prompt and completion.
216
+ */
217
+ type LanguageModelUsage = {
218
+ /**
219
+ The number of tokens used in the prompt.
220
+ */
221
+ promptTokens: number;
222
+ /**
223
+ The number of tokens used in the completion.
224
+ */
225
+ completionTokens: number;
226
+ /**
227
+ The total number of tokens used (promptTokens + completionTokens).
228
+ */
229
+ totalTokens: number;
230
+ };
231
+
232
232
  /**
233
233
  Data content. Can either be a base64-encoded string, a Uint8Array, an ArrayBuffer, or a Buffer.
234
234
  */
@@ -172,24 +172,6 @@ type CallSettings = {
172
172
  headers?: Record<string, string | undefined>;
173
173
  };
174
174
 
175
- /**
176
- Represents the number of tokens used in a prompt and completion.
177
- */
178
- type LanguageModelUsage = {
179
- /**
180
- The number of tokens used in the prompt.
181
- */
182
- promptTokens: number;
183
- /**
184
- The number of tokens used in the completion.
185
- */
186
- completionTokens: number;
187
- /**
188
- The total number of tokens used (promptTokens + completionTokens).
189
- */
190
- totalTokens: number;
191
- };
192
-
193
175
  /**
194
176
  Reason why a language model finished generating a response.
195
177
 
@@ -227,6 +209,24 @@ functionality that can be fully encapsulated in the provider.
227
209
  */
228
210
  type ProviderMetadata = LanguageModelV1ProviderMetadata;
229
211
 
212
+ /**
213
+ Represents the number of tokens used in a prompt and completion.
214
+ */
215
+ type LanguageModelUsage = {
216
+ /**
217
+ The number of tokens used in the prompt.
218
+ */
219
+ promptTokens: number;
220
+ /**
221
+ The number of tokens used in the completion.
222
+ */
223
+ completionTokens: number;
224
+ /**
225
+ The total number of tokens used (promptTokens + completionTokens).
226
+ */
227
+ totalTokens: number;
228
+ };
229
+
230
230
  /**
231
231
  Data content. Can either be a base64-encoded string, a Uint8Array, an ArrayBuffer, or a Buffer.
232
232
  */
@@ -1180,11 +1180,14 @@ function standardizePrompt({
1180
1180
  }
1181
1181
 
1182
1182
  // core/types/usage.ts
1183
- function calculateLanguageModelUsage(usage) {
1183
+ function calculateLanguageModelUsage({
1184
+ promptTokens,
1185
+ completionTokens
1186
+ }) {
1184
1187
  return {
1185
- promptTokens: usage.promptTokens,
1186
- completionTokens: usage.completionTokens,
1187
- totalTokens: usage.promptTokens + usage.completionTokens
1188
+ promptTokens,
1189
+ completionTokens,
1190
+ totalTokens: promptTokens + completionTokens
1188
1191
  };
1189
1192
  }
1190
1193