ai 4.0.0-canary.6 → 4.0.0-canary.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/dist/index.d.mts +36 -160
- package/dist/index.d.ts +36 -160
- package/dist/index.js +11 -36
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +11 -33
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
- package/rsc/dist/index.d.ts +18 -18
- package/rsc/dist/rsc-server.d.mts +18 -18
- package/rsc/dist/rsc-server.mjs +7 -4
- package/rsc/dist/rsc-server.mjs.map +1 -1
package/package.json
CHANGED
package/rsc/dist/index.d.ts
CHANGED
@@ -174,24 +174,6 @@ type CallSettings = {
|
|
174
174
|
headers?: Record<string, string | undefined>;
|
175
175
|
};
|
176
176
|
|
177
|
-
/**
|
178
|
-
Represents the number of tokens used in a prompt and completion.
|
179
|
-
*/
|
180
|
-
type LanguageModelUsage = {
|
181
|
-
/**
|
182
|
-
The number of tokens used in the prompt.
|
183
|
-
*/
|
184
|
-
promptTokens: number;
|
185
|
-
/**
|
186
|
-
The number of tokens used in the completion.
|
187
|
-
*/
|
188
|
-
completionTokens: number;
|
189
|
-
/**
|
190
|
-
The total number of tokens used (promptTokens + completionTokens).
|
191
|
-
*/
|
192
|
-
totalTokens: number;
|
193
|
-
};
|
194
|
-
|
195
177
|
/**
|
196
178
|
Reason why a language model finished generating a response.
|
197
179
|
|
@@ -229,6 +211,24 @@ functionality that can be fully encapsulated in the provider.
|
|
229
211
|
*/
|
230
212
|
type ProviderMetadata = LanguageModelV1ProviderMetadata;
|
231
213
|
|
214
|
+
/**
|
215
|
+
Represents the number of tokens used in a prompt and completion.
|
216
|
+
*/
|
217
|
+
type LanguageModelUsage = {
|
218
|
+
/**
|
219
|
+
The number of tokens used in the prompt.
|
220
|
+
*/
|
221
|
+
promptTokens: number;
|
222
|
+
/**
|
223
|
+
The number of tokens used in the completion.
|
224
|
+
*/
|
225
|
+
completionTokens: number;
|
226
|
+
/**
|
227
|
+
The total number of tokens used (promptTokens + completionTokens).
|
228
|
+
*/
|
229
|
+
totalTokens: number;
|
230
|
+
};
|
231
|
+
|
232
232
|
/**
|
233
233
|
Data content. Can either be a base64-encoded string, a Uint8Array, an ArrayBuffer, or a Buffer.
|
234
234
|
*/
|
@@ -172,24 +172,6 @@ type CallSettings = {
|
|
172
172
|
headers?: Record<string, string | undefined>;
|
173
173
|
};
|
174
174
|
|
175
|
-
/**
|
176
|
-
Represents the number of tokens used in a prompt and completion.
|
177
|
-
*/
|
178
|
-
type LanguageModelUsage = {
|
179
|
-
/**
|
180
|
-
The number of tokens used in the prompt.
|
181
|
-
*/
|
182
|
-
promptTokens: number;
|
183
|
-
/**
|
184
|
-
The number of tokens used in the completion.
|
185
|
-
*/
|
186
|
-
completionTokens: number;
|
187
|
-
/**
|
188
|
-
The total number of tokens used (promptTokens + completionTokens).
|
189
|
-
*/
|
190
|
-
totalTokens: number;
|
191
|
-
};
|
192
|
-
|
193
175
|
/**
|
194
176
|
Reason why a language model finished generating a response.
|
195
177
|
|
@@ -227,6 +209,24 @@ functionality that can be fully encapsulated in the provider.
|
|
227
209
|
*/
|
228
210
|
type ProviderMetadata = LanguageModelV1ProviderMetadata;
|
229
211
|
|
212
|
+
/**
|
213
|
+
Represents the number of tokens used in a prompt and completion.
|
214
|
+
*/
|
215
|
+
type LanguageModelUsage = {
|
216
|
+
/**
|
217
|
+
The number of tokens used in the prompt.
|
218
|
+
*/
|
219
|
+
promptTokens: number;
|
220
|
+
/**
|
221
|
+
The number of tokens used in the completion.
|
222
|
+
*/
|
223
|
+
completionTokens: number;
|
224
|
+
/**
|
225
|
+
The total number of tokens used (promptTokens + completionTokens).
|
226
|
+
*/
|
227
|
+
totalTokens: number;
|
228
|
+
};
|
229
|
+
|
230
230
|
/**
|
231
231
|
Data content. Can either be a base64-encoded string, a Uint8Array, an ArrayBuffer, or a Buffer.
|
232
232
|
*/
|
package/rsc/dist/rsc-server.mjs
CHANGED
@@ -1180,11 +1180,14 @@ function standardizePrompt({
|
|
1180
1180
|
}
|
1181
1181
|
|
1182
1182
|
// core/types/usage.ts
|
1183
|
-
function calculateLanguageModelUsage(
|
1183
|
+
function calculateLanguageModelUsage({
|
1184
|
+
promptTokens,
|
1185
|
+
completionTokens
|
1186
|
+
}) {
|
1184
1187
|
return {
|
1185
|
-
promptTokens
|
1186
|
-
completionTokens
|
1187
|
-
totalTokens:
|
1188
|
+
promptTokens,
|
1189
|
+
completionTokens,
|
1190
|
+
totalTokens: promptTokens + completionTokens
|
1188
1191
|
};
|
1189
1192
|
}
|
1190
1193
|
|