ai 5.0.0-canary.11 → 5.0.0-canary.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +8 -0
- package/dist/index.d.mts +2 -34
- package/dist/index.d.ts +2 -34
- package/dist/index.js +0 -15
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +0 -15
- package/dist/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/index.mjs
CHANGED
@@ -4174,7 +4174,6 @@ async function generateObject({
|
|
4174
4174
|
let warnings;
|
4175
4175
|
let response;
|
4176
4176
|
let request;
|
4177
|
-
let logprobs;
|
4178
4177
|
let resultProviderMetadata;
|
4179
4178
|
const standardizedPrompt = standardizePrompt({
|
4180
4179
|
prompt: { system, prompt, messages },
|
@@ -4274,7 +4273,6 @@ async function generateObject({
|
|
4274
4273
|
finishReason = generateResult.finishReason;
|
4275
4274
|
usage = generateResult.usage;
|
4276
4275
|
warnings = generateResult.warnings;
|
4277
|
-
logprobs = generateResult.logprobs;
|
4278
4276
|
resultProviderMetadata = generateResult.providerMetadata;
|
4279
4277
|
request = (_a17 = generateResult.request) != null ? _a17 : {};
|
4280
4278
|
response = generateResult.responseData;
|
@@ -4348,7 +4346,6 @@ async function generateObject({
|
|
4348
4346
|
warnings,
|
4349
4347
|
request,
|
4350
4348
|
response,
|
4351
|
-
logprobs,
|
4352
4349
|
providerMetadata: resultProviderMetadata
|
4353
4350
|
});
|
4354
4351
|
}
|
@@ -4363,7 +4360,6 @@ var DefaultGenerateObjectResult = class {
|
|
4363
4360
|
this.providerMetadata = options.providerMetadata;
|
4364
4361
|
this.response = options.response;
|
4365
4362
|
this.request = options.request;
|
4366
|
-
this.logprobs = options.logprobs;
|
4367
4363
|
}
|
4368
4364
|
toJsonResponse(init) {
|
4369
4365
|
var _a17;
|
@@ -5632,7 +5628,6 @@ async function generateText({
|
|
5632
5628
|
finishReason: currentModelResponse.finishReason,
|
5633
5629
|
usage: currentUsage,
|
5634
5630
|
warnings: currentModelResponse.warnings,
|
5635
|
-
logprobs: currentModelResponse.logprobs,
|
5636
5631
|
request: (_b = currentModelResponse.request) != null ? _b : {},
|
5637
5632
|
response: {
|
5638
5633
|
...currentModelResponse.response,
|
@@ -5695,7 +5690,6 @@ async function generateText({
|
|
5695
5690
|
...currentModelResponse.response,
|
5696
5691
|
messages: responseMessages
|
5697
5692
|
},
|
5698
|
-
logprobs: currentModelResponse.logprobs,
|
5699
5693
|
steps,
|
5700
5694
|
providerMetadata: currentModelResponse.providerMetadata
|
5701
5695
|
});
|
@@ -5792,7 +5786,6 @@ var DefaultGenerateTextResult = class {
|
|
5792
5786
|
this.response = options.response;
|
5793
5787
|
this.steps = options.steps;
|
5794
5788
|
this.providerMetadata = options.providerMetadata;
|
5795
|
-
this.logprobs = options.logprobs;
|
5796
5789
|
this.outputResolver = options.outputResolver;
|
5797
5790
|
this.sources = options.sources;
|
5798
5791
|
}
|
@@ -6326,7 +6319,6 @@ function runToolsTransformation({
|
|
6326
6319
|
finishChunk = {
|
6327
6320
|
type: "finish",
|
6328
6321
|
finishReason: chunk.finishReason,
|
6329
|
-
logprobs: chunk.logprobs,
|
6330
6322
|
usage: calculateLanguageModelUsage2(chunk.usage),
|
6331
6323
|
providerMetadata: chunk.providerMetadata
|
6332
6324
|
};
|
@@ -6648,7 +6640,6 @@ var DefaultStreamTextResult = class {
|
|
6648
6640
|
finishReason: part.finishReason,
|
6649
6641
|
usage: part.usage,
|
6650
6642
|
warnings: part.warnings,
|
6651
|
-
logprobs: part.logprobs,
|
6652
6643
|
request: part.request,
|
6653
6644
|
response: {
|
6654
6645
|
...part.response,
|
@@ -6712,7 +6703,6 @@ var DefaultStreamTextResult = class {
|
|
6712
6703
|
self.stepsPromise.resolve(recordedSteps);
|
6713
6704
|
await (onFinish == null ? void 0 : onFinish({
|
6714
6705
|
finishReason,
|
6715
|
-
logprobs: void 0,
|
6716
6706
|
usage,
|
6717
6707
|
text: recordedFullText,
|
6718
6708
|
reasoningText: lastStep.reasoningText,
|
@@ -6919,7 +6909,6 @@ var DefaultStreamTextResult = class {
|
|
6919
6909
|
let stepFirstChunk = true;
|
6920
6910
|
let stepText = "";
|
6921
6911
|
let fullStepText = stepType2 === "continue" ? previousStepText : "";
|
6922
|
-
let stepLogProbs;
|
6923
6912
|
let stepResponse = {
|
6924
6913
|
id: generateId3(),
|
6925
6914
|
timestamp: currentDate(),
|
@@ -7044,7 +7033,6 @@ var DefaultStreamTextResult = class {
|
|
7044
7033
|
stepUsage = chunk.usage;
|
7045
7034
|
stepFinishReason = chunk.finishReason;
|
7046
7035
|
stepProviderMetadata = chunk.providerMetadata;
|
7047
|
-
stepLogProbs = chunk.logprobs;
|
7048
7036
|
const msToFinish = now2() - startTimestampMs;
|
7049
7037
|
doStreamSpan.addEvent("ai.stream.finish");
|
7050
7038
|
doStreamSpan.setAttributes({
|
@@ -7132,7 +7120,6 @@ var DefaultStreamTextResult = class {
|
|
7132
7120
|
finishReason: stepFinishReason,
|
7133
7121
|
usage: stepUsage,
|
7134
7122
|
providerMetadata: stepProviderMetadata,
|
7135
|
-
logprobs: stepLogProbs,
|
7136
7123
|
request: stepRequest,
|
7137
7124
|
response: {
|
7138
7125
|
...stepResponse,
|
@@ -7149,7 +7136,6 @@ var DefaultStreamTextResult = class {
|
|
7149
7136
|
finishReason: stepFinishReason,
|
7150
7137
|
usage: combinedUsage,
|
7151
7138
|
providerMetadata: stepProviderMetadata,
|
7152
|
-
logprobs: stepLogProbs,
|
7153
7139
|
response: {
|
7154
7140
|
...stepResponse,
|
7155
7141
|
headers: response == null ? void 0 : response.headers
|
@@ -7951,7 +7937,6 @@ function simulateStreamingMiddleware() {
|
|
7951
7937
|
type: "finish",
|
7952
7938
|
finishReason: result.finishReason,
|
7953
7939
|
usage: result.usage,
|
7954
|
-
logprobs: result.logprobs,
|
7955
7940
|
providerMetadata: result.providerMetadata
|
7956
7941
|
});
|
7957
7942
|
controller.close();
|