ai 5.0.0-canary.11 → 5.0.0-canary.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -4174,7 +4174,6 @@ async function generateObject({
4174
4174
  let warnings;
4175
4175
  let response;
4176
4176
  let request;
4177
- let logprobs;
4178
4177
  let resultProviderMetadata;
4179
4178
  const standardizedPrompt = standardizePrompt({
4180
4179
  prompt: { system, prompt, messages },
@@ -4274,7 +4273,6 @@ async function generateObject({
4274
4273
  finishReason = generateResult.finishReason;
4275
4274
  usage = generateResult.usage;
4276
4275
  warnings = generateResult.warnings;
4277
- logprobs = generateResult.logprobs;
4278
4276
  resultProviderMetadata = generateResult.providerMetadata;
4279
4277
  request = (_a17 = generateResult.request) != null ? _a17 : {};
4280
4278
  response = generateResult.responseData;
@@ -4348,7 +4346,6 @@ async function generateObject({
4348
4346
  warnings,
4349
4347
  request,
4350
4348
  response,
4351
- logprobs,
4352
4349
  providerMetadata: resultProviderMetadata
4353
4350
  });
4354
4351
  }
@@ -4363,7 +4360,6 @@ var DefaultGenerateObjectResult = class {
4363
4360
  this.providerMetadata = options.providerMetadata;
4364
4361
  this.response = options.response;
4365
4362
  this.request = options.request;
4366
- this.logprobs = options.logprobs;
4367
4363
  }
4368
4364
  toJsonResponse(init) {
4369
4365
  var _a17;
@@ -5632,7 +5628,6 @@ async function generateText({
5632
5628
  finishReason: currentModelResponse.finishReason,
5633
5629
  usage: currentUsage,
5634
5630
  warnings: currentModelResponse.warnings,
5635
- logprobs: currentModelResponse.logprobs,
5636
5631
  request: (_b = currentModelResponse.request) != null ? _b : {},
5637
5632
  response: {
5638
5633
  ...currentModelResponse.response,
@@ -5695,7 +5690,6 @@ async function generateText({
5695
5690
  ...currentModelResponse.response,
5696
5691
  messages: responseMessages
5697
5692
  },
5698
- logprobs: currentModelResponse.logprobs,
5699
5693
  steps,
5700
5694
  providerMetadata: currentModelResponse.providerMetadata
5701
5695
  });
@@ -5792,7 +5786,6 @@ var DefaultGenerateTextResult = class {
5792
5786
  this.response = options.response;
5793
5787
  this.steps = options.steps;
5794
5788
  this.providerMetadata = options.providerMetadata;
5795
- this.logprobs = options.logprobs;
5796
5789
  this.outputResolver = options.outputResolver;
5797
5790
  this.sources = options.sources;
5798
5791
  }
@@ -6326,7 +6319,6 @@ function runToolsTransformation({
6326
6319
  finishChunk = {
6327
6320
  type: "finish",
6328
6321
  finishReason: chunk.finishReason,
6329
- logprobs: chunk.logprobs,
6330
6322
  usage: calculateLanguageModelUsage2(chunk.usage),
6331
6323
  providerMetadata: chunk.providerMetadata
6332
6324
  };
@@ -6648,7 +6640,6 @@ var DefaultStreamTextResult = class {
6648
6640
  finishReason: part.finishReason,
6649
6641
  usage: part.usage,
6650
6642
  warnings: part.warnings,
6651
- logprobs: part.logprobs,
6652
6643
  request: part.request,
6653
6644
  response: {
6654
6645
  ...part.response,
@@ -6712,7 +6703,6 @@ var DefaultStreamTextResult = class {
6712
6703
  self.stepsPromise.resolve(recordedSteps);
6713
6704
  await (onFinish == null ? void 0 : onFinish({
6714
6705
  finishReason,
6715
- logprobs: void 0,
6716
6706
  usage,
6717
6707
  text: recordedFullText,
6718
6708
  reasoningText: lastStep.reasoningText,
@@ -6919,7 +6909,6 @@ var DefaultStreamTextResult = class {
6919
6909
  let stepFirstChunk = true;
6920
6910
  let stepText = "";
6921
6911
  let fullStepText = stepType2 === "continue" ? previousStepText : "";
6922
- let stepLogProbs;
6923
6912
  let stepResponse = {
6924
6913
  id: generateId3(),
6925
6914
  timestamp: currentDate(),
@@ -7044,7 +7033,6 @@ var DefaultStreamTextResult = class {
7044
7033
  stepUsage = chunk.usage;
7045
7034
  stepFinishReason = chunk.finishReason;
7046
7035
  stepProviderMetadata = chunk.providerMetadata;
7047
- stepLogProbs = chunk.logprobs;
7048
7036
  const msToFinish = now2() - startTimestampMs;
7049
7037
  doStreamSpan.addEvent("ai.stream.finish");
7050
7038
  doStreamSpan.setAttributes({
@@ -7132,7 +7120,6 @@ var DefaultStreamTextResult = class {
7132
7120
  finishReason: stepFinishReason,
7133
7121
  usage: stepUsage,
7134
7122
  providerMetadata: stepProviderMetadata,
7135
- logprobs: stepLogProbs,
7136
7123
  request: stepRequest,
7137
7124
  response: {
7138
7125
  ...stepResponse,
@@ -7149,7 +7136,6 @@ var DefaultStreamTextResult = class {
7149
7136
  finishReason: stepFinishReason,
7150
7137
  usage: combinedUsage,
7151
7138
  providerMetadata: stepProviderMetadata,
7152
- logprobs: stepLogProbs,
7153
7139
  response: {
7154
7140
  ...stepResponse,
7155
7141
  headers: response == null ? void 0 : response.headers
@@ -7951,7 +7937,6 @@ function simulateStreamingMiddleware() {
7951
7937
  type: "finish",
7952
7938
  finishReason: result.finishReason,
7953
7939
  usage: result.usage,
7954
- logprobs: result.logprobs,
7955
7940
  providerMetadata: result.providerMetadata
7956
7941
  });
7957
7942
  controller.close();