@ai-sdk/google 3.0.0-beta.67 → 3.0.0-beta.68

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -7,7 +7,7 @@ import {
7
7
  } from "@ai-sdk/provider-utils";
8
8
 
9
9
  // src/version.ts
10
- var VERSION = true ? "3.0.0-beta.67" : "0.0.0-test";
10
+ var VERSION = true ? "3.0.0-beta.68" : "0.0.0-test";
11
11
 
12
12
  // src/google-generative-ai-embedding-model.ts
13
13
  import {
@@ -211,6 +211,45 @@ import {
211
211
  } from "@ai-sdk/provider-utils";
212
212
  import { z as z5 } from "zod/v4";
213
213
 
214
+ // src/convert-google-generative-ai-usage.ts
215
+ function convertGoogleGenerativeAIUsage(usage) {
216
+ var _a, _b, _c, _d;
217
+ if (usage == null) {
218
+ return {
219
+ inputTokens: {
220
+ total: void 0,
221
+ noCache: void 0,
222
+ cacheRead: void 0,
223
+ cacheWrite: void 0
224
+ },
225
+ outputTokens: {
226
+ total: void 0,
227
+ text: void 0,
228
+ reasoning: void 0
229
+ },
230
+ raw: void 0
231
+ };
232
+ }
233
+ const promptTokens = (_a = usage.promptTokenCount) != null ? _a : 0;
234
+ const candidatesTokens = (_b = usage.candidatesTokenCount) != null ? _b : 0;
235
+ const cachedContentTokens = (_c = usage.cachedContentTokenCount) != null ? _c : 0;
236
+ const thoughtsTokens = (_d = usage.thoughtsTokenCount) != null ? _d : 0;
237
+ return {
238
+ inputTokens: {
239
+ total: promptTokens,
240
+ noCache: promptTokens - cachedContentTokens,
241
+ cacheRead: cachedContentTokens,
242
+ cacheWrite: void 0
243
+ },
244
+ outputTokens: {
245
+ total: candidatesTokens + thoughtsTokens,
246
+ text: candidatesTokens,
247
+ reasoning: thoughtsTokens
248
+ },
249
+ raw: usage
250
+ };
251
+ }
252
+
214
253
  // src/convert-json-schema-to-openapi-schema.ts
215
254
  function convertJSONSchemaToOpenAPISchema(jsonSchema) {
216
255
  if (jsonSchema == null || isEmptyObjectSchema(jsonSchema)) {
@@ -901,7 +940,7 @@ var GoogleGenerativeAILanguageModel = class {
901
940
  };
902
941
  }
903
942
  async doGenerate(options) {
904
- var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m;
943
+ var _a, _b, _c, _d, _e, _f, _g, _h;
905
944
  const { args, warnings } = await this.getArgs(options);
906
945
  const body = JSON.stringify(args);
907
946
  const mergedHeaders = combineHeaders2(
@@ -987,20 +1026,14 @@ var GoogleGenerativeAILanguageModel = class {
987
1026
  finishReason: candidate.finishReason,
988
1027
  hasToolCalls: content.some((part) => part.type === "tool-call")
989
1028
  }),
990
- usage: {
991
- inputTokens: (_e = usageMetadata == null ? void 0 : usageMetadata.promptTokenCount) != null ? _e : void 0,
992
- outputTokens: (_f = usageMetadata == null ? void 0 : usageMetadata.candidatesTokenCount) != null ? _f : void 0,
993
- totalTokens: (_g = usageMetadata == null ? void 0 : usageMetadata.totalTokenCount) != null ? _g : void 0,
994
- reasoningTokens: (_h = usageMetadata == null ? void 0 : usageMetadata.thoughtsTokenCount) != null ? _h : void 0,
995
- cachedInputTokens: (_i = usageMetadata == null ? void 0 : usageMetadata.cachedContentTokenCount) != null ? _i : void 0
996
- },
1029
+ usage: convertGoogleGenerativeAIUsage(usageMetadata),
997
1030
  warnings,
998
1031
  providerMetadata: {
999
1032
  google: {
1000
- promptFeedback: (_j = response.promptFeedback) != null ? _j : null,
1001
- groundingMetadata: (_k = candidate.groundingMetadata) != null ? _k : null,
1002
- urlContextMetadata: (_l = candidate.urlContextMetadata) != null ? _l : null,
1003
- safetyRatings: (_m = candidate.safetyRatings) != null ? _m : null,
1033
+ promptFeedback: (_e = response.promptFeedback) != null ? _e : null,
1034
+ groundingMetadata: (_f = candidate.groundingMetadata) != null ? _f : null,
1035
+ urlContextMetadata: (_g = candidate.urlContextMetadata) != null ? _g : null,
1036
+ safetyRatings: (_h = candidate.safetyRatings) != null ? _h : null,
1004
1037
  usageMetadata: usageMetadata != null ? usageMetadata : null
1005
1038
  }
1006
1039
  },
@@ -1031,11 +1064,7 @@ var GoogleGenerativeAILanguageModel = class {
1031
1064
  fetch: this.config.fetch
1032
1065
  });
1033
1066
  let finishReason = "unknown";
1034
- const usage = {
1035
- inputTokens: void 0,
1036
- outputTokens: void 0,
1037
- totalTokens: void 0
1038
- };
1067
+ let usage = void 0;
1039
1068
  let providerMetadata = void 0;
1040
1069
  const generateId3 = this.config.generateId;
1041
1070
  let hasToolCalls = false;
@@ -1051,7 +1080,7 @@ var GoogleGenerativeAILanguageModel = class {
1051
1080
  controller.enqueue({ type: "stream-start", warnings });
1052
1081
  },
1053
1082
  transform(chunk, controller) {
1054
- var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l;
1083
+ var _a, _b, _c, _d, _e, _f, _g;
1055
1084
  if (options.includeRawChunks) {
1056
1085
  controller.enqueue({ type: "raw", rawValue: chunk.rawValue });
1057
1086
  }
@@ -1062,13 +1091,9 @@ var GoogleGenerativeAILanguageModel = class {
1062
1091
  const value = chunk.value;
1063
1092
  const usageMetadata = value.usageMetadata;
1064
1093
  if (usageMetadata != null) {
1065
- usage.inputTokens = (_a = usageMetadata.promptTokenCount) != null ? _a : void 0;
1066
- usage.outputTokens = (_b = usageMetadata.candidatesTokenCount) != null ? _b : void 0;
1067
- usage.totalTokens = (_c = usageMetadata.totalTokenCount) != null ? _c : void 0;
1068
- usage.reasoningTokens = (_d = usageMetadata.thoughtsTokenCount) != null ? _d : void 0;
1069
- usage.cachedInputTokens = (_e = usageMetadata.cachedContentTokenCount) != null ? _e : void 0;
1094
+ usage = usageMetadata;
1070
1095
  }
1071
- const candidate = (_f = value.candidates) == null ? void 0 : _f[0];
1096
+ const candidate = (_a = value.candidates) == null ? void 0 : _a[0];
1072
1097
  if (candidate == null) {
1073
1098
  return;
1074
1099
  }
@@ -1086,9 +1111,9 @@ var GoogleGenerativeAILanguageModel = class {
1086
1111
  }
1087
1112
  }
1088
1113
  if (content != null) {
1089
- const parts = (_g = content.parts) != null ? _g : [];
1114
+ const parts = (_b = content.parts) != null ? _b : [];
1090
1115
  for (const part of parts) {
1091
- if ("executableCode" in part && ((_h = part.executableCode) == null ? void 0 : _h.code)) {
1116
+ if ("executableCode" in part && ((_c = part.executableCode) == null ? void 0 : _c.code)) {
1092
1117
  const toolCallId = generateId3();
1093
1118
  lastCodeExecutionToolCallId = toolCallId;
1094
1119
  controller.enqueue({
@@ -1220,10 +1245,10 @@ var GoogleGenerativeAILanguageModel = class {
1220
1245
  });
1221
1246
  providerMetadata = {
1222
1247
  google: {
1223
- promptFeedback: (_i = value.promptFeedback) != null ? _i : null,
1224
- groundingMetadata: (_j = candidate.groundingMetadata) != null ? _j : null,
1225
- urlContextMetadata: (_k = candidate.urlContextMetadata) != null ? _k : null,
1226
- safetyRatings: (_l = candidate.safetyRatings) != null ? _l : null
1248
+ promptFeedback: (_d = value.promptFeedback) != null ? _d : null,
1249
+ groundingMetadata: (_e = candidate.groundingMetadata) != null ? _e : null,
1250
+ urlContextMetadata: (_f = candidate.urlContextMetadata) != null ? _f : null,
1251
+ safetyRatings: (_g = candidate.safetyRatings) != null ? _g : null
1227
1252
  }
1228
1253
  };
1229
1254
  if (usageMetadata != null) {
@@ -1247,7 +1272,7 @@ var GoogleGenerativeAILanguageModel = class {
1247
1272
  controller.enqueue({
1248
1273
  type: "finish",
1249
1274
  finishReason,
1250
- usage,
1275
+ usage: convertGoogleGenerativeAIUsage(usage),
1251
1276
  providerMetadata
1252
1277
  });
1253
1278
  }