@ai-sdk/google 3.0.0-beta.67 → 3.0.0-beta.68
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +9 -0
- package/dist/index.js +57 -32
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +57 -32
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +56 -31
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +56 -31
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/index.mjs
CHANGED
|
@@ -7,7 +7,7 @@ import {
|
|
|
7
7
|
} from "@ai-sdk/provider-utils";
|
|
8
8
|
|
|
9
9
|
// src/version.ts
|
|
10
|
-
var VERSION = true ? "3.0.0-beta.
|
|
10
|
+
var VERSION = true ? "3.0.0-beta.68" : "0.0.0-test";
|
|
11
11
|
|
|
12
12
|
// src/google-generative-ai-embedding-model.ts
|
|
13
13
|
import {
|
|
@@ -211,6 +211,45 @@ import {
|
|
|
211
211
|
} from "@ai-sdk/provider-utils";
|
|
212
212
|
import { z as z5 } from "zod/v4";
|
|
213
213
|
|
|
214
|
+
// src/convert-google-generative-ai-usage.ts
|
|
215
|
+
function convertGoogleGenerativeAIUsage(usage) {
|
|
216
|
+
var _a, _b, _c, _d;
|
|
217
|
+
if (usage == null) {
|
|
218
|
+
return {
|
|
219
|
+
inputTokens: {
|
|
220
|
+
total: void 0,
|
|
221
|
+
noCache: void 0,
|
|
222
|
+
cacheRead: void 0,
|
|
223
|
+
cacheWrite: void 0
|
|
224
|
+
},
|
|
225
|
+
outputTokens: {
|
|
226
|
+
total: void 0,
|
|
227
|
+
text: void 0,
|
|
228
|
+
reasoning: void 0
|
|
229
|
+
},
|
|
230
|
+
raw: void 0
|
|
231
|
+
};
|
|
232
|
+
}
|
|
233
|
+
const promptTokens = (_a = usage.promptTokenCount) != null ? _a : 0;
|
|
234
|
+
const candidatesTokens = (_b = usage.candidatesTokenCount) != null ? _b : 0;
|
|
235
|
+
const cachedContentTokens = (_c = usage.cachedContentTokenCount) != null ? _c : 0;
|
|
236
|
+
const thoughtsTokens = (_d = usage.thoughtsTokenCount) != null ? _d : 0;
|
|
237
|
+
return {
|
|
238
|
+
inputTokens: {
|
|
239
|
+
total: promptTokens,
|
|
240
|
+
noCache: promptTokens - cachedContentTokens,
|
|
241
|
+
cacheRead: cachedContentTokens,
|
|
242
|
+
cacheWrite: void 0
|
|
243
|
+
},
|
|
244
|
+
outputTokens: {
|
|
245
|
+
total: candidatesTokens + thoughtsTokens,
|
|
246
|
+
text: candidatesTokens,
|
|
247
|
+
reasoning: thoughtsTokens
|
|
248
|
+
},
|
|
249
|
+
raw: usage
|
|
250
|
+
};
|
|
251
|
+
}
|
|
252
|
+
|
|
214
253
|
// src/convert-json-schema-to-openapi-schema.ts
|
|
215
254
|
function convertJSONSchemaToOpenAPISchema(jsonSchema) {
|
|
216
255
|
if (jsonSchema == null || isEmptyObjectSchema(jsonSchema)) {
|
|
@@ -901,7 +940,7 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
901
940
|
};
|
|
902
941
|
}
|
|
903
942
|
async doGenerate(options) {
|
|
904
|
-
var _a, _b, _c, _d, _e, _f, _g, _h
|
|
943
|
+
var _a, _b, _c, _d, _e, _f, _g, _h;
|
|
905
944
|
const { args, warnings } = await this.getArgs(options);
|
|
906
945
|
const body = JSON.stringify(args);
|
|
907
946
|
const mergedHeaders = combineHeaders2(
|
|
@@ -987,20 +1026,14 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
987
1026
|
finishReason: candidate.finishReason,
|
|
988
1027
|
hasToolCalls: content.some((part) => part.type === "tool-call")
|
|
989
1028
|
}),
|
|
990
|
-
usage:
|
|
991
|
-
inputTokens: (_e = usageMetadata == null ? void 0 : usageMetadata.promptTokenCount) != null ? _e : void 0,
|
|
992
|
-
outputTokens: (_f = usageMetadata == null ? void 0 : usageMetadata.candidatesTokenCount) != null ? _f : void 0,
|
|
993
|
-
totalTokens: (_g = usageMetadata == null ? void 0 : usageMetadata.totalTokenCount) != null ? _g : void 0,
|
|
994
|
-
reasoningTokens: (_h = usageMetadata == null ? void 0 : usageMetadata.thoughtsTokenCount) != null ? _h : void 0,
|
|
995
|
-
cachedInputTokens: (_i = usageMetadata == null ? void 0 : usageMetadata.cachedContentTokenCount) != null ? _i : void 0
|
|
996
|
-
},
|
|
1029
|
+
usage: convertGoogleGenerativeAIUsage(usageMetadata),
|
|
997
1030
|
warnings,
|
|
998
1031
|
providerMetadata: {
|
|
999
1032
|
google: {
|
|
1000
|
-
promptFeedback: (
|
|
1001
|
-
groundingMetadata: (
|
|
1002
|
-
urlContextMetadata: (
|
|
1003
|
-
safetyRatings: (
|
|
1033
|
+
promptFeedback: (_e = response.promptFeedback) != null ? _e : null,
|
|
1034
|
+
groundingMetadata: (_f = candidate.groundingMetadata) != null ? _f : null,
|
|
1035
|
+
urlContextMetadata: (_g = candidate.urlContextMetadata) != null ? _g : null,
|
|
1036
|
+
safetyRatings: (_h = candidate.safetyRatings) != null ? _h : null,
|
|
1004
1037
|
usageMetadata: usageMetadata != null ? usageMetadata : null
|
|
1005
1038
|
}
|
|
1006
1039
|
},
|
|
@@ -1031,11 +1064,7 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1031
1064
|
fetch: this.config.fetch
|
|
1032
1065
|
});
|
|
1033
1066
|
let finishReason = "unknown";
|
|
1034
|
-
|
|
1035
|
-
inputTokens: void 0,
|
|
1036
|
-
outputTokens: void 0,
|
|
1037
|
-
totalTokens: void 0
|
|
1038
|
-
};
|
|
1067
|
+
let usage = void 0;
|
|
1039
1068
|
let providerMetadata = void 0;
|
|
1040
1069
|
const generateId3 = this.config.generateId;
|
|
1041
1070
|
let hasToolCalls = false;
|
|
@@ -1051,7 +1080,7 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1051
1080
|
controller.enqueue({ type: "stream-start", warnings });
|
|
1052
1081
|
},
|
|
1053
1082
|
transform(chunk, controller) {
|
|
1054
|
-
var _a, _b, _c, _d, _e, _f, _g
|
|
1083
|
+
var _a, _b, _c, _d, _e, _f, _g;
|
|
1055
1084
|
if (options.includeRawChunks) {
|
|
1056
1085
|
controller.enqueue({ type: "raw", rawValue: chunk.rawValue });
|
|
1057
1086
|
}
|
|
@@ -1062,13 +1091,9 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1062
1091
|
const value = chunk.value;
|
|
1063
1092
|
const usageMetadata = value.usageMetadata;
|
|
1064
1093
|
if (usageMetadata != null) {
|
|
1065
|
-
usage
|
|
1066
|
-
usage.outputTokens = (_b = usageMetadata.candidatesTokenCount) != null ? _b : void 0;
|
|
1067
|
-
usage.totalTokens = (_c = usageMetadata.totalTokenCount) != null ? _c : void 0;
|
|
1068
|
-
usage.reasoningTokens = (_d = usageMetadata.thoughtsTokenCount) != null ? _d : void 0;
|
|
1069
|
-
usage.cachedInputTokens = (_e = usageMetadata.cachedContentTokenCount) != null ? _e : void 0;
|
|
1094
|
+
usage = usageMetadata;
|
|
1070
1095
|
}
|
|
1071
|
-
const candidate = (
|
|
1096
|
+
const candidate = (_a = value.candidates) == null ? void 0 : _a[0];
|
|
1072
1097
|
if (candidate == null) {
|
|
1073
1098
|
return;
|
|
1074
1099
|
}
|
|
@@ -1086,9 +1111,9 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1086
1111
|
}
|
|
1087
1112
|
}
|
|
1088
1113
|
if (content != null) {
|
|
1089
|
-
const parts = (
|
|
1114
|
+
const parts = (_b = content.parts) != null ? _b : [];
|
|
1090
1115
|
for (const part of parts) {
|
|
1091
|
-
if ("executableCode" in part && ((
|
|
1116
|
+
if ("executableCode" in part && ((_c = part.executableCode) == null ? void 0 : _c.code)) {
|
|
1092
1117
|
const toolCallId = generateId3();
|
|
1093
1118
|
lastCodeExecutionToolCallId = toolCallId;
|
|
1094
1119
|
controller.enqueue({
|
|
@@ -1220,10 +1245,10 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1220
1245
|
});
|
|
1221
1246
|
providerMetadata = {
|
|
1222
1247
|
google: {
|
|
1223
|
-
promptFeedback: (
|
|
1224
|
-
groundingMetadata: (
|
|
1225
|
-
urlContextMetadata: (
|
|
1226
|
-
safetyRatings: (
|
|
1248
|
+
promptFeedback: (_d = value.promptFeedback) != null ? _d : null,
|
|
1249
|
+
groundingMetadata: (_e = candidate.groundingMetadata) != null ? _e : null,
|
|
1250
|
+
urlContextMetadata: (_f = candidate.urlContextMetadata) != null ? _f : null,
|
|
1251
|
+
safetyRatings: (_g = candidate.safetyRatings) != null ? _g : null
|
|
1227
1252
|
}
|
|
1228
1253
|
};
|
|
1229
1254
|
if (usageMetadata != null) {
|
|
@@ -1247,7 +1272,7 @@ var GoogleGenerativeAILanguageModel = class {
|
|
|
1247
1272
|
controller.enqueue({
|
|
1248
1273
|
type: "finish",
|
|
1249
1274
|
finishReason,
|
|
1250
|
-
usage,
|
|
1275
|
+
usage: convertGoogleGenerativeAIUsage(usage),
|
|
1251
1276
|
providerMetadata
|
|
1252
1277
|
});
|
|
1253
1278
|
}
|