@ai-sdk/google 3.0.56 → 3.0.58

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,17 @@
1
1
  # @ai-sdk/google
2
2
 
3
+ ## 3.0.58
4
+
5
+ ### Patch Changes
6
+
7
+ - f20ba77: feat(provider/google): preserve per-modality token details in usage data
8
+
9
+ ## 3.0.57
10
+
11
+ ### Patch Changes
12
+
13
+ - 0f2b2f1: fix(provider/google): fix Gemini service tier enum after upstream update
14
+
3
15
  ## 3.0.56
4
16
 
5
17
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -39,7 +39,7 @@ declare const googleLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
39
39
  longitude: number;
40
40
  } | undefined;
41
41
  } | undefined;
42
- serviceTier?: "SERVICE_TIER_STANDARD" | "SERVICE_TIER_FLEX" | "SERVICE_TIER_PRIORITY" | undefined;
42
+ serviceTier?: "standard" | "flex" | "priority" | undefined;
43
43
  }>;
44
44
  type GoogleLanguageModelOptions = InferSchema<typeof googleLanguageModelOptions>;
45
45
 
@@ -158,6 +158,14 @@ declare const responseSchema: _ai_sdk_provider_utils.LazySchema<{
158
158
  candidatesTokenCount?: number | null | undefined;
159
159
  totalTokenCount?: number | null | undefined;
160
160
  trafficType?: string | null | undefined;
161
+ promptTokensDetails?: {
162
+ modality: string;
163
+ tokenCount: number;
164
+ }[] | null | undefined;
165
+ candidatesTokensDetails?: {
166
+ modality: string;
167
+ tokenCount: number;
168
+ }[] | null | undefined;
161
169
  } | null | undefined;
162
170
  promptFeedback?: {
163
171
  blockReason?: string | null | undefined;
package/dist/index.d.ts CHANGED
@@ -39,7 +39,7 @@ declare const googleLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
39
39
  longitude: number;
40
40
  } | undefined;
41
41
  } | undefined;
42
- serviceTier?: "SERVICE_TIER_STANDARD" | "SERVICE_TIER_FLEX" | "SERVICE_TIER_PRIORITY" | undefined;
42
+ serviceTier?: "standard" | "flex" | "priority" | undefined;
43
43
  }>;
44
44
  type GoogleLanguageModelOptions = InferSchema<typeof googleLanguageModelOptions>;
45
45
 
@@ -158,6 +158,14 @@ declare const responseSchema: _ai_sdk_provider_utils.LazySchema<{
158
158
  candidatesTokenCount?: number | null | undefined;
159
159
  totalTokenCount?: number | null | undefined;
160
160
  trafficType?: string | null | undefined;
161
+ promptTokensDetails?: {
162
+ modality: string;
163
+ tokenCount: number;
164
+ }[] | null | undefined;
165
+ candidatesTokensDetails?: {
166
+ modality: string;
167
+ tokenCount: number;
168
+ }[] | null | undefined;
161
169
  } | null | undefined;
162
170
  promptFeedback?: {
163
171
  blockReason?: string | null | undefined;
package/dist/index.js CHANGED
@@ -30,7 +30,7 @@ module.exports = __toCommonJS(index_exports);
30
30
  var import_provider_utils16 = require("@ai-sdk/provider-utils");
31
31
 
32
32
  // src/version.ts
33
- var VERSION = true ? "3.0.56" : "0.0.0-test";
33
+ var VERSION = true ? "3.0.58" : "0.0.0-test";
34
34
 
35
35
  // src/google-generative-ai-embedding-model.ts
36
36
  var import_provider = require("@ai-sdk/provider");
@@ -822,11 +822,7 @@ var googleLanguageModelOptions = (0, import_provider_utils5.lazySchema)(
822
822
  /**
823
823
  * Optional. The service tier to use for the request.
824
824
  */
825
- serviceTier: import_v44.z.enum([
826
- "SERVICE_TIER_STANDARD",
827
- "SERVICE_TIER_FLEX",
828
- "SERVICE_TIER_PRIORITY"
829
- ]).optional()
825
+ serviceTier: import_v44.z.enum(["standard", "flex", "priority"]).optional()
830
826
  })
831
827
  )
832
828
  );
@@ -1947,6 +1943,12 @@ var getSafetyRatingSchema = () => import_v45.z.object({
1947
1943
  severityScore: import_v45.z.number().nullish(),
1948
1944
  blocked: import_v45.z.boolean().nullish()
1949
1945
  });
1946
+ var tokenDetailsSchema = import_v45.z.array(
1947
+ import_v45.z.object({
1948
+ modality: import_v45.z.string(),
1949
+ tokenCount: import_v45.z.number()
1950
+ })
1951
+ ).nullish();
1950
1952
  var usageSchema = import_v45.z.object({
1951
1953
  cachedContentTokenCount: import_v45.z.number().nullish(),
1952
1954
  thoughtsTokenCount: import_v45.z.number().nullish(),
@@ -1954,7 +1956,10 @@ var usageSchema = import_v45.z.object({
1954
1956
  candidatesTokenCount: import_v45.z.number().nullish(),
1955
1957
  totalTokenCount: import_v45.z.number().nullish(),
1956
1958
  // https://cloud.google.com/vertex-ai/generative-ai/docs/reference/rest/v1/GenerateContentResponse#TrafficType
1957
- trafficType: import_v45.z.string().nullish()
1959
+ trafficType: import_v45.z.string().nullish(),
1960
+ // https://ai.google.dev/api/generate-content#Modality
1961
+ promptTokensDetails: tokenDetailsSchema,
1962
+ candidatesTokensDetails: tokenDetailsSchema
1958
1963
  });
1959
1964
  var getUrlContextMetadataSchema = () => import_v45.z.object({
1960
1965
  urlMetadata: import_v45.z.array(