@ai-sdk/google-vertex 4.0.68 → 4.0.69

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,14 @@
1
1
  # @ai-sdk/google-vertex
2
2
 
3
+ ## 4.0.69
4
+
5
+ ### Patch Changes
6
+
7
+ - 64a8fae: chore: remove obsolete model IDs for Anthropic, Google, OpenAI, xAI
8
+ - Updated dependencies [64a8fae]
9
+ - @ai-sdk/anthropic@3.0.51
10
+ - @ai-sdk/google@3.0.35
11
+
3
12
  ## 4.0.68
4
13
 
5
14
  ### Patch Changes
package/README.md CHANGED
@@ -26,14 +26,14 @@ The Google Vertex provider has two different authentication implementations depe
26
26
 
27
27
  ### Node.js Runtime
28
28
 
29
- The Node.js runtime is the default runtime supported by the AI SDK. You can use the default provider instance to generate text with the `gemini-1.5-flash` model like this:
29
+ The Node.js runtime is the default runtime supported by the AI SDK. You can use the default provider instance to generate text with the `gemini-2.5-flash` model like this:
30
30
 
31
31
  ```ts
32
32
  import { vertex } from '@ai-sdk/google-vertex';
33
33
  import { generateText } from 'ai';
34
34
 
35
35
  const { text } = await generateText({
36
- model: vertex('gemini-1.5-flash'),
36
+ model: vertex('gemini-2.5-flash'),
37
37
  prompt: 'Write a vegetarian lasagna recipe.',
38
38
  });
39
39
  ```
@@ -44,14 +44,14 @@ This provider supports all standard Google Cloud authentication options through
44
44
 
45
45
  The Edge runtime is supported through the `@ai-sdk/google-vertex/edge` module. Note the additional sub-module path `/edge` required to differentiate the Edge provider from the Node.js provider.
46
46
 
47
- You can use the default provider instance to generate text with the `gemini-1.5-flash` model like this:
47
+ You can use the default provider instance to generate text with the `gemini-2.5-flash` model like this:
48
48
 
49
49
  ```ts
50
50
  import { vertex } from '@ai-sdk/google-vertex/edge';
51
51
  import { generateText } from 'ai';
52
52
 
53
53
  const { text } = await generateText({
54
- model: vertex('gemini-1.5-flash'),
54
+ model: vertex('gemini-2.5-flash'),
55
55
  prompt: 'Write a vegetarian lasagna recipe.',
56
56
  });
57
57
  ```
@@ -160,7 +160,7 @@ const customProvider = createVertex({
160
160
  });
161
161
 
162
162
  const { text } = await generateText({
163
- model: customProvider('gemini-1.5-flash'),
163
+ model: customProvider('gemini-2.5-flash'),
164
164
  prompt: 'Write a vegetarian lasagna recipe.',
165
165
  });
166
166
  ```
@@ -183,7 +183,7 @@ const customProvider = createVertex({
183
183
  });
184
184
 
185
185
  const { text } = await generateText({
186
- model: customProvider('gemini-1.5-flash'),
186
+ model: customProvider('gemini-2.5-flash'),
187
187
  prompt: 'Write a vegetarian lasagna recipe.',
188
188
  });
189
189
  ```
@@ -32,7 +32,7 @@ var import_provider_utils3 = require("@ai-sdk/provider-utils");
32
32
  var import_provider_utils = require("@ai-sdk/provider-utils");
33
33
 
34
34
  // src/version.ts
35
- var VERSION = true ? "4.0.68" : "0.0.0-test";
35
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
36
36
 
37
37
  // src/edge/google-vertex-auth-edge.ts
38
38
  var loadCredentials = async () => {
@@ -10,7 +10,7 @@ import {
10
10
  } from "@ai-sdk/provider-utils";
11
11
 
12
12
  // src/version.ts
13
- var VERSION = true ? "4.0.68" : "0.0.0-test";
13
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
14
14
 
15
15
  // src/edge/google-vertex-auth-edge.ts
16
16
  var loadCredentials = async () => {
@@ -33,7 +33,7 @@ var import_internal3 = require("@ai-sdk/google/internal");
33
33
  var import_provider_utils5 = require("@ai-sdk/provider-utils");
34
34
 
35
35
  // src/version.ts
36
- var VERSION = true ? "4.0.68" : "0.0.0-test";
36
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
37
37
 
38
38
  // src/google-vertex-embedding-model.ts
39
39
  var import_provider = require("@ai-sdk/provider");
@@ -14,7 +14,7 @@ import {
14
14
  } from "@ai-sdk/provider-utils";
15
15
 
16
16
  // src/version.ts
17
- var VERSION = true ? "4.0.68" : "0.0.0-test";
17
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
18
18
 
19
19
  // src/google-vertex-embedding-model.ts
20
20
  import {
package/dist/index.js CHANGED
@@ -55,7 +55,7 @@ var import_internal3 = require("@ai-sdk/google/internal");
55
55
  var import_provider_utils5 = require("@ai-sdk/provider-utils");
56
56
 
57
57
  // src/version.ts
58
- var VERSION = true ? "4.0.68" : "0.0.0-test";
58
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
59
59
 
60
60
  // src/google-vertex-embedding-model.ts
61
61
  var import_provider = require("@ai-sdk/provider");
package/dist/index.mjs CHANGED
@@ -35,7 +35,7 @@ import {
35
35
  } from "@ai-sdk/provider-utils";
36
36
 
37
37
  // src/version.ts
38
- var VERSION = true ? "4.0.68" : "0.0.0-test";
38
+ var VERSION = true ? "4.0.69" : "0.0.0-test";
39
39
 
40
40
  // src/google-vertex-embedding-model.ts
41
41
  import {
@@ -251,10 +251,10 @@ const vertex = createVertex({
251
251
  ### Language Models
252
252
 
253
253
  You can create models that call the Vertex API using the provider instance.
254
- The first argument is the model id, e.g. `gemini-1.5-pro`.
254
+ The first argument is the model id, e.g. `gemini-2.5-pro`.
255
255
 
256
256
  ```ts
257
- const model = vertex('gemini-1.5-pro');
257
+ const model = vertex('gemini-2.5-pro');
258
258
  ```
259
259
 
260
260
  <Note>
@@ -271,7 +271,7 @@ an options argument:
271
271
  import { vertex } from '@ai-sdk/google-vertex';
272
272
  import { type GoogleLanguageModelOptions } from '@ai-sdk/google';
273
273
 
274
- const model = vertex('gemini-1.5-pro');
274
+ const model = vertex('gemini-2.5-pro');
275
275
 
276
276
  await generateText({
277
277
  model,
@@ -351,7 +351,7 @@ import { vertex } from '@ai-sdk/google-vertex';
351
351
  import { generateText } from 'ai';
352
352
 
353
353
  const { text } = await generateText({
354
- model: vertex('gemini-1.5-pro'),
354
+ model: vertex('gemini-2.5-pro'),
355
355
  prompt: 'Write a vegetarian lasagna recipe for 4 people.',
356
356
  });
357
357
  ```
@@ -525,7 +525,7 @@ import { vertex } from '@ai-sdk/google-vertex';
525
525
  import { generateText } from 'ai';
526
526
 
527
527
  const { text } = await generateText({
528
- model: vertex('gemini-1.5-pro'),
528
+ model: vertex('gemini-2.5-pro'),
529
529
  messages: [
530
530
  {
531
531
  role: 'user',
@@ -721,7 +721,7 @@ import { type GoogleLanguageModelOptions } from '@ai-sdk/google';
721
721
  import { generateText, Output } from 'ai';
722
722
 
723
723
  const result = await generateText({
724
- model: vertex('gemini-1.5-pro'),
724
+ model: vertex('gemini-2.5-pro'),
725
725
  providerOptions: {
726
726
  vertex: {
727
727
  structuredOutputs: false,
@@ -760,8 +760,6 @@ The following Zod features are known to not work with Google Vertex:
760
760
  | `gemini-2.5-pro` | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
761
761
  | `gemini-2.5-flash` | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
762
762
  | `gemini-2.0-flash-001` | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
763
- | `gemini-1.5-flash` | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
764
- | `gemini-1.5-pro` | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
765
763
 
766
764
  <Note>
767
765
  The table above lists popular models. Please see the [Google Vertex AI
@@ -775,7 +773,7 @@ The following Zod features are known to not work with Google Vertex:
775
773
  You can create models that call the Google Vertex AI embeddings API using the `.embeddingModel()` factory method:
776
774
 
777
775
  ```ts
778
- const model = vertex.embeddingModel('text-embedding-004');
776
+ const model = vertex.embeddingModel('text-embedding-005');
779
777
  ```
780
778
 
781
779
  Google Vertex AI embedding models support additional settings. You can pass them as an options argument:
@@ -787,7 +785,7 @@ import {
787
785
  } from '@ai-sdk/google-vertex';
788
786
  import { embed } from 'ai';
789
787
 
790
- const model = vertex.embeddingModel('text-embedding-004');
788
+ const model = vertex.embeddingModel('text-embedding-005');
791
789
 
792
790
  const { embedding } = await embed({
793
791
  model,
@@ -833,7 +831,7 @@ The following optional provider options are available for Google Vertex AI embed
833
831
 
834
832
  | Model | Max Values Per Call | Parallel Calls |
835
833
  | -------------------- | ------------------- | ------------------- |
836
- | `text-embedding-004` | 2048 | <Check size={18} /> |
834
+ | `text-embedding-005` | 2048 | <Check size={18} /> |
837
835
 
838
836
  <Note>
839
837
  The table above lists popular models. You can also pass any available provider
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@ai-sdk/google-vertex",
3
- "version": "4.0.68",
3
+ "version": "4.0.69",
4
4
  "license": "Apache-2.0",
5
5
  "sideEffects": false,
6
6
  "main": "./dist/index.js",
@@ -48,10 +48,10 @@
48
48
  },
49
49
  "dependencies": {
50
50
  "google-auth-library": "^10.5.0",
51
- "@ai-sdk/anthropic": "3.0.50",
52
- "@ai-sdk/google": "3.0.34",
53
- "@ai-sdk/provider": "3.0.8",
54
- "@ai-sdk/provider-utils": "4.0.16"
51
+ "@ai-sdk/anthropic": "3.0.51",
52
+ "@ai-sdk/google": "3.0.35",
53
+ "@ai-sdk/provider-utils": "4.0.16",
54
+ "@ai-sdk/provider": "3.0.8"
55
55
  },
56
56
  "devDependencies": {
57
57
  "@types/node": "20.17.24",