@ai-sdk/openai-compatible 2.0.36 → 2.0.38

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,20 @@
1
1
  # @ai-sdk/openai-compatible
2
2
 
3
+ ## 2.0.38
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [0469aed]
8
+ - @ai-sdk/provider-utils@4.0.22
9
+
10
+ ## 2.0.37
11
+
12
+ ### Patch Changes
13
+
14
+ - 055cd68: fix: publish v6 to latest npm dist tag
15
+ - Updated dependencies [055cd68]
16
+ - @ai-sdk/provider-utils@4.0.21
17
+
3
18
  ## 2.0.36
4
19
 
5
20
  ### Patch Changes
@@ -265,13 +280,13 @@
265
280
  Before
266
281
 
267
282
  ```ts
268
- model.textEmbeddingModel('my-model-id');
283
+ model.textEmbeddingModel("my-model-id");
269
284
  ```
270
285
 
271
286
  After
272
287
 
273
288
  ```ts
274
- model.embeddingModel('my-model-id');
289
+ model.embeddingModel("my-model-id");
275
290
  ```
276
291
 
277
292
  - 2625a04: feat(openai); update spec for mcp approval
@@ -486,13 +501,13 @@
486
501
  Before
487
502
 
488
503
  ```ts
489
- model.textEmbeddingModel('my-model-id');
504
+ model.textEmbeddingModel("my-model-id");
490
505
  ```
491
506
 
492
507
  After
493
508
 
494
509
  ```ts
495
- model.embeddingModel('my-model-id');
510
+ model.embeddingModel("my-model-id");
496
511
  ```
497
512
 
498
513
  - Updated dependencies [8d9e8ad]
@@ -928,7 +943,7 @@
928
943
 
929
944
  ```js
930
945
  await generateImage({
931
- model: luma.image('photon-flash-1', {
946
+ model: luma.image("photon-flash-1", {
932
947
  maxImagesPerCall: 5,
933
948
  pollIntervalMillis: 500,
934
949
  }),
@@ -941,7 +956,7 @@
941
956
 
942
957
  ```js
943
958
  await generateImage({
944
- model: luma.image('photon-flash-1'),
959
+ model: luma.image("photon-flash-1"),
945
960
  prompt,
946
961
  n: 10,
947
962
  maxImagesPerCall: 5,
@@ -1210,7 +1225,7 @@
1210
1225
 
1211
1226
  ```js
1212
1227
  await generateImage({
1213
- model: luma.image('photon-flash-1', {
1228
+ model: luma.image("photon-flash-1", {
1214
1229
  maxImagesPerCall: 5,
1215
1230
  pollIntervalMillis: 500,
1216
1231
  }),
@@ -1223,7 +1238,7 @@
1223
1238
 
1224
1239
  ```js
1225
1240
  await generateImage({
1226
- model: luma.image('photon-flash-1'),
1241
+ model: luma.image("photon-flash-1"),
1227
1242
  prompt,
1228
1243
  n: 10,
1229
1244
  maxImagesPerCall: 5,
package/dist/index.js CHANGED
@@ -18,8 +18,8 @@ var __copyProps = (to, from, except, desc) => {
18
18
  var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
19
 
20
20
  // src/index.ts
21
- var src_exports = {};
22
- __export(src_exports, {
21
+ var index_exports = {};
22
+ __export(index_exports, {
23
23
  OpenAICompatibleChatLanguageModel: () => OpenAICompatibleChatLanguageModel,
24
24
  OpenAICompatibleCompletionLanguageModel: () => OpenAICompatibleCompletionLanguageModel,
25
25
  OpenAICompatibleEmbeddingModel: () => OpenAICompatibleEmbeddingModel,
@@ -27,7 +27,7 @@ __export(src_exports, {
27
27
  VERSION: () => VERSION,
28
28
  createOpenAICompatible: () => createOpenAICompatible
29
29
  });
30
- module.exports = __toCommonJS(src_exports);
30
+ module.exports = __toCommonJS(index_exports);
31
31
 
32
32
  // src/chat/openai-compatible-chat-language-model.ts
33
33
  var import_provider3 = require("@ai-sdk/provider");
@@ -1665,7 +1665,7 @@ function toCamelCase(str) {
1665
1665
  var import_provider_utils6 = require("@ai-sdk/provider-utils");
1666
1666
 
1667
1667
  // src/version.ts
1668
- var VERSION = true ? "2.0.36" : "0.0.0-test";
1668
+ var VERSION = true ? "2.0.38" : "0.0.0-test";
1669
1669
 
1670
1670
  // src/openai-compatible-provider.ts
1671
1671
  function createOpenAICompatible(options) {