supercompat 2.30.1 → 2.32.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.cts CHANGED
@@ -3,8 +3,6 @@ import OpenAI, { AzureOpenAI } from 'openai';
3
3
  import { Mistral } from '@mistralai/mistralai';
4
4
  import Anthropic from '@anthropic-ai/sdk';
5
5
  import { MessageWithRun, StorageAdapterArgs } from './types/index.cjs';
6
- import * as openai_streaming_mjs from 'openai/streaming.mjs';
7
- import * as openai_resources_beta_assistants_mjs from 'openai/resources/beta/assistants.mjs';
8
6
  import * as openai_resources_beta_threads_runs_runs_mjs from 'openai/resources/beta/threads/runs/runs.mjs';
9
7
  import * as openai_resources_beta_threads_messages_mjs from 'openai/resources/beta/threads/messages.mjs';
10
8
  import * as openai_resources_beta_threads_threads_mjs from 'openai/resources/beta/threads/threads.mjs';
@@ -162,31 +160,28 @@ declare const prismaStorageAdapter: ({ prisma, }: {
162
160
  requestHandlers: {
163
161
  '^/(?:v1|/?openai)/threads$': {
164
162
  post: (options?: openai_core_mjs.RequestOptions<unknown> | undefined) => Promise<Response & {
165
- json: () => Promise<{
166
- (body?: openai_resources_beta_threads_threads_mjs.ThreadCreateParams | undefined, options?: openai_core_mjs.RequestOptions<unknown> | undefined): openai_core_mjs.APIPromise<openai_resources_beta_threads_threads_mjs.Thread>;
167
- (options?: openai_core_mjs.RequestOptions<unknown> | undefined): openai_core_mjs.APIPromise<openai_resources_beta_threads_threads_mjs.Thread>;
168
- }>;
163
+ json: () => Promise<openai_resources_beta_threads_threads_mjs.Threads["create"]>;
169
164
  }>;
170
165
  };
171
166
  "^/(?:v1|/?openai)/threads/([^/]+)/messages$": {
172
167
  post: (urlString: string, options: any) => Promise<Response & {
173
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
168
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
174
169
  }>;
175
170
  get: (urlString: string) => Promise<Response & {
176
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
171
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
177
172
  }>;
178
173
  };
179
174
  "^/(?:v1|/?openai)/threads/([^/]+)/runs$": {
180
175
  get: (urlString: string) => Promise<Response & {
181
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
176
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
182
177
  }>;
183
178
  post: (urlString: string, options: any) => Promise<Response & {
184
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_runs_runs_mjs.Run | openai_streaming_mjs.Stream<openai_resources_beta_assistants_mjs.AssistantStreamEvent>>>;
179
+ json: () => Promise<ReturnType<openai_resources_beta_threads_runs_runs_mjs.Runs["create"]>>;
185
180
  }>;
186
181
  };
187
182
  "^/(?:v1|/?openai)/threads/([^/]+)/runs/([^/]+)$": {
188
183
  get: (urlString: string) => Promise<Response & {
189
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_runs_runs_mjs.Run>>;
184
+ json: () => Promise<ReturnType<openai_resources_beta_threads_runs_runs_mjs.Runs["retrieve"]>>;
190
185
  }>;
191
186
  };
192
187
  "^/(?:v1|/?openai)/threads/([^/]+)/runs/([^/]+)/steps$": {
package/dist/index.d.ts CHANGED
@@ -3,8 +3,6 @@ import OpenAI, { AzureOpenAI } from 'openai';
3
3
  import { Mistral } from '@mistralai/mistralai';
4
4
  import Anthropic from '@anthropic-ai/sdk';
5
5
  import { MessageWithRun, StorageAdapterArgs } from './types/index.js';
6
- import * as openai_streaming_mjs from 'openai/streaming.mjs';
7
- import * as openai_resources_beta_assistants_mjs from 'openai/resources/beta/assistants.mjs';
8
6
  import * as openai_resources_beta_threads_runs_runs_mjs from 'openai/resources/beta/threads/runs/runs.mjs';
9
7
  import * as openai_resources_beta_threads_messages_mjs from 'openai/resources/beta/threads/messages.mjs';
10
8
  import * as openai_resources_beta_threads_threads_mjs from 'openai/resources/beta/threads/threads.mjs';
@@ -162,31 +160,28 @@ declare const prismaStorageAdapter: ({ prisma, }: {
162
160
  requestHandlers: {
163
161
  '^/(?:v1|/?openai)/threads$': {
164
162
  post: (options?: openai_core_mjs.RequestOptions<unknown> | undefined) => Promise<Response & {
165
- json: () => Promise<{
166
- (body?: openai_resources_beta_threads_threads_mjs.ThreadCreateParams | undefined, options?: openai_core_mjs.RequestOptions<unknown> | undefined): openai_core_mjs.APIPromise<openai_resources_beta_threads_threads_mjs.Thread>;
167
- (options?: openai_core_mjs.RequestOptions<unknown> | undefined): openai_core_mjs.APIPromise<openai_resources_beta_threads_threads_mjs.Thread>;
168
- }>;
163
+ json: () => Promise<openai_resources_beta_threads_threads_mjs.Threads["create"]>;
169
164
  }>;
170
165
  };
171
166
  "^/(?:v1|/?openai)/threads/([^/]+)/messages$": {
172
167
  post: (urlString: string, options: any) => Promise<Response & {
173
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
168
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
174
169
  }>;
175
170
  get: (urlString: string) => Promise<Response & {
176
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
171
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
177
172
  }>;
178
173
  };
179
174
  "^/(?:v1|/?openai)/threads/([^/]+)/runs$": {
180
175
  get: (urlString: string) => Promise<Response & {
181
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_messages_mjs.Message>>;
176
+ json: () => Promise<ReturnType<openai_resources_beta_threads_messages_mjs.Messages["create"]>>;
182
177
  }>;
183
178
  post: (urlString: string, options: any) => Promise<Response & {
184
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_runs_runs_mjs.Run | openai_streaming_mjs.Stream<openai_resources_beta_assistants_mjs.AssistantStreamEvent>>>;
179
+ json: () => Promise<ReturnType<openai_resources_beta_threads_runs_runs_mjs.Runs["create"]>>;
185
180
  }>;
186
181
  };
187
182
  "^/(?:v1|/?openai)/threads/([^/]+)/runs/([^/]+)$": {
188
183
  get: (urlString: string) => Promise<Response & {
189
- json: () => Promise<openai_core_mjs.APIPromise<openai_resources_beta_threads_runs_runs_mjs.Run>>;
184
+ json: () => Promise<ReturnType<openai_resources_beta_threads_runs_runs_mjs.Runs["retrieve"]>>;
190
185
  }>;
191
186
  };
192
187
  "^/(?:v1|/?openai)/threads/([^/]+)/runs/([^/]+)/steps$": {
package/dist/index.js CHANGED
@@ -1636,56 +1636,61 @@ var perplexityClientAdapter = function(param) {
1636
1636
  };
1637
1637
  };
1638
1638
  // src/adapters/client/anthropicClientAdapter/models/get.ts
1639
- var models6 = [
1640
- "claude-3-7-sonnet-latest",
1641
- "claude-3-7-sonnet-20250219",
1642
- "claude-3-5-haiku-latest",
1643
- "claude-3-5-sonnet-20241022",
1644
- "claude-3-5-sonnet-20240620",
1645
- "claude-3-opus-20240229",
1646
- "claude-3-sonnet-20240229",
1647
- "claude-3-5-haiku-20241022",
1648
- "claude-3-haiku-20240307"
1649
- ];
1650
1639
  var get5 = function(param) {
1651
1640
  var anthropic = param.anthropic;
1652
1641
  return /*#__PURE__*/ function() {
1653
1642
  var _ref = _async_to_generator(function(_url, _options) {
1643
+ var response, error;
1654
1644
  return _ts_generator(this, function(_state) {
1655
- try {
1656
- return [
1657
- 2,
1658
- new Response(JSON.stringify({
1659
- type: "list",
1660
- data: models6.map(function(model) {
1661
- return {
1662
- id: model,
1663
- object: "model"
1664
- };
1645
+ switch(_state.label){
1646
+ case 0:
1647
+ _state.trys.push([
1648
+ 0,
1649
+ 2,
1650
+ ,
1651
+ 3
1652
+ ]);
1653
+ return [
1654
+ 4,
1655
+ anthropic.models.list()
1656
+ ];
1657
+ case 1:
1658
+ response = _state.sent();
1659
+ return [
1660
+ 2,
1661
+ new Response(JSON.stringify({
1662
+ type: "list",
1663
+ data: response.data.map(function(model) {
1664
+ return {
1665
+ id: model.id,
1666
+ object: "model"
1667
+ };
1668
+ })
1669
+ }), {
1670
+ status: 200,
1671
+ headers: {
1672
+ "Content-Type": "application/json"
1673
+ }
1665
1674
  })
1666
- }), {
1667
- status: 200,
1668
- headers: {
1669
- "Content-Type": "application/json"
1670
- }
1671
- })
1672
- ];
1673
- } catch (error) {
1674
- return [
1675
- 2,
1676
- new Response(JSON.stringify({
1677
- error: error
1678
- }), {
1679
- status: 500,
1680
- headers: {
1681
- "Content-Type": "application/json"
1682
- }
1683
- })
1684
- ];
1675
+ ];
1676
+ case 2:
1677
+ error = _state.sent();
1678
+ return [
1679
+ 2,
1680
+ new Response(JSON.stringify({
1681
+ error: error
1682
+ }), {
1683
+ status: 500,
1684
+ headers: {
1685
+ "Content-Type": "application/json"
1686
+ }
1687
+ })
1688
+ ];
1689
+ case 3:
1690
+ return [
1691
+ 2
1692
+ ];
1685
1693
  }
1686
- return [
1687
- 2
1688
- ];
1689
1694
  });
1690
1695
  });
1691
1696
  return function(_url, _options) {
@@ -1694,7 +1699,7 @@ var get5 = function(param) {
1694
1699
  }();
1695
1700
  };
1696
1701
  // src/adapters/client/anthropicClientAdapter/models/index.ts
1697
- var models7 = function(param) {
1702
+ var models6 = function(param) {
1698
1703
  var anthropic = param.anthropic;
1699
1704
  return {
1700
1705
  get: get5({
@@ -2086,7 +2091,7 @@ var anthropicClientAdapter = function(param) {
2086
2091
  return {
2087
2092
  client: anthropic,
2088
2093
  requestHandlers: {
2089
- "^/v1/models$": models7({
2094
+ "^/v1/models$": models6({
2090
2095
  anthropic: anthropic
2091
2096
  }),
2092
2097
  "^/v1/chat/completions$": completions5({
@@ -2155,7 +2160,7 @@ var get6 = function(param) {
2155
2160
  }();
2156
2161
  };
2157
2162
  // src/adapters/client/togetherClientAdapter/models/index.ts
2158
- var models8 = function(param) {
2163
+ var models7 = function(param) {
2159
2164
  var together = param.together;
2160
2165
  return {
2161
2166
  get: get6({
@@ -2350,7 +2355,7 @@ var togetherClientAdapter = function(param) {
2350
2355
  return {
2351
2356
  client: together,
2352
2357
  requestHandlers: {
2353
- "^/v1/models$": models8({
2358
+ "^/v1/models$": models7({
2354
2359
  together: together
2355
2360
  }),
2356
2361
  "^/(?:v1|/?openai)/chat/completions$": completions6({
@@ -2360,7 +2365,7 @@ var togetherClientAdapter = function(param) {
2360
2365
  };
2361
2366
  };
2362
2367
  // src/adapters/client/googleClientAdapter/models/get.ts
2363
- var models9 = [
2368
+ var models8 = [
2364
2369
  "gemini-2.5-flash-preview-04-17",
2365
2370
  "gemini-2.5-pro-preview-03-25",
2366
2371
  "gemini-2.0-flash",
@@ -2379,7 +2384,7 @@ var get7 = function(param) {
2379
2384
  2,
2380
2385
  new Response(JSON.stringify({
2381
2386
  type: "list",
2382
- data: models9.map(function(model) {
2387
+ data: models8.map(function(model) {
2383
2388
  return {
2384
2389
  id: model,
2385
2390
  object: "model"
@@ -2416,7 +2421,7 @@ var get7 = function(param) {
2416
2421
  }();
2417
2422
  };
2418
2423
  // src/adapters/client/googleClientAdapter/models/index.ts
2419
- var models10 = function(param) {
2424
+ var models9 = function(param) {
2420
2425
  var google = param.google;
2421
2426
  return {
2422
2427
  get: get7({
@@ -2644,7 +2649,7 @@ var googleClientAdapter = function(param) {
2644
2649
  return {
2645
2650
  client: google,
2646
2651
  requestHandlers: {
2647
- "^/v1/models$": models10({
2652
+ "^/v1/models$": models9({
2648
2653
  google: google
2649
2654
  }),
2650
2655
  "^/(?:v1|/?openai)/chat/completions$": completions7({
@@ -2654,7 +2659,7 @@ var googleClientAdapter = function(param) {
2654
2659
  };
2655
2660
  };
2656
2661
  // src/adapters/client/humirisClientAdapter/models/get.ts
2657
- var models11 = [
2662
+ var models10 = [
2658
2663
  "Humiris/humiris-moai"
2659
2664
  ];
2660
2665
  var get8 = function(param) {
@@ -2667,7 +2672,7 @@ var get8 = function(param) {
2667
2672
  2,
2668
2673
  new Response(JSON.stringify({
2669
2674
  type: "list",
2670
- data: models11.map(function(model) {
2675
+ data: models10.map(function(model) {
2671
2676
  return {
2672
2677
  id: model,
2673
2678
  object: "model"
@@ -2704,7 +2709,7 @@ var get8 = function(param) {
2704
2709
  }();
2705
2710
  };
2706
2711
  // src/adapters/client/humirisClientAdapter/models/index.ts
2707
- var models12 = function(param) {
2712
+ var models11 = function(param) {
2708
2713
  var humiris = param.humiris;
2709
2714
  return {
2710
2715
  get: get8({
@@ -2837,7 +2842,7 @@ var humirisClientAdapter = function(param) {
2837
2842
  return {
2838
2843
  client: humiris,
2839
2844
  requestHandlers: {
2840
- "^/v1/models$": models12({
2845
+ "^/v1/models$": models11({
2841
2846
  humiris: humiris
2842
2847
  }),
2843
2848
  "^/(?:v1|/?openai)/chat/completions$": completions8({
@@ -2902,7 +2907,7 @@ var get9 = function(param) {
2902
2907
  }();
2903
2908
  };
2904
2909
  // src/adapters/client/ollamaClientAdapter/models/index.ts
2905
- var models13 = function(param) {
2910
+ var models12 = function(param) {
2906
2911
  var ollama = param.ollama;
2907
2912
  return {
2908
2913
  get: get9({
@@ -3097,7 +3102,7 @@ var ollamaClientAdapter = function(param) {
3097
3102
  return {
3098
3103
  client: ollama,
3099
3104
  requestHandlers: {
3100
- "^/v1/models$": models13({
3105
+ "^/v1/models$": models12({
3101
3106
  ollama: ollama
3102
3107
  }),
3103
3108
  "^/(?:v1|/?openai)/chat/completions$": completions9({
@@ -3797,14 +3802,14 @@ var post11 = function(param) {
3797
3802
  var prisma = param.prisma;
3798
3803
  return /*#__PURE__*/ function() {
3799
3804
  var _ref = _async_to_generator(function(urlString, options) {
3800
- var url, _url_pathname_match, threadId, body, role, content, message;
3805
+ var url, _url_pathname_match, threadId, body, role, content, metadata, message;
3801
3806
  return _ts_generator(this, function(_state) {
3802
3807
  switch(_state.label){
3803
3808
  case 0:
3804
3809
  url = new URL(urlString);
3805
3810
  _url_pathname_match = _sliced_to_array(url.pathname.match(new RegExp(messagesRegexp)), 2), threadId = _url_pathname_match[1];
3806
3811
  body = JSON.parse(options.body);
3807
- role = body.role, content = body.content;
3812
+ role = body.role, content = body.content, metadata = body.metadata;
3808
3813
  return [
3809
3814
  4,
3810
3815
  prisma.message.create({
@@ -3813,7 +3818,8 @@ var post11 = function(param) {
3813
3818
  content: messageContentBlocks({
3814
3819
  content: content
3815
3820
  }),
3816
- role: role === "user" ? "USER" : "ASSISTANT"
3821
+ role: role === "user" ? "USER" : "ASSISTANT",
3822
+ metadata: metadata || {}
3817
3823
  }
3818
3824
  })
3819
3825
  ];