@draht/ai 2026.3.2 → 2026.3.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (63) hide show
  1. package/README.md +31 -28
  2. package/bedrock-provider.d.ts +1 -0
  3. package/bedrock-provider.js +1 -0
  4. package/dist/bedrock-provider.d.ts +5 -0
  5. package/dist/bedrock-provider.d.ts.map +1 -0
  6. package/dist/bedrock-provider.js +6 -0
  7. package/dist/bedrock-provider.js.map +1 -0
  8. package/dist/cli.js +0 -0
  9. package/dist/env-api-keys.d.ts.map +1 -1
  10. package/dist/env-api-keys.js +8 -3
  11. package/dist/env-api-keys.js.map +1 -1
  12. package/dist/index.d.ts +1 -1
  13. package/dist/index.d.ts.map +1 -1
  14. package/dist/index.js +0 -1
  15. package/dist/index.js.map +1 -1
  16. package/dist/models.generated.d.ts +175 -3
  17. package/dist/models.generated.d.ts.map +1 -1
  18. package/dist/models.generated.js +230 -58
  19. package/dist/models.generated.js.map +1 -1
  20. package/dist/oauth.d.ts +2 -0
  21. package/dist/oauth.d.ts.map +1 -0
  22. package/dist/oauth.js +2 -0
  23. package/dist/oauth.js.map +1 -0
  24. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  25. package/dist/providers/amazon-bedrock.js +15 -3
  26. package/dist/providers/amazon-bedrock.js.map +1 -1
  27. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  28. package/dist/providers/google-gemini-cli.js +12 -3
  29. package/dist/providers/google-gemini-cli.js.map +1 -1
  30. package/dist/providers/google-vertex.d.ts.map +1 -1
  31. package/dist/providers/google-vertex.js +2 -2
  32. package/dist/providers/google-vertex.js.map +1 -1
  33. package/dist/providers/google.d.ts.map +1 -1
  34. package/dist/providers/google.js +2 -2
  35. package/dist/providers/google.js.map +1 -1
  36. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  37. package/dist/providers/openai-codex-responses.js +4 -2
  38. package/dist/providers/openai-codex-responses.js.map +1 -1
  39. package/dist/providers/openai-completions.d.ts.map +1 -1
  40. package/dist/providers/openai-completions.js +16 -1
  41. package/dist/providers/openai-completions.js.map +1 -1
  42. package/dist/providers/register-builtins.d.ts +7 -0
  43. package/dist/providers/register-builtins.d.ts.map +1 -1
  44. package/dist/providers/register-builtins.js +72 -3
  45. package/dist/providers/register-builtins.js.map +1 -1
  46. package/dist/stream.d.ts +0 -1
  47. package/dist/stream.d.ts.map +1 -1
  48. package/dist/stream.js +0 -1
  49. package/dist/stream.js.map +1 -1
  50. package/dist/types.d.ts +3 -1
  51. package/dist/types.d.ts.map +1 -1
  52. package/dist/types.js.map +1 -1
  53. package/dist/utils/oauth/index.d.ts +0 -1
  54. package/dist/utils/oauth/index.d.ts.map +1 -1
  55. package/dist/utils/oauth/index.js +0 -2
  56. package/dist/utils/oauth/index.js.map +1 -1
  57. package/oauth.d.ts +1 -0
  58. package/oauth.js +1 -0
  59. package/package.json +23 -5
  60. package/dist/utils/http-proxy.d.ts +0 -2
  61. package/dist/utils/http-proxy.d.ts.map +0 -1
  62. package/dist/utils/http-proxy.js +0 -15
  63. package/dist/utils/http-proxy.js.map +0 -1
@@ -3168,6 +3168,23 @@ export const MODELS = {
3168
3168
  contextWindow: 1000000,
3169
3169
  maxTokens: 64000,
3170
3170
  },
3171
+ "gemini-3.1-flash-lite-preview": {
3172
+ id: "gemini-3.1-flash-lite-preview",
3173
+ name: "Gemini 3.1 Flash Lite Preview",
3174
+ api: "google-generative-ai",
3175
+ provider: "google",
3176
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
3177
+ reasoning: true,
3178
+ input: ["text", "image"],
3179
+ cost: {
3180
+ input: 0,
3181
+ output: 0,
3182
+ cacheRead: 0,
3183
+ cacheWrite: 0,
3184
+ },
3185
+ contextWindow: 1048576,
3186
+ maxTokens: 65536,
3187
+ },
3171
3188
  "gemini-3.1-pro-preview": {
3172
3189
  id: "gemini-3.1-pro-preview",
3173
3190
  name: "Gemini 3.1 Pro Preview",
@@ -3357,9 +3374,9 @@ export const MODELS = {
3357
3374
  contextWindow: 1048576,
3358
3375
  maxTokens: 65535,
3359
3376
  },
3360
- "gemini-3-pro-high": {
3361
- id: "gemini-3-pro-high",
3362
- name: "Gemini 3 Pro High (Antigravity)",
3377
+ "gemini-3.1-pro-high": {
3378
+ id: "gemini-3.1-pro-high",
3379
+ name: "Gemini 3.1 Pro High (Antigravity)",
3363
3380
  api: "google-gemini-cli",
3364
3381
  provider: "google-antigravity",
3365
3382
  baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
@@ -3374,9 +3391,9 @@ export const MODELS = {
3374
3391
  contextWindow: 1048576,
3375
3392
  maxTokens: 65535,
3376
3393
  },
3377
- "gemini-3-pro-low": {
3378
- id: "gemini-3-pro-low",
3379
- name: "Gemini 3 Pro Low (Antigravity)",
3394
+ "gemini-3.1-pro-low": {
3395
+ id: "gemini-3.1-pro-low",
3396
+ name: "Gemini 3.1 Pro Low (Antigravity)",
3380
3397
  api: "google-gemini-cli",
3381
3398
  provider: "google-antigravity",
3382
3399
  baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
@@ -6152,6 +6169,59 @@ export const MODELS = {
6152
6169
  maxTokens: 131072,
6153
6170
  },
6154
6171
  },
6172
+ "opencode-go": {
6173
+ "glm-5": {
6174
+ id: "glm-5",
6175
+ name: "GLM-5",
6176
+ api: "openai-completions",
6177
+ provider: "opencode-go",
6178
+ baseUrl: "https://opencode.ai/zen/go/v1",
6179
+ reasoning: true,
6180
+ input: ["text"],
6181
+ cost: {
6182
+ input: 1,
6183
+ output: 3.2,
6184
+ cacheRead: 0.2,
6185
+ cacheWrite: 0,
6186
+ },
6187
+ contextWindow: 204800,
6188
+ maxTokens: 131072,
6189
+ },
6190
+ "kimi-k2.5": {
6191
+ id: "kimi-k2.5",
6192
+ name: "Kimi K2.5",
6193
+ api: "openai-completions",
6194
+ provider: "opencode-go",
6195
+ baseUrl: "https://opencode.ai/zen/go/v1",
6196
+ reasoning: true,
6197
+ input: ["text", "image"],
6198
+ cost: {
6199
+ input: 0.6,
6200
+ output: 3,
6201
+ cacheRead: 0.1,
6202
+ cacheWrite: 0,
6203
+ },
6204
+ contextWindow: 262144,
6205
+ maxTokens: 65536,
6206
+ },
6207
+ "minimax-m2.5": {
6208
+ id: "minimax-m2.5",
6209
+ name: "MiniMax M2.5",
6210
+ api: "anthropic-messages",
6211
+ provider: "opencode-go",
6212
+ baseUrl: "https://opencode.ai/zen/go",
6213
+ reasoning: true,
6214
+ input: ["text"],
6215
+ cost: {
6216
+ input: 0.3,
6217
+ output: 1.2,
6218
+ cacheRead: 0.03,
6219
+ cacheWrite: 0,
6220
+ },
6221
+ contextWindow: 204800,
6222
+ maxTokens: 131072,
6223
+ },
6224
+ },
6155
6225
  "openrouter": {
6156
6226
  "ai21/jamba-large-1.7": {
6157
6227
  id: "ai21/jamba-large-1.7",
@@ -6867,6 +6937,23 @@ export const MODELS = {
6867
6937
  contextWindow: 163840,
6868
6938
  maxTokens: 65536,
6869
6939
  },
6940
+ "essentialai/rnj-1-instruct": {
6941
+ id: "essentialai/rnj-1-instruct",
6942
+ name: "EssentialAI: Rnj 1 Instruct",
6943
+ api: "openai-completions",
6944
+ provider: "openrouter",
6945
+ baseUrl: "https://openrouter.ai/api/v1",
6946
+ reasoning: false,
6947
+ input: ["text"],
6948
+ cost: {
6949
+ input: 0.15,
6950
+ output: 0.15,
6951
+ cacheRead: 0,
6952
+ cacheWrite: 0,
6953
+ },
6954
+ contextWindow: 32768,
6955
+ maxTokens: 4096,
6956
+ },
6870
6957
  "google/gemini-2.0-flash-001": {
6871
6958
  id: "google/gemini-2.0-flash-001",
6872
6959
  name: "Google: Gemini 2.0 Flash",
@@ -6950,7 +7037,7 @@ export const MODELS = {
6950
7037
  cacheWrite: 0.08333333333333334,
6951
7038
  },
6952
7039
  contextWindow: 1048576,
6953
- maxTokens: 65535,
7040
+ maxTokens: 65536,
6954
7041
  },
6955
7042
  "google/gemini-2.5-pro": {
6956
7043
  id: "google/gemini-2.5-pro",
@@ -7018,7 +7105,7 @@ export const MODELS = {
7018
7105
  cacheWrite: 0.08333333333333334,
7019
7106
  },
7020
7107
  contextWindow: 1048576,
7021
- maxTokens: 65535,
7108
+ maxTokens: 65536,
7022
7109
  },
7023
7110
  "google/gemini-3-pro-preview": {
7024
7111
  id: "google/gemini-3-pro-preview",
@@ -7037,6 +7124,23 @@ export const MODELS = {
7037
7124
  contextWindow: 1048576,
7038
7125
  maxTokens: 65536,
7039
7126
  },
7127
+ "google/gemini-3.1-flash-lite-preview": {
7128
+ id: "google/gemini-3.1-flash-lite-preview",
7129
+ name: "Google: Gemini 3.1 Flash Lite Preview",
7130
+ api: "openai-completions",
7131
+ provider: "openrouter",
7132
+ baseUrl: "https://openrouter.ai/api/v1",
7133
+ reasoning: true,
7134
+ input: ["text", "image"],
7135
+ cost: {
7136
+ input: 0.25,
7137
+ output: 1.5,
7138
+ cacheRead: 0.024999999999999998,
7139
+ cacheWrite: 0.08333333333333334,
7140
+ },
7141
+ contextWindow: 1048576,
7142
+ maxTokens: 65536,
7143
+ },
7040
7144
  "google/gemini-3.1-pro-preview": {
7041
7145
  id: "google/gemini-3.1-pro-preview",
7042
7146
  name: "Google: Gemini 3.1 Pro Preview",
@@ -7115,8 +7219,8 @@ export const MODELS = {
7115
7219
  input: ["text"],
7116
7220
  cost: {
7117
7221
  input: 0.25,
7118
- output: 1,
7119
- cacheRead: 0,
7222
+ output: 0.75,
7223
+ cacheRead: 0.024999999999999998,
7120
7224
  cacheWrite: 0,
7121
7225
  },
7122
7226
  contextWindow: 128000,
@@ -7132,8 +7236,8 @@ export const MODELS = {
7132
7236
  input: ["text"],
7133
7237
  cost: {
7134
7238
  input: 0.25,
7135
- output: 1,
7136
- cacheRead: 0,
7239
+ output: 0.75,
7240
+ cacheRead: 0.024999999999999998,
7137
7241
  cacheWrite: 0,
7138
7242
  },
7139
7243
  contextWindow: 128000,
@@ -7171,7 +7275,7 @@ export const MODELS = {
7171
7275
  cacheWrite: 0,
7172
7276
  },
7173
7277
  contextWindow: 131072,
7174
- maxTokens: 32768,
7278
+ maxTokens: 131072,
7175
7279
  },
7176
7280
  "meta-llama/llama-3-8b-instruct": {
7177
7281
  id: "meta-llama/llama-3-8b-instruct",
@@ -7341,7 +7445,7 @@ export const MODELS = {
7341
7445
  cacheWrite: 0,
7342
7446
  },
7343
7447
  contextWindow: 196608,
7344
- maxTokens: 65536,
7448
+ maxTokens: 196608,
7345
7449
  },
7346
7450
  "minimax/minimax-m2.1": {
7347
7451
  id: "minimax/minimax-m2.1",
@@ -8584,6 +8688,23 @@ export const MODELS = {
8584
8688
  contextWindow: 400000,
8585
8689
  maxTokens: 128000,
8586
8690
  },
8691
+ "openai/gpt-5.3-chat": {
8692
+ id: "openai/gpt-5.3-chat",
8693
+ name: "OpenAI: GPT-5.3 Chat",
8694
+ api: "openai-completions",
8695
+ provider: "openrouter",
8696
+ baseUrl: "https://openrouter.ai/api/v1",
8697
+ reasoning: false,
8698
+ input: ["text", "image"],
8699
+ cost: {
8700
+ input: 1.75,
8701
+ output: 14,
8702
+ cacheRead: 0.175,
8703
+ cacheWrite: 0,
8704
+ },
8705
+ contextWindow: 128000,
8706
+ maxTokens: 16384,
8707
+ },
8587
8708
  "openai/gpt-5.3-codex": {
8588
8709
  id: "openai/gpt-5.3-codex",
8589
8710
  name: "OpenAI: GPT-5.3-Codex",
@@ -8950,9 +9071,9 @@ export const MODELS = {
8950
9071
  reasoning: false,
8951
9072
  input: ["text"],
8952
9073
  cost: {
8953
- input: 1.5999999999999999,
8954
- output: 6.3999999999999995,
8955
- cacheRead: 0.32,
9074
+ input: 1.04,
9075
+ output: 4.16,
9076
+ cacheRead: 0.20800000000000002,
8956
9077
  cacheWrite: 0,
8957
9078
  },
8958
9079
  contextWindow: 32768,
@@ -8984,8 +9105,8 @@ export const MODELS = {
8984
9105
  reasoning: false,
8985
9106
  input: ["text"],
8986
9107
  cost: {
8987
- input: 0.39999999999999997,
8988
- output: 1.2,
9108
+ input: 0.26,
9109
+ output: 0.78,
8989
9110
  cacheRead: 0,
8990
9111
  cacheWrite: 0,
8991
9112
  },
@@ -9001,8 +9122,8 @@ export const MODELS = {
9001
9122
  reasoning: true,
9002
9123
  input: ["text"],
9003
9124
  cost: {
9004
- input: 0.39999999999999997,
9005
- output: 1.2,
9125
+ input: 0.26,
9126
+ output: 0.78,
9006
9127
  cacheRead: 0,
9007
9128
  cacheWrite: 0,
9008
9129
  },
@@ -9018,9 +9139,9 @@ export const MODELS = {
9018
9139
  reasoning: false,
9019
9140
  input: ["text"],
9020
9141
  cost: {
9021
- input: 0.049999999999999996,
9022
- output: 0.19999999999999998,
9023
- cacheRead: 0.01,
9142
+ input: 0.0325,
9143
+ output: 0.13,
9144
+ cacheRead: 0.006500000000000001,
9024
9145
  cacheWrite: 0,
9025
9146
  },
9026
9147
  contextWindow: 131072,
@@ -9256,9 +9377,9 @@ export const MODELS = {
9256
9377
  reasoning: false,
9257
9378
  input: ["text"],
9258
9379
  cost: {
9259
- input: 0.3,
9260
- output: 1.5,
9261
- cacheRead: 0.06,
9380
+ input: 0.195,
9381
+ output: 0.975,
9382
+ cacheRead: 0.039,
9262
9383
  cacheWrite: 0,
9263
9384
  },
9264
9385
  contextWindow: 1000000,
@@ -9290,9 +9411,9 @@ export const MODELS = {
9290
9411
  reasoning: false,
9291
9412
  input: ["text"],
9292
9413
  cost: {
9293
- input: 1,
9294
- output: 5,
9295
- cacheRead: 0.19999999999999998,
9414
+ input: 0.65,
9415
+ output: 3.25,
9416
+ cacheRead: 0.13,
9296
9417
  cacheWrite: 0,
9297
9418
  },
9298
9419
  contextWindow: 1000000,
@@ -9358,8 +9479,8 @@ export const MODELS = {
9358
9479
  reasoning: true,
9359
9480
  input: ["text"],
9360
9481
  cost: {
9361
- input: 1.2,
9362
- output: 6,
9482
+ input: 0.78,
9483
+ output: 3.9,
9363
9484
  cacheRead: 0,
9364
9485
  cacheWrite: 0,
9365
9486
  },
@@ -9545,8 +9666,8 @@ export const MODELS = {
9545
9666
  reasoning: true,
9546
9667
  input: ["text", "image"],
9547
9668
  cost: {
9548
- input: 0.39999999999999997,
9549
- output: 3.1999999999999997,
9669
+ input: 0.26,
9670
+ output: 2.08,
9550
9671
  cacheRead: 0,
9551
9672
  cacheWrite: 0,
9552
9673
  },
@@ -9562,8 +9683,8 @@ export const MODELS = {
9562
9683
  reasoning: true,
9563
9684
  input: ["text", "image"],
9564
9685
  cost: {
9565
- input: 0.3,
9566
- output: 2.4,
9686
+ input: 0.195,
9687
+ output: 1.56,
9567
9688
  cacheRead: 0,
9568
9689
  cacheWrite: 0,
9569
9690
  },
@@ -9579,8 +9700,8 @@ export const MODELS = {
9579
9700
  reasoning: true,
9580
9701
  input: ["text", "image"],
9581
9702
  cost: {
9582
- input: 0.25,
9583
- output: 2,
9703
+ input: 0.1625,
9704
+ output: 1.3,
9584
9705
  cacheRead: 0,
9585
9706
  cacheWrite: 0,
9586
9707
  },
@@ -9596,9 +9717,9 @@ export const MODELS = {
9596
9717
  reasoning: true,
9597
9718
  input: ["text", "image"],
9598
9719
  cost: {
9599
- input: 0.55,
9600
- output: 3.5,
9601
- cacheRead: 0.55,
9720
+ input: 0.39,
9721
+ output: 2.34,
9722
+ cacheRead: 0,
9602
9723
  cacheWrite: 0,
9603
9724
  },
9604
9725
  contextWindow: 262144,
@@ -9630,8 +9751,8 @@ export const MODELS = {
9630
9751
  reasoning: true,
9631
9752
  input: ["text", "image"],
9632
9753
  cost: {
9633
- input: 0.39999999999999997,
9634
- output: 2.4,
9754
+ input: 0.26,
9755
+ output: 1.56,
9635
9756
  cacheRead: 0,
9636
9757
  cacheWrite: 0,
9637
9758
  },
@@ -9791,18 +9912,18 @@ export const MODELS = {
9791
9912
  contextWindow: 163840,
9792
9913
  maxTokens: 163840,
9793
9914
  },
9794
- "upstage/solar-pro-3:free": {
9795
- id: "upstage/solar-pro-3:free",
9796
- name: "Upstage: Solar Pro 3 (free)",
9915
+ "upstage/solar-pro-3": {
9916
+ id: "upstage/solar-pro-3",
9917
+ name: "Upstage: Solar Pro 3",
9797
9918
  api: "openai-completions",
9798
9919
  provider: "openrouter",
9799
9920
  baseUrl: "https://openrouter.ai/api/v1",
9800
9921
  reasoning: true,
9801
9922
  input: ["text"],
9802
9923
  cost: {
9803
- input: 0,
9804
- output: 0,
9805
- cacheRead: 0,
9924
+ input: 0.15,
9925
+ output: 0.6,
9926
+ cacheRead: 0.015,
9806
9927
  cacheWrite: 0,
9807
9928
  },
9808
9929
  contextWindow: 128000,
@@ -10055,13 +10176,13 @@ export const MODELS = {
10055
10176
  reasoning: true,
10056
10177
  input: ["text"],
10057
10178
  cost: {
10058
- input: 0.35,
10059
- output: 1.71,
10179
+ input: 0.39,
10180
+ output: 1.9,
10060
10181
  cacheRead: 0,
10061
10182
  cacheWrite: 0,
10062
10183
  },
10063
- contextWindow: 202752,
10064
- maxTokens: 131072,
10184
+ contextWindow: 204800,
10185
+ maxTokens: 204800,
10065
10186
  },
10066
10187
  "z-ai/glm-4.6:exacto": {
10067
10188
  id: "z-ai/glm-4.6:exacto",
@@ -10140,13 +10261,13 @@ export const MODELS = {
10140
10261
  reasoning: true,
10141
10262
  input: ["text"],
10142
10263
  cost: {
10143
- input: 0.95,
10144
- output: 2.5500000000000003,
10145
- cacheRead: 0.19999999999999998,
10264
+ input: 0.7999999999999999,
10265
+ output: 2.56,
10266
+ cacheRead: 0.16,
10146
10267
  cacheWrite: 0,
10147
10268
  },
10148
- contextWindow: 204800,
10149
- maxTokens: 131072,
10269
+ contextWindow: 202752,
10270
+ maxTokens: 4096,
10150
10271
  },
10151
10272
  },
10152
10273
  "vercel-ai-gateway": {
@@ -10864,6 +10985,23 @@ export const MODELS = {
10864
10985
  contextWindow: 1000000,
10865
10986
  maxTokens: 64000,
10866
10987
  },
10988
+ "google/gemini-3.1-flash-lite-preview": {
10989
+ id: "google/gemini-3.1-flash-lite-preview",
10990
+ name: "Gemini 3.1 Flash Lite Preview",
10991
+ api: "anthropic-messages",
10992
+ provider: "vercel-ai-gateway",
10993
+ baseUrl: "https://ai-gateway.vercel.sh",
10994
+ reasoning: true,
10995
+ input: ["text", "image"],
10996
+ cost: {
10997
+ input: 0.25,
10998
+ output: 1.5,
10999
+ cacheRead: 0,
11000
+ cacheWrite: 0,
11001
+ },
11002
+ contextWindow: 1000000,
11003
+ maxTokens: 65000,
11004
+ },
10867
11005
  "google/gemini-3.1-pro-preview": {
10868
11006
  id: "google/gemini-3.1-pro-preview",
10869
11007
  name: "Gemini 3.1 Pro Preview",
@@ -10881,6 +11019,23 @@ export const MODELS = {
10881
11019
  contextWindow: 1000000,
10882
11020
  maxTokens: 64000,
10883
11021
  },
11022
+ "inception/mercury-2": {
11023
+ id: "inception/mercury-2",
11024
+ name: "Mercury 2",
11025
+ api: "anthropic-messages",
11026
+ provider: "vercel-ai-gateway",
11027
+ baseUrl: "https://ai-gateway.vercel.sh",
11028
+ reasoning: true,
11029
+ input: ["text"],
11030
+ cost: {
11031
+ input: 0.25,
11032
+ output: 0.75,
11033
+ cacheRead: 0.024999999999999998,
11034
+ cacheWrite: 0,
11035
+ },
11036
+ contextWindow: 128000,
11037
+ maxTokens: 128000,
11038
+ },
10884
11039
  "inception/mercury-coder-small": {
10885
11040
  id: "inception/mercury-coder-small",
10886
11041
  name: "Mercury Coder Small Beta",
@@ -11782,6 +11937,23 @@ export const MODELS = {
11782
11937
  contextWindow: 400000,
11783
11938
  maxTokens: 128000,
11784
11939
  },
11940
+ "openai/gpt-5.3-chat": {
11941
+ id: "openai/gpt-5.3-chat",
11942
+ name: "GPT-5.3 Chat",
11943
+ api: "anthropic-messages",
11944
+ provider: "vercel-ai-gateway",
11945
+ baseUrl: "https://ai-gateway.vercel.sh",
11946
+ reasoning: true,
11947
+ input: ["text", "image"],
11948
+ cost: {
11949
+ input: 1.75,
11950
+ output: 14,
11951
+ cacheRead: 0.175,
11952
+ cacheWrite: 0,
11953
+ },
11954
+ contextWindow: 128000,
11955
+ maxTokens: 16384,
11956
+ },
11785
11957
  "openai/gpt-5.3-codex": {
11786
11958
  id: "openai/gpt-5.3-codex",
11787
11959
  name: "GPT 5.3 Codex",