@draht/ai 2026.3.2 → 2026.3.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -6867,6 +6867,23 @@ export const MODELS = {
6867
6867
  contextWindow: 163840,
6868
6868
  maxTokens: 65536,
6869
6869
  },
6870
+ "essentialai/rnj-1-instruct": {
6871
+ id: "essentialai/rnj-1-instruct",
6872
+ name: "EssentialAI: Rnj 1 Instruct",
6873
+ api: "openai-completions",
6874
+ provider: "openrouter",
6875
+ baseUrl: "https://openrouter.ai/api/v1",
6876
+ reasoning: false,
6877
+ input: ["text"],
6878
+ cost: {
6879
+ input: 0.15,
6880
+ output: 0.15,
6881
+ cacheRead: 0,
6882
+ cacheWrite: 0,
6883
+ },
6884
+ contextWindow: 32768,
6885
+ maxTokens: 4096,
6886
+ },
6870
6887
  "google/gemini-2.0-flash-001": {
6871
6888
  id: "google/gemini-2.0-flash-001",
6872
6889
  name: "Google: Gemini 2.0 Flash",
@@ -6950,7 +6967,7 @@ export const MODELS = {
6950
6967
  cacheWrite: 0.08333333333333334,
6951
6968
  },
6952
6969
  contextWindow: 1048576,
6953
- maxTokens: 65535,
6970
+ maxTokens: 65536,
6954
6971
  },
6955
6972
  "google/gemini-2.5-pro": {
6956
6973
  id: "google/gemini-2.5-pro",
@@ -7018,7 +7035,7 @@ export const MODELS = {
7018
7035
  cacheWrite: 0.08333333333333334,
7019
7036
  },
7020
7037
  contextWindow: 1048576,
7021
- maxTokens: 65535,
7038
+ maxTokens: 65536,
7022
7039
  },
7023
7040
  "google/gemini-3-pro-preview": {
7024
7041
  id: "google/gemini-3-pro-preview",
@@ -7037,6 +7054,23 @@ export const MODELS = {
7037
7054
  contextWindow: 1048576,
7038
7055
  maxTokens: 65536,
7039
7056
  },
7057
+ "google/gemini-3.1-flash-lite-preview": {
7058
+ id: "google/gemini-3.1-flash-lite-preview",
7059
+ name: "Google: Gemini 3.1 Flash Lite Preview",
7060
+ api: "openai-completions",
7061
+ provider: "openrouter",
7062
+ baseUrl: "https://openrouter.ai/api/v1",
7063
+ reasoning: true,
7064
+ input: ["text", "image"],
7065
+ cost: {
7066
+ input: 0.25,
7067
+ output: 1.5,
7068
+ cacheRead: 0.024999999999999998,
7069
+ cacheWrite: 0.08333333333333334,
7070
+ },
7071
+ contextWindow: 1048576,
7072
+ maxTokens: 65536,
7073
+ },
7040
7074
  "google/gemini-3.1-pro-preview": {
7041
7075
  id: "google/gemini-3.1-pro-preview",
7042
7076
  name: "Google: Gemini 3.1 Pro Preview",
@@ -7171,7 +7205,7 @@ export const MODELS = {
7171
7205
  cacheWrite: 0,
7172
7206
  },
7173
7207
  contextWindow: 131072,
7174
- maxTokens: 32768,
7208
+ maxTokens: 131072,
7175
7209
  },
7176
7210
  "meta-llama/llama-3-8b-instruct": {
7177
7211
  id: "meta-llama/llama-3-8b-instruct",
@@ -7341,7 +7375,7 @@ export const MODELS = {
7341
7375
  cacheWrite: 0,
7342
7376
  },
7343
7377
  contextWindow: 196608,
7344
- maxTokens: 65536,
7378
+ maxTokens: 196608,
7345
7379
  },
7346
7380
  "minimax/minimax-m2.1": {
7347
7381
  id: "minimax/minimax-m2.1",
@@ -9545,9 +9579,9 @@ export const MODELS = {
9545
9579
  reasoning: true,
9546
9580
  input: ["text", "image"],
9547
9581
  cost: {
9548
- input: 0.39999999999999997,
9549
- output: 3.1999999999999997,
9550
- cacheRead: 0,
9582
+ input: 0.3,
9583
+ output: 2.4,
9584
+ cacheRead: 0.3,
9551
9585
  cacheWrite: 0,
9552
9586
  },
9553
9587
  contextWindow: 262144,
@@ -9562,9 +9596,9 @@ export const MODELS = {
9562
9596
  reasoning: true,
9563
9597
  input: ["text", "image"],
9564
9598
  cost: {
9565
- input: 0.3,
9566
- output: 2.4,
9567
- cacheRead: 0,
9599
+ input: 0.27,
9600
+ output: 2.16,
9601
+ cacheRead: 0.27,
9568
9602
  cacheWrite: 0,
9569
9603
  },
9570
9604
  contextWindow: 262144,
@@ -9579,9 +9613,9 @@ export const MODELS = {
9579
9613
  reasoning: true,
9580
9614
  input: ["text", "image"],
9581
9615
  cost: {
9582
- input: 0.25,
9583
- output: 2,
9584
- cacheRead: 0,
9616
+ input: 0.22499999999999998,
9617
+ output: 1.7999999999999998,
9618
+ cacheRead: 0.22499999999999998,
9585
9619
  cacheWrite: 0,
9586
9620
  },
9587
9621
  contextWindow: 262144,
@@ -9791,18 +9825,18 @@ export const MODELS = {
9791
9825
  contextWindow: 163840,
9792
9826
  maxTokens: 163840,
9793
9827
  },
9794
- "upstage/solar-pro-3:free": {
9795
- id: "upstage/solar-pro-3:free",
9796
- name: "Upstage: Solar Pro 3 (free)",
9828
+ "upstage/solar-pro-3": {
9829
+ id: "upstage/solar-pro-3",
9830
+ name: "Upstage: Solar Pro 3",
9797
9831
  api: "openai-completions",
9798
9832
  provider: "openrouter",
9799
9833
  baseUrl: "https://openrouter.ai/api/v1",
9800
9834
  reasoning: true,
9801
9835
  input: ["text"],
9802
9836
  cost: {
9803
- input: 0,
9804
- output: 0,
9805
- cacheRead: 0,
9837
+ input: 0.15,
9838
+ output: 0.6,
9839
+ cacheRead: 0.015,
9806
9840
  cacheWrite: 0,
9807
9841
  },
9808
9842
  contextWindow: 128000,
@@ -10140,13 +10174,13 @@ export const MODELS = {
10140
10174
  reasoning: true,
10141
10175
  input: ["text"],
10142
10176
  cost: {
10143
- input: 0.95,
10144
- output: 2.5500000000000003,
10145
- cacheRead: 0.19999999999999998,
10177
+ input: 0.7999999999999999,
10178
+ output: 2.56,
10179
+ cacheRead: 0.16,
10146
10180
  cacheWrite: 0,
10147
10181
  },
10148
- contextWindow: 204800,
10149
- maxTokens: 131072,
10182
+ contextWindow: 202752,
10183
+ maxTokens: 4096,
10150
10184
  },
10151
10185
  },
10152
10186
  "vercel-ai-gateway": {
@@ -10864,6 +10898,23 @@ export const MODELS = {
10864
10898
  contextWindow: 1000000,
10865
10899
  maxTokens: 64000,
10866
10900
  },
10901
+ "google/gemini-3.1-flash-lite-preview": {
10902
+ id: "google/gemini-3.1-flash-lite-preview",
10903
+ name: "Gemini 3.1 Flash Lite Preview",
10904
+ api: "anthropic-messages",
10905
+ provider: "vercel-ai-gateway",
10906
+ baseUrl: "https://ai-gateway.vercel.sh",
10907
+ reasoning: true,
10908
+ input: ["text", "image"],
10909
+ cost: {
10910
+ input: 0.25,
10911
+ output: 1.5,
10912
+ cacheRead: 0,
10913
+ cacheWrite: 0,
10914
+ },
10915
+ contextWindow: 1000000,
10916
+ maxTokens: 65000,
10917
+ },
10867
10918
  "google/gemini-3.1-pro-preview": {
10868
10919
  id: "google/gemini-3.1-pro-preview",
10869
10920
  name: "Gemini 3.1 Pro Preview",
@@ -11782,6 +11833,23 @@ export const MODELS = {
11782
11833
  contextWindow: 400000,
11783
11834
  maxTokens: 128000,
11784
11835
  },
11836
+ "openai/gpt-5.3-chat": {
11837
+ id: "openai/gpt-5.3-chat",
11838
+ name: "GPT-5.3 Chat",
11839
+ api: "anthropic-messages",
11840
+ provider: "vercel-ai-gateway",
11841
+ baseUrl: "https://ai-gateway.vercel.sh",
11842
+ reasoning: true,
11843
+ input: ["text", "image"],
11844
+ cost: {
11845
+ input: 1.75,
11846
+ output: 14,
11847
+ cacheRead: 0.175,
11848
+ cacheWrite: 0,
11849
+ },
11850
+ contextWindow: 128000,
11851
+ maxTokens: 16384,
11852
+ },
11785
11853
  "openai/gpt-5.3-codex": {
11786
11854
  id: "openai/gpt-5.3-codex",
11787
11855
  name: "GPT 5.3 Codex",