@mariozechner/pi-ai 0.27.2 → 0.27.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2871,6 +2871,40 @@ export const MODELS = {
2871
2871
  },
2872
2872
  },
2873
2873
  "openrouter": {
2874
+ "bytedance-seed/seed-1.6-flash": {
2875
+ id: "bytedance-seed/seed-1.6-flash",
2876
+ name: "ByteDance Seed: Seed 1.6 Flash",
2877
+ api: "openai-completions",
2878
+ provider: "openrouter",
2879
+ baseUrl: "https://openrouter.ai/api/v1",
2880
+ reasoning: true,
2881
+ input: ["text", "image"],
2882
+ cost: {
2883
+ input: 0.075,
2884
+ output: 0.3,
2885
+ cacheRead: 0,
2886
+ cacheWrite: 0,
2887
+ },
2888
+ contextWindow: 262144,
2889
+ maxTokens: 16384,
2890
+ },
2891
+ "bytedance-seed/seed-1.6": {
2892
+ id: "bytedance-seed/seed-1.6",
2893
+ name: "ByteDance Seed: Seed 1.6",
2894
+ api: "openai-completions",
2895
+ provider: "openrouter",
2896
+ baseUrl: "https://openrouter.ai/api/v1",
2897
+ reasoning: true,
2898
+ input: ["text", "image"],
2899
+ cost: {
2900
+ input: 0.25,
2901
+ output: 2,
2902
+ cacheRead: 0,
2903
+ cacheWrite: 0,
2904
+ },
2905
+ contextWindow: 262144,
2906
+ maxTokens: 32768,
2907
+ },
2874
2908
  "minimax/minimax-m2.1": {
2875
2909
  id: "minimax/minimax-m2.1",
2876
2910
  name: "MiniMax: MiniMax M2.1",
@@ -2897,13 +2931,13 @@ export const MODELS = {
2897
2931
  reasoning: true,
2898
2932
  input: ["text"],
2899
2933
  cost: {
2900
- input: 0.44,
2901
- output: 1.74,
2934
+ input: 0.39999999999999997,
2935
+ output: 1.5,
2902
2936
  cacheRead: 0,
2903
2937
  cacheWrite: 0,
2904
2938
  },
2905
2939
  contextWindow: 202752,
2906
- maxTokens: 131072,
2940
+ maxTokens: 65535,
2907
2941
  },
2908
2942
  "google/gemini-3-flash-preview": {
2909
2943
  id: "google/gemini-3-flash-preview",
@@ -3271,13 +3305,13 @@ export const MODELS = {
3271
3305
  reasoning: true,
3272
3306
  input: ["text"],
3273
3307
  cost: {
3274
- input: 0.26,
3275
- output: 0.38,
3308
+ input: 0.224,
3309
+ output: 0.32,
3276
3310
  cacheRead: 0,
3277
3311
  cacheWrite: 0,
3278
3312
  },
3279
3313
  contextWindow: 163840,
3280
- maxTokens: 65536,
3314
+ maxTokens: 4096,
3281
3315
  },
3282
3316
  "prime-intellect/intellect-3": {
3283
3317
  id: "prime-intellect/intellect-3",
@@ -4347,7 +4381,7 @@ export const MODELS = {
4347
4381
  cacheRead: 0,
4348
4382
  cacheWrite: 0,
4349
4383
  },
4350
- contextWindow: 8192,
4384
+ contextWindow: 32768,
4351
4385
  maxTokens: 7168,
4352
4386
  },
4353
4387
  "openai/gpt-4o-audio-preview": {
@@ -6067,9 +6101,9 @@ export const MODELS = {
6067
6101
  contextWindow: 32768,
6068
6102
  maxTokens: 4096,
6069
6103
  },
6070
- "anthropic/claude-3.5-haiku": {
6071
- id: "anthropic/claude-3.5-haiku",
6072
- name: "Anthropic: Claude 3.5 Haiku",
6104
+ "anthropic/claude-3.5-haiku-20241022": {
6105
+ id: "anthropic/claude-3.5-haiku-20241022",
6106
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
6073
6107
  api: "openai-completions",
6074
6108
  provider: "openrouter",
6075
6109
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6084,9 +6118,9 @@ export const MODELS = {
6084
6118
  contextWindow: 200000,
6085
6119
  maxTokens: 8192,
6086
6120
  },
6087
- "anthropic/claude-3.5-haiku-20241022": {
6088
- id: "anthropic/claude-3.5-haiku-20241022",
6089
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
6121
+ "anthropic/claude-3.5-haiku": {
6122
+ id: "anthropic/claude-3.5-haiku",
6123
+ name: "Anthropic: Claude 3.5 Haiku",
6090
6124
  api: "openai-completions",
6091
6125
  provider: "openrouter",
6092
6126
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6322,21 +6356,21 @@ export const MODELS = {
6322
6356
  contextWindow: 128000,
6323
6357
  maxTokens: 16384,
6324
6358
  },
6325
- "meta-llama/llama-3.1-405b-instruct": {
6326
- id: "meta-llama/llama-3.1-405b-instruct",
6327
- name: "Meta: Llama 3.1 405B Instruct",
6359
+ "meta-llama/llama-3.1-70b-instruct": {
6360
+ id: "meta-llama/llama-3.1-70b-instruct",
6361
+ name: "Meta: Llama 3.1 70B Instruct",
6328
6362
  api: "openai-completions",
6329
6363
  provider: "openrouter",
6330
6364
  baseUrl: "https://openrouter.ai/api/v1",
6331
6365
  reasoning: false,
6332
6366
  input: ["text"],
6333
6367
  cost: {
6334
- input: 3.5,
6335
- output: 3.5,
6368
+ input: 0.39999999999999997,
6369
+ output: 0.39999999999999997,
6336
6370
  cacheRead: 0,
6337
6371
  cacheWrite: 0,
6338
6372
  },
6339
- contextWindow: 130815,
6373
+ contextWindow: 131072,
6340
6374
  maxTokens: 4096,
6341
6375
  },
6342
6376
  "meta-llama/llama-3.1-8b-instruct": {
@@ -6356,21 +6390,21 @@ export const MODELS = {
6356
6390
  contextWindow: 131072,
6357
6391
  maxTokens: 16384,
6358
6392
  },
6359
- "meta-llama/llama-3.1-70b-instruct": {
6360
- id: "meta-llama/llama-3.1-70b-instruct",
6361
- name: "Meta: Llama 3.1 70B Instruct",
6393
+ "meta-llama/llama-3.1-405b-instruct": {
6394
+ id: "meta-llama/llama-3.1-405b-instruct",
6395
+ name: "Meta: Llama 3.1 405B Instruct",
6362
6396
  api: "openai-completions",
6363
6397
  provider: "openrouter",
6364
6398
  baseUrl: "https://openrouter.ai/api/v1",
6365
6399
  reasoning: false,
6366
6400
  input: ["text"],
6367
6401
  cost: {
6368
- input: 0.39999999999999997,
6369
- output: 0.39999999999999997,
6402
+ input: 3.5,
6403
+ output: 3.5,
6370
6404
  cacheRead: 0,
6371
6405
  cacheWrite: 0,
6372
6406
  },
6373
- contextWindow: 131072,
6407
+ contextWindow: 10000,
6374
6408
  maxTokens: 4096,
6375
6409
  },
6376
6410
  "mistralai/mistral-nemo": {
@@ -6509,23 +6543,6 @@ export const MODELS = {
6509
6543
  contextWindow: 128000,
6510
6544
  maxTokens: 4096,
6511
6545
  },
6512
- "openai/gpt-4o-2024-05-13": {
6513
- id: "openai/gpt-4o-2024-05-13",
6514
- name: "OpenAI: GPT-4o (2024-05-13)",
6515
- api: "openai-completions",
6516
- provider: "openrouter",
6517
- baseUrl: "https://openrouter.ai/api/v1",
6518
- reasoning: false,
6519
- input: ["text", "image"],
6520
- cost: {
6521
- input: 5,
6522
- output: 15,
6523
- cacheRead: 0,
6524
- cacheWrite: 0,
6525
- },
6526
- contextWindow: 128000,
6527
- maxTokens: 4096,
6528
- },
6529
6546
  "openai/gpt-4o": {
6530
6547
  id: "openai/gpt-4o",
6531
6548
  name: "OpenAI: GPT-4o",
@@ -6560,6 +6577,23 @@ export const MODELS = {
6560
6577
  contextWindow: 128000,
6561
6578
  maxTokens: 64000,
6562
6579
  },
6580
+ "openai/gpt-4o-2024-05-13": {
6581
+ id: "openai/gpt-4o-2024-05-13",
6582
+ name: "OpenAI: GPT-4o (2024-05-13)",
6583
+ api: "openai-completions",
6584
+ provider: "openrouter",
6585
+ baseUrl: "https://openrouter.ai/api/v1",
6586
+ reasoning: false,
6587
+ input: ["text", "image"],
6588
+ cost: {
6589
+ input: 5,
6590
+ output: 15,
6591
+ cacheRead: 0,
6592
+ cacheWrite: 0,
6593
+ },
6594
+ contextWindow: 128000,
6595
+ maxTokens: 4096,
6596
+ },
6563
6597
  "meta-llama/llama-3-70b-instruct": {
6564
6598
  id: "meta-llama/llama-3-70b-instruct",
6565
6599
  name: "Meta: Llama 3 70B Instruct",
@@ -6679,38 +6713,38 @@ export const MODELS = {
6679
6713
  contextWindow: 128000,
6680
6714
  maxTokens: 4096,
6681
6715
  },
6682
- "openai/gpt-3.5-turbo-0613": {
6683
- id: "openai/gpt-3.5-turbo-0613",
6684
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6716
+ "openai/gpt-4-turbo-preview": {
6717
+ id: "openai/gpt-4-turbo-preview",
6718
+ name: "OpenAI: GPT-4 Turbo Preview",
6685
6719
  api: "openai-completions",
6686
6720
  provider: "openrouter",
6687
6721
  baseUrl: "https://openrouter.ai/api/v1",
6688
6722
  reasoning: false,
6689
6723
  input: ["text"],
6690
6724
  cost: {
6691
- input: 1,
6692
- output: 2,
6725
+ input: 10,
6726
+ output: 30,
6693
6727
  cacheRead: 0,
6694
6728
  cacheWrite: 0,
6695
6729
  },
6696
- contextWindow: 4095,
6730
+ contextWindow: 128000,
6697
6731
  maxTokens: 4096,
6698
6732
  },
6699
- "openai/gpt-4-turbo-preview": {
6700
- id: "openai/gpt-4-turbo-preview",
6701
- name: "OpenAI: GPT-4 Turbo Preview",
6733
+ "openai/gpt-3.5-turbo-0613": {
6734
+ id: "openai/gpt-3.5-turbo-0613",
6735
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6702
6736
  api: "openai-completions",
6703
6737
  provider: "openrouter",
6704
6738
  baseUrl: "https://openrouter.ai/api/v1",
6705
6739
  reasoning: false,
6706
6740
  input: ["text"],
6707
6741
  cost: {
6708
- input: 10,
6709
- output: 30,
6742
+ input: 1,
6743
+ output: 2,
6710
6744
  cacheRead: 0,
6711
6745
  cacheWrite: 0,
6712
6746
  },
6713
- contextWindow: 128000,
6747
+ contextWindow: 4095,
6714
6748
  maxTokens: 4096,
6715
6749
  },
6716
6750
  "mistralai/mistral-tiny": {
@@ -6781,9 +6815,9 @@ export const MODELS = {
6781
6815
  contextWindow: 16385,
6782
6816
  maxTokens: 4096,
6783
6817
  },
6784
- "openai/gpt-4": {
6785
- id: "openai/gpt-4",
6786
- name: "OpenAI: GPT-4",
6818
+ "openai/gpt-4-0314": {
6819
+ id: "openai/gpt-4-0314",
6820
+ name: "OpenAI: GPT-4 (older v0314)",
6787
6821
  api: "openai-completions",
6788
6822
  provider: "openrouter",
6789
6823
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6798,9 +6832,9 @@ export const MODELS = {
6798
6832
  contextWindow: 8191,
6799
6833
  maxTokens: 4096,
6800
6834
  },
6801
- "openai/gpt-4-0314": {
6802
- id: "openai/gpt-4-0314",
6803
- name: "OpenAI: GPT-4 (older v0314)",
6835
+ "openai/gpt-4": {
6836
+ id: "openai/gpt-4",
6837
+ name: "OpenAI: GPT-4",
6804
6838
  api: "openai-completions",
6805
6839
  provider: "openrouter",
6806
6840
  baseUrl: "https://openrouter.ai/api/v1",