@lobehub/chat 1.51.7 → 1.51.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (43) hide show
  1. package/CHANGELOG.md +25 -0
  2. package/README.ja-JP.md +8 -8
  3. package/README.md +8 -8
  4. package/README.zh-CN.md +8 -8
  5. package/changelog/v1.json +9 -0
  6. package/package.json +1 -1
  7. package/src/app/(backend)/webapi/chat/models/[provider]/route.ts +1 -1
  8. package/src/libs/agent-runtime/ai360/index.ts +8 -1
  9. package/src/libs/agent-runtime/anthropic/index.ts +2 -1
  10. package/src/libs/agent-runtime/baichuan/index.ts +1 -1
  11. package/src/libs/agent-runtime/cloudflare/index.test.ts +0 -117
  12. package/src/libs/agent-runtime/cloudflare/index.ts +32 -11
  13. package/src/libs/agent-runtime/deepseek/index.ts +4 -1
  14. package/src/libs/agent-runtime/fireworksai/index.ts +8 -1
  15. package/src/libs/agent-runtime/giteeai/index.ts +9 -1
  16. package/src/libs/agent-runtime/github/index.test.ts +5 -16
  17. package/src/libs/agent-runtime/github/index.ts +31 -33
  18. package/src/libs/agent-runtime/google/index.ts +2 -1
  19. package/src/libs/agent-runtime/groq/index.ts +7 -1
  20. package/src/libs/agent-runtime/higress/index.ts +2 -1
  21. package/src/libs/agent-runtime/huggingface/index.ts +10 -1
  22. package/src/libs/agent-runtime/hunyuan/index.ts +3 -1
  23. package/src/libs/agent-runtime/internlm/index.ts +3 -1
  24. package/src/libs/agent-runtime/mistral/index.ts +2 -1
  25. package/src/libs/agent-runtime/moonshot/index.ts +3 -1
  26. package/src/libs/agent-runtime/novita/__snapshots__/index.test.ts.snap +48 -12
  27. package/src/libs/agent-runtime/novita/index.ts +9 -1
  28. package/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap +70 -66
  29. package/src/libs/agent-runtime/openai/index.ts +37 -0
  30. package/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap +172 -4
  31. package/src/libs/agent-runtime/openrouter/index.ts +17 -2
  32. package/src/libs/agent-runtime/qwen/index.ts +10 -1
  33. package/src/libs/agent-runtime/sensenova/index.ts +3 -1
  34. package/src/libs/agent-runtime/siliconcloud/index.ts +10 -1
  35. package/src/libs/agent-runtime/stepfun/index.ts +3 -1
  36. package/src/libs/agent-runtime/togetherai/__snapshots__/index.test.ts.snap +1309 -5
  37. package/src/libs/agent-runtime/togetherai/index.test.ts +0 -13
  38. package/src/libs/agent-runtime/togetherai/index.ts +25 -20
  39. package/src/libs/agent-runtime/utils/cloudflareHelpers.test.ts +0 -99
  40. package/src/libs/agent-runtime/utils/cloudflareHelpers.ts +0 -70
  41. package/src/libs/agent-runtime/xai/index.ts +3 -1
  42. package/src/libs/agent-runtime/zeroone/index.ts +3 -1
  43. package/src/libs/agent-runtime/zhipu/index.ts +3 -1
@@ -14,6 +14,7 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co
14
14
  "functionCall": false,
15
15
  "id": "cohere/command-r-03-2024",
16
16
  "maxTokens": 4000,
17
+ "reasoning": false,
17
18
  "vision": false,
18
19
  },
19
20
  {
@@ -28,6 +29,7 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co
28
29
  "functionCall": false,
29
30
  "id": "cohere/command-r-plus-04-2024",
30
31
  "maxTokens": 4000,
32
+ "reasoning": false,
31
33
  "vision": false,
32
34
  },
33
35
  {
@@ -42,6 +44,7 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co
42
44
  "functionCall": false,
43
45
  "id": "cohere/command-r-plus-08-2024",
44
46
  "maxTokens": 4000,
47
+ "reasoning": false,
45
48
  "vision": false,
46
49
  },
47
50
  {
@@ -56,6 +59,7 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co
56
59
  "functionCall": false,
57
60
  "id": "cohere/command-r-08-2024",
58
61
  "maxTokens": 4000,
62
+ "reasoning": false,
59
63
  "vision": false,
60
64
  },
61
65
  {
@@ -72,6 +76,7 @@ Note: This model is experimental and not suited for production use-cases. It may
72
76
  "functionCall": false,
73
77
  "id": "google/gemini-flash-8b-1.5-exp",
74
78
  "maxTokens": 32768,
79
+ "reasoning": false,
75
80
  "vision": true,
76
81
  },
77
82
  {
@@ -88,6 +93,7 @@ Note: This model is experimental and not suited for production use-cases. It may
88
93
  "functionCall": false,
89
94
  "id": "google/gemini-flash-1.5-exp",
90
95
  "maxTokens": 32768,
96
+ "reasoning": false,
91
97
  "vision": true,
92
98
  },
93
99
  {
@@ -98,6 +104,7 @@ Note: This model is experimental and not suited for production use-cases. It may
98
104
  "functionCall": false,
99
105
  "id": "sao10k/l3.1-euryale-70b",
100
106
  "maxTokens": undefined,
107
+ "reasoning": false,
101
108
  "vision": false,
102
109
  },
103
110
  {
@@ -114,6 +121,7 @@ Read their [announcement](https://www.ai21.com/blog/announcing-jamba-model-famil
114
121
  "functionCall": false,
115
122
  "id": "ai21/jamba-1-5-large",
116
123
  "maxTokens": 4096,
124
+ "reasoning": false,
117
125
  "vision": false,
118
126
  },
119
127
  {
@@ -130,6 +138,7 @@ Read their [announcement](https://www.ai21.com/blog/announcing-jamba-model-famil
130
138
  "functionCall": false,
131
139
  "id": "ai21/jamba-1-5-mini",
132
140
  "maxTokens": 4096,
141
+ "reasoning": false,
133
142
  "vision": false,
134
143
  },
135
144
  {
@@ -142,6 +151,7 @@ The models underwent a rigorous enhancement process, incorporating both supervis
142
151
  "functionCall": false,
143
152
  "id": "microsoft/phi-3.5-mini-128k-instruct",
144
153
  "maxTokens": undefined,
154
+ "reasoning": false,
145
155
  "vision": false,
146
156
  },
147
157
  {
@@ -156,6 +166,7 @@ The Hermes 3 series builds and expands on the Hermes 2 set of capabilities, incl
156
166
  "functionCall": true,
157
167
  "id": "nousresearch/hermes-3-llama-3.1-70b",
158
168
  "maxTokens": undefined,
169
+ "reasoning": false,
159
170
  "vision": false,
160
171
  },
161
172
  {
@@ -172,6 +183,7 @@ Hermes 3 is competitive, if not superior, to Llama-3.1 Instruct models at genera
172
183
  "functionCall": true,
173
184
  "id": "nousresearch/hermes-3-llama-3.1-405b",
174
185
  "maxTokens": undefined,
186
+ "reasoning": false,
175
187
  "vision": false,
176
188
  },
177
189
  {
@@ -190,6 +202,7 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr
190
202
  "functionCall": true,
191
203
  "id": "nousresearch/hermes-3-llama-3.1-405b:extended",
192
204
  "maxTokens": undefined,
205
+ "reasoning": false,
193
206
  "vision": false,
194
207
  },
195
208
  {
@@ -200,6 +213,7 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr
200
213
  "functionCall": false,
201
214
  "id": "perplexity/llama-3.1-sonar-huge-128k-online",
202
215
  "maxTokens": undefined,
216
+ "reasoning": false,
203
217
  "vision": false,
204
218
  },
205
219
  {
@@ -208,10 +222,11 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr
208
222
 
209
223
  Note: This model is experimental and not suited for production use-cases. It may be removed or redirected to another model in the future.",
210
224
  "displayName": "OpenAI: ChatGPT-4o",
211
- "enabled": true,
225
+ "enabled": false,
212
226
  "functionCall": false,
213
227
  "id": "openai/chatgpt-4o-latest",
214
228
  "maxTokens": 16384,
229
+ "reasoning": false,
215
230
  "vision": false,
216
231
  },
217
232
  {
@@ -226,6 +241,7 @@ For best results, use with Llama 3 Instruct context template, temperature 1.4, a
226
241
  "functionCall": false,
227
242
  "id": "sao10k/l3-lunaris-8b",
228
243
  "maxTokens": undefined,
244
+ "reasoning": false,
229
245
  "vision": false,
230
246
  },
231
247
  {
@@ -238,6 +254,7 @@ Although more similar to Magnum overall, the model remains very creative, with a
238
254
  "functionCall": false,
239
255
  "id": "aetherwiing/mn-starcannon-12b",
240
256
  "maxTokens": undefined,
257
+ "reasoning": false,
241
258
  "vision": false,
242
259
  },
243
260
  {
@@ -252,6 +269,7 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp
252
269
  "functionCall": false,
253
270
  "id": "openai/gpt-4o-2024-08-06",
254
271
  "maxTokens": 16384,
272
+ "reasoning": false,
255
273
  "vision": false,
256
274
  },
257
275
  {
@@ -266,6 +284,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
266
284
  "functionCall": false,
267
285
  "id": "meta-llama/llama-3.1-405b",
268
286
  "maxTokens": undefined,
287
+ "reasoning": false,
269
288
  "vision": false,
270
289
  },
271
290
  {
@@ -280,6 +299,7 @@ Check out the model's [HuggingFace page](https://huggingface.co/nothingiisreal/M
280
299
  "functionCall": false,
281
300
  "id": "nothingiisreal/mn-celeste-12b",
282
301
  "maxTokens": undefined,
302
+ "reasoning": false,
283
303
  "vision": false,
284
304
  },
285
305
  {
@@ -296,6 +316,7 @@ Note: This model is experimental and not suited for production use-cases. It may
296
316
  "functionCall": false,
297
317
  "id": "google/gemini-pro-1.5-exp",
298
318
  "maxTokens": 32768,
319
+ "reasoning": false,
299
320
  "vision": true,
300
321
  },
301
322
  {
@@ -308,6 +329,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
308
329
  "functionCall": false,
309
330
  "id": "perplexity/llama-3.1-sonar-large-128k-online",
310
331
  "maxTokens": undefined,
332
+ "reasoning": false,
311
333
  "vision": false,
312
334
  },
313
335
  {
@@ -320,6 +342,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
320
342
  "functionCall": false,
321
343
  "id": "perplexity/llama-3.1-sonar-large-128k-chat",
322
344
  "maxTokens": undefined,
345
+ "reasoning": false,
323
346
  "vision": false,
324
347
  },
325
348
  {
@@ -332,6 +355,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
332
355
  "functionCall": false,
333
356
  "id": "perplexity/llama-3.1-sonar-small-128k-online",
334
357
  "maxTokens": undefined,
358
+ "reasoning": false,
335
359
  "vision": false,
336
360
  },
337
361
  {
@@ -344,6 +368,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
344
368
  "functionCall": false,
345
369
  "id": "perplexity/llama-3.1-sonar-small-128k-chat",
346
370
  "maxTokens": undefined,
371
+ "reasoning": false,
347
372
  "vision": false,
348
373
  },
349
374
  {
@@ -358,6 +383,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
358
383
  "functionCall": false,
359
384
  "id": "meta-llama/llama-3.1-70b-instruct",
360
385
  "maxTokens": undefined,
386
+ "reasoning": false,
361
387
  "vision": false,
362
388
  },
363
389
  {
@@ -374,6 +400,7 @@ _These are free, rate-limited endpoints for [Llama 3.1 8B Instruct](/models/meta
374
400
  "functionCall": false,
375
401
  "id": "meta-llama/llama-3.1-8b-instruct:free",
376
402
  "maxTokens": 4096,
403
+ "reasoning": false,
377
404
  "vision": false,
378
405
  },
379
406
  {
@@ -388,6 +415,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
388
415
  "functionCall": false,
389
416
  "id": "meta-llama/llama-3.1-8b-instruct",
390
417
  "maxTokens": undefined,
418
+ "reasoning": false,
391
419
  "vision": false,
392
420
  },
393
421
  {
@@ -404,6 +432,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
404
432
  "functionCall": false,
405
433
  "id": "meta-llama/llama-3.1-405b-instruct",
406
434
  "maxTokens": undefined,
435
+ "reasoning": false,
407
436
  "vision": false,
408
437
  },
409
438
  {
@@ -418,6 +447,7 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m
418
447
  "functionCall": true,
419
448
  "id": "cognitivecomputations/dolphin-llama-3-70b",
420
449
  "maxTokens": undefined,
450
+ "reasoning": false,
421
451
  "vision": false,
422
452
  },
423
453
  {
@@ -434,6 +464,7 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m
434
464
  "functionCall": false,
435
465
  "id": "mistralai/codestral-mamba",
436
466
  "maxTokens": undefined,
467
+ "reasoning": false,
437
468
  "vision": false,
438
469
  },
439
470
  {
@@ -444,10 +475,11 @@ The model is multilingual, supporting English, French, German, Spanish, Italian,
444
475
 
445
476
  It supports function calling and is released under the Apache 2.0 license.",
446
477
  "displayName": "Mistral: Mistral Nemo",
447
- "enabled": false,
478
+ "enabled": true,
448
479
  "functionCall": true,
449
480
  "id": "mistralai/mistral-nemo",
450
481
  "maxTokens": undefined,
482
+ "reasoning": false,
451
483
  "vision": false,
452
484
  },
453
485
  {
@@ -464,6 +496,7 @@ Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advanci
464
496
  "functionCall": false,
465
497
  "id": "openai/gpt-4o-mini-2024-07-18",
466
498
  "maxTokens": 16384,
499
+ "reasoning": false,
467
500
  "vision": false,
468
501
  },
469
502
  {
@@ -480,6 +513,7 @@ Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advanci
480
513
  "functionCall": false,
481
514
  "id": "openai/gpt-4o-mini",
482
515
  "maxTokens": 16384,
516
+ "reasoning": false,
483
517
  "vision": false,
484
518
  },
485
519
  {
@@ -498,6 +532,7 @@ _These are free, rate-limited endpoints for [Qwen 2 7B Instruct](/models/qwen/qw
498
532
  "functionCall": false,
499
533
  "id": "qwen/qwen-2-7b-instruct:free",
500
534
  "maxTokens": 4096,
535
+ "reasoning": false,
501
536
  "vision": false,
502
537
  },
503
538
  {
@@ -514,6 +549,7 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug
514
549
  "functionCall": false,
515
550
  "id": "qwen/qwen-2-7b-instruct",
516
551
  "maxTokens": undefined,
552
+ "reasoning": false,
517
553
  "vision": false,
518
554
  },
519
555
  {
@@ -528,6 +564,7 @@ See the [launch announcement](https://blog.google/technology/developers/google-g
528
564
  "functionCall": false,
529
565
  "id": "google/gemma-2-27b-it",
530
566
  "maxTokens": undefined,
567
+ "reasoning": false,
531
568
  "vision": false,
532
569
  },
533
570
  {
@@ -540,6 +577,7 @@ The model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-i
540
577
  "functionCall": false,
541
578
  "id": "alpindale/magnum-72b",
542
579
  "maxTokens": 1024,
580
+ "reasoning": false,
543
581
  "vision": false,
544
582
  },
545
583
  {
@@ -552,6 +590,7 @@ Hermes-2 Θ (theta) was specifically designed with a few capabilities in mind: e
552
590
  "functionCall": false,
553
591
  "id": "nousresearch/hermes-2-theta-llama-3-8b",
554
592
  "maxTokens": 2048,
593
+ "reasoning": false,
555
594
  "vision": false,
556
595
  },
557
596
  {
@@ -568,6 +607,7 @@ _These are free, rate-limited endpoints for [Gemma 2 9B](/models/google/gemma-2-
568
607
  "functionCall": false,
569
608
  "id": "google/gemma-2-9b-it:free",
570
609
  "maxTokens": 2048,
610
+ "reasoning": false,
571
611
  "vision": false,
572
612
  },
573
613
  {
@@ -582,6 +622,7 @@ See the [launch announcement](https://blog.google/technology/developers/google-g
582
622
  "functionCall": false,
583
623
  "id": "google/gemma-2-9b-it",
584
624
  "maxTokens": undefined,
625
+ "reasoning": false,
585
626
  "vision": false,
586
627
  },
587
628
  {
@@ -597,6 +638,7 @@ Compared to older Stheno version, this model is trained on:
597
638
  "functionCall": false,
598
639
  "id": "sao10k/l3-stheno-8b",
599
640
  "maxTokens": undefined,
641
+ "reasoning": false,
600
642
  "vision": false,
601
643
  },
602
644
  {
@@ -614,6 +656,7 @@ Jamba has a knowledge cutoff of February 2024.",
614
656
  "functionCall": false,
615
657
  "id": "ai21/jamba-instruct",
616
658
  "maxTokens": 4096,
659
+ "reasoning": false,
617
660
  "vision": false,
618
661
  },
619
662
  {
@@ -631,6 +674,7 @@ Jamba has a knowledge cutoff of February 2024.",
631
674
  "functionCall": true,
632
675
  "id": "anthropic/claude-3.5-sonnet",
633
676
  "maxTokens": 8192,
677
+ "reasoning": false,
634
678
  "vision": true,
635
679
  },
636
680
  {
@@ -650,6 +694,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
650
694
  "functionCall": true,
651
695
  "id": "anthropic/claude-3.5-sonnet:beta",
652
696
  "maxTokens": 8192,
697
+ "reasoning": false,
653
698
  "vision": true,
654
699
  },
655
700
  {
@@ -666,6 +711,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
666
711
  "functionCall": false,
667
712
  "id": "sao10k/l3-euryale-70b",
668
713
  "maxTokens": undefined,
714
+ "reasoning": false,
669
715
  "vision": false,
670
716
  },
671
717
  {
@@ -680,6 +726,7 @@ For 128k context length, try [Phi-3 Medium 128K](/models/microsoft/phi-3-medium-
680
726
  "functionCall": false,
681
727
  "id": "microsoft/phi-3-medium-4k-instruct",
682
728
  "maxTokens": undefined,
729
+ "reasoning": false,
683
730
  "vision": false,
684
731
  },
685
732
  {
@@ -696,6 +743,7 @@ The model is uncensored and is stripped of alignment and bias. It requires an ex
696
743
  "functionCall": false,
697
744
  "id": "cognitivecomputations/dolphin-mixtral-8x22b",
698
745
  "maxTokens": undefined,
746
+ "reasoning": false,
699
747
  "vision": false,
700
748
  },
701
749
  {
@@ -712,6 +760,7 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug
712
760
  "functionCall": false,
713
761
  "id": "qwen/qwen-2-72b-instruct",
714
762
  "maxTokens": undefined,
763
+ "reasoning": false,
715
764
  "vision": false,
716
765
  },
717
766
  {
@@ -729,6 +778,7 @@ It outperforms many similarly sized models including [Llama 3 8B Instruct](/mode
729
778
  "functionCall": false,
730
779
  "id": "openchat/openchat-8b",
731
780
  "maxTokens": undefined,
781
+ "reasoning": false,
732
782
  "vision": false,
733
783
  },
734
784
  {
@@ -739,6 +789,7 @@ It outperforms many similarly sized models including [Llama 3 8B Instruct](/mode
739
789
  "functionCall": false,
740
790
  "id": "nousresearch/hermes-2-pro-llama-3-8b",
741
791
  "maxTokens": undefined,
792
+ "reasoning": false,
742
793
  "vision": false,
743
794
  },
744
795
  {
@@ -757,6 +808,7 @@ NOTE: Support for function calling depends on the provider.",
757
808
  "functionCall": true,
758
809
  "id": "mistralai/mistral-7b-instruct-v0.3",
759
810
  "maxTokens": undefined,
811
+ "reasoning": false,
760
812
  "vision": false,
761
813
  },
762
814
  {
@@ -771,6 +823,7 @@ _These are free, rate-limited endpoints for [Mistral 7B Instruct](/models/mistra
771
823
  "functionCall": false,
772
824
  "id": "mistralai/mistral-7b-instruct:free",
773
825
  "maxTokens": 4096,
826
+ "reasoning": false,
774
827
  "vision": false,
775
828
  },
776
829
  {
@@ -783,6 +836,7 @@ _These are free, rate-limited endpoints for [Mistral 7B Instruct](/models/mistra
783
836
  "functionCall": false,
784
837
  "id": "mistralai/mistral-7b-instruct",
785
838
  "maxTokens": undefined,
839
+ "reasoning": false,
786
840
  "vision": false,
787
841
  },
788
842
  {
@@ -797,6 +851,7 @@ _These are higher-throughput endpoints for [Mistral 7B Instruct](/models/mistral
797
851
  "functionCall": false,
798
852
  "id": "mistralai/mistral-7b-instruct:nitro",
799
853
  "maxTokens": undefined,
854
+ "reasoning": false,
800
855
  "vision": false,
801
856
  },
802
857
  {
@@ -811,6 +866,7 @@ _These are free, rate-limited endpoints for [Phi-3 Mini 128K Instruct](/models/m
811
866
  "functionCall": false,
812
867
  "id": "microsoft/phi-3-mini-128k-instruct:free",
813
868
  "maxTokens": 4096,
869
+ "reasoning": false,
814
870
  "vision": false,
815
871
  },
816
872
  {
@@ -823,6 +879,7 @@ At time of release, Phi-3 Medium demonstrated state-of-the-art performance among
823
879
  "functionCall": false,
824
880
  "id": "microsoft/phi-3-mini-128k-instruct",
825
881
  "maxTokens": undefined,
882
+ "reasoning": false,
826
883
  "vision": false,
827
884
  },
828
885
  {
@@ -839,6 +896,7 @@ _These are free, rate-limited endpoints for [Phi-3 Medium 128K Instruct](/models
839
896
  "functionCall": false,
840
897
  "id": "microsoft/phi-3-medium-128k-instruct:free",
841
898
  "maxTokens": 4096,
899
+ "reasoning": false,
842
900
  "vision": false,
843
901
  },
844
902
  {
@@ -853,6 +911,7 @@ For 4k context length, try [Phi-3 Medium 4K](/models/microsoft/phi-3-medium-4k-i
853
911
  "functionCall": false,
854
912
  "id": "microsoft/phi-3-medium-128k-instruct",
855
913
  "maxTokens": undefined,
914
+ "reasoning": false,
856
915
  "vision": false,
857
916
  },
858
917
  {
@@ -867,6 +926,7 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m
867
926
  "functionCall": false,
868
927
  "id": "neversleep/llama-3-lumimaid-70b",
869
928
  "maxTokens": 2048,
929
+ "reasoning": false,
870
930
  "vision": false,
871
931
  },
872
932
  {
@@ -883,6 +943,7 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d
883
943
  "functionCall": false,
884
944
  "id": "google/gemini-flash-1.5",
885
945
  "maxTokens": 32768,
946
+ "reasoning": false,
886
947
  "vision": true,
887
948
  },
888
949
  {
@@ -895,6 +956,7 @@ The original V1 model was trained from scratch on 2T tokens, with a composition
895
956
  "functionCall": false,
896
957
  "id": "deepseek/deepseek-coder",
897
958
  "maxTokens": 4096,
959
+ "reasoning": false,
898
960
  "vision": false,
899
961
  },
900
962
  {
@@ -909,6 +971,7 @@ DeepSeek-V2 achieves remarkable performance on both standard benchmarks and open
909
971
  "functionCall": false,
910
972
  "id": "deepseek/deepseek-chat",
911
973
  "maxTokens": 4096,
974
+ "reasoning": false,
912
975
  "vision": false,
913
976
  },
914
977
  {
@@ -921,6 +984,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
921
984
  "functionCall": false,
922
985
  "id": "perplexity/llama-3-sonar-large-32k-online",
923
986
  "maxTokens": undefined,
987
+ "reasoning": false,
924
988
  "vision": false,
925
989
  },
926
990
  {
@@ -933,6 +997,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
933
997
  "functionCall": false,
934
998
  "id": "perplexity/llama-3-sonar-large-32k-chat",
935
999
  "maxTokens": undefined,
1000
+ "reasoning": false,
936
1001
  "vision": false,
937
1002
  },
938
1003
  {
@@ -945,6 +1010,7 @@ This is the online version of the [offline chat model](/models/perplexity/llama-
945
1010
  "functionCall": false,
946
1011
  "id": "perplexity/llama-3-sonar-small-32k-online",
947
1012
  "maxTokens": undefined,
1013
+ "reasoning": false,
948
1014
  "vision": false,
949
1015
  },
950
1016
  {
@@ -957,6 +1023,7 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama-
957
1023
  "functionCall": false,
958
1024
  "id": "perplexity/llama-3-sonar-small-32k-chat",
959
1025
  "maxTokens": undefined,
1026
+ "reasoning": false,
960
1027
  "vision": false,
961
1028
  },
962
1029
  {
@@ -975,6 +1042,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
975
1042
  "functionCall": false,
976
1043
  "id": "meta-llama/llama-guard-2-8b",
977
1044
  "maxTokens": undefined,
1045
+ "reasoning": false,
978
1046
  "vision": false,
979
1047
  },
980
1048
  {
@@ -987,6 +1055,7 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp
987
1055
  "functionCall": false,
988
1056
  "id": "openai/gpt-4o-2024-05-13",
989
1057
  "maxTokens": 4096,
1058
+ "reasoning": false,
990
1059
  "vision": false,
991
1060
  },
992
1061
  {
@@ -999,6 +1068,7 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp
999
1068
  "functionCall": false,
1000
1069
  "id": "openai/gpt-4o",
1001
1070
  "maxTokens": 4096,
1071
+ "reasoning": false,
1002
1072
  "vision": false,
1003
1073
  },
1004
1074
  {
@@ -1011,6 +1081,7 @@ _These are extended-context endpoints for [GPT-4o](/models/openai/gpt-4o). They
1011
1081
  "functionCall": false,
1012
1082
  "id": "openai/gpt-4o:extended",
1013
1083
  "maxTokens": 64000,
1084
+ "reasoning": false,
1014
1085
  "vision": false,
1015
1086
  },
1016
1087
  {
@@ -1029,6 +1100,7 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug
1029
1100
  "functionCall": false,
1030
1101
  "id": "qwen/qwen-72b-chat",
1031
1102
  "maxTokens": undefined,
1103
+ "reasoning": false,
1032
1104
  "vision": false,
1033
1105
  },
1034
1106
  {
@@ -1047,6 +1119,7 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug
1047
1119
  "functionCall": false,
1048
1120
  "id": "qwen/qwen-110b-chat",
1049
1121
  "maxTokens": undefined,
1122
+ "reasoning": false,
1050
1123
  "vision": false,
1051
1124
  },
1052
1125
  {
@@ -1061,6 +1134,7 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m
1061
1134
  "functionCall": false,
1062
1135
  "id": "neversleep/llama-3-lumimaid-8b",
1063
1136
  "maxTokens": undefined,
1137
+ "reasoning": false,
1064
1138
  "vision": false,
1065
1139
  },
1066
1140
  {
@@ -1077,6 +1151,7 @@ _These are extended-context endpoints for [Llama 3 Lumimaid 8B](/models/neversle
1077
1151
  "functionCall": false,
1078
1152
  "id": "neversleep/llama-3-lumimaid-8b:extended",
1079
1153
  "maxTokens": 2048,
1154
+ "reasoning": false,
1080
1155
  "vision": false,
1081
1156
  },
1082
1157
  {
@@ -1089,6 +1164,7 @@ If you submit a raw prompt, you can use Alpaca or Vicuna formats.",
1089
1164
  "functionCall": false,
1090
1165
  "id": "sao10k/fimbulvetr-11b-v2",
1091
1166
  "maxTokens": 2048,
1167
+ "reasoning": false,
1092
1168
  "vision": false,
1093
1169
  },
1094
1170
  {
@@ -1103,6 +1179,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
1103
1179
  "functionCall": false,
1104
1180
  "id": "meta-llama/llama-3-70b-instruct",
1105
1181
  "maxTokens": undefined,
1182
+ "reasoning": false,
1106
1183
  "vision": false,
1107
1184
  },
1108
1185
  {
@@ -1119,6 +1196,7 @@ _These are higher-throughput endpoints for [Llama 3 70B Instruct](/models/meta-l
1119
1196
  "functionCall": false,
1120
1197
  "id": "meta-llama/llama-3-70b-instruct:nitro",
1121
1198
  "maxTokens": undefined,
1199
+ "reasoning": false,
1122
1200
  "vision": false,
1123
1201
  },
1124
1202
  {
@@ -1135,6 +1213,7 @@ _These are free, rate-limited endpoints for [Llama 3 8B Instruct](/models/meta-l
1135
1213
  "functionCall": false,
1136
1214
  "id": "meta-llama/llama-3-8b-instruct:free",
1137
1215
  "maxTokens": 4096,
1216
+ "reasoning": false,
1138
1217
  "vision": false,
1139
1218
  },
1140
1219
  {
@@ -1149,6 +1228,7 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta
1149
1228
  "functionCall": false,
1150
1229
  "id": "meta-llama/llama-3-8b-instruct",
1151
1230
  "maxTokens": undefined,
1231
+ "reasoning": false,
1152
1232
  "vision": false,
1153
1233
  },
1154
1234
  {
@@ -1165,6 +1245,7 @@ _These are higher-throughput endpoints for [Llama 3 8B Instruct](/models/meta-ll
1165
1245
  "functionCall": false,
1166
1246
  "id": "meta-llama/llama-3-8b-instruct:nitro",
1167
1247
  "maxTokens": undefined,
1248
+ "reasoning": false,
1168
1249
  "vision": false,
1169
1250
  },
1170
1251
  {
@@ -1181,6 +1262,7 @@ _These are extended-context endpoints for [Llama 3 8B Instruct](/models/meta-lla
1181
1262
  "functionCall": false,
1182
1263
  "id": "meta-llama/llama-3-8b-instruct:extended",
1183
1264
  "maxTokens": 2048,
1265
+ "reasoning": false,
1184
1266
  "vision": false,
1185
1267
  },
1186
1268
  {
@@ -1197,6 +1279,7 @@ See benchmarks on the launch announcement [here](https://mistral.ai/news/mixtral
1197
1279
  "functionCall": false,
1198
1280
  "id": "mistralai/mixtral-8x22b-instruct",
1199
1281
  "maxTokens": undefined,
1282
+ "reasoning": false,
1200
1283
  "vision": false,
1201
1284
  },
1202
1285
  {
@@ -1213,6 +1296,7 @@ To read more about the model release, [click here](https://wizardlm.github.io/Wi
1213
1296
  "functionCall": false,
1214
1297
  "id": "microsoft/wizardlm-2-7b",
1215
1298
  "maxTokens": undefined,
1299
+ "reasoning": false,
1216
1300
  "vision": false,
1217
1301
  },
1218
1302
  {
@@ -1229,6 +1313,7 @@ To read more about the model release, [click here](https://wizardlm.github.io/Wi
1229
1313
  "functionCall": false,
1230
1314
  "id": "microsoft/wizardlm-2-8x22b",
1231
1315
  "maxTokens": undefined,
1316
+ "reasoning": false,
1232
1317
  "vision": false,
1233
1318
  },
1234
1319
  {
@@ -1254,6 +1339,7 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d
1254
1339
  "functionCall": false,
1255
1340
  "id": "google/gemini-pro-1.5",
1256
1341
  "maxTokens": 32768,
1342
+ "reasoning": false,
1257
1343
  "vision": true,
1258
1344
  },
1259
1345
  {
@@ -1266,6 +1352,7 @@ Training data: up to December 2023.",
1266
1352
  "functionCall": true,
1267
1353
  "id": "openai/gpt-4-turbo",
1268
1354
  "maxTokens": 4096,
1355
+ "reasoning": false,
1269
1356
  "vision": true,
1270
1357
  },
1271
1358
  {
@@ -1276,10 +1363,11 @@ It offers multilingual support for ten key languages to facilitate global busine
1276
1363
 
1277
1364
  Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.cohere.com/docs/c4ai-acceptable-use-policy).",
1278
1365
  "displayName": "Cohere: Command R+",
1279
- "enabled": true,
1366
+ "enabled": false,
1280
1367
  "functionCall": false,
1281
1368
  "id": "cohere/command-r-plus",
1282
1369
  "maxTokens": 4000,
1370
+ "reasoning": false,
1283
1371
  "vision": false,
1284
1372
  },
1285
1373
  {
@@ -1296,6 +1384,7 @@ See the launch announcement and benchmark results [here](https://www.databricks.
1296
1384
  "functionCall": false,
1297
1385
  "id": "databricks/dbrx-instruct",
1298
1386
  "maxTokens": undefined,
1387
+ "reasoning": false,
1299
1388
  "vision": false,
1300
1389
  },
1301
1390
  {
@@ -1308,6 +1397,7 @@ Descending from earlier versions of Midnight Rose and [Wizard Tulu Dolphin 70B](
1308
1397
  "functionCall": false,
1309
1398
  "id": "sophosympatheia/midnight-rose-70b",
1310
1399
  "maxTokens": undefined,
1400
+ "reasoning": false,
1311
1401
  "vision": false,
1312
1402
  },
1313
1403
  {
@@ -1318,10 +1408,11 @@ Read the launch post [here](https://txt.cohere.com/command-r/).
1318
1408
 
1319
1409
  Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.cohere.com/docs/c4ai-acceptable-use-policy).",
1320
1410
  "displayName": "Cohere: Command R",
1321
- "enabled": true,
1411
+ "enabled": false,
1322
1412
  "functionCall": false,
1323
1413
  "id": "cohere/command-r",
1324
1414
  "maxTokens": 4000,
1415
+ "reasoning": false,
1325
1416
  "vision": false,
1326
1417
  },
1327
1418
  {
@@ -1334,6 +1425,7 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co
1334
1425
  "functionCall": false,
1335
1426
  "id": "cohere/command",
1336
1427
  "maxTokens": 4000,
1428
+ "reasoning": false,
1337
1429
  "vision": false,
1338
1430
  },
1339
1431
  {
@@ -1349,6 +1441,7 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c
1349
1441
  "functionCall": false,
1350
1442
  "id": "anthropic/claude-3-haiku",
1351
1443
  "maxTokens": 4096,
1444
+ "reasoning": false,
1352
1445
  "vision": true,
1353
1446
  },
1354
1447
  {
@@ -1366,6 +1459,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1366
1459
  "functionCall": false,
1367
1460
  "id": "anthropic/claude-3-haiku:beta",
1368
1461
  "maxTokens": 4096,
1462
+ "reasoning": false,
1369
1463
  "vision": true,
1370
1464
  },
1371
1465
  {
@@ -1380,6 +1474,7 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c
1380
1474
  "functionCall": false,
1381
1475
  "id": "anthropic/claude-3-sonnet",
1382
1476
  "maxTokens": 4096,
1477
+ "reasoning": false,
1383
1478
  "vision": true,
1384
1479
  },
1385
1480
  {
@@ -1396,6 +1491,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1396
1491
  "functionCall": false,
1397
1492
  "id": "anthropic/claude-3-sonnet:beta",
1398
1493
  "maxTokens": 4096,
1494
+ "reasoning": false,
1399
1495
  "vision": true,
1400
1496
  },
1401
1497
  {
@@ -1410,6 +1506,7 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c
1410
1506
  "functionCall": false,
1411
1507
  "id": "anthropic/claude-3-opus",
1412
1508
  "maxTokens": 4096,
1509
+ "reasoning": false,
1413
1510
  "vision": true,
1414
1511
  },
1415
1512
  {
@@ -1426,6 +1523,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1426
1523
  "functionCall": false,
1427
1524
  "id": "anthropic/claude-3-opus:beta",
1428
1525
  "maxTokens": 4096,
1526
+ "reasoning": false,
1429
1527
  "vision": true,
1430
1528
  },
1431
1529
  {
@@ -1438,6 +1536,7 @@ It is fluent in English, French, Spanish, German, and Italian, with high grammat
1438
1536
  "functionCall": false,
1439
1537
  "id": "mistralai/mistral-large",
1440
1538
  "maxTokens": undefined,
1539
+ "reasoning": false,
1441
1540
  "vision": false,
1442
1541
  },
1443
1542
  {
@@ -1450,6 +1549,7 @@ It is fluent in English, French, Spanish, German, and Italian, with high grammat
1450
1549
  "functionCall": true,
1451
1550
  "id": "openai/gpt-4-turbo-preview",
1452
1551
  "maxTokens": 4096,
1552
+ "reasoning": false,
1453
1553
  "vision": false,
1454
1554
  },
1455
1555
  {
@@ -1462,6 +1562,7 @@ Training data up to Sep 2021.",
1462
1562
  "functionCall": false,
1463
1563
  "id": "openai/gpt-3.5-turbo-0613",
1464
1564
  "maxTokens": 4096,
1565
+ "reasoning": false,
1465
1566
  "vision": false,
1466
1567
  },
1467
1568
  {
@@ -1476,6 +1577,7 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope
1476
1577
  "functionCall": false,
1477
1578
  "id": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
1478
1579
  "maxTokens": undefined,
1580
+ "reasoning": false,
1479
1581
  "vision": false,
1480
1582
  },
1481
1583
  {
@@ -1486,6 +1588,7 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope
1486
1588
  "functionCall": false,
1487
1589
  "id": "mistralai/mistral-medium",
1488
1590
  "maxTokens": undefined,
1591
+ "reasoning": false,
1489
1592
  "vision": false,
1490
1593
  },
1491
1594
  {
@@ -1497,6 +1600,7 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope
1497
1600
  "functionCall": false,
1498
1601
  "id": "mistralai/mistral-small",
1499
1602
  "maxTokens": undefined,
1603
+ "reasoning": false,
1500
1604
  "vision": false,
1501
1605
  },
1502
1606
  {
@@ -1507,6 +1611,7 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope
1507
1611
  "functionCall": false,
1508
1612
  "id": "mistralai/mistral-tiny",
1509
1613
  "maxTokens": undefined,
1614
+ "reasoning": false,
1510
1615
  "vision": false,
1511
1616
  },
1512
1617
  {
@@ -1517,6 +1622,7 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope
1517
1622
  "functionCall": false,
1518
1623
  "id": "austism/chronos-hermes-13b",
1519
1624
  "maxTokens": undefined,
1625
+ "reasoning": false,
1520
1626
  "vision": false,
1521
1627
  },
1522
1628
  {
@@ -1529,6 +1635,7 @@ Nous-Hermes 2 on Yi 34B outperforms all Nous-Hermes & Open-Hermes models of the
1529
1635
  "functionCall": false,
1530
1636
  "id": "nousresearch/nous-hermes-yi-34b",
1531
1637
  "maxTokens": undefined,
1638
+ "reasoning": false,
1532
1639
  "vision": false,
1533
1640
  },
1534
1641
  {
@@ -1545,6 +1652,7 @@ An improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruc
1545
1652
  "functionCall": false,
1546
1653
  "id": "mistralai/mistral-7b-instruct-v0.2",
1547
1654
  "maxTokens": undefined,
1655
+ "reasoning": false,
1548
1656
  "vision": false,
1549
1657
  },
1550
1658
  {
@@ -1559,6 +1667,7 @@ The model is uncensored and is stripped of alignment and bias. It requires an ex
1559
1667
  "functionCall": false,
1560
1668
  "id": "cognitivecomputations/dolphin-mixtral-8x7b",
1561
1669
  "maxTokens": undefined,
1670
+ "reasoning": false,
1562
1671
  "vision": false,
1563
1672
  },
1564
1673
  {
@@ -1575,6 +1684,7 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d
1575
1684
  "functionCall": false,
1576
1685
  "id": "google/gemini-pro-vision",
1577
1686
  "maxTokens": 8192,
1687
+ "reasoning": false,
1578
1688
  "vision": true,
1579
1689
  },
1580
1690
  {
@@ -1589,6 +1699,7 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d
1589
1699
  "functionCall": false,
1590
1700
  "id": "google/gemini-pro",
1591
1701
  "maxTokens": 32768,
1702
+ "reasoning": false,
1592
1703
  "vision": false,
1593
1704
  },
1594
1705
  {
@@ -1601,6 +1712,7 @@ Instruct model fine-tuned by Mistral. #moe",
1601
1712
  "functionCall": false,
1602
1713
  "id": "mistralai/mixtral-8x7b-instruct",
1603
1714
  "maxTokens": undefined,
1715
+ "reasoning": false,
1604
1716
  "vision": false,
1605
1717
  },
1606
1718
  {
@@ -1615,6 +1727,7 @@ _These are higher-throughput endpoints for [Mixtral 8x7B Instruct](/models/mistr
1615
1727
  "functionCall": false,
1616
1728
  "id": "mistralai/mixtral-8x7b-instruct:nitro",
1617
1729
  "maxTokens": undefined,
1730
+ "reasoning": false,
1618
1731
  "vision": false,
1619
1732
  },
1620
1733
  {
@@ -1627,6 +1740,7 @@ _These are higher-throughput endpoints for [Mixtral 8x7B Instruct](/models/mistr
1627
1740
  "functionCall": false,
1628
1741
  "id": "mistralai/mixtral-8x7b",
1629
1742
  "maxTokens": undefined,
1743
+ "reasoning": false,
1630
1744
  "vision": false,
1631
1745
  },
1632
1746
  {
@@ -1639,6 +1753,7 @@ StripedHyena uses a new architecture that competes with traditional Transformers
1639
1753
  "functionCall": false,
1640
1754
  "id": "togethercomputer/stripedhyena-nous-7b",
1641
1755
  "maxTokens": undefined,
1756
+ "reasoning": false,
1642
1757
  "vision": false,
1643
1758
  },
1644
1759
  {
@@ -1655,6 +1770,7 @@ _These are free, rate-limited endpoints for [MythoMist 7B](/models/gryphe/mythom
1655
1770
  "functionCall": false,
1656
1771
  "id": "gryphe/mythomist-7b:free",
1657
1772
  "maxTokens": 4096,
1773
+ "reasoning": false,
1658
1774
  "vision": false,
1659
1775
  },
1660
1776
  {
@@ -1669,6 +1785,7 @@ It combines [Neural Chat 7B](/models/intel/neural-chat-7b), Airoboros 7b, [Toppy
1669
1785
  "functionCall": false,
1670
1786
  "id": "gryphe/mythomist-7b",
1671
1787
  "maxTokens": 2048,
1788
+ "reasoning": false,
1672
1789
  "vision": false,
1673
1790
  },
1674
1791
  {
@@ -1686,6 +1803,7 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o
1686
1803
  "functionCall": false,
1687
1804
  "id": "openchat/openchat-7b:free",
1688
1805
  "maxTokens": 4096,
1806
+ "reasoning": false,
1689
1807
  "vision": false,
1690
1808
  },
1691
1809
  {
@@ -1701,6 +1819,7 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o
1701
1819
  "functionCall": false,
1702
1820
  "id": "openchat/openchat-7b",
1703
1821
  "maxTokens": undefined,
1822
+ "reasoning": false,
1704
1823
  "vision": false,
1705
1824
  },
1706
1825
  {
@@ -1713,6 +1832,7 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o
1713
1832
  "functionCall": false,
1714
1833
  "id": "neversleep/noromaid-20b",
1715
1834
  "maxTokens": 2048,
1835
+ "reasoning": false,
1716
1836
  "vision": false,
1717
1837
  },
1718
1838
  {
@@ -1723,6 +1843,7 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o
1723
1843
  "functionCall": false,
1724
1844
  "id": "anthropic/claude-instant-1.1",
1725
1845
  "maxTokens": 2048,
1846
+ "reasoning": false,
1726
1847
  "vision": false,
1727
1848
  },
1728
1849
  {
@@ -1733,6 +1854,7 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o
1733
1854
  "functionCall": false,
1734
1855
  "id": "anthropic/claude-2.1",
1735
1856
  "maxTokens": 4096,
1857
+ "reasoning": false,
1736
1858
  "vision": false,
1737
1859
  },
1738
1860
  {
@@ -1745,6 +1867,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1745
1867
  "functionCall": false,
1746
1868
  "id": "anthropic/claude-2.1:beta",
1747
1869
  "maxTokens": 4096,
1870
+ "reasoning": false,
1748
1871
  "vision": false,
1749
1872
  },
1750
1873
  {
@@ -1755,6 +1878,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1755
1878
  "functionCall": false,
1756
1879
  "id": "anthropic/claude-2",
1757
1880
  "maxTokens": 4096,
1881
+ "reasoning": false,
1758
1882
  "vision": false,
1759
1883
  },
1760
1884
  {
@@ -1767,6 +1891,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
1767
1891
  "functionCall": false,
1768
1892
  "id": "anthropic/claude-2:beta",
1769
1893
  "maxTokens": 4096,
1894
+ "reasoning": false,
1770
1895
  "vision": false,
1771
1896
  },
1772
1897
  {
@@ -1778,6 +1903,7 @@ Potentially the most interesting finding from training on a good ratio (est. of
1778
1903
  "functionCall": false,
1779
1904
  "id": "teknium/openhermes-2.5-mistral-7b",
1780
1905
  "maxTokens": undefined,
1906
+ "reasoning": false,
1781
1907
  "vision": false,
1782
1908
  },
1783
1909
  {
@@ -1792,6 +1918,7 @@ Potentially the most interesting finding from training on a good ratio (est. of
1792
1918
  "functionCall": false,
1793
1919
  "id": "openai/gpt-4-vision-preview",
1794
1920
  "maxTokens": 4096,
1921
+ "reasoning": false,
1795
1922
  "vision": true,
1796
1923
  },
1797
1924
  {
@@ -1805,6 +1932,7 @@ A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative
1805
1932
  "functionCall": false,
1806
1933
  "id": "lizpreciatior/lzlv-70b-fp16-hf",
1807
1934
  "maxTokens": undefined,
1935
+ "reasoning": false,
1808
1936
  "vision": false,
1809
1937
  },
1810
1938
  {
@@ -1821,6 +1949,7 @@ Credits to
1821
1949
  "functionCall": false,
1822
1950
  "id": "alpindale/goliath-120b",
1823
1951
  "maxTokens": 400,
1952
+ "reasoning": false,
1824
1953
  "vision": false,
1825
1954
  },
1826
1955
  {
@@ -1841,6 +1970,7 @@ _These are free, rate-limited endpoints for [Toppy M 7B](/models/undi95/toppy-m-
1841
1970
  "functionCall": false,
1842
1971
  "id": "undi95/toppy-m-7b:free",
1843
1972
  "maxTokens": 2048,
1973
+ "reasoning": false,
1844
1974
  "vision": false,
1845
1975
  },
1846
1976
  {
@@ -1859,6 +1989,7 @@ List of merged models:
1859
1989
  "functionCall": false,
1860
1990
  "id": "undi95/toppy-m-7b",
1861
1991
  "maxTokens": undefined,
1992
+ "reasoning": false,
1862
1993
  "vision": false,
1863
1994
  },
1864
1995
  {
@@ -1879,6 +2010,7 @@ _These are higher-throughput endpoints for [Toppy M 7B](/models/undi95/toppy-m-7
1879
2010
  "functionCall": false,
1880
2011
  "id": "undi95/toppy-m-7b:nitro",
1881
2012
  "maxTokens": undefined,
2013
+ "reasoning": false,
1882
2014
  "vision": false,
1883
2015
  },
1884
2016
  {
@@ -1891,6 +2023,7 @@ A major redesign of this router is coming soon. Stay tuned on [Discord](https://
1891
2023
  "functionCall": false,
1892
2024
  "id": "openrouter/auto",
1893
2025
  "maxTokens": undefined,
2026
+ "reasoning": false,
1894
2027
  "vision": false,
1895
2028
  },
1896
2029
  {
@@ -1903,6 +2036,7 @@ Training data: up to April 2023.",
1903
2036
  "functionCall": true,
1904
2037
  "id": "openai/gpt-4-1106-preview",
1905
2038
  "maxTokens": 4096,
2039
+ "reasoning": false,
1906
2040
  "vision": true,
1907
2041
  },
1908
2042
  {
@@ -1913,6 +2047,7 @@ Training data: up to April 2023.",
1913
2047
  "functionCall": true,
1914
2048
  "id": "openai/gpt-3.5-turbo-1106",
1915
2049
  "maxTokens": 4096,
2050
+ "reasoning": false,
1916
2051
  "vision": false,
1917
2052
  },
1918
2053
  {
@@ -1923,6 +2058,7 @@ Training data: up to April 2023.",
1923
2058
  "functionCall": false,
1924
2059
  "id": "google/palm-2-codechat-bison-32k",
1925
2060
  "maxTokens": 32768,
2061
+ "reasoning": false,
1926
2062
  "vision": false,
1927
2063
  },
1928
2064
  {
@@ -1933,6 +2069,7 @@ Training data: up to April 2023.",
1933
2069
  "functionCall": false,
1934
2070
  "id": "google/palm-2-chat-bison-32k",
1935
2071
  "maxTokens": 32768,
2072
+ "reasoning": false,
1936
2073
  "vision": false,
1937
2074
  },
1938
2075
  {
@@ -1945,6 +2082,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1945
2082
  "functionCall": false,
1946
2083
  "id": "jondurbin/airoboros-l2-70b",
1947
2084
  "maxTokens": undefined,
2085
+ "reasoning": false,
1948
2086
  "vision": false,
1949
2087
  },
1950
2088
  {
@@ -1955,6 +2093,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1955
2093
  "functionCall": false,
1956
2094
  "id": "xwin-lm/xwin-lm-70b",
1957
2095
  "maxTokens": 400,
2096
+ "reasoning": false,
1958
2097
  "vision": false,
1959
2098
  },
1960
2099
  {
@@ -1965,6 +2104,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1965
2104
  "functionCall": false,
1966
2105
  "id": "mistralai/mistral-7b-instruct-v0.1",
1967
2106
  "maxTokens": undefined,
2107
+ "reasoning": false,
1968
2108
  "vision": false,
1969
2109
  },
1970
2110
  {
@@ -1975,6 +2115,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1975
2115
  "functionCall": false,
1976
2116
  "id": "openai/gpt-3.5-turbo-instruct",
1977
2117
  "maxTokens": 4096,
2118
+ "reasoning": false,
1978
2119
  "vision": false,
1979
2120
  },
1980
2121
  {
@@ -1985,6 +2126,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1985
2126
  "functionCall": false,
1986
2127
  "id": "pygmalionai/mythalion-13b",
1987
2128
  "maxTokens": 400,
2129
+ "reasoning": false,
1988
2130
  "vision": false,
1989
2131
  },
1990
2132
  {
@@ -1995,6 +2137,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
1995
2137
  "functionCall": false,
1996
2138
  "id": "openai/gpt-4-32k-0314",
1997
2139
  "maxTokens": 4096,
2140
+ "reasoning": false,
1998
2141
  "vision": false,
1999
2142
  },
2000
2143
  {
@@ -2005,6 +2148,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
2005
2148
  "functionCall": false,
2006
2149
  "id": "openai/gpt-4-32k",
2007
2150
  "maxTokens": 4096,
2151
+ "reasoning": false,
2008
2152
  "vision": false,
2009
2153
  },
2010
2154
  {
@@ -2015,6 +2159,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
2015
2159
  "functionCall": false,
2016
2160
  "id": "openai/gpt-3.5-turbo-16k",
2017
2161
  "maxTokens": 4096,
2162
+ "reasoning": false,
2018
2163
  "vision": false,
2019
2164
  },
2020
2165
  {
@@ -2025,6 +2170,7 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin
2025
2170
  "functionCall": false,
2026
2171
  "id": "nousresearch/nous-hermes-llama2-13b",
2027
2172
  "maxTokens": undefined,
2173
+ "reasoning": false,
2028
2174
  "vision": false,
2029
2175
  },
2030
2176
  {
@@ -2037,6 +2183,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2037
2183
  "functionCall": false,
2038
2184
  "id": "huggingfaceh4/zephyr-7b-beta:free",
2039
2185
  "maxTokens": 2048,
2186
+ "reasoning": false,
2040
2187
  "vision": false,
2041
2188
  },
2042
2189
  {
@@ -2047,6 +2194,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2047
2194
  "functionCall": false,
2048
2195
  "id": "mancer/weaver",
2049
2196
  "maxTokens": 1000,
2197
+ "reasoning": false,
2050
2198
  "vision": false,
2051
2199
  },
2052
2200
  {
@@ -2057,6 +2205,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2057
2205
  "functionCall": false,
2058
2206
  "id": "anthropic/claude-instant-1.0",
2059
2207
  "maxTokens": 4096,
2208
+ "reasoning": false,
2060
2209
  "vision": false,
2061
2210
  },
2062
2211
  {
@@ -2067,6 +2216,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2067
2216
  "functionCall": false,
2068
2217
  "id": "anthropic/claude-1.2",
2069
2218
  "maxTokens": 4096,
2219
+ "reasoning": false,
2070
2220
  "vision": false,
2071
2221
  },
2072
2222
  {
@@ -2077,6 +2227,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2077
2227
  "functionCall": false,
2078
2228
  "id": "anthropic/claude-1",
2079
2229
  "maxTokens": 4096,
2230
+ "reasoning": false,
2080
2231
  "vision": false,
2081
2232
  },
2082
2233
  {
@@ -2087,6 +2238,7 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze
2087
2238
  "functionCall": false,
2088
2239
  "id": "anthropic/claude-instant-1",
2089
2240
  "maxTokens": 4096,
2241
+ "reasoning": false,
2090
2242
  "vision": false,
2091
2243
  },
2092
2244
  {
@@ -2099,6 +2251,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
2099
2251
  "functionCall": false,
2100
2252
  "id": "anthropic/claude-instant-1:beta",
2101
2253
  "maxTokens": 4096,
2254
+ "reasoning": false,
2102
2255
  "vision": false,
2103
2256
  },
2104
2257
  {
@@ -2109,6 +2262,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
2109
2262
  "functionCall": false,
2110
2263
  "id": "anthropic/claude-2.0",
2111
2264
  "maxTokens": 4096,
2265
+ "reasoning": false,
2112
2266
  "vision": false,
2113
2267
  },
2114
2268
  {
@@ -2121,6 +2275,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
2121
2275
  "functionCall": false,
2122
2276
  "id": "anthropic/claude-2.0:beta",
2123
2277
  "maxTokens": 4096,
2278
+ "reasoning": false,
2124
2279
  "vision": false,
2125
2280
  },
2126
2281
  {
@@ -2131,6 +2286,7 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that
2131
2286
  "functionCall": false,
2132
2287
  "id": "undi95/remm-slerp-l2-13b",
2133
2288
  "maxTokens": 400,
2289
+ "reasoning": false,
2134
2290
  "vision": false,
2135
2291
  },
2136
2292
  {
@@ -2143,6 +2299,7 @@ _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-s
2143
2299
  "functionCall": false,
2144
2300
  "id": "undi95/remm-slerp-l2-13b:extended",
2145
2301
  "maxTokens": 400,
2302
+ "reasoning": false,
2146
2303
  "vision": false,
2147
2304
  },
2148
2305
  {
@@ -2153,6 +2310,7 @@ _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-s
2153
2310
  "functionCall": false,
2154
2311
  "id": "google/palm-2-codechat-bison",
2155
2312
  "maxTokens": 4096,
2313
+ "reasoning": false,
2156
2314
  "vision": false,
2157
2315
  },
2158
2316
  {
@@ -2163,6 +2321,7 @@ _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-s
2163
2321
  "functionCall": false,
2164
2322
  "id": "google/palm-2-chat-bison",
2165
2323
  "maxTokens": 4096,
2324
+ "reasoning": false,
2166
2325
  "vision": false,
2167
2326
  },
2168
2327
  {
@@ -2173,6 +2332,7 @@ _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-s
2173
2332
  "functionCall": false,
2174
2333
  "id": "gryphe/mythomax-l2-13b",
2175
2334
  "maxTokens": undefined,
2335
+ "reasoning": false,
2176
2336
  "vision": false,
2177
2337
  },
2178
2338
  {
@@ -2185,6 +2345,7 @@ _These are higher-throughput endpoints for [MythoMax 13B](/models/gryphe/mythoma
2185
2345
  "functionCall": false,
2186
2346
  "id": "gryphe/mythomax-l2-13b:nitro",
2187
2347
  "maxTokens": undefined,
2348
+ "reasoning": false,
2188
2349
  "vision": false,
2189
2350
  },
2190
2351
  {
@@ -2197,6 +2358,7 @@ _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax
2197
2358
  "functionCall": false,
2198
2359
  "id": "gryphe/mythomax-l2-13b:extended",
2199
2360
  "maxTokens": 400,
2361
+ "reasoning": false,
2200
2362
  "vision": false,
2201
2363
  },
2202
2364
  {
@@ -2207,6 +2369,7 @@ _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax
2207
2369
  "functionCall": false,
2208
2370
  "id": "meta-llama/llama-2-13b-chat",
2209
2371
  "maxTokens": undefined,
2372
+ "reasoning": false,
2210
2373
  "vision": false,
2211
2374
  },
2212
2375
  {
@@ -2217,6 +2380,7 @@ _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax
2217
2380
  "functionCall": false,
2218
2381
  "id": "openai/gpt-4-0314",
2219
2382
  "maxTokens": 4096,
2383
+ "reasoning": false,
2220
2384
  "vision": false,
2221
2385
  },
2222
2386
  {
@@ -2227,6 +2391,7 @@ _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax
2227
2391
  "functionCall": false,
2228
2392
  "id": "openai/gpt-4",
2229
2393
  "maxTokens": 4096,
2394
+ "reasoning": false,
2230
2395
  "vision": true,
2231
2396
  },
2232
2397
  {
@@ -2239,6 +2404,7 @@ Training data up to Sep 2021.",
2239
2404
  "functionCall": false,
2240
2405
  "id": "openai/gpt-3.5-turbo-0301",
2241
2406
  "maxTokens": 4096,
2407
+ "reasoning": false,
2242
2408
  "vision": false,
2243
2409
  },
2244
2410
  {
@@ -2251,6 +2417,7 @@ This version has a higher accuracy at responding in requested formats and a fix
2251
2417
  "functionCall": true,
2252
2418
  "id": "openai/gpt-3.5-turbo-0125",
2253
2419
  "maxTokens": 4096,
2420
+ "reasoning": false,
2254
2421
  "vision": false,
2255
2422
  },
2256
2423
  {
@@ -2263,6 +2430,7 @@ Training data up to Sep 2021.",
2263
2430
  "functionCall": false,
2264
2431
  "id": "openai/gpt-3.5-turbo",
2265
2432
  "maxTokens": 4096,
2433
+ "reasoning": false,
2266
2434
  "vision": false,
2267
2435
  },
2268
2436
  ]