modelpedia 0.0.2 → 0.0.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +141 -0
- package/dist/index.cjs +1 -101
- package/dist/index.d.cts +10 -2
- package/dist/index.d.mts +10 -2
- package/dist/index.mjs +1 -92
- package/dist/providers/alibaba.cjs +4 -835
- package/dist/providers/alibaba.d.cts +1 -1
- package/dist/providers/alibaba.d.mts +1 -1
- package/dist/providers/alibaba.mjs +4 -833
- package/dist/providers/amazon.cjs +4 -1484
- package/dist/providers/amazon.d.cts +1 -1
- package/dist/providers/amazon.d.mts +1 -1
- package/dist/providers/amazon.mjs +4 -1482
- package/dist/providers/anthropic.cjs +3 -698
- package/dist/providers/anthropic.d.cts +1 -1
- package/dist/providers/anthropic.d.mts +1 -1
- package/dist/providers/anthropic.mjs +3 -696
- package/dist/providers/azure.cjs +6 -1491
- package/dist/providers/azure.d.cts +1 -1
- package/dist/providers/azure.d.mts +1 -1
- package/dist/providers/azure.mjs +6 -1489
- package/dist/providers/baseten.cjs +3 -279
- package/dist/providers/baseten.d.cts +1 -1
- package/dist/providers/baseten.d.mts +1 -1
- package/dist/providers/baseten.mjs +3 -277
- package/dist/providers/cerebras.cjs +4 -145
- package/dist/providers/cerebras.d.cts +1 -1
- package/dist/providers/cerebras.d.mts +1 -1
- package/dist/providers/cerebras.mjs +4 -143
- package/dist/providers/cloudflare-ai-gateway.cjs +4 -0
- package/dist/providers/cloudflare-ai-gateway.d.cts +7 -0
- package/dist/providers/cloudflare-ai-gateway.d.mts +7 -0
- package/dist/providers/cloudflare-ai-gateway.mjs +4 -0
- package/dist/providers/cloudflare-workers-ai.cjs +4 -0
- package/dist/providers/cloudflare-workers-ai.d.cts +7 -0
- package/dist/providers/cloudflare-workers-ai.d.mts +7 -0
- package/dist/providers/cloudflare-workers-ai.mjs +4 -0
- package/dist/providers/cohere.cjs +5 -525
- package/dist/providers/cohere.d.cts +1 -1
- package/dist/providers/cohere.d.mts +1 -1
- package/dist/providers/cohere.mjs +5 -523
- package/dist/providers/cursor.cjs +1 -856
- package/dist/providers/cursor.d.cts +1 -1
- package/dist/providers/cursor.d.mts +1 -1
- package/dist/providers/cursor.mjs +1 -854
- package/dist/providers/deepseek.cjs +3 -64
- package/dist/providers/deepseek.d.cts +1 -1
- package/dist/providers/deepseek.d.mts +1 -1
- package/dist/providers/deepseek.mjs +3 -62
- package/dist/providers/fireworks.cjs +3 -3794
- package/dist/providers/fireworks.d.cts +1 -1
- package/dist/providers/fireworks.d.mts +1 -1
- package/dist/providers/fireworks.mjs +3 -3792
- package/dist/providers/google.cjs +6 -972
- package/dist/providers/google.d.cts +1 -1
- package/dist/providers/google.d.mts +1 -1
- package/dist/providers/google.mjs +6 -970
- package/dist/providers/groq.cjs +3 -353
- package/dist/providers/groq.d.cts +1 -1
- package/dist/providers/groq.d.mts +1 -1
- package/dist/providers/groq.mjs +3 -351
- package/dist/providers/huggingface.cjs +3 -1022
- package/dist/providers/huggingface.d.cts +1 -1
- package/dist/providers/huggingface.d.mts +1 -1
- package/dist/providers/huggingface.mjs +3 -1020
- package/dist/providers/meta.cjs +3 -481
- package/dist/providers/meta.d.cts +1 -1
- package/dist/providers/meta.d.mts +1 -1
- package/dist/providers/meta.mjs +3 -479
- package/dist/providers/minimax.cjs +3 -156
- package/dist/providers/minimax.d.cts +1 -1
- package/dist/providers/minimax.d.mts +1 -1
- package/dist/providers/minimax.mjs +3 -154
- package/dist/providers/mistral.cjs +3 -1594
- package/dist/providers/mistral.d.cts +1 -1
- package/dist/providers/mistral.d.mts +1 -1
- package/dist/providers/mistral.mjs +3 -1592
- package/dist/providers/moonshot.cjs +3 -138
- package/dist/providers/moonshot.d.cts +1 -1
- package/dist/providers/moonshot.d.mts +1 -1
- package/dist/providers/moonshot.mjs +3 -136
- package/dist/providers/nvidia.cjs +3 -2040
- package/dist/providers/nvidia.d.cts +1 -1
- package/dist/providers/nvidia.d.mts +1 -1
- package/dist/providers/nvidia.mjs +3 -2038
- package/dist/providers/ollama.cjs +3 -326
- package/dist/providers/ollama.d.cts +1 -1
- package/dist/providers/ollama.d.mts +1 -1
- package/dist/providers/ollama.mjs +3 -324
- package/dist/providers/openai.cjs +3 -9079
- package/dist/providers/openai.d.cts +1 -1
- package/dist/providers/openai.d.mts +1 -1
- package/dist/providers/openai.mjs +3 -9077
- package/dist/providers/opencode.cjs +3 -367
- package/dist/providers/opencode.d.cts +1 -1
- package/dist/providers/opencode.d.mts +1 -1
- package/dist/providers/opencode.mjs +3 -365
- package/dist/providers/openrouter.cjs +21 -7835
- package/dist/providers/openrouter.d.cts +1 -1
- package/dist/providers/openrouter.d.mts +1 -1
- package/dist/providers/openrouter.mjs +21 -7833
- package/dist/providers/perplexity.cjs +3 -311
- package/dist/providers/perplexity.d.cts +1 -1
- package/dist/providers/perplexity.d.mts +1 -1
- package/dist/providers/perplexity.mjs +3 -309
- package/dist/providers/qwen.cjs +3 -835
- package/dist/providers/qwen.d.cts +1 -1
- package/dist/providers/qwen.d.mts +1 -1
- package/dist/providers/qwen.mjs +3 -833
- package/dist/providers/together.cjs +4 -360
- package/dist/providers/together.d.cts +1 -1
- package/dist/providers/together.d.mts +1 -1
- package/dist/providers/together.mjs +4 -358
- package/dist/providers/vercel.cjs +3 -4037
- package/dist/providers/vercel.d.cts +1 -1
- package/dist/providers/vercel.d.mts +1 -1
- package/dist/providers/vercel.mjs +3 -4035
- package/dist/providers/vertex.cjs +10 -1072
- package/dist/providers/vertex.d.cts +1 -1
- package/dist/providers/vertex.d.mts +1 -1
- package/dist/providers/vertex.mjs +10 -1070
- package/dist/providers/xai.cjs +3 -301
- package/dist/providers/xai.d.cts +1 -1
- package/dist/providers/xai.d.mts +1 -1
- package/dist/providers/xai.mjs +3 -299
- package/dist/providers/zai.cjs +3 -243
- package/dist/providers/zai.d.cts +1 -1
- package/dist/providers/zai.d.mts +1 -1
- package/dist/providers/zai.mjs +3 -241
- package/dist/types.d.cts +80 -8
- package/dist/types.d.mts +80 -8
- package/package.json +3 -3
- package/dist/providers/cloudflare.cjs +0 -826
- package/dist/providers/cloudflare.d.cts +0 -7
- package/dist/providers/cloudflare.d.mts +0 -7
- package/dist/providers/cloudflare.mjs +0 -824
|
@@ -1,1022 +1,3 @@
|
|
|
1
|
-
Object.defineProperty(exports,
|
|
2
|
-
|
|
3
|
-
const provider = {
|
|
4
|
-
"id": "huggingface",
|
|
5
|
-
"name": "Hugging Face",
|
|
6
|
-
"region": "US",
|
|
7
|
-
"url": "https://huggingface.co",
|
|
8
|
-
"api_url": "https://router.huggingface.co/v1",
|
|
9
|
-
"docs_url": "https://huggingface.co/docs/api-inference",
|
|
10
|
-
"pricing_url": "https://huggingface.co/pricing",
|
|
11
|
-
"icon": "<svg xmlns=\"http://www.w3.org/2000/svg\" viewBox=\"0 0 24 24\" fill=\"currentColor\" fill-rule=\"evenodd\">\n <path d=\"M16.781 3.277c2.997 1.704 4.844 4.851 4.844 8.258 0 .995-.155 1.955-.443 2.857a1.332 1.332 0 011.125.4 1.41 1.41 0 01.2 1.723c.204.165.352.385.428.632l.017.062c.06.222.12.69-.2 1.166.244.37.279.836.093 1.236-.255.57-.893 1.018-2.128 1.5l-.202.078-.131.048c-.478.173-.89.295-1.061.345l-.086.024c-.89.243-1.808.375-2.732.394-1.32 0-2.3-.36-2.923-1.067a9.852 9.852 0 01-3.18.018C9.778 21.647 8.802 22 7.494 22a11.249 11.249 0 01-2.541-.343l-.221-.06-.273-.08a16.574 16.574 0 01-1.175-.405c-1.237-.483-1.875-.93-2.13-1.501-.186-.4-.151-.867.093-1.236a1.42 1.42 0 01-.2-1.166c.069-.273.226-.516.447-.694a1.41 1.41 0 01.2-1.722c.233-.248.557-.391.917-.407l.078-.001a9.385 9.385 0 01-.44-2.85c0-3.407 1.847-6.554 4.844-8.258a9.822 9.822 0 019.687 0zM4.188 14.758c.125.687 2.357 2.35 2.14 2.707-.19.315-.796-.239-.948-.386l-.041-.04-.168-.147c-.561-.479-2.304-1.9-2.74-1.432-.43.46.119.859 1.055 1.42l.784.467.136.083c1.045.643 1.12.84.95 1.113-.188.295-3.07-2.1-3.34-1.083-.27 1.011 2.942 1.304 2.744 2.006-.2.7-2.265-1.324-2.685-.537-.425.79 2.913 1.718 2.94 1.725l.16.04.175.042c1.227.284 3.565.65 4.435-.604.673-.973.64-1.709-.248-2.61l-.057-.057c-.945-.928-1.495-2.288-1.495-2.288l-.017-.058-.025-.072c-.082-.22-.284-.639-.63-.584-.46.073-.798 1.21.12 1.933l.05.038c.977.721-.195 1.21-.573.534l-.058-.104-.143-.25c-.463-.799-1.282-2.111-1.739-2.397-.532-.332-.907-.148-.782.541zm14.842-.541c-.533.335-1.563 2.074-1.94 2.751a.613.613 0 01-.687.302.436.436 0 01-.176-.098.303.303 0 01-.049-.06l-.014-.028-.008-.02-.007-.019-.003-.013-.003-.017a.289.289 0 01-.004-.048c0-.12.071-.266.25-.427.026-.024.054-.047.084-.07l.047-.036c.022-.016.043-.032.063-.049.883-.71.573-1.81.131-1.917l-.031-.006-.056-.004a.368.368 0 00-.062.006l-.028.005-.042.014-.039.017-.028.015-.028.019-.036.027-.023.02c-.173.158-.273.428-.31.542l-.016.054s-.53 1.309-1.439 2.234l-.054.054c-.365.358-.596.69-.702 1.018-.143.437-.066.868.21 1.353.055.097.117.195.187.296.882 1.275 3.282.876 4.494.59l.286-.07.25-.074c.276-.084.736-.233 1.2-.42l.188-.077.065-.028.064-.028.124-.056.081-.038c.529-.252.964-.543.994-.827l.001-.036a.299.299 0 00-.037-.139c-.094-.176-.271-.212-.491-.168l-.045.01c-.044.01-.09.024-.136.04l-.097.035-.054.022c-.559.23-1.238.705-1.607.745h.006a.452.452 0 01-.05.003h-.024l-.024-.003-.023-.005c-.068-.016-.116-.06-.14-.142a.22.22 0 01-.005-.1c.062-.345.958-.595 1.713-.91l.066-.028c.528-.224.97-.483.985-.832v-.04a.47.47 0 00-.016-.098c-.048-.18-.175-.251-.36-.251-.785 0-2.55 1.36-2.92 1.36-.025 0-.048-.007-.058-.024a.6.6 0 01-.046-.088c-.1-.238.068-.462 1.06-1.066l.209-.126c.538-.32 1.01-.588 1.341-.831.29-.212.475-.406.503-.6l.003-.028c.008-.113-.038-.227-.147-.344a.266.266 0 00-.07-.054l-.034-.015-.013-.005a.403.403 0 00-.13-.02c-.162 0-.369.07-.595.18-.637.313-1.431.952-1.826 1.285l-.249.215-.033.033c-.08.078-.288.27-.493.386l-.071.037-.041.019a.535.535 0 01-.122.036h.005a.346.346 0 01-.031.003l.01-.001-.013.001c-.079.005-.145-.021-.19-.095a.113.113 0 01-.014-.065c.027-.465 2.034-1.991 2.152-2.642l.009-.048c.1-.65-.271-.817-.791-.493zM11.938 2.984c-4.798 0-8.688 3.829-8.688 8.55 0 .692.083 1.364.24 2.008l.008-.009c.252-.298.612-.46 1.017-.46.355.008.699.117.993.312.22.14.465.384.715.694.261-.372.69-.598 1.15-.605.852 0 1.367.728 1.562 1.383l.047.105.06.127c.192.396.595 1.139 1.143 1.68 1.06 1.04 1.324 2.115.8 3.266a8.865 8.865 0 002.024-.014c-.505-1.12-.26-2.17.74-3.186l.066-.066c.695-.684 1.157-1.69 1.252-1.912.195-.655.708-1.383 1.56-1.383.46.007.889.233 1.15.605.25-.31.495-.553.718-.694a1.87 1.87 0 01.99-.312c.357 0 .682.126.925.36.14-.61.215-1.245.215-1.898 0-4.722-3.89-8.55-8.687-8.55zm1.857 8.926l.439-.212c.553-.264.89-.383.89.152 0 1.093-.771 3.208-3.155 3.262h-.184c-2.325-.052-3.116-2.06-3.156-3.175l-.001-.087c0-1.107 1.452.586 3.25.586.716 0 1.379-.272 1.917-.526zm4.017-3.143c.45 0 .813.358.813.8 0 .441-.364.8-.813.8a.806.806 0 01-.812-.8c0-.442.364-.8.812-.8zm-11.624 0c.448 0 .812.358.812.8 0 .441-.364.8-.812.8a.806.806 0 01-.813-.8c0-.442.364-.8.813-.8zm7.79-.841c.32-.384.846-.54 1.33-.394.483.146.83.564.878 1.06.048.495-.212.97-.659 1.203-.322.168-.447-.477-.767-.585l.002-.003c-.287-.098-.772.362-.925.079a1.215 1.215 0 01.14-1.36zm-4.323 0c.322.384.377.92.14 1.36-.152.283-.64-.177-.925-.079l.003.003c-.108.036-.194.134-.273.24l-.118.165c-.11.15-.22.262-.377.18a1.226 1.226 0 01-.658-1.204c.048-.495.395-.913.878-1.059a1.262 1.262 0 011.33.394z\"/>\n</svg>",
|
|
12
|
-
"models": [
|
|
13
|
-
{
|
|
14
|
-
"id": "allenai/Olmo-3-7B-Instruct",
|
|
15
|
-
"name": "Olmo-3-7B-Instruct",
|
|
16
|
-
"created_by": "allenai",
|
|
17
|
-
"source": "official",
|
|
18
|
-
"last_updated": "2026-03-21",
|
|
19
|
-
"capabilities": { "streaming": true }
|
|
20
|
-
},
|
|
21
|
-
{
|
|
22
|
-
"id": "allenai/Olmo-3.1-32B-Instruct",
|
|
23
|
-
"name": "Olmo-3.1-32B-Instruct",
|
|
24
|
-
"created_by": "allenai",
|
|
25
|
-
"source": "official",
|
|
26
|
-
"last_updated": "2026-03-21",
|
|
27
|
-
"capabilities": { "streaming": true }
|
|
28
|
-
},
|
|
29
|
-
{
|
|
30
|
-
"id": "argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1",
|
|
31
|
-
"name": "Llama-3.2-1B-Instruct-APIGen-FC-v0.1",
|
|
32
|
-
"created_by": "argilla",
|
|
33
|
-
"source": "official",
|
|
34
|
-
"last_updated": "2026-03-21",
|
|
35
|
-
"capabilities": { "streaming": true }
|
|
36
|
-
},
|
|
37
|
-
{
|
|
38
|
-
"id": "argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC",
|
|
39
|
-
"name": "Llama-3.2-1B-Instruct-v2-FC",
|
|
40
|
-
"created_by": "argilla-warehouse",
|
|
41
|
-
"source": "official",
|
|
42
|
-
"last_updated": "2026-03-21",
|
|
43
|
-
"capabilities": { "streaming": true }
|
|
44
|
-
},
|
|
45
|
-
{
|
|
46
|
-
"id": "DeepMount00/Llama-3-8b-Ita",
|
|
47
|
-
"name": "Llama-3-8b-Ita",
|
|
48
|
-
"created_by": "deepmount00",
|
|
49
|
-
"source": "official",
|
|
50
|
-
"last_updated": "2026-03-21",
|
|
51
|
-
"capabilities": { "streaming": true }
|
|
52
|
-
},
|
|
53
|
-
{
|
|
54
|
-
"id": "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B",
|
|
55
|
-
"name": "DeepSeek-R1-0528-Qwen3-8B",
|
|
56
|
-
"created_by": "deepseek",
|
|
57
|
-
"source": "official",
|
|
58
|
-
"last_updated": "2026-03-21",
|
|
59
|
-
"capabilities": { "streaming": true }
|
|
60
|
-
},
|
|
61
|
-
{
|
|
62
|
-
"id": "Efficient-Large-Model/gemma-2-2b-it",
|
|
63
|
-
"name": "gemma-2-2b-it",
|
|
64
|
-
"created_by": "efficient-large-model",
|
|
65
|
-
"source": "official",
|
|
66
|
-
"last_updated": "2026-03-21",
|
|
67
|
-
"family": "gemma-2",
|
|
68
|
-
"capabilities": { "streaming": true }
|
|
69
|
-
},
|
|
70
|
-
{
|
|
71
|
-
"id": "EssentialAI/rnj-1-instruct",
|
|
72
|
-
"name": "rnj-1-instruct",
|
|
73
|
-
"created_by": "essentialai",
|
|
74
|
-
"source": "official",
|
|
75
|
-
"last_updated": "2026-03-21",
|
|
76
|
-
"capabilities": { "streaming": true }
|
|
77
|
-
},
|
|
78
|
-
{
|
|
79
|
-
"id": "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3",
|
|
80
|
-
"name": "Meta-Llama-3-8B-Instruct-abliterated-v3",
|
|
81
|
-
"created_by": "failspy",
|
|
82
|
-
"source": "official",
|
|
83
|
-
"last_updated": "2026-03-21",
|
|
84
|
-
"capabilities": { "streaming": true }
|
|
85
|
-
},
|
|
86
|
-
{
|
|
87
|
-
"id": "Featherless-Chat-Models/Mistral-7B-Instruct-v0.2",
|
|
88
|
-
"name": "Mistral-7B-Instruct-v0.2",
|
|
89
|
-
"created_by": "featherless-chat-models",
|
|
90
|
-
"source": "official",
|
|
91
|
-
"last_updated": "2026-03-21",
|
|
92
|
-
"capabilities": { "streaming": true }
|
|
93
|
-
},
|
|
94
|
-
{
|
|
95
|
-
"id": "FlagAlpha/Llama3-Chinese-8B-Instruct",
|
|
96
|
-
"name": "Llama3-Chinese-8B-Instruct",
|
|
97
|
-
"created_by": "flagalpha",
|
|
98
|
-
"source": "official",
|
|
99
|
-
"last_updated": "2026-03-21",
|
|
100
|
-
"capabilities": { "streaming": true }
|
|
101
|
-
},
|
|
102
|
-
{
|
|
103
|
-
"id": "google/gemma-2-2b-jpn-it",
|
|
104
|
-
"name": "gemma-2-2b-jpn-it",
|
|
105
|
-
"created_by": "google",
|
|
106
|
-
"source": "official",
|
|
107
|
-
"last_updated": "2026-03-21",
|
|
108
|
-
"family": "gemma-2",
|
|
109
|
-
"capabilities": { "streaming": true }
|
|
110
|
-
},
|
|
111
|
-
{
|
|
112
|
-
"id": "google/gemma-2-9b-it",
|
|
113
|
-
"name": "gemma-2-9b-it",
|
|
114
|
-
"created_by": "google",
|
|
115
|
-
"source": "official",
|
|
116
|
-
"last_updated": "2026-03-21",
|
|
117
|
-
"family": "gemma-2",
|
|
118
|
-
"capabilities": { "streaming": true }
|
|
119
|
-
},
|
|
120
|
-
{
|
|
121
|
-
"id": "GraySwanAI/Llama-3-8B-Instruct-RR",
|
|
122
|
-
"name": "Llama-3-8B-Instruct-RR",
|
|
123
|
-
"created_by": "grayswanai",
|
|
124
|
-
"source": "official",
|
|
125
|
-
"last_updated": "2026-03-21",
|
|
126
|
-
"capabilities": { "streaming": true }
|
|
127
|
-
},
|
|
128
|
-
{
|
|
129
|
-
"id": "GritLM/GritLM-7B",
|
|
130
|
-
"name": "GritLM-7B",
|
|
131
|
-
"created_by": "gritlm",
|
|
132
|
-
"source": "official",
|
|
133
|
-
"last_updated": "2026-03-21",
|
|
134
|
-
"capabilities": { "streaming": true }
|
|
135
|
-
},
|
|
136
|
-
{
|
|
137
|
-
"id": "huihui-ai/Qwen2.5-72B-Instruct-abliterated",
|
|
138
|
-
"name": "Qwen2.5-72B-Instruct-abliterated",
|
|
139
|
-
"created_by": "huihui-ai",
|
|
140
|
-
"source": "official",
|
|
141
|
-
"last_updated": "2026-03-21",
|
|
142
|
-
"capabilities": { "streaming": true }
|
|
143
|
-
},
|
|
144
|
-
{
|
|
145
|
-
"id": "instruction-pretrain/finance-Llama3-8B",
|
|
146
|
-
"name": "finance-Llama3-8B",
|
|
147
|
-
"created_by": "instruction-pretrain",
|
|
148
|
-
"source": "official",
|
|
149
|
-
"last_updated": "2026-03-21",
|
|
150
|
-
"capabilities": { "streaming": true }
|
|
151
|
-
},
|
|
152
|
-
{
|
|
153
|
-
"id": "Intel/neural-chat-7b-v3-3",
|
|
154
|
-
"name": "neural-chat-7b-v3-3",
|
|
155
|
-
"created_by": "intel",
|
|
156
|
-
"source": "official",
|
|
157
|
-
"last_updated": "2026-03-21",
|
|
158
|
-
"capabilities": { "streaming": true }
|
|
159
|
-
},
|
|
160
|
-
{
|
|
161
|
-
"id": "invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp",
|
|
162
|
-
"name": "Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp",
|
|
163
|
-
"created_by": "invalid-coder",
|
|
164
|
-
"source": "official",
|
|
165
|
-
"last_updated": "2026-03-21",
|
|
166
|
-
"capabilities": { "streaming": true }
|
|
167
|
-
},
|
|
168
|
-
{
|
|
169
|
-
"id": "m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix",
|
|
170
|
-
"name": "Mistral-Nemo-Instruct-2407-lenient-chatfix",
|
|
171
|
-
"created_by": "m8than",
|
|
172
|
-
"source": "official",
|
|
173
|
-
"last_updated": "2026-03-21",
|
|
174
|
-
"capabilities": { "streaming": true }
|
|
175
|
-
},
|
|
176
|
-
{
|
|
177
|
-
"id": "meta-llama/Llama-2-13b-chat-hf",
|
|
178
|
-
"name": "Llama-2-13b-chat-hf",
|
|
179
|
-
"created_by": "meta",
|
|
180
|
-
"source": "official",
|
|
181
|
-
"last_updated": "2026-03-21",
|
|
182
|
-
"capabilities": { "streaming": true }
|
|
183
|
-
},
|
|
184
|
-
{
|
|
185
|
-
"id": "meta-llama/Llama-3.1-70B-Instruct",
|
|
186
|
-
"name": "Llama-3.1-70B-Instruct",
|
|
187
|
-
"created_by": "meta",
|
|
188
|
-
"source": "official",
|
|
189
|
-
"last_updated": "2026-03-21",
|
|
190
|
-
"capabilities": { "streaming": true }
|
|
191
|
-
},
|
|
192
|
-
{
|
|
193
|
-
"id": "meta-llama/Llama-3.1-8B-Instruct",
|
|
194
|
-
"name": "Llama-3.1-8B-Instruct",
|
|
195
|
-
"created_by": "meta",
|
|
196
|
-
"source": "official",
|
|
197
|
-
"last_updated": "2026-03-21",
|
|
198
|
-
"capabilities": { "streaming": true }
|
|
199
|
-
},
|
|
200
|
-
{
|
|
201
|
-
"id": "meta-llama/Llama-3.2-1B-Instruct",
|
|
202
|
-
"name": "Llama-3.2-1B-Instruct",
|
|
203
|
-
"created_by": "meta",
|
|
204
|
-
"source": "official",
|
|
205
|
-
"last_updated": "2026-03-21",
|
|
206
|
-
"capabilities": { "streaming": true }
|
|
207
|
-
},
|
|
208
|
-
{
|
|
209
|
-
"id": "meta-llama/Llama-3.3-70B-Instruct",
|
|
210
|
-
"name": "Llama-3.3-70B-Instruct",
|
|
211
|
-
"created_by": "meta",
|
|
212
|
-
"source": "official",
|
|
213
|
-
"last_updated": "2026-03-21",
|
|
214
|
-
"capabilities": { "streaming": true }
|
|
215
|
-
},
|
|
216
|
-
{
|
|
217
|
-
"id": "meta-llama/Meta-Llama-3-70B-Instruct",
|
|
218
|
-
"name": "Meta-Llama-3-70B-Instruct",
|
|
219
|
-
"created_by": "meta",
|
|
220
|
-
"source": "official",
|
|
221
|
-
"last_updated": "2026-03-21",
|
|
222
|
-
"capabilities": { "streaming": true }
|
|
223
|
-
},
|
|
224
|
-
{
|
|
225
|
-
"id": "meta-llama/Meta-Llama-3-8B-Instruct",
|
|
226
|
-
"name": "Meta-Llama-3-8B-Instruct",
|
|
227
|
-
"created_by": "meta",
|
|
228
|
-
"source": "official",
|
|
229
|
-
"last_updated": "2026-03-21",
|
|
230
|
-
"capabilities": { "streaming": true }
|
|
231
|
-
},
|
|
232
|
-
{
|
|
233
|
-
"id": "mistralai/Mistral-7B-Instruct-v0.2",
|
|
234
|
-
"name": "Mistral-7B-Instruct-v0.2",
|
|
235
|
-
"created_by": "mistral",
|
|
236
|
-
"source": "official",
|
|
237
|
-
"last_updated": "2026-03-21",
|
|
238
|
-
"capabilities": { "streaming": true }
|
|
239
|
-
},
|
|
240
|
-
{
|
|
241
|
-
"id": "mlabonne/NeuralDaredevil-8B-abliterated",
|
|
242
|
-
"name": "NeuralDaredevil-8B-abliterated",
|
|
243
|
-
"created_by": "mlabonne",
|
|
244
|
-
"source": "official",
|
|
245
|
-
"last_updated": "2026-03-21",
|
|
246
|
-
"capabilities": { "streaming": true }
|
|
247
|
-
},
|
|
248
|
-
{
|
|
249
|
-
"id": "moonshotai/Kimi-K2-Instruct-0905",
|
|
250
|
-
"name": "Kimi-K2-Instruct-0905",
|
|
251
|
-
"created_by": "moonshot",
|
|
252
|
-
"source": "official",
|
|
253
|
-
"last_updated": "2026-03-21",
|
|
254
|
-
"capabilities": { "streaming": true }
|
|
255
|
-
},
|
|
256
|
-
{
|
|
257
|
-
"id": "moonshotai/Kimi-K2-Instruct",
|
|
258
|
-
"name": "Kimi-K2-Instruct",
|
|
259
|
-
"created_by": "moonshot",
|
|
260
|
-
"source": "official",
|
|
261
|
-
"last_updated": "2026-03-21",
|
|
262
|
-
"capabilities": { "streaming": true }
|
|
263
|
-
},
|
|
264
|
-
{
|
|
265
|
-
"id": "moonshotai/Kimi-K2-Thinking",
|
|
266
|
-
"name": "Kimi-K2-Thinking",
|
|
267
|
-
"created_by": "moonshot",
|
|
268
|
-
"source": "official",
|
|
269
|
-
"last_updated": "2026-03-21",
|
|
270
|
-
"capabilities": { "streaming": true }
|
|
271
|
-
},
|
|
272
|
-
{
|
|
273
|
-
"id": "NousResearch/Meta-Llama-3-70B-Instruct",
|
|
274
|
-
"name": "Meta-Llama-3-70B-Instruct",
|
|
275
|
-
"created_by": "nousresearch",
|
|
276
|
-
"source": "official",
|
|
277
|
-
"last_updated": "2026-03-21",
|
|
278
|
-
"capabilities": { "streaming": true }
|
|
279
|
-
},
|
|
280
|
-
{
|
|
281
|
-
"id": "NousResearch/Meta-Llama-3-8B-Instruct",
|
|
282
|
-
"name": "Meta-Llama-3-8B-Instruct",
|
|
283
|
-
"created_by": "nousresearch",
|
|
284
|
-
"source": "official",
|
|
285
|
-
"last_updated": "2026-03-21",
|
|
286
|
-
"capabilities": { "streaming": true }
|
|
287
|
-
},
|
|
288
|
-
{
|
|
289
|
-
"id": "NousResearch/Meta-Llama-3.1-8B-Instruct",
|
|
290
|
-
"name": "Meta-Llama-3.1-8B-Instruct",
|
|
291
|
-
"created_by": "nousresearch",
|
|
292
|
-
"source": "official",
|
|
293
|
-
"last_updated": "2026-03-21",
|
|
294
|
-
"capabilities": { "streaming": true }
|
|
295
|
-
},
|
|
296
|
-
{
|
|
297
|
-
"id": "nvidia/Llama-3.1-8B-Instruct-FP8",
|
|
298
|
-
"name": "Llama-3.1-8B-Instruct-FP8",
|
|
299
|
-
"created_by": "nvidia",
|
|
300
|
-
"source": "official",
|
|
301
|
-
"last_updated": "2026-03-21",
|
|
302
|
-
"capabilities": { "streaming": true }
|
|
303
|
-
},
|
|
304
|
-
{
|
|
305
|
-
"id": "openai/gpt-oss-120b",
|
|
306
|
-
"name": "gpt-oss-120b",
|
|
307
|
-
"created_by": "openai",
|
|
308
|
-
"source": "official",
|
|
309
|
-
"last_updated": "2026-03-21",
|
|
310
|
-
"family": "gpt-oss",
|
|
311
|
-
"capabilities": { "streaming": true }
|
|
312
|
-
},
|
|
313
|
-
{
|
|
314
|
-
"id": "openai/gpt-oss-20b",
|
|
315
|
-
"name": "gpt-oss-20b",
|
|
316
|
-
"created_by": "openai",
|
|
317
|
-
"source": "official",
|
|
318
|
-
"last_updated": "2026-03-21",
|
|
319
|
-
"family": "gpt-oss",
|
|
320
|
-
"capabilities": { "streaming": true }
|
|
321
|
-
},
|
|
322
|
-
{
|
|
323
|
-
"id": "openai/gpt-oss-safeguard-20b",
|
|
324
|
-
"name": "gpt-oss-safeguard-20b",
|
|
325
|
-
"created_by": "openai",
|
|
326
|
-
"source": "official",
|
|
327
|
-
"last_updated": "2026-03-21",
|
|
328
|
-
"family": "gpt-oss",
|
|
329
|
-
"capabilities": { "streaming": true },
|
|
330
|
-
"model_type": "moderation"
|
|
331
|
-
},
|
|
332
|
-
{
|
|
333
|
-
"id": "openchat/openchat-3.6-8b-20240522",
|
|
334
|
-
"name": "openchat-3.6-8b-20240522",
|
|
335
|
-
"created_by": "openchat",
|
|
336
|
-
"source": "official",
|
|
337
|
-
"last_updated": "2026-03-21",
|
|
338
|
-
"alias": "openchat/openchat-3.6-8b",
|
|
339
|
-
"capabilities": { "streaming": true }
|
|
340
|
-
},
|
|
341
|
-
{
|
|
342
|
-
"id": "openchat/openchat-3.6-8b",
|
|
343
|
-
"name": "openchat-3.6-8b-20240522",
|
|
344
|
-
"created_by": "openchat",
|
|
345
|
-
"source": "official",
|
|
346
|
-
"last_updated": "2026-03-21",
|
|
347
|
-
"capabilities": { "streaming": true },
|
|
348
|
-
"snapshots": ["openchat/openchat-3.6-8b-20240522"]
|
|
349
|
-
},
|
|
350
|
-
{
|
|
351
|
-
"id": "Orion-zhen/Qwen2.5-7B-Instruct-Uncensored",
|
|
352
|
-
"name": "Qwen2.5-7B-Instruct-Uncensored",
|
|
353
|
-
"created_by": "orion-zhen",
|
|
354
|
-
"source": "official",
|
|
355
|
-
"last_updated": "2026-03-21",
|
|
356
|
-
"capabilities": { "streaming": true }
|
|
357
|
-
},
|
|
358
|
-
{
|
|
359
|
-
"id": "PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1",
|
|
360
|
-
"name": "Llama-3-Patronus-Lynx-8B-Instruct-v1.1",
|
|
361
|
-
"created_by": "patronusai",
|
|
362
|
-
"source": "official",
|
|
363
|
-
"last_updated": "2026-03-21",
|
|
364
|
-
"capabilities": { "streaming": true }
|
|
365
|
-
},
|
|
366
|
-
{
|
|
367
|
-
"id": "Qwen/Qwen1.5-1.8B-Chat",
|
|
368
|
-
"name": "Qwen1.5-1.8B-Chat",
|
|
369
|
-
"created_by": "qwen",
|
|
370
|
-
"source": "official",
|
|
371
|
-
"last_updated": "2026-03-21",
|
|
372
|
-
"capabilities": { "streaming": true }
|
|
373
|
-
},
|
|
374
|
-
{
|
|
375
|
-
"id": "Qwen/Qwen2-1.5B-Instruct",
|
|
376
|
-
"name": "Qwen2-1.5B-Instruct",
|
|
377
|
-
"created_by": "qwen",
|
|
378
|
-
"source": "official",
|
|
379
|
-
"last_updated": "2026-03-21",
|
|
380
|
-
"capabilities": { "streaming": true }
|
|
381
|
-
},
|
|
382
|
-
{
|
|
383
|
-
"id": "Qwen/Qwen2-72B-Instruct",
|
|
384
|
-
"name": "Qwen2-72B-Instruct",
|
|
385
|
-
"created_by": "qwen",
|
|
386
|
-
"source": "official",
|
|
387
|
-
"last_updated": "2026-03-21",
|
|
388
|
-
"capabilities": { "streaming": true }
|
|
389
|
-
},
|
|
390
|
-
{
|
|
391
|
-
"id": "Qwen/Qwen2-7B-Instruct",
|
|
392
|
-
"name": "Qwen2-7B-Instruct",
|
|
393
|
-
"created_by": "qwen",
|
|
394
|
-
"source": "official",
|
|
395
|
-
"last_updated": "2026-03-21",
|
|
396
|
-
"capabilities": { "streaming": true }
|
|
397
|
-
},
|
|
398
|
-
{
|
|
399
|
-
"id": "Qwen/Qwen2.5-1.5B-Instruct",
|
|
400
|
-
"name": "Qwen2.5-1.5B-Instruct",
|
|
401
|
-
"created_by": "qwen",
|
|
402
|
-
"source": "official",
|
|
403
|
-
"last_updated": "2026-03-21",
|
|
404
|
-
"capabilities": { "streaming": true }
|
|
405
|
-
},
|
|
406
|
-
{
|
|
407
|
-
"id": "Qwen/Qwen2.5-14B-Instruct-1M",
|
|
408
|
-
"name": "Qwen2.5-14B-Instruct-1M",
|
|
409
|
-
"created_by": "qwen",
|
|
410
|
-
"source": "official",
|
|
411
|
-
"last_updated": "2026-03-21",
|
|
412
|
-
"capabilities": { "streaming": true }
|
|
413
|
-
},
|
|
414
|
-
{
|
|
415
|
-
"id": "Qwen/Qwen2.5-14B-Instruct",
|
|
416
|
-
"name": "Qwen2.5-14B-Instruct",
|
|
417
|
-
"created_by": "qwen",
|
|
418
|
-
"source": "official",
|
|
419
|
-
"last_updated": "2026-03-21",
|
|
420
|
-
"capabilities": { "streaming": true }
|
|
421
|
-
},
|
|
422
|
-
{
|
|
423
|
-
"id": "Qwen/Qwen2.5-32B-Instruct",
|
|
424
|
-
"name": "Qwen2.5-32B-Instruct",
|
|
425
|
-
"created_by": "qwen",
|
|
426
|
-
"source": "official",
|
|
427
|
-
"last_updated": "2026-03-21",
|
|
428
|
-
"capabilities": { "streaming": true }
|
|
429
|
-
},
|
|
430
|
-
{
|
|
431
|
-
"id": "Qwen/Qwen2.5-72B-Instruct",
|
|
432
|
-
"name": "Qwen2.5-72B-Instruct",
|
|
433
|
-
"created_by": "qwen",
|
|
434
|
-
"source": "official",
|
|
435
|
-
"last_updated": "2026-03-21",
|
|
436
|
-
"capabilities": { "streaming": true }
|
|
437
|
-
},
|
|
438
|
-
{
|
|
439
|
-
"id": "Qwen/Qwen2.5-7B-Instruct-1M",
|
|
440
|
-
"name": "Qwen2.5-7B-Instruct-1M",
|
|
441
|
-
"created_by": "qwen",
|
|
442
|
-
"source": "official",
|
|
443
|
-
"last_updated": "2026-03-21",
|
|
444
|
-
"capabilities": { "streaming": true }
|
|
445
|
-
},
|
|
446
|
-
{
|
|
447
|
-
"id": "Qwen/Qwen2.5-7B-Instruct",
|
|
448
|
-
"name": "Qwen2.5-7B-Instruct",
|
|
449
|
-
"created_by": "qwen",
|
|
450
|
-
"source": "official",
|
|
451
|
-
"last_updated": "2026-03-21",
|
|
452
|
-
"capabilities": { "streaming": true }
|
|
453
|
-
},
|
|
454
|
-
{
|
|
455
|
-
"id": "Qwen/Qwen2.5-Coder-1.5B-Instruct",
|
|
456
|
-
"name": "Qwen2.5-Coder-1.5B-Instruct",
|
|
457
|
-
"created_by": "qwen",
|
|
458
|
-
"source": "official",
|
|
459
|
-
"last_updated": "2026-03-21",
|
|
460
|
-
"capabilities": { "streaming": true }
|
|
461
|
-
},
|
|
462
|
-
{
|
|
463
|
-
"id": "Qwen/Qwen2.5-Coder-1.5B",
|
|
464
|
-
"name": "Qwen2.5-Coder-1.5B",
|
|
465
|
-
"created_by": "qwen",
|
|
466
|
-
"source": "official",
|
|
467
|
-
"last_updated": "2026-03-21",
|
|
468
|
-
"capabilities": { "streaming": true }
|
|
469
|
-
},
|
|
470
|
-
{
|
|
471
|
-
"id": "Qwen/Qwen2.5-Coder-14B-Instruct",
|
|
472
|
-
"name": "Qwen2.5-Coder-14B-Instruct",
|
|
473
|
-
"created_by": "qwen",
|
|
474
|
-
"source": "official",
|
|
475
|
-
"last_updated": "2026-03-21",
|
|
476
|
-
"capabilities": { "streaming": true }
|
|
477
|
-
},
|
|
478
|
-
{
|
|
479
|
-
"id": "Qwen/Qwen2.5-Coder-14B",
|
|
480
|
-
"name": "Qwen2.5-Coder-14B",
|
|
481
|
-
"created_by": "qwen",
|
|
482
|
-
"source": "official",
|
|
483
|
-
"last_updated": "2026-03-21",
|
|
484
|
-
"capabilities": { "streaming": true }
|
|
485
|
-
},
|
|
486
|
-
{
|
|
487
|
-
"id": "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
488
|
-
"name": "Qwen2.5-Coder-32B-Instruct",
|
|
489
|
-
"created_by": "qwen",
|
|
490
|
-
"source": "official",
|
|
491
|
-
"last_updated": "2026-03-21",
|
|
492
|
-
"capabilities": { "streaming": true }
|
|
493
|
-
},
|
|
494
|
-
{
|
|
495
|
-
"id": "Qwen/Qwen2.5-Coder-32B",
|
|
496
|
-
"name": "Qwen2.5-Coder-32B",
|
|
497
|
-
"created_by": "qwen",
|
|
498
|
-
"source": "official",
|
|
499
|
-
"last_updated": "2026-03-21",
|
|
500
|
-
"capabilities": { "streaming": true }
|
|
501
|
-
},
|
|
502
|
-
{
|
|
503
|
-
"id": "Qwen/Qwen2.5-Coder-3B-Instruct",
|
|
504
|
-
"name": "Qwen2.5-Coder-3B-Instruct",
|
|
505
|
-
"created_by": "qwen",
|
|
506
|
-
"source": "official",
|
|
507
|
-
"last_updated": "2026-03-21",
|
|
508
|
-
"capabilities": { "streaming": true }
|
|
509
|
-
},
|
|
510
|
-
{
|
|
511
|
-
"id": "Qwen/Qwen2.5-Coder-7B-Instruct",
|
|
512
|
-
"name": "Qwen2.5-Coder-7B-Instruct",
|
|
513
|
-
"created_by": "qwen",
|
|
514
|
-
"source": "official",
|
|
515
|
-
"last_updated": "2026-03-21",
|
|
516
|
-
"capabilities": { "streaming": true }
|
|
517
|
-
},
|
|
518
|
-
{
|
|
519
|
-
"id": "Qwen/Qwen2.5-Coder-7B",
|
|
520
|
-
"name": "Qwen2.5-Coder-7B",
|
|
521
|
-
"created_by": "qwen",
|
|
522
|
-
"source": "official",
|
|
523
|
-
"last_updated": "2026-03-21",
|
|
524
|
-
"capabilities": { "streaming": true }
|
|
525
|
-
},
|
|
526
|
-
{
|
|
527
|
-
"id": "Qwen/Qwen2.5-Math-1.5B-Instruct",
|
|
528
|
-
"name": "Qwen2.5-Math-1.5B-Instruct",
|
|
529
|
-
"created_by": "qwen",
|
|
530
|
-
"source": "official",
|
|
531
|
-
"last_updated": "2026-03-21",
|
|
532
|
-
"capabilities": { "streaming": true }
|
|
533
|
-
},
|
|
534
|
-
{
|
|
535
|
-
"id": "Qwen/Qwen2.5-Math-7B-Instruct",
|
|
536
|
-
"name": "Qwen2.5-Math-7B-Instruct",
|
|
537
|
-
"created_by": "qwen",
|
|
538
|
-
"source": "official",
|
|
539
|
-
"last_updated": "2026-03-21",
|
|
540
|
-
"capabilities": { "streaming": true }
|
|
541
|
-
},
|
|
542
|
-
{
|
|
543
|
-
"id": "Qwen/Qwen3-1.7B",
|
|
544
|
-
"name": "Qwen3-1.7B",
|
|
545
|
-
"created_by": "qwen",
|
|
546
|
-
"source": "official",
|
|
547
|
-
"last_updated": "2026-03-21",
|
|
548
|
-
"capabilities": { "streaming": true }
|
|
549
|
-
},
|
|
550
|
-
{
|
|
551
|
-
"id": "Qwen/Qwen3-14B-Base",
|
|
552
|
-
"name": "Qwen3-14B-Base",
|
|
553
|
-
"created_by": "qwen",
|
|
554
|
-
"source": "official",
|
|
555
|
-
"last_updated": "2026-03-21",
|
|
556
|
-
"capabilities": { "streaming": true }
|
|
557
|
-
},
|
|
558
|
-
{
|
|
559
|
-
"id": "Qwen/Qwen3-14B",
|
|
560
|
-
"name": "Qwen3-14B",
|
|
561
|
-
"created_by": "qwen",
|
|
562
|
-
"source": "official",
|
|
563
|
-
"last_updated": "2026-03-21",
|
|
564
|
-
"capabilities": { "streaming": true }
|
|
565
|
-
},
|
|
566
|
-
{
|
|
567
|
-
"id": "Qwen/Qwen3-235B-A22B-Instruct-2507",
|
|
568
|
-
"name": "Qwen3-235B-A22B-Instruct-2507",
|
|
569
|
-
"created_by": "qwen",
|
|
570
|
-
"source": "official",
|
|
571
|
-
"last_updated": "2026-03-21",
|
|
572
|
-
"alias": "Qwen/Qwen3-235B-A22B-Instruct",
|
|
573
|
-
"capabilities": { "streaming": true }
|
|
574
|
-
},
|
|
575
|
-
{
|
|
576
|
-
"id": "Qwen/Qwen3-235B-A22B-Instruct",
|
|
577
|
-
"name": "Qwen3-235B-A22B-Instruct-2507",
|
|
578
|
-
"created_by": "qwen",
|
|
579
|
-
"source": "official",
|
|
580
|
-
"last_updated": "2026-03-21",
|
|
581
|
-
"capabilities": { "streaming": true },
|
|
582
|
-
"snapshots": ["Qwen/Qwen3-235B-A22B-Instruct-2507"]
|
|
583
|
-
},
|
|
584
|
-
{
|
|
585
|
-
"id": "Qwen/Qwen3-235B-A22B-Thinking-2507",
|
|
586
|
-
"name": "Qwen3-235B-A22B-Thinking-2507",
|
|
587
|
-
"created_by": "qwen",
|
|
588
|
-
"source": "official",
|
|
589
|
-
"last_updated": "2026-03-21",
|
|
590
|
-
"alias": "Qwen/Qwen3-235B-A22B-Thinking",
|
|
591
|
-
"capabilities": { "streaming": true }
|
|
592
|
-
},
|
|
593
|
-
{
|
|
594
|
-
"id": "Qwen/Qwen3-235B-A22B-Thinking",
|
|
595
|
-
"name": "Qwen3-235B-A22B-Thinking-2507",
|
|
596
|
-
"created_by": "qwen",
|
|
597
|
-
"source": "official",
|
|
598
|
-
"last_updated": "2026-03-21",
|
|
599
|
-
"capabilities": { "streaming": true },
|
|
600
|
-
"snapshots": ["Qwen/Qwen3-235B-A22B-Thinking-2507"]
|
|
601
|
-
},
|
|
602
|
-
{
|
|
603
|
-
"id": "Qwen/Qwen3-235B-A22B",
|
|
604
|
-
"name": "Qwen3-235B-A22B",
|
|
605
|
-
"created_by": "qwen",
|
|
606
|
-
"source": "official",
|
|
607
|
-
"last_updated": "2026-03-21",
|
|
608
|
-
"capabilities": { "streaming": true }
|
|
609
|
-
},
|
|
610
|
-
{
|
|
611
|
-
"id": "Qwen/Qwen3-30B-A3B-Instruct-2507",
|
|
612
|
-
"name": "Qwen3-30B-A3B-Instruct-2507",
|
|
613
|
-
"created_by": "qwen",
|
|
614
|
-
"source": "official",
|
|
615
|
-
"last_updated": "2026-03-21",
|
|
616
|
-
"alias": "Qwen/Qwen3-30B-A3B-Instruct",
|
|
617
|
-
"capabilities": { "streaming": true }
|
|
618
|
-
},
|
|
619
|
-
{
|
|
620
|
-
"id": "Qwen/Qwen3-30B-A3B-Instruct",
|
|
621
|
-
"name": "Qwen3-30B-A3B-Instruct-2507",
|
|
622
|
-
"created_by": "qwen",
|
|
623
|
-
"source": "official",
|
|
624
|
-
"last_updated": "2026-03-21",
|
|
625
|
-
"capabilities": { "streaming": true },
|
|
626
|
-
"snapshots": ["Qwen/Qwen3-30B-A3B-Instruct-2507"]
|
|
627
|
-
},
|
|
628
|
-
{
|
|
629
|
-
"id": "Qwen/Qwen3-30B-A3B",
|
|
630
|
-
"name": "Qwen3-30B-A3B",
|
|
631
|
-
"created_by": "qwen",
|
|
632
|
-
"source": "official",
|
|
633
|
-
"last_updated": "2026-03-21",
|
|
634
|
-
"capabilities": { "streaming": true }
|
|
635
|
-
},
|
|
636
|
-
{
|
|
637
|
-
"id": "Qwen/Qwen3-32B",
|
|
638
|
-
"name": "Qwen3-32B",
|
|
639
|
-
"created_by": "qwen",
|
|
640
|
-
"source": "official",
|
|
641
|
-
"last_updated": "2026-03-21",
|
|
642
|
-
"capabilities": { "streaming": true }
|
|
643
|
-
},
|
|
644
|
-
{
|
|
645
|
-
"id": "Qwen/Qwen3-4B-Instruct-2507",
|
|
646
|
-
"name": "Qwen3-4B-Instruct-2507",
|
|
647
|
-
"created_by": "qwen",
|
|
648
|
-
"source": "official",
|
|
649
|
-
"last_updated": "2026-03-21",
|
|
650
|
-
"alias": "Qwen/Qwen3-4B-Instruct",
|
|
651
|
-
"capabilities": { "streaming": true }
|
|
652
|
-
},
|
|
653
|
-
{
|
|
654
|
-
"id": "Qwen/Qwen3-4B-Instruct",
|
|
655
|
-
"name": "Qwen3-4B-Instruct-2507",
|
|
656
|
-
"created_by": "qwen",
|
|
657
|
-
"source": "official",
|
|
658
|
-
"last_updated": "2026-03-21",
|
|
659
|
-
"capabilities": { "streaming": true },
|
|
660
|
-
"snapshots": ["Qwen/Qwen3-4B-Instruct-2507"]
|
|
661
|
-
},
|
|
662
|
-
{
|
|
663
|
-
"id": "Qwen/Qwen3-4B-Thinking-2507",
|
|
664
|
-
"name": "Qwen3-4B-Thinking-2507",
|
|
665
|
-
"created_by": "qwen",
|
|
666
|
-
"source": "official",
|
|
667
|
-
"last_updated": "2026-03-21",
|
|
668
|
-
"alias": "Qwen/Qwen3-4B-Thinking",
|
|
669
|
-
"capabilities": { "streaming": true }
|
|
670
|
-
},
|
|
671
|
-
{
|
|
672
|
-
"id": "Qwen/Qwen3-4B-Thinking",
|
|
673
|
-
"name": "Qwen3-4B-Thinking-2507",
|
|
674
|
-
"created_by": "qwen",
|
|
675
|
-
"source": "official",
|
|
676
|
-
"last_updated": "2026-03-21",
|
|
677
|
-
"capabilities": { "streaming": true },
|
|
678
|
-
"snapshots": ["Qwen/Qwen3-4B-Thinking-2507"]
|
|
679
|
-
},
|
|
680
|
-
{
|
|
681
|
-
"id": "Qwen/Qwen3-8B-Base",
|
|
682
|
-
"name": "Qwen3-8B-Base",
|
|
683
|
-
"created_by": "qwen",
|
|
684
|
-
"source": "official",
|
|
685
|
-
"last_updated": "2026-03-21",
|
|
686
|
-
"capabilities": { "streaming": true }
|
|
687
|
-
},
|
|
688
|
-
{
|
|
689
|
-
"id": "Qwen/Qwen3-8B",
|
|
690
|
-
"name": "Qwen3-8B",
|
|
691
|
-
"created_by": "qwen",
|
|
692
|
-
"source": "official",
|
|
693
|
-
"last_updated": "2026-03-21",
|
|
694
|
-
"capabilities": { "streaming": true }
|
|
695
|
-
},
|
|
696
|
-
{
|
|
697
|
-
"id": "Qwen/Qwen3-Coder-30B-A3B-Instruct",
|
|
698
|
-
"name": "Qwen3-Coder-30B-A3B-Instruct",
|
|
699
|
-
"created_by": "qwen",
|
|
700
|
-
"source": "official",
|
|
701
|
-
"last_updated": "2026-03-21",
|
|
702
|
-
"capabilities": { "streaming": true }
|
|
703
|
-
},
|
|
704
|
-
{
|
|
705
|
-
"id": "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8",
|
|
706
|
-
"name": "Qwen3-Coder-480B-A35B-Instruct-FP8",
|
|
707
|
-
"created_by": "qwen",
|
|
708
|
-
"source": "official",
|
|
709
|
-
"last_updated": "2026-03-21",
|
|
710
|
-
"capabilities": { "streaming": true }
|
|
711
|
-
},
|
|
712
|
-
{
|
|
713
|
-
"id": "Qwen/Qwen3-Coder-480B-A35B-Instruct",
|
|
714
|
-
"name": "Qwen3-Coder-480B-A35B-Instruct",
|
|
715
|
-
"created_by": "qwen",
|
|
716
|
-
"source": "official",
|
|
717
|
-
"last_updated": "2026-03-21",
|
|
718
|
-
"capabilities": { "streaming": true }
|
|
719
|
-
},
|
|
720
|
-
{
|
|
721
|
-
"id": "Qwen/Qwen3-Coder-Next-FP8",
|
|
722
|
-
"name": "Qwen3-Coder-Next-FP8",
|
|
723
|
-
"created_by": "qwen",
|
|
724
|
-
"source": "official",
|
|
725
|
-
"last_updated": "2026-03-21",
|
|
726
|
-
"capabilities": { "streaming": true }
|
|
727
|
-
},
|
|
728
|
-
{
|
|
729
|
-
"id": "Qwen/Qwen3-Coder-Next",
|
|
730
|
-
"name": "Qwen3-Coder-Next",
|
|
731
|
-
"created_by": "qwen",
|
|
732
|
-
"source": "official",
|
|
733
|
-
"last_updated": "2026-03-21",
|
|
734
|
-
"capabilities": { "streaming": true }
|
|
735
|
-
},
|
|
736
|
-
{
|
|
737
|
-
"id": "Qwen/Qwen3-Next-80B-A3B-Instruct",
|
|
738
|
-
"name": "Qwen3-Next-80B-A3B-Instruct",
|
|
739
|
-
"created_by": "qwen",
|
|
740
|
-
"source": "official",
|
|
741
|
-
"last_updated": "2026-03-21",
|
|
742
|
-
"capabilities": { "streaming": true }
|
|
743
|
-
},
|
|
744
|
-
{
|
|
745
|
-
"id": "Qwen/Qwen3-Next-80B-A3B-Thinking",
|
|
746
|
-
"name": "Qwen3-Next-80B-A3B-Thinking",
|
|
747
|
-
"created_by": "qwen",
|
|
748
|
-
"source": "official",
|
|
749
|
-
"last_updated": "2026-03-21",
|
|
750
|
-
"capabilities": { "streaming": true }
|
|
751
|
-
},
|
|
752
|
-
{
|
|
753
|
-
"id": "RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV",
|
|
754
|
-
"name": "Meta-Llama-3-8B-Instruct-FP8-KV",
|
|
755
|
-
"created_by": "redhatai",
|
|
756
|
-
"source": "official",
|
|
757
|
-
"last_updated": "2026-03-21",
|
|
758
|
-
"capabilities": { "streaming": true }
|
|
759
|
-
},
|
|
760
|
-
{
|
|
761
|
-
"id": "SakanaAI/TinySwallow-1.5B-Instruct",
|
|
762
|
-
"name": "TinySwallow-1.5B-Instruct",
|
|
763
|
-
"created_by": "sakanaai",
|
|
764
|
-
"source": "official",
|
|
765
|
-
"last_updated": "2026-03-21",
|
|
766
|
-
"capabilities": { "streaming": true }
|
|
767
|
-
},
|
|
768
|
-
{
|
|
769
|
-
"id": "speakleash/Bielik-7B-Instruct-v0.1",
|
|
770
|
-
"name": "Bielik-7B-Instruct-v0.1",
|
|
771
|
-
"created_by": "speakleash",
|
|
772
|
-
"source": "official",
|
|
773
|
-
"last_updated": "2026-03-21",
|
|
774
|
-
"capabilities": { "streaming": true }
|
|
775
|
-
},
|
|
776
|
-
{
|
|
777
|
-
"id": "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
|
|
778
|
-
"name": "LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
|
|
779
|
-
"created_by": "swap-uniba",
|
|
780
|
-
"source": "official",
|
|
781
|
-
"last_updated": "2026-03-21",
|
|
782
|
-
"capabilities": { "streaming": true }
|
|
783
|
-
},
|
|
784
|
-
{
|
|
785
|
-
"id": "swiss-ai/Apertus-8B-Instruct-2509",
|
|
786
|
-
"name": "Apertus-8B-Instruct-2509",
|
|
787
|
-
"created_by": "swiss-ai",
|
|
788
|
-
"source": "official",
|
|
789
|
-
"last_updated": "2026-03-21",
|
|
790
|
-
"alias": "swiss-ai/Apertus-8B-Instruct",
|
|
791
|
-
"capabilities": { "streaming": true }
|
|
792
|
-
},
|
|
793
|
-
{
|
|
794
|
-
"id": "swiss-ai/Apertus-8B-Instruct",
|
|
795
|
-
"name": "Apertus-8B-Instruct-2509",
|
|
796
|
-
"created_by": "swiss-ai",
|
|
797
|
-
"source": "official",
|
|
798
|
-
"last_updated": "2026-03-21",
|
|
799
|
-
"capabilities": { "streaming": true },
|
|
800
|
-
"snapshots": ["swiss-ai/Apertus-8B-Instruct-2509"]
|
|
801
|
-
},
|
|
802
|
-
{
|
|
803
|
-
"id": "tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1",
|
|
804
|
-
"name": "Llama-3-Swallow-8B-Instruct-v0.1",
|
|
805
|
-
"created_by": "tokyotech-llm",
|
|
806
|
-
"source": "official",
|
|
807
|
-
"last_updated": "2026-03-21",
|
|
808
|
-
"capabilities": { "streaming": true }
|
|
809
|
-
},
|
|
810
|
-
{
|
|
811
|
-
"id": "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5",
|
|
812
|
-
"name": "Llama-3.1-Swallow-8B-Instruct-v0.5",
|
|
813
|
-
"created_by": "tokyotech-llm",
|
|
814
|
-
"source": "official",
|
|
815
|
-
"last_updated": "2026-03-21",
|
|
816
|
-
"capabilities": { "streaming": true }
|
|
817
|
-
},
|
|
818
|
-
{
|
|
819
|
-
"id": "unsloth/gemma-2-9b-it",
|
|
820
|
-
"name": "gemma-2-9b-it",
|
|
821
|
-
"created_by": "unsloth",
|
|
822
|
-
"source": "official",
|
|
823
|
-
"last_updated": "2026-03-21",
|
|
824
|
-
"family": "gemma-2",
|
|
825
|
-
"capabilities": { "streaming": true }
|
|
826
|
-
},
|
|
827
|
-
{
|
|
828
|
-
"id": "unsloth/llama-3-8b-Instruct",
|
|
829
|
-
"name": "llama-3-8b-Instruct",
|
|
830
|
-
"created_by": "unsloth",
|
|
831
|
-
"source": "official",
|
|
832
|
-
"last_updated": "2026-03-21",
|
|
833
|
-
"family": "llama-3",
|
|
834
|
-
"capabilities": { "streaming": true }
|
|
835
|
-
},
|
|
836
|
-
{
|
|
837
|
-
"id": "unsloth/Llama-3.1-8B-Instruct",
|
|
838
|
-
"name": "Llama-3.1-8B-Instruct",
|
|
839
|
-
"created_by": "unsloth",
|
|
840
|
-
"source": "official",
|
|
841
|
-
"last_updated": "2026-03-21",
|
|
842
|
-
"capabilities": { "streaming": true }
|
|
843
|
-
},
|
|
844
|
-
{
|
|
845
|
-
"id": "unsloth/Meta-Llama-3.1-8B-Instruct",
|
|
846
|
-
"name": "Meta-Llama-3.1-8B-Instruct",
|
|
847
|
-
"created_by": "unsloth",
|
|
848
|
-
"source": "official",
|
|
849
|
-
"last_updated": "2026-03-21",
|
|
850
|
-
"capabilities": { "streaming": true }
|
|
851
|
-
},
|
|
852
|
-
{
|
|
853
|
-
"id": "unsloth/Mistral-Small-24B-Instruct-2501",
|
|
854
|
-
"name": "Mistral-Small-24B-Instruct-2501",
|
|
855
|
-
"created_by": "unsloth",
|
|
856
|
-
"source": "official",
|
|
857
|
-
"last_updated": "2026-03-21",
|
|
858
|
-
"alias": "unsloth/Mistral-Small-24B-Instruct",
|
|
859
|
-
"capabilities": { "streaming": true }
|
|
860
|
-
},
|
|
861
|
-
{
|
|
862
|
-
"id": "unsloth/Mistral-Small-24B-Instruct",
|
|
863
|
-
"name": "Mistral-Small-24B-Instruct-2501",
|
|
864
|
-
"created_by": "unsloth",
|
|
865
|
-
"source": "official",
|
|
866
|
-
"last_updated": "2026-03-21",
|
|
867
|
-
"capabilities": { "streaming": true },
|
|
868
|
-
"snapshots": ["unsloth/Mistral-Small-24B-Instruct-2501"]
|
|
869
|
-
},
|
|
870
|
-
{
|
|
871
|
-
"id": "unsloth/Qwen2.5-14B-Instruct",
|
|
872
|
-
"name": "Qwen2.5-14B-Instruct",
|
|
873
|
-
"created_by": "unsloth",
|
|
874
|
-
"source": "official",
|
|
875
|
-
"last_updated": "2026-03-21",
|
|
876
|
-
"capabilities": { "streaming": true }
|
|
877
|
-
},
|
|
878
|
-
{
|
|
879
|
-
"id": "unsloth/Qwen2.5-32B-Instruct",
|
|
880
|
-
"name": "Qwen2.5-32B-Instruct",
|
|
881
|
-
"created_by": "unsloth",
|
|
882
|
-
"source": "official",
|
|
883
|
-
"last_updated": "2026-03-21",
|
|
884
|
-
"capabilities": { "streaming": true }
|
|
885
|
-
},
|
|
886
|
-
{
|
|
887
|
-
"id": "unsloth/Qwen2.5-7B-Instruct",
|
|
888
|
-
"name": "Qwen2.5-7B-Instruct",
|
|
889
|
-
"created_by": "unsloth",
|
|
890
|
-
"source": "official",
|
|
891
|
-
"last_updated": "2026-03-21",
|
|
892
|
-
"capabilities": { "streaming": true }
|
|
893
|
-
},
|
|
894
|
-
{
|
|
895
|
-
"id": "unsloth/Qwen2.5-Coder-32B-Instruct",
|
|
896
|
-
"name": "Qwen2.5-Coder-32B-Instruct",
|
|
897
|
-
"created_by": "unsloth",
|
|
898
|
-
"source": "official",
|
|
899
|
-
"last_updated": "2026-03-21",
|
|
900
|
-
"capabilities": { "streaming": true }
|
|
901
|
-
},
|
|
902
|
-
{
|
|
903
|
-
"id": "unsloth/Qwen3-8B",
|
|
904
|
-
"name": "Qwen3-8B",
|
|
905
|
-
"created_by": "unsloth",
|
|
906
|
-
"source": "official",
|
|
907
|
-
"last_updated": "2026-03-21",
|
|
908
|
-
"capabilities": { "streaming": true }
|
|
909
|
-
},
|
|
910
|
-
{
|
|
911
|
-
"id": "unsloth/tinyllama-chat",
|
|
912
|
-
"name": "tinyllama-chat",
|
|
913
|
-
"created_by": "unsloth",
|
|
914
|
-
"source": "official",
|
|
915
|
-
"last_updated": "2026-03-21",
|
|
916
|
-
"capabilities": { "streaming": true }
|
|
917
|
-
},
|
|
918
|
-
{
|
|
919
|
-
"id": "VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct",
|
|
920
|
-
"name": "Llama-3-SauerkrautLM-8b-Instruct",
|
|
921
|
-
"created_by": "vagosolutions",
|
|
922
|
-
"source": "official",
|
|
923
|
-
"last_updated": "2026-03-21",
|
|
924
|
-
"capabilities": { "streaming": true }
|
|
925
|
-
},
|
|
926
|
-
{
|
|
927
|
-
"id": "zai-org/GLM-4-9B-0414",
|
|
928
|
-
"name": "GLM-4-9B-0414",
|
|
929
|
-
"created_by": "zhipu",
|
|
930
|
-
"source": "official",
|
|
931
|
-
"last_updated": "2026-03-21",
|
|
932
|
-
"capabilities": { "streaming": true }
|
|
933
|
-
},
|
|
934
|
-
{
|
|
935
|
-
"id": "zai-org/GLM-4.5-Air-FP8",
|
|
936
|
-
"name": "GLM-4.5-Air-FP8",
|
|
937
|
-
"created_by": "zhipu",
|
|
938
|
-
"source": "official",
|
|
939
|
-
"last_updated": "2026-03-21",
|
|
940
|
-
"capabilities": { "streaming": true }
|
|
941
|
-
},
|
|
942
|
-
{
|
|
943
|
-
"id": "zai-org/GLM-4.5-Air",
|
|
944
|
-
"name": "GLM-4.5-Air",
|
|
945
|
-
"created_by": "zhipu",
|
|
946
|
-
"source": "official",
|
|
947
|
-
"last_updated": "2026-03-21",
|
|
948
|
-
"capabilities": { "streaming": true }
|
|
949
|
-
},
|
|
950
|
-
{
|
|
951
|
-
"id": "zai-org/GLM-4.5",
|
|
952
|
-
"name": "GLM-4.5",
|
|
953
|
-
"created_by": "zhipu",
|
|
954
|
-
"source": "official",
|
|
955
|
-
"last_updated": "2026-03-21",
|
|
956
|
-
"capabilities": { "streaming": true }
|
|
957
|
-
},
|
|
958
|
-
{
|
|
959
|
-
"id": "zai-org/GLM-4.6-FP8",
|
|
960
|
-
"name": "GLM-4.6-FP8",
|
|
961
|
-
"created_by": "zhipu",
|
|
962
|
-
"source": "official",
|
|
963
|
-
"last_updated": "2026-03-21",
|
|
964
|
-
"capabilities": { "streaming": true }
|
|
965
|
-
},
|
|
966
|
-
{
|
|
967
|
-
"id": "zai-org/GLM-4.6",
|
|
968
|
-
"name": "GLM-4.6",
|
|
969
|
-
"created_by": "zhipu",
|
|
970
|
-
"source": "official",
|
|
971
|
-
"last_updated": "2026-03-21",
|
|
972
|
-
"capabilities": { "streaming": true }
|
|
973
|
-
},
|
|
974
|
-
{
|
|
975
|
-
"id": "zai-org/GLM-4.7-Flash",
|
|
976
|
-
"name": "GLM-4.7-Flash",
|
|
977
|
-
"created_by": "zhipu",
|
|
978
|
-
"source": "official",
|
|
979
|
-
"last_updated": "2026-03-21",
|
|
980
|
-
"capabilities": { "streaming": true }
|
|
981
|
-
},
|
|
982
|
-
{
|
|
983
|
-
"id": "zai-org/GLM-4.7-FP8",
|
|
984
|
-
"name": "GLM-4.7-FP8",
|
|
985
|
-
"created_by": "zhipu",
|
|
986
|
-
"source": "official",
|
|
987
|
-
"last_updated": "2026-03-21",
|
|
988
|
-
"capabilities": { "streaming": true }
|
|
989
|
-
},
|
|
990
|
-
{
|
|
991
|
-
"id": "zai-org/GLM-4.7",
|
|
992
|
-
"name": "GLM-4.7",
|
|
993
|
-
"created_by": "zhipu",
|
|
994
|
-
"source": "official",
|
|
995
|
-
"last_updated": "2026-03-21",
|
|
996
|
-
"capabilities": { "streaming": true }
|
|
997
|
-
},
|
|
998
|
-
{
|
|
999
|
-
"id": "zai-org/GLM-5",
|
|
1000
|
-
"name": "GLM-5",
|
|
1001
|
-
"created_by": "zhipu",
|
|
1002
|
-
"source": "official",
|
|
1003
|
-
"last_updated": "2026-03-21",
|
|
1004
|
-
"capabilities": { "streaming": true }
|
|
1005
|
-
},
|
|
1006
|
-
{
|
|
1007
|
-
"id": "zai-org/GLM-Z1-32B-0414",
|
|
1008
|
-
"name": "GLM-Z1-32B-0414",
|
|
1009
|
-
"created_by": "zhipu",
|
|
1010
|
-
"source": "official",
|
|
1011
|
-
"last_updated": "2026-03-21",
|
|
1012
|
-
"capabilities": { "streaming": true }
|
|
1013
|
-
}
|
|
1014
|
-
]
|
|
1015
|
-
};
|
|
1016
|
-
const models = provider.models.map((m) => ({
|
|
1017
|
-
...m,
|
|
1018
|
-
provider: provider.id
|
|
1019
|
-
}));
|
|
1020
|
-
//#endregion
|
|
1021
|
-
exports.models = models;
|
|
1022
|
-
exports.provider = provider;
|
|
1
|
+
Object.defineProperty(exports,Symbol.toStringTag,{value:`Module`});const e={id:`huggingface`,name:`Hugging Face`,region:`US`,headquarters:`New York, NY`,founded:2016,url:`https://huggingface.co`,api_url:`https://router.huggingface.co/v1`,docs_url:`https://huggingface.co/docs/api-inference`,pricing_url:`https://huggingface.co/pricing`,description:`The AI community platform hosting open-source models with serverless inference endpoints.`,type:`aggregator`,sdk:{python:`huggingface_hub`,javascript:`@huggingface/inference`},openai_compatible:!0,free_tier:!0,github_url:`https://github.com/huggingface`,models_url:`https://huggingface.co/models`,twitter_url:`https://x.com/huggingface`,discord_url:`https://discord.com/invite/hugging-face-879548962464493619`,blog_url:`https://huggingface.co/blog`,terms_url:`https://huggingface.co/terms-of-service`,support_url:`https://huggingface.co/support`,icon:`<svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24" fill="currentColor" fill-rule="evenodd">
|
|
2
|
+
<path d="M16.781 3.277c2.997 1.704 4.844 4.851 4.844 8.258 0 .995-.155 1.955-.443 2.857a1.332 1.332 0 011.125.4 1.41 1.41 0 01.2 1.723c.204.165.352.385.428.632l.017.062c.06.222.12.69-.2 1.166.244.37.279.836.093 1.236-.255.57-.893 1.018-2.128 1.5l-.202.078-.131.048c-.478.173-.89.295-1.061.345l-.086.024c-.89.243-1.808.375-2.732.394-1.32 0-2.3-.36-2.923-1.067a9.852 9.852 0 01-3.18.018C9.778 21.647 8.802 22 7.494 22a11.249 11.249 0 01-2.541-.343l-.221-.06-.273-.08a16.574 16.574 0 01-1.175-.405c-1.237-.483-1.875-.93-2.13-1.501-.186-.4-.151-.867.093-1.236a1.42 1.42 0 01-.2-1.166c.069-.273.226-.516.447-.694a1.41 1.41 0 01.2-1.722c.233-.248.557-.391.917-.407l.078-.001a9.385 9.385 0 01-.44-2.85c0-3.407 1.847-6.554 4.844-8.258a9.822 9.822 0 019.687 0zM4.188 14.758c.125.687 2.357 2.35 2.14 2.707-.19.315-.796-.239-.948-.386l-.041-.04-.168-.147c-.561-.479-2.304-1.9-2.74-1.432-.43.46.119.859 1.055 1.42l.784.467.136.083c1.045.643 1.12.84.95 1.113-.188.295-3.07-2.1-3.34-1.083-.27 1.011 2.942 1.304 2.744 2.006-.2.7-2.265-1.324-2.685-.537-.425.79 2.913 1.718 2.94 1.725l.16.04.175.042c1.227.284 3.565.65 4.435-.604.673-.973.64-1.709-.248-2.61l-.057-.057c-.945-.928-1.495-2.288-1.495-2.288l-.017-.058-.025-.072c-.082-.22-.284-.639-.63-.584-.46.073-.798 1.21.12 1.933l.05.038c.977.721-.195 1.21-.573.534l-.058-.104-.143-.25c-.463-.799-1.282-2.111-1.739-2.397-.532-.332-.907-.148-.782.541zm14.842-.541c-.533.335-1.563 2.074-1.94 2.751a.613.613 0 01-.687.302.436.436 0 01-.176-.098.303.303 0 01-.049-.06l-.014-.028-.008-.02-.007-.019-.003-.013-.003-.017a.289.289 0 01-.004-.048c0-.12.071-.266.25-.427.026-.024.054-.047.084-.07l.047-.036c.022-.016.043-.032.063-.049.883-.71.573-1.81.131-1.917l-.031-.006-.056-.004a.368.368 0 00-.062.006l-.028.005-.042.014-.039.017-.028.015-.028.019-.036.027-.023.02c-.173.158-.273.428-.31.542l-.016.054s-.53 1.309-1.439 2.234l-.054.054c-.365.358-.596.69-.702 1.018-.143.437-.066.868.21 1.353.055.097.117.195.187.296.882 1.275 3.282.876 4.494.59l.286-.07.25-.074c.276-.084.736-.233 1.2-.42l.188-.077.065-.028.064-.028.124-.056.081-.038c.529-.252.964-.543.994-.827l.001-.036a.299.299 0 00-.037-.139c-.094-.176-.271-.212-.491-.168l-.045.01c-.044.01-.09.024-.136.04l-.097.035-.054.022c-.559.23-1.238.705-1.607.745h.006a.452.452 0 01-.05.003h-.024l-.024-.003-.023-.005c-.068-.016-.116-.06-.14-.142a.22.22 0 01-.005-.1c.062-.345.958-.595 1.713-.91l.066-.028c.528-.224.97-.483.985-.832v-.04a.47.47 0 00-.016-.098c-.048-.18-.175-.251-.36-.251-.785 0-2.55 1.36-2.92 1.36-.025 0-.048-.007-.058-.024a.6.6 0 01-.046-.088c-.1-.238.068-.462 1.06-1.066l.209-.126c.538-.32 1.01-.588 1.341-.831.29-.212.475-.406.503-.6l.003-.028c.008-.113-.038-.227-.147-.344a.266.266 0 00-.07-.054l-.034-.015-.013-.005a.403.403 0 00-.13-.02c-.162 0-.369.07-.595.18-.637.313-1.431.952-1.826 1.285l-.249.215-.033.033c-.08.078-.288.27-.493.386l-.071.037-.041.019a.535.535 0 01-.122.036h.005a.346.346 0 01-.031.003l.01-.001-.013.001c-.079.005-.145-.021-.19-.095a.113.113 0 01-.014-.065c.027-.465 2.034-1.991 2.152-2.642l.009-.048c.1-.65-.271-.817-.791-.493zM11.938 2.984c-4.798 0-8.688 3.829-8.688 8.55 0 .692.083 1.364.24 2.008l.008-.009c.252-.298.612-.46 1.017-.46.355.008.699.117.993.312.22.14.465.384.715.694.261-.372.69-.598 1.15-.605.852 0 1.367.728 1.562 1.383l.047.105.06.127c.192.396.595 1.139 1.143 1.68 1.06 1.04 1.324 2.115.8 3.266a8.865 8.865 0 002.024-.014c-.505-1.12-.26-2.17.74-3.186l.066-.066c.695-.684 1.157-1.69 1.252-1.912.195-.655.708-1.383 1.56-1.383.46.007.889.233 1.15.605.25-.31.495-.553.718-.694a1.87 1.87 0 01.99-.312c.357 0 .682.126.925.36.14-.61.215-1.245.215-1.898 0-4.722-3.89-8.55-8.687-8.55zm1.857 8.926l.439-.212c.553-.264.89-.383.89.152 0 1.093-.771 3.208-3.155 3.262h-.184c-2.325-.052-3.116-2.06-3.156-3.175l-.001-.087c0-1.107 1.452.586 3.25.586.716 0 1.379-.272 1.917-.526zm4.017-3.143c.45 0 .813.358.813.8 0 .441-.364.8-.813.8a.806.806 0 01-.812-.8c0-.442.364-.8.812-.8zm-11.624 0c.448 0 .812.358.812.8 0 .441-.364.8-.812.8a.806.806 0 01-.813-.8c0-.442.364-.8.813-.8zm7.79-.841c.32-.384.846-.54 1.33-.394.483.146.83.564.878 1.06.048.495-.212.97-.659 1.203-.322.168-.447-.477-.767-.585l.002-.003c-.287-.098-.772.362-.925.079a1.215 1.215 0 01.14-1.36zm-4.323 0c.322.384.377.92.14 1.36-.152.283-.64-.177-.925-.079l.003.003c-.108.036-.194.134-.273.24l-.118.165c-.11.15-.22.262-.377.18a1.226 1.226 0 01-.658-1.204c.048-.495.395-.913.878-1.059a1.262 1.262 0 011.33.394z"/>
|
|
3
|
+
</svg>`,models:[{id:`aaditya/Llama3-OpenBioLLM-8B`,name:`Llama3-OpenBioLLM-8B`,created_by:`aaditya`,source:`official`,last_updated:`2026-03-23`,page_url:`https://huggingface.co/aaditya/Llama3-OpenBioLLM-8B`,model_type:`chat`,capabilities:{streaming:!0},description:`Introducing OpenBioLLM-8B: A State-of-the-Art Open Source Biomedical Large Language Model`,release_date:`2024-04-20`,context_window:8192,license:`llama3`,tagline:`Introducing OpenBioLLM-8B: A State-of-the-Art Open Source Biomedical Large Language Model`,architecture:`transformer`,open_weight:!0},{id:`allenai/Olmo-3-7B-Instruct`,name:`Olmo-3-7B-Instruct`,created_by:`allenai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:0,page_url:`https://huggingface.co/allenai/Olmo-3-7B-Instruct`,release_date:`2025-11-19`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`allenai/Olmo-3.1-32B-Instruct`,name:`Olmo-3.1-32B-Instruct`,created_by:`allenai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:32,page_url:`https://huggingface.co/allenai/Olmo-3.1-32B-Instruct`,release_date:`2025-12-10`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,name:`Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,created_by:`argilla`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1,page_url:`https://huggingface.co/argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,description:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on [argilla-warehouse/apigen-synth-trl](https://huggingface.co/datasets/argilla-warehouse/apigen-synth-trl) dataset, a version of [argilla/Synth-APIGen-v0.1](https://huggingface.co/datasets/argilla-warehouse/Synth-APIGen-v0.1) ready to do SFT on top of it. It has been trained using [TRL](https://github.com/huggingface/trl).`,release_date:`2024-10-07`,context_window:131072,license:`apache-2.0`,tagline:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on [argilla-warehouse/apigen-synth-trl](https://huggingface.co/datasets/argilla-warehouse/apigen-synth-trl) dataset, a version of [argilla/Synth-APIGen-v0.1](https://huggingface.co/datasets/argilla-warehouse/Synth-APIGen-v0.1) ready to do SFT on top of it.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC`,name:`Llama-3.2-1B-Instruct-v2-FC`,created_by:`argilla-warehouse`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1,page_url:`https://huggingface.co/argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC`,description:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the [argilla-warehouse/apigen-smollm-trl-FC](https://huggingface.co/datasets/argilla-warehouse/apigen-smollm-trl-FC) dataset. It has been trained using [TRL](https://github.com/huggingface/trl).`,release_date:`2024-10-20`,context_window:131072,tagline:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the [argilla-warehouse/apigen-smollm-trl-FC](https://huggingface.co/datasets/argilla-warehouse/apigen-smollm-trl-FC) dataset.`,architecture:`transformer`,tools:[`function_calling`]},{id:`DeepMount00/Llama-3-8b-Ita`,name:`Llama-3-8b-Ita`,created_by:`deepmount00`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/DeepMount00/Llama-3-8b-Ita`,description:`**💡 Found this resource helpful?** Creating and maintaining open source AI models and datasets requires significant computational resources. If this work has been valuable to you, consider [supporting my research](https://buymeacoffee.com/michele.montebovi) to help me continue building tools that benefit the entire AI community. Every contribution directly funds more open source innovation! ☕`,release_date:`2024-05-01`,context_window:8192,license:`llama3`,tagline:`**💡 Found this resource helpful?** Creating and maintaining open source AI models and datasets requires significant computational resources.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`deepseek-ai/DeepSeek-R1-0528-Qwen3-8B`,name:`DeepSeek-R1-0528-Qwen3-8B`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`reasoning`,parameters:8.2,page_url:`https://huggingface.co/deepseek-ai/DeepSeek-R1-0528-Qwen3-8B`,release_date:`2025-05-29`,context_window:131072,license:`mit`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Efficient-Large-Model/gemma-2-2b-it`,name:`gemma-2-2b-it`,created_by:`efficient-large-model`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2.6,page_url:`https://huggingface.co/Efficient-Large-Model/gemma-2-2b-it`,description:`**Model Page**: [Gemma](https://ai.google.dev/gemma/docs/base)`,release_date:`2024-12-12`,context_window:8192,license:`gemma`,tagline:`**Model Page**: [Gemma](https://ai.google.dev/gemma/docs/base)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`EssentialAI/rnj-1-instruct`,name:`rnj-1-instruct`,created_by:`essentialai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},page_url:`https://huggingface.co/EssentialAI/rnj-1-instruct`,description:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models. These models perform well across a range of programming languages and boast strong agentic capabilities (e.g., inside agentic frameworks like mini-SWE-agent), while also excelling at tool-calling. They additionally exhibit strong capabilities in math and science.`,release_date:`2025-12-04`,context_window:32768,license:`apache-2.0`,parameters:8.3,tagline:`style="vertical-align: middle;" src="https://img.shields.io/badge/%F0%9F%8C%90%20Website-essential.ai-4b9fe1?color=4b9fe1&logoColor=white"/>`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`failspy/Meta-Llama-3-8B-Instruct-abliterated-v3`,name:`Meta-Llama-3-8B-Instruct-abliterated-v3`,created_by:`failspy`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3`,description:`[My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon](https://huggingface.co/failspy/llama-3-70B-Instruct-abliterated/blob/main/ortho_cookbook.ipynb)`,release_date:`2024-05-20`,context_window:8192,license:`llama3`,tagline:`[My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon](https://huggingface.co/failspy/llama-3-70B-Instruct-abliterated/blob/main/ortho_cookbook.ipynb)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Featherless-Chat-Models/Mistral-7B-Instruct-v0.2`,name:`Mistral-7B-Instruct-v0.2`,created_by:`featherless-chat-models`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.2,page_url:`https://huggingface.co/Featherless-Chat-Models/Mistral-7B-Instruct-v0.2`,release_date:`2025-05-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`FlagAlpha/Llama3-Chinese-8B-Instruct`,name:`Llama3-Chinese-8B-Instruct`,created_by:`flagalpha`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/FlagAlpha/Llama3-Chinese-8B-Instruct`,description:`Llama3-Chinese-8B-Instruct基于Llama3-8B中文微调对话模型,由Llama中文社区和AtomEcho(原子回声)联合研发,我们会持续提供更新的模型参数,模型训练过程见 [https://llama.family](https://llama.family)。`,release_date:`2024-04-23`,context_window:8192,license:`apache-2.0`,tagline:`Llama3-Chinese-8B-Instruct基于Llama3-8B中文微调对话模型,由Llama中文社区和AtomEcho(原子回声)联合研发,我们会持续提供更新的模型参数,模型训练过程见 [https://llama.family](https://llama.family)。`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`google/gemma-2-2b-jpn-it`,name:`gemma-2-2b-jpn-it`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2.6,page_url:`https://huggingface.co/google/gemma-2-2b-jpn-it`,release_date:`2024-09-25`,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`google/gemma-2-9b-it`,name:`gemma-2-9b-it`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:9.2,page_url:`https://huggingface.co/google/gemma-2-9b-it`,release_date:`2024-06-24`,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`GraySwanAI/Llama-3-8B-Instruct-RR`,name:`Llama-3-8B-Instruct-RR`,created_by:`grayswanai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/GraySwanAI/Llama-3-8B-Instruct-RR`,description:`Llama-3-8B-Instruct-RR is a Llama-3 model with circuit breakers inserted using Representation Rerouting (RR).`,release_date:`2024-07-08`,context_window:8192,tagline:`Llama-3-8B-Instruct-RR is a Llama-3 model with circuit breakers inserted using Representation Rerouting (RR).`,architecture:`transformer`,tools:[`function_calling`]},{id:`GritLM/GritLM-7B`,name:`GritLM-7B`,created_by:`gritlm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:7.2,page_url:`https://huggingface.co/GritLM/GritLM-7B`,description:`> GritLM is a generative representational instruction tuned language model. It unifies text representation (embedding) and text generation into a single model achieving state-of-the-art performance on both types of tasks.`,release_date:`2024-02-11`,context_window:32768,license:`apache-2.0`,tagline:`> GritLM is a generative representational instruction tuned language model.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`huihui-ai/Qwen2.5-72B-Instruct-abliterated`,name:`Qwen2.5-72B-Instruct-abliterated`,created_by:`huihui-ai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/huihui-ai/Qwen2.5-72B-Instruct-abliterated`,description:`This is an uncensored version of [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it). This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.`,release_date:`2024-10-26`,context_window:32768,license:`other`,tagline:`This is an uncensored version of [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`instruction-pretrain/finance-Llama3-8B`,name:`finance-Llama3-8B`,created_by:`instruction-pretrain`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},parameters:8,page_url:`https://huggingface.co/instruction-pretrain/finance-Llama3-8B`,description:`This repo contains the **finance model developed from Llama3-8B** in our paper [Instruction Pre-Training: Language Models are Supervised Multitask Learners](https://huggingface.co/papers/2406.14491).`,release_date:`2024-06-18`,context_window:8192,license:`llama3`,tagline:`This repo contains the **finance model developed from Llama3-8B** in our paper [Instruction Pre-Training: Language Models are Supervised Multitask Learners](https://huggingface.co/papers/2406.14491).`,architecture:`transformer`,open_weight:!0},{id:`Intel/neural-chat-7b-v3-3`,name:`neural-chat-7b-v3-3`,created_by:`intel`,source:`official`,last_updated:`2026-03-22`,capabilities:{streaming:!0,tool_call:!0},parameters:7.2,page_url:`https://huggingface.co/Intel/neural-chat-7b-v3-3`,release_date:`2023-12-09`,context_window:32768,license:`apache-2.0`},{id:`invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,name:`Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,created_by:`invalid-coder`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:11,page_url:`https://huggingface.co/invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,description:`NeuralPipe-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [jeonsworld/CarbonVillain-en-10.7B-v2](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v2) * [kyujinpy/Sakura-SOLAR-Instruct](https://huggingface.co/kyujinpy/Sakura-SOLAR-Instruct)`,release_date:`2024-01-10`,context_window:4096,license:`apache-2.0`,tagline:`NeuralPipe-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [jeonsworld/CarbonVillain-en-10.7B`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix`,name:`Mistral-Nemo-Instruct-2407-lenient-chatfix`,created_by:`m8than`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,page_url:`https://huggingface.co/m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix`,description:`It's just mistral nemo 2407 with a less strict chat format.`,release_date:`2025-05-06`,context_window:131072,license:`apache-2.0`,parameters:12,tagline:`It's just mistral nemo 2407 with a less strict chat format.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`meta-llama/Llama-2-13b-chat-hf`,name:`Llama-2-13b-chat-hf`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-2`,model_type:`chat`,parameters:13,page_url:`https://huggingface.co/meta-llama/Llama-2-13b-chat-hf`,release_date:`2023-07-13`,license:`llama2`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:4096,max_output_tokens:4096,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.1-70B-Instruct`,name:`Llama-3.1-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct`,release_date:`2024-07-16`,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.1-8B-Instruct`,name:`Llama-3.1-8B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct`,release_date:`2024-07-18`,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.2-1B-Instruct`,name:`Llama-3.2-1B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1.2,page_url:`https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct`,release_date:`2024-09-18`,license:`llama3.2`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.3-70B-Instruct`,name:`Llama-3.3-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct`,release_date:`2024-11-26`,license:`llama3.3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Meta-Llama-3-70B-Instruct`,name:`Meta-Llama-3-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct`,release_date:`2024-04-17`,license:`llama3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:8192,max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Meta-Llama-3-8B-Instruct`,name:`Meta-Llama-3-8B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct`,release_date:`2024-04-17`,license:`llama3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:8192,max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`mistralai/Mistral-7B-Instruct-v0.2`,name:`Mistral-7B-Instruct-v0.2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.2,page_url:`https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2`,release_date:`2023-12-11`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated`,name:`Meta-Llama-3.1-8B-Instruct-abliterated`,created_by:`mlabonne`,source:`official`,last_updated:`2026-03-23`,family:`Llama-3.1`,page_url:`https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated`,model_type:`chat`,capabilities:{streaming:!0,tool_call:!0},description:`This is an uncensored version of Llama 3.1 8B Instruct created with abliteration (see [this article](https://huggingface.co/blog/mlabonne/abliteration) to know more about it).`,release_date:`2024-07-24`,context_window:131072,license:`llama3.1`,parameters:8,tagline:`This is an uncensored version of Llama 3.1 8B Instruct created with abliteration (see [this article](https://huggingface.co/blog/mlabonne/abliteration) to know more about it).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`mlabonne/NeuralDaredevil-8B-abliterated`,name:`NeuralDaredevil-8B-abliterated`,created_by:`mlabonne`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated`,description:`This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated), trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k). The DPO fine-tuning successfully recovers the performance loss due to the abliteration process, making it an excellent uncensored model.`,release_date:`2024-05-27`,context_window:8192,license:`llama3`,tagline:`This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated), trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Instruct-0905`,name:`Kimi-K2-Instruct-0905`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Instruct-0905`,release_date:`2025-09-03`,context_window:262144,license:`other`,parameters:1026,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Instruct`,name:`Kimi-K2-Instruct`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Instruct`,release_date:`2025-07-11`,context_window:131072,license:`other`,parameters:1026,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Thinking`,name:`Kimi-K2-Thinking`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Thinking`,release_date:`2025-11-04`,context_window:262144,license:`other`,parameters:1058,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],description:`Long-term thinking, multi-step tool usage, complex problem solving.`,tagline:`Long-term thinking, multi-step tool usage, complex problem solving.`},{id:`NousResearch/Meta-Llama-3-70B-Instruct`,name:`Meta-Llama-3-70B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3-70B-Instruct`,release_date:`2024-04-19`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`NousResearch/Meta-Llama-3-8B-Instruct`,name:`Meta-Llama-3-8B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct`,release_date:`2024-04-18`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`NousResearch/Meta-Llama-3.1-8B-Instruct`,name:`Meta-Llama-3.1-8B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct`,release_date:`2024-07-24`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`nvidia/Llama-3.1-8B-Instruct-FP8`,name:`Llama-3.1-8B-Instruct-FP8`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/nvidia/Llama-3.1-8B-Instruct-FP8`,release_date:`2024-08-29`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`openai/gpt-oss-120b`,name:`gpt-oss-120b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0},model_type:`chat`,parameters:120,page_url:`https://huggingface.co/openai/gpt-oss-120b`,description:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-04`,context_window:131072,license:`apache-2.0`,tagline:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:131072,knowledge_cutoff:`2024-05`,modalities:{input:[`text`],output:[`text`]}},{id:`openai/gpt-oss-20b`,name:`gpt-oss-20b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0},model_type:`chat`,parameters:22,page_url:`https://huggingface.co/openai/gpt-oss-20b`,description:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-04`,context_window:131072,license:`apache-2.0`,tagline:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:131072,knowledge_cutoff:`2024-05`,modalities:{input:[`text`],output:[`text`]}},{id:`openai/gpt-oss-safeguard-20b`,name:`gpt-oss-safeguard-20b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0},model_type:`moderation`,parameters:20,page_url:`https://huggingface.co/openai/gpt-oss-safeguard-20b`,description:"`gpt-oss-safeguard-120b` and `gpt-oss-safeguard-20b` are safety reasoning models built-upon gpt-oss. With these models, you can classify text content based on safety policies that you provide and perform a suite of foundational safety tasks. These models are intended for safety use cases. For other applications, we recommend using [gpt-oss models](https://huggingface.co/collections/openai/gpt-oss).",release_date:`2025-09-18`,context_window:131072,license:`apache-2.0`,tagline:"`gpt-oss-safeguard-120b` and `gpt-oss-safeguard-20b` are safety reasoning models built-upon gpt-oss.",architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`openchat/openchat-3.6-8b-20240522`,name:`openchat-3.6-8b-20240522`,created_by:`openchat`,source:`official`,last_updated:`2026-03-23`,alias:`openchat/openchat-3.6-8b`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/openchat/openchat-3.6-8b-20240522`,description:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM.`,release_date:`2024-05-07`,context_window:8192,license:`llama3`,tagline:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`openchat/openchat-3.6-8b`,name:`openchat-3.6-8b-20240522`,created_by:`openchat`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`openchat/openchat-3.6-8b-20240522`],parameters:8,page_url:`https://huggingface.co/openchat/openchat-3.6-8b-20240522`,description:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM.`,release_date:`2024-05-07`,context_window:8192,license:`llama3`,tagline:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Orion-zhen/Qwen2.5-7B-Instruct-Uncensored`,name:`Qwen2.5-7B-Instruct-Uncensored`,created_by:`orion-zhen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored`,description:`This model is an uncensored fine-tune version of Qwen2.5-7B-Instruct. However, I can still notice that though uncensored, the model fails to generate detailed descriptions on certain extreme scenarios, which might be associated with deletion on some pretrain datasets in Qwen's pretraining stage.`,release_date:`2024-09-26`,context_window:32768,license:`gpl-3.0`,tagline:`This model is an uncensored fine-tune version of Qwen2.5-7B-Instruct.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,name:`Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,created_by:`patronusai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,description:`Lynx is an open-source hallucination evaluation model. Patronus-Lynx-8B-Instruct-v1.1 was trained on a mix of datasets including CovidQA, PubmedQA, DROP, RAGTruth. The datasets contain a mix of hand-annotated and synthetic data. The maximum sequence length is 128000 tokens.`,release_date:`2024-07-24`,context_window:131072,license:`cc-by-nc-4.0`,tagline:`Lynx is an open-source hallucination evaluation model.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen1.5-1.8B-Chat`,name:`Qwen1.5-1.8B-Chat`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.8,page_url:`https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat`,release_date:`2024-01-30`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2-1.5B-Instruct`,name:`Qwen2-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2-1.5B-Instruct`,release_date:`2024-06-03`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2-72B-Instruct`,name:`Qwen2-72B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/Qwen/Qwen2-72B-Instruct`,release_date:`2024-05-28`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],max_output_tokens:6144,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2-7B-Instruct`,name:`Qwen2-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2-7B-Instruct`,release_date:`2024-06-04`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-1.5B-Instruct`,name:`Qwen2.5-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-14B-Instruct-1M`,name:`Qwen2.5-14B-Instruct-1M`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-14B-Instruct-1M`,release_date:`2025-01-23`,context_window:101e4,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-14B-Instruct`,name:`Qwen2.5-14B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:14,page_url:`https://huggingface.co/Qwen/Qwen2.5-14B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-32B-Instruct`,name:`Qwen2.5-32B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:32,page_url:`https://huggingface.co/Qwen/Qwen2.5-32B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-72B-Instruct`,name:`Qwen2.5-72B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/Qwen/Qwen2.5-72B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-7B-Instruct-1M`,name:`Qwen2.5-7B-Instruct-1M`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-1M`,release_date:`2025-01-23`,context_window:101e4,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-7B-Instruct`,name:`Qwen2.5-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-7B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-Coder-1.5B-Instruct`,name:`Qwen2.5-Coder-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B-Instruct`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-1.5B`,name:`Qwen2.5-Coder-1.5B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-14B-Instruct`,name:`Qwen2.5-Coder-14B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-14B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-14B`,name:`Qwen2.5-Coder-14B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-14B`,release_date:`2024-11-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-32B-Instruct`,name:`Qwen2.5-Coder-32B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-32B`,name:`Qwen2.5-Coder-32B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-32B`,release_date:`2024-11-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-3B-Instruct`,name:`Qwen2.5-Coder-3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:3.1,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-7B-Instruct`,name:`Qwen2.5-Coder-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-7B`,name:`Qwen2.5-Coder-7B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-7B`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Math-1.5B-Instruct`,name:`Qwen2.5-Math-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Math-1.5B-Instruct`,description:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR. We do not recommend using this series of models for other tasks. > </b> > </div>`,release_date:`2024-09-16`,context_window:4096,license:`apache-2.0`,tagline:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Math-7B-Instruct`,name:`Qwen2.5-Math-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Math-7B-Instruct`,description:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR. We do not recommend using this series of models for other tasks. > </b> > </div>`,release_date:`2024-09-19`,context_window:4096,license:`apache-2.0`,tagline:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen3-1.7B`,name:`Qwen3-1.7B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2,page_url:`https://huggingface.co/Qwen/Qwen3-1.7B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-14B-Base`,name:`Qwen3-14B-Base`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen3-14B-Base`,release_date:`2025-04-28`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-14B`,name:`Qwen3-14B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:14,page_url:`https://huggingface.co/Qwen/Qwen3-14B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Instruct-2507`,name:`Qwen3-235B-A22B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-235B-A22B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507`,release_date:`2025-07-21`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Instruct`,name:`Qwen3-235B-A22B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-235B-A22B-Instruct-2507`],model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507`,release_date:`2025-07-21`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Thinking-2507`,name:`Qwen3-235B-A22B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-235B-A22B-Thinking`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507`,release_date:`2025-07-25`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Thinking`,name:`Qwen3-235B-A22B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-235B-A22B-Thinking-2507`],model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507`,release_date:`2025-07-25`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B`,name:`Qwen3-235B-A22B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:15,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B-Instruct-2507`,name:`Qwen3-30B-A3B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-30B-A3B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507`,release_date:`2025-07-28`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B-Instruct`,name:`Qwen3-30B-A3B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-30B-A3B-Instruct-2507`],model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507`,release_date:`2025-07-28`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B`,name:`Qwen3-30B-A3B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-32B`,name:`Qwen3-32B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:32,page_url:`https://huggingface.co/Qwen/Qwen3-32B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Instruct-2507`,name:`Qwen3-4B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-4B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Instruct`,name:`Qwen3-4B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-4B-Instruct-2507`],model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Thinking-2507`,name:`Qwen3-4B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-4B-Thinking`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Thinking`,name:`Qwen3-4B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-4B-Thinking-2507`],model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-8B-Base`,name:`Qwen3-8B-Base`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8.2,page_url:`https://huggingface.co/Qwen/Qwen3-8B-Base`,release_date:`2025-04-28`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-8B`,name:`Qwen3-8B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8.2,page_url:`https://huggingface.co/Qwen/Qwen3-8B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-30B-A3B-Instruct`,name:`Qwen3-Coder-30B-A3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:31,active_parameters:1.9,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct`,release_date:`2025-07-31`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8`,name:`Qwen3-Coder-480B-A35B-Instruct-FP8`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:480,active_parameters:35,release_date:`2025-07-22`,context_window:262144,license:`apache-2.0`,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-480B-A35B-Instruct`,name:`Qwen3-Coder-480B-A35B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:480,active_parameters:24,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct`,release_date:`2025-07-22`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-Next-FP8`,name:`Qwen3-Coder-Next-FP8`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-Next-FP8`,release_date:`2026-02-01`,context_window:262144,license:`apache-2.0`,parameters:80,active_parameters:1.6,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-Next`,name:`Qwen3-Coder-Next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,release_date:`2026-01-30`,context_window:262144,license:`apache-2.0`,parameters:80,active_parameters:1.6,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-Next`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:65536,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen3-Next-80B-A3B-Instruct`,name:`Qwen3-Next-80B-A3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:81,active_parameters:1.6,page_url:`https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct`,description:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI). We are excited to share our latest advancements in addressing these demands, centered on improving scaling efficiency through innovative model architecture. We call this next-generation foundation models **Qwen3-Next**.`,release_date:`2025-09-09`,context_window:262144,license:`apache-2.0`,tagline:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI).`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Next-80B-A3B-Thinking`,name:`Qwen3-Next-80B-A3B-Thinking`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:80,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking`,description:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI). We are excited to share our latest advancements in addressing these demands, centered on improving scaling efficiency through innovative model architecture. We call this next-generation foundation models **Qwen3-Next**.`,release_date:`2025-09-09`,context_window:262144,license:`apache-2.0`,tagline:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI).`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV`,name:`Meta-Llama-3-8B-Instruct-FP8-KV`,created_by:`redhatai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV`,release_date:`2024-05-20`,context_window:8192,architecture:`transformer`,tools:[`function_calling`]},{id:`SakanaAI/TinySwallow-1.5B-Instruct`,name:`TinySwallow-1.5B-Instruct`,created_by:`sakanaai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:1.5,page_url:`https://huggingface.co/SakanaAI/TinySwallow-1.5B-Instruct`,description:`🤗 [Models](https://huggingface.co/SakanaAI) | 📚 [Paper](https://arxiv.org/abs/2501.16937) | 📝 [Blog](https://sakana.ai/taid-jp/) | 🐦 [Twitter](https://twitter.com/SakanaAILabs)`,release_date:`2025-01-07`,context_window:32768,license:`apache-2.0`,tagline:`🤗 [Models](https://huggingface.co/SakanaAI) | 📚 [Paper](https://arxiv.org/abs/2501.16937) | 📝 [Blog](https://sakana.ai/taid-jp/) | 🐦 [Twitter](https://twitter.com/SakanaAILabs)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`speakleash/Bielik-7B-Instruct-v0.1`,name:`Bielik-7B-Instruct-v0.1`,created_by:`speakleash`,source:`official`,last_updated:`2026-03-21`,capabilities:{streaming:!0},parameters:7},{id:`swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,name:`LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,created_by:`swap-uniba`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,description:`📣 New MODEL FAMILY❗ [https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA](https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA)`,release_date:`2024-04-29`,context_window:8192,license:`llama3`,tagline:`📣 New MODEL FAMILY❗ [https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA](https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`swiss-ai/Apertus-8B-Instruct-2509`,name:`Apertus-8B-Instruct-2509`,created_by:`swiss-ai`,source:`official`,last_updated:`2026-03-23`,alias:`swiss-ai/Apertus-8B-Instruct`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/swiss-ai/Apertus-8B-Instruct-2509`,release_date:`2025-08-13`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`swiss-ai/Apertus-8B-Instruct`,name:`Apertus-8B-Instruct-2509`,created_by:`swiss-ai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`swiss-ai/Apertus-8B-Instruct-2509`],parameters:8,page_url:`https://huggingface.co/swiss-ai/Apertus-8B-Instruct-2509`,release_date:`2025-08-13`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1`,name:`Llama-3-Swallow-8B-Instruct-v0.1`,created_by:`tokyotech-llm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1`,description:`Our Swallow model has undergone continual pre-training from the [Llama 3 family](https://huggingface.co/collections/meta-llama/meta-llama-3-66214712577ca38149ebb2b6), primarily with the addition of Japanese language data. The Instruct versions use supervised fine-tuning (SFT) and Chat Vector. Links to other models can be found in the index.`,release_date:`2024-06-26`,context_window:8192,license:`llama3`,tagline:`Our Swallow model has undergone continual pre-training from the [Llama 3 family](https://huggingface.co/collections/meta-llama/meta-llama-3-66214712577ca38149ebb2b6), primarily with the addition of Japanese language data.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5`,name:`Llama-3.1-Swallow-8B-Instruct-v0.5`,created_by:`tokyotech-llm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5`,description:`Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models. Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.`,release_date:`2025-06-12`,context_window:131072,license:`llama3.3`,tagline:`Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/gemma-2-9b-it`,name:`gemma-2-9b-it`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:9.2,page_url:`https://huggingface.co/unsloth/gemma-2-9b-it`,release_date:`2024-07-03`,context_window:8192,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/llama-3-8b-Instruct`,name:`llama-3-8b-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,family:`llama-3`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/llama-3-8b-Instruct`,description:"Directly quantized 4bit model with `bitsandbytes`.",release_date:`2024-04-18`,context_window:8192,license:`llama3`,tagline:"Directly quantized 4bit model with `bitsandbytes`.",architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Llama-3.1-8B-Instruct`,name:`Llama-3.1-8B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Llama-3.1-8B-Instruct`,release_date:`2025-02-15`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Meta-Llama-3.1-8B-Instruct`,name:`Meta-Llama-3.1-8B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Meta-Llama-3.1-8B-Instruct`,description:`We have a free Google Colab Tesla T4 notebook for Llama 3.1 (8B) here: https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRnrIOjZAuxp?usp=sharing`,release_date:`2024-07-23`,context_window:131072,license:`llama3.1`,tagline:`We have a free Google Colab Tesla T4 notebook for Llama 3.1 (8B) here: https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRnrIOjZAuxp?usp=sharing`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Mistral-Small-24B-Instruct-2501`,name:`Mistral-Small-24B-Instruct-2501`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,alias:`unsloth/Mistral-Small-24B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:24,page_url:`https://huggingface.co/unsloth/Mistral-Small-24B-Instruct-2501`,description:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,release_date:`2025-01-30`,context_window:32768,license:`apache-2.0`,tagline:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Mistral-Small-24B-Instruct`,name:`Mistral-Small-24B-Instruct-2501`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`unsloth/Mistral-Small-24B-Instruct-2501`],model_type:`chat`,parameters:24,page_url:`https://huggingface.co/unsloth/Mistral-Small-24B-Instruct-2501`,description:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,release_date:`2025-01-30`,context_window:32768,license:`apache-2.0`,tagline:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-14B-Instruct`,name:`Qwen2.5-14B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/unsloth/Qwen2.5-14B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-23`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-32B-Instruct`,name:`Qwen2.5-32B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:33,page_url:`https://huggingface.co/unsloth/Qwen2.5-32B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-23`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-7B-Instruct`,name:`Qwen2.5-7B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/unsloth/Qwen2.5-7B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-Coder-32B-Instruct`,name:`Qwen2.5-Coder-32B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/unsloth/Qwen2.5-Coder-32B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-11-12`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen3-8B`,name:`Qwen3-8B`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Qwen3-8B`,release_date:`2025-04-28`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/tinyllama-chat`,name:`tinyllama-chat`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},page_url:`https://huggingface.co/unsloth/tinyllama-chat`,description:`[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/Discord%20button.png" width="200"/>](https://discord.gg/u54VK8m8tk) [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/buy%20me%20a%20coffee%20button.png" width="200"/>](https://ko-fi.com/unsloth) [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)`,release_date:`2024-02-14`,context_window:2048,license:`apache-2.0`,parameters:1.1,tagline:`[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/Discord%20button.png" width="200"/>](https://discord.gg/u54VK8m8tk) [<img src="https://raw.githubusercontent.com/unslothai/un`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct`,name:`Llama-3-SauerkrautLM-8b-Instruct`,created_by:`vagosolutions`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct`,release_date:`2024-04-19`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4-9B-0414`,name:`GLM-4-9B-0414`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4`,model_type:`chat`,parameters:9.4,page_url:`https://huggingface.co/zai-org/GLM-4-9B-0414`,release_date:`2025-04-07`,context_window:32768,license:`mit`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5-Air-FP8`,name:`GLM-4.5-Air-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5-Air-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5-Air`,name:`GLM-4.5-Air`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5-Air`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,parameters:110,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5`,name:`GLM-4.5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,parameters:358,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.6-FP8`,name:`GLM-4.6-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.6`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.6-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-09-29`,context_window:202752,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.6`,name:`GLM-4.6`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.6`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.6`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-09-29`,context_window:202752,license:`mit`,parameters:357,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7-Flash`,name:`GLM-4.7-Flash`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7-Flash`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2026-01-19`,context_window:202752,license:`mit`,parameters:31,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7-FP8`,name:`GLM-4.7-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-12-22`,context_window:202752,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7`,name:`GLM-4.7`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-12-22`,context_window:202752,license:`mit`,parameters:358,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-5`,name:`GLM-5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-5`,description:`👋 Join our <a href="https://raw.githubusercontent.com/zai-org/GLM-5/refs/heads/main/resources/wechat.png" target="_blank">WeChat</a> or <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2026-02-11`,context_window:202752,license:`mit`,parameters:754,tagline:`👋 Join our <a href="https://raw.githubusercontent.com/zai-org/GLM-5/refs/heads/main/resources/wechat.png" target="_blank">WeChat</a> or <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-Z1-32B-0414`,name:`GLM-Z1-32B-0414`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-21`,capabilities:{streaming:!0},model_type:`chat`,family:`glm`,parameters:32}]},t=e.models.map(t=>({...t,provider:e.id}));exports.models=t,exports.provider=e;
|