modelpedia 0.0.2 → 0.0.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (136) hide show
  1. package/README.md +141 -0
  2. package/dist/index.cjs +1 -101
  3. package/dist/index.d.cts +10 -2
  4. package/dist/index.d.mts +10 -2
  5. package/dist/index.mjs +1 -92
  6. package/dist/providers/alibaba.cjs +4 -835
  7. package/dist/providers/alibaba.d.cts +1 -1
  8. package/dist/providers/alibaba.d.mts +1 -1
  9. package/dist/providers/alibaba.mjs +4 -833
  10. package/dist/providers/amazon.cjs +4 -1484
  11. package/dist/providers/amazon.d.cts +1 -1
  12. package/dist/providers/amazon.d.mts +1 -1
  13. package/dist/providers/amazon.mjs +4 -1482
  14. package/dist/providers/anthropic.cjs +3 -698
  15. package/dist/providers/anthropic.d.cts +1 -1
  16. package/dist/providers/anthropic.d.mts +1 -1
  17. package/dist/providers/anthropic.mjs +3 -696
  18. package/dist/providers/azure.cjs +6 -1491
  19. package/dist/providers/azure.d.cts +1 -1
  20. package/dist/providers/azure.d.mts +1 -1
  21. package/dist/providers/azure.mjs +6 -1489
  22. package/dist/providers/baseten.cjs +3 -279
  23. package/dist/providers/baseten.d.cts +1 -1
  24. package/dist/providers/baseten.d.mts +1 -1
  25. package/dist/providers/baseten.mjs +3 -277
  26. package/dist/providers/cerebras.cjs +4 -145
  27. package/dist/providers/cerebras.d.cts +1 -1
  28. package/dist/providers/cerebras.d.mts +1 -1
  29. package/dist/providers/cerebras.mjs +4 -143
  30. package/dist/providers/cloudflare-ai-gateway.cjs +4 -0
  31. package/dist/providers/cloudflare-ai-gateway.d.cts +7 -0
  32. package/dist/providers/cloudflare-ai-gateway.d.mts +7 -0
  33. package/dist/providers/cloudflare-ai-gateway.mjs +4 -0
  34. package/dist/providers/cloudflare-workers-ai.cjs +4 -0
  35. package/dist/providers/cloudflare-workers-ai.d.cts +7 -0
  36. package/dist/providers/cloudflare-workers-ai.d.mts +7 -0
  37. package/dist/providers/cloudflare-workers-ai.mjs +4 -0
  38. package/dist/providers/cohere.cjs +5 -525
  39. package/dist/providers/cohere.d.cts +1 -1
  40. package/dist/providers/cohere.d.mts +1 -1
  41. package/dist/providers/cohere.mjs +5 -523
  42. package/dist/providers/cursor.cjs +1 -856
  43. package/dist/providers/cursor.d.cts +1 -1
  44. package/dist/providers/cursor.d.mts +1 -1
  45. package/dist/providers/cursor.mjs +1 -854
  46. package/dist/providers/deepseek.cjs +3 -64
  47. package/dist/providers/deepseek.d.cts +1 -1
  48. package/dist/providers/deepseek.d.mts +1 -1
  49. package/dist/providers/deepseek.mjs +3 -62
  50. package/dist/providers/fireworks.cjs +3 -3794
  51. package/dist/providers/fireworks.d.cts +1 -1
  52. package/dist/providers/fireworks.d.mts +1 -1
  53. package/dist/providers/fireworks.mjs +3 -3792
  54. package/dist/providers/google.cjs +6 -972
  55. package/dist/providers/google.d.cts +1 -1
  56. package/dist/providers/google.d.mts +1 -1
  57. package/dist/providers/google.mjs +6 -970
  58. package/dist/providers/groq.cjs +3 -353
  59. package/dist/providers/groq.d.cts +1 -1
  60. package/dist/providers/groq.d.mts +1 -1
  61. package/dist/providers/groq.mjs +3 -351
  62. package/dist/providers/huggingface.cjs +3 -1022
  63. package/dist/providers/huggingface.d.cts +1 -1
  64. package/dist/providers/huggingface.d.mts +1 -1
  65. package/dist/providers/huggingface.mjs +3 -1020
  66. package/dist/providers/meta.cjs +3 -481
  67. package/dist/providers/meta.d.cts +1 -1
  68. package/dist/providers/meta.d.mts +1 -1
  69. package/dist/providers/meta.mjs +3 -479
  70. package/dist/providers/minimax.cjs +3 -156
  71. package/dist/providers/minimax.d.cts +1 -1
  72. package/dist/providers/minimax.d.mts +1 -1
  73. package/dist/providers/minimax.mjs +3 -154
  74. package/dist/providers/mistral.cjs +3 -1594
  75. package/dist/providers/mistral.d.cts +1 -1
  76. package/dist/providers/mistral.d.mts +1 -1
  77. package/dist/providers/mistral.mjs +3 -1592
  78. package/dist/providers/moonshot.cjs +3 -138
  79. package/dist/providers/moonshot.d.cts +1 -1
  80. package/dist/providers/moonshot.d.mts +1 -1
  81. package/dist/providers/moonshot.mjs +3 -136
  82. package/dist/providers/nvidia.cjs +3 -2040
  83. package/dist/providers/nvidia.d.cts +1 -1
  84. package/dist/providers/nvidia.d.mts +1 -1
  85. package/dist/providers/nvidia.mjs +3 -2038
  86. package/dist/providers/ollama.cjs +3 -326
  87. package/dist/providers/ollama.d.cts +1 -1
  88. package/dist/providers/ollama.d.mts +1 -1
  89. package/dist/providers/ollama.mjs +3 -324
  90. package/dist/providers/openai.cjs +3 -9079
  91. package/dist/providers/openai.d.cts +1 -1
  92. package/dist/providers/openai.d.mts +1 -1
  93. package/dist/providers/openai.mjs +3 -9077
  94. package/dist/providers/opencode.cjs +3 -367
  95. package/dist/providers/opencode.d.cts +1 -1
  96. package/dist/providers/opencode.d.mts +1 -1
  97. package/dist/providers/opencode.mjs +3 -365
  98. package/dist/providers/openrouter.cjs +21 -7835
  99. package/dist/providers/openrouter.d.cts +1 -1
  100. package/dist/providers/openrouter.d.mts +1 -1
  101. package/dist/providers/openrouter.mjs +21 -7833
  102. package/dist/providers/perplexity.cjs +3 -311
  103. package/dist/providers/perplexity.d.cts +1 -1
  104. package/dist/providers/perplexity.d.mts +1 -1
  105. package/dist/providers/perplexity.mjs +3 -309
  106. package/dist/providers/qwen.cjs +3 -835
  107. package/dist/providers/qwen.d.cts +1 -1
  108. package/dist/providers/qwen.d.mts +1 -1
  109. package/dist/providers/qwen.mjs +3 -833
  110. package/dist/providers/together.cjs +4 -360
  111. package/dist/providers/together.d.cts +1 -1
  112. package/dist/providers/together.d.mts +1 -1
  113. package/dist/providers/together.mjs +4 -358
  114. package/dist/providers/vercel.cjs +3 -4037
  115. package/dist/providers/vercel.d.cts +1 -1
  116. package/dist/providers/vercel.d.mts +1 -1
  117. package/dist/providers/vercel.mjs +3 -4035
  118. package/dist/providers/vertex.cjs +10 -1072
  119. package/dist/providers/vertex.d.cts +1 -1
  120. package/dist/providers/vertex.d.mts +1 -1
  121. package/dist/providers/vertex.mjs +10 -1070
  122. package/dist/providers/xai.cjs +3 -301
  123. package/dist/providers/xai.d.cts +1 -1
  124. package/dist/providers/xai.d.mts +1 -1
  125. package/dist/providers/xai.mjs +3 -299
  126. package/dist/providers/zai.cjs +3 -243
  127. package/dist/providers/zai.d.cts +1 -1
  128. package/dist/providers/zai.d.mts +1 -1
  129. package/dist/providers/zai.mjs +3 -241
  130. package/dist/types.d.cts +80 -8
  131. package/dist/types.d.mts +80 -8
  132. package/package.json +3 -3
  133. package/dist/providers/cloudflare.cjs +0 -826
  134. package/dist/providers/cloudflare.d.cts +0 -7
  135. package/dist/providers/cloudflare.d.mts +0 -7
  136. package/dist/providers/cloudflare.mjs +0 -824
@@ -1,1022 +1,3 @@
1
- Object.defineProperty(exports, Symbol.toStringTag, { value: "Module" });
2
- //#region src/providers/huggingface.ts
3
- const provider = {
4
- "id": "huggingface",
5
- "name": "Hugging Face",
6
- "region": "US",
7
- "url": "https://huggingface.co",
8
- "api_url": "https://router.huggingface.co/v1",
9
- "docs_url": "https://huggingface.co/docs/api-inference",
10
- "pricing_url": "https://huggingface.co/pricing",
11
- "icon": "<svg xmlns=\"http://www.w3.org/2000/svg\" viewBox=\"0 0 24 24\" fill=\"currentColor\" fill-rule=\"evenodd\">\n <path d=\"M16.781 3.277c2.997 1.704 4.844 4.851 4.844 8.258 0 .995-.155 1.955-.443 2.857a1.332 1.332 0 011.125.4 1.41 1.41 0 01.2 1.723c.204.165.352.385.428.632l.017.062c.06.222.12.69-.2 1.166.244.37.279.836.093 1.236-.255.57-.893 1.018-2.128 1.5l-.202.078-.131.048c-.478.173-.89.295-1.061.345l-.086.024c-.89.243-1.808.375-2.732.394-1.32 0-2.3-.36-2.923-1.067a9.852 9.852 0 01-3.18.018C9.778 21.647 8.802 22 7.494 22a11.249 11.249 0 01-2.541-.343l-.221-.06-.273-.08a16.574 16.574 0 01-1.175-.405c-1.237-.483-1.875-.93-2.13-1.501-.186-.4-.151-.867.093-1.236a1.42 1.42 0 01-.2-1.166c.069-.273.226-.516.447-.694a1.41 1.41 0 01.2-1.722c.233-.248.557-.391.917-.407l.078-.001a9.385 9.385 0 01-.44-2.85c0-3.407 1.847-6.554 4.844-8.258a9.822 9.822 0 019.687 0zM4.188 14.758c.125.687 2.357 2.35 2.14 2.707-.19.315-.796-.239-.948-.386l-.041-.04-.168-.147c-.561-.479-2.304-1.9-2.74-1.432-.43.46.119.859 1.055 1.42l.784.467.136.083c1.045.643 1.12.84.95 1.113-.188.295-3.07-2.1-3.34-1.083-.27 1.011 2.942 1.304 2.744 2.006-.2.7-2.265-1.324-2.685-.537-.425.79 2.913 1.718 2.94 1.725l.16.04.175.042c1.227.284 3.565.65 4.435-.604.673-.973.64-1.709-.248-2.61l-.057-.057c-.945-.928-1.495-2.288-1.495-2.288l-.017-.058-.025-.072c-.082-.22-.284-.639-.63-.584-.46.073-.798 1.21.12 1.933l.05.038c.977.721-.195 1.21-.573.534l-.058-.104-.143-.25c-.463-.799-1.282-2.111-1.739-2.397-.532-.332-.907-.148-.782.541zm14.842-.541c-.533.335-1.563 2.074-1.94 2.751a.613.613 0 01-.687.302.436.436 0 01-.176-.098.303.303 0 01-.049-.06l-.014-.028-.008-.02-.007-.019-.003-.013-.003-.017a.289.289 0 01-.004-.048c0-.12.071-.266.25-.427.026-.024.054-.047.084-.07l.047-.036c.022-.016.043-.032.063-.049.883-.71.573-1.81.131-1.917l-.031-.006-.056-.004a.368.368 0 00-.062.006l-.028.005-.042.014-.039.017-.028.015-.028.019-.036.027-.023.02c-.173.158-.273.428-.31.542l-.016.054s-.53 1.309-1.439 2.234l-.054.054c-.365.358-.596.69-.702 1.018-.143.437-.066.868.21 1.353.055.097.117.195.187.296.882 1.275 3.282.876 4.494.59l.286-.07.25-.074c.276-.084.736-.233 1.2-.42l.188-.077.065-.028.064-.028.124-.056.081-.038c.529-.252.964-.543.994-.827l.001-.036a.299.299 0 00-.037-.139c-.094-.176-.271-.212-.491-.168l-.045.01c-.044.01-.09.024-.136.04l-.097.035-.054.022c-.559.23-1.238.705-1.607.745h.006a.452.452 0 01-.05.003h-.024l-.024-.003-.023-.005c-.068-.016-.116-.06-.14-.142a.22.22 0 01-.005-.1c.062-.345.958-.595 1.713-.91l.066-.028c.528-.224.97-.483.985-.832v-.04a.47.47 0 00-.016-.098c-.048-.18-.175-.251-.36-.251-.785 0-2.55 1.36-2.92 1.36-.025 0-.048-.007-.058-.024a.6.6 0 01-.046-.088c-.1-.238.068-.462 1.06-1.066l.209-.126c.538-.32 1.01-.588 1.341-.831.29-.212.475-.406.503-.6l.003-.028c.008-.113-.038-.227-.147-.344a.266.266 0 00-.07-.054l-.034-.015-.013-.005a.403.403 0 00-.13-.02c-.162 0-.369.07-.595.18-.637.313-1.431.952-1.826 1.285l-.249.215-.033.033c-.08.078-.288.27-.493.386l-.071.037-.041.019a.535.535 0 01-.122.036h.005a.346.346 0 01-.031.003l.01-.001-.013.001c-.079.005-.145-.021-.19-.095a.113.113 0 01-.014-.065c.027-.465 2.034-1.991 2.152-2.642l.009-.048c.1-.65-.271-.817-.791-.493zM11.938 2.984c-4.798 0-8.688 3.829-8.688 8.55 0 .692.083 1.364.24 2.008l.008-.009c.252-.298.612-.46 1.017-.46.355.008.699.117.993.312.22.14.465.384.715.694.261-.372.69-.598 1.15-.605.852 0 1.367.728 1.562 1.383l.047.105.06.127c.192.396.595 1.139 1.143 1.68 1.06 1.04 1.324 2.115.8 3.266a8.865 8.865 0 002.024-.014c-.505-1.12-.26-2.17.74-3.186l.066-.066c.695-.684 1.157-1.69 1.252-1.912.195-.655.708-1.383 1.56-1.383.46.007.889.233 1.15.605.25-.31.495-.553.718-.694a1.87 1.87 0 01.99-.312c.357 0 .682.126.925.36.14-.61.215-1.245.215-1.898 0-4.722-3.89-8.55-8.687-8.55zm1.857 8.926l.439-.212c.553-.264.89-.383.89.152 0 1.093-.771 3.208-3.155 3.262h-.184c-2.325-.052-3.116-2.06-3.156-3.175l-.001-.087c0-1.107 1.452.586 3.25.586.716 0 1.379-.272 1.917-.526zm4.017-3.143c.45 0 .813.358.813.8 0 .441-.364.8-.813.8a.806.806 0 01-.812-.8c0-.442.364-.8.812-.8zm-11.624 0c.448 0 .812.358.812.8 0 .441-.364.8-.812.8a.806.806 0 01-.813-.8c0-.442.364-.8.813-.8zm7.79-.841c.32-.384.846-.54 1.33-.394.483.146.83.564.878 1.06.048.495-.212.97-.659 1.203-.322.168-.447-.477-.767-.585l.002-.003c-.287-.098-.772.362-.925.079a1.215 1.215 0 01.14-1.36zm-4.323 0c.322.384.377.92.14 1.36-.152.283-.64-.177-.925-.079l.003.003c-.108.036-.194.134-.273.24l-.118.165c-.11.15-.22.262-.377.18a1.226 1.226 0 01-.658-1.204c.048-.495.395-.913.878-1.059a1.262 1.262 0 011.33.394z\"/>\n</svg>",
12
- "models": [
13
- {
14
- "id": "allenai/Olmo-3-7B-Instruct",
15
- "name": "Olmo-3-7B-Instruct",
16
- "created_by": "allenai",
17
- "source": "official",
18
- "last_updated": "2026-03-21",
19
- "capabilities": { "streaming": true }
20
- },
21
- {
22
- "id": "allenai/Olmo-3.1-32B-Instruct",
23
- "name": "Olmo-3.1-32B-Instruct",
24
- "created_by": "allenai",
25
- "source": "official",
26
- "last_updated": "2026-03-21",
27
- "capabilities": { "streaming": true }
28
- },
29
- {
30
- "id": "argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1",
31
- "name": "Llama-3.2-1B-Instruct-APIGen-FC-v0.1",
32
- "created_by": "argilla",
33
- "source": "official",
34
- "last_updated": "2026-03-21",
35
- "capabilities": { "streaming": true }
36
- },
37
- {
38
- "id": "argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC",
39
- "name": "Llama-3.2-1B-Instruct-v2-FC",
40
- "created_by": "argilla-warehouse",
41
- "source": "official",
42
- "last_updated": "2026-03-21",
43
- "capabilities": { "streaming": true }
44
- },
45
- {
46
- "id": "DeepMount00/Llama-3-8b-Ita",
47
- "name": "Llama-3-8b-Ita",
48
- "created_by": "deepmount00",
49
- "source": "official",
50
- "last_updated": "2026-03-21",
51
- "capabilities": { "streaming": true }
52
- },
53
- {
54
- "id": "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B",
55
- "name": "DeepSeek-R1-0528-Qwen3-8B",
56
- "created_by": "deepseek",
57
- "source": "official",
58
- "last_updated": "2026-03-21",
59
- "capabilities": { "streaming": true }
60
- },
61
- {
62
- "id": "Efficient-Large-Model/gemma-2-2b-it",
63
- "name": "gemma-2-2b-it",
64
- "created_by": "efficient-large-model",
65
- "source": "official",
66
- "last_updated": "2026-03-21",
67
- "family": "gemma-2",
68
- "capabilities": { "streaming": true }
69
- },
70
- {
71
- "id": "EssentialAI/rnj-1-instruct",
72
- "name": "rnj-1-instruct",
73
- "created_by": "essentialai",
74
- "source": "official",
75
- "last_updated": "2026-03-21",
76
- "capabilities": { "streaming": true }
77
- },
78
- {
79
- "id": "failspy/Meta-Llama-3-8B-Instruct-abliterated-v3",
80
- "name": "Meta-Llama-3-8B-Instruct-abliterated-v3",
81
- "created_by": "failspy",
82
- "source": "official",
83
- "last_updated": "2026-03-21",
84
- "capabilities": { "streaming": true }
85
- },
86
- {
87
- "id": "Featherless-Chat-Models/Mistral-7B-Instruct-v0.2",
88
- "name": "Mistral-7B-Instruct-v0.2",
89
- "created_by": "featherless-chat-models",
90
- "source": "official",
91
- "last_updated": "2026-03-21",
92
- "capabilities": { "streaming": true }
93
- },
94
- {
95
- "id": "FlagAlpha/Llama3-Chinese-8B-Instruct",
96
- "name": "Llama3-Chinese-8B-Instruct",
97
- "created_by": "flagalpha",
98
- "source": "official",
99
- "last_updated": "2026-03-21",
100
- "capabilities": { "streaming": true }
101
- },
102
- {
103
- "id": "google/gemma-2-2b-jpn-it",
104
- "name": "gemma-2-2b-jpn-it",
105
- "created_by": "google",
106
- "source": "official",
107
- "last_updated": "2026-03-21",
108
- "family": "gemma-2",
109
- "capabilities": { "streaming": true }
110
- },
111
- {
112
- "id": "google/gemma-2-9b-it",
113
- "name": "gemma-2-9b-it",
114
- "created_by": "google",
115
- "source": "official",
116
- "last_updated": "2026-03-21",
117
- "family": "gemma-2",
118
- "capabilities": { "streaming": true }
119
- },
120
- {
121
- "id": "GraySwanAI/Llama-3-8B-Instruct-RR",
122
- "name": "Llama-3-8B-Instruct-RR",
123
- "created_by": "grayswanai",
124
- "source": "official",
125
- "last_updated": "2026-03-21",
126
- "capabilities": { "streaming": true }
127
- },
128
- {
129
- "id": "GritLM/GritLM-7B",
130
- "name": "GritLM-7B",
131
- "created_by": "gritlm",
132
- "source": "official",
133
- "last_updated": "2026-03-21",
134
- "capabilities": { "streaming": true }
135
- },
136
- {
137
- "id": "huihui-ai/Qwen2.5-72B-Instruct-abliterated",
138
- "name": "Qwen2.5-72B-Instruct-abliterated",
139
- "created_by": "huihui-ai",
140
- "source": "official",
141
- "last_updated": "2026-03-21",
142
- "capabilities": { "streaming": true }
143
- },
144
- {
145
- "id": "instruction-pretrain/finance-Llama3-8B",
146
- "name": "finance-Llama3-8B",
147
- "created_by": "instruction-pretrain",
148
- "source": "official",
149
- "last_updated": "2026-03-21",
150
- "capabilities": { "streaming": true }
151
- },
152
- {
153
- "id": "Intel/neural-chat-7b-v3-3",
154
- "name": "neural-chat-7b-v3-3",
155
- "created_by": "intel",
156
- "source": "official",
157
- "last_updated": "2026-03-21",
158
- "capabilities": { "streaming": true }
159
- },
160
- {
161
- "id": "invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp",
162
- "name": "Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp",
163
- "created_by": "invalid-coder",
164
- "source": "official",
165
- "last_updated": "2026-03-21",
166
- "capabilities": { "streaming": true }
167
- },
168
- {
169
- "id": "m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix",
170
- "name": "Mistral-Nemo-Instruct-2407-lenient-chatfix",
171
- "created_by": "m8than",
172
- "source": "official",
173
- "last_updated": "2026-03-21",
174
- "capabilities": { "streaming": true }
175
- },
176
- {
177
- "id": "meta-llama/Llama-2-13b-chat-hf",
178
- "name": "Llama-2-13b-chat-hf",
179
- "created_by": "meta",
180
- "source": "official",
181
- "last_updated": "2026-03-21",
182
- "capabilities": { "streaming": true }
183
- },
184
- {
185
- "id": "meta-llama/Llama-3.1-70B-Instruct",
186
- "name": "Llama-3.1-70B-Instruct",
187
- "created_by": "meta",
188
- "source": "official",
189
- "last_updated": "2026-03-21",
190
- "capabilities": { "streaming": true }
191
- },
192
- {
193
- "id": "meta-llama/Llama-3.1-8B-Instruct",
194
- "name": "Llama-3.1-8B-Instruct",
195
- "created_by": "meta",
196
- "source": "official",
197
- "last_updated": "2026-03-21",
198
- "capabilities": { "streaming": true }
199
- },
200
- {
201
- "id": "meta-llama/Llama-3.2-1B-Instruct",
202
- "name": "Llama-3.2-1B-Instruct",
203
- "created_by": "meta",
204
- "source": "official",
205
- "last_updated": "2026-03-21",
206
- "capabilities": { "streaming": true }
207
- },
208
- {
209
- "id": "meta-llama/Llama-3.3-70B-Instruct",
210
- "name": "Llama-3.3-70B-Instruct",
211
- "created_by": "meta",
212
- "source": "official",
213
- "last_updated": "2026-03-21",
214
- "capabilities": { "streaming": true }
215
- },
216
- {
217
- "id": "meta-llama/Meta-Llama-3-70B-Instruct",
218
- "name": "Meta-Llama-3-70B-Instruct",
219
- "created_by": "meta",
220
- "source": "official",
221
- "last_updated": "2026-03-21",
222
- "capabilities": { "streaming": true }
223
- },
224
- {
225
- "id": "meta-llama/Meta-Llama-3-8B-Instruct",
226
- "name": "Meta-Llama-3-8B-Instruct",
227
- "created_by": "meta",
228
- "source": "official",
229
- "last_updated": "2026-03-21",
230
- "capabilities": { "streaming": true }
231
- },
232
- {
233
- "id": "mistralai/Mistral-7B-Instruct-v0.2",
234
- "name": "Mistral-7B-Instruct-v0.2",
235
- "created_by": "mistral",
236
- "source": "official",
237
- "last_updated": "2026-03-21",
238
- "capabilities": { "streaming": true }
239
- },
240
- {
241
- "id": "mlabonne/NeuralDaredevil-8B-abliterated",
242
- "name": "NeuralDaredevil-8B-abliterated",
243
- "created_by": "mlabonne",
244
- "source": "official",
245
- "last_updated": "2026-03-21",
246
- "capabilities": { "streaming": true }
247
- },
248
- {
249
- "id": "moonshotai/Kimi-K2-Instruct-0905",
250
- "name": "Kimi-K2-Instruct-0905",
251
- "created_by": "moonshot",
252
- "source": "official",
253
- "last_updated": "2026-03-21",
254
- "capabilities": { "streaming": true }
255
- },
256
- {
257
- "id": "moonshotai/Kimi-K2-Instruct",
258
- "name": "Kimi-K2-Instruct",
259
- "created_by": "moonshot",
260
- "source": "official",
261
- "last_updated": "2026-03-21",
262
- "capabilities": { "streaming": true }
263
- },
264
- {
265
- "id": "moonshotai/Kimi-K2-Thinking",
266
- "name": "Kimi-K2-Thinking",
267
- "created_by": "moonshot",
268
- "source": "official",
269
- "last_updated": "2026-03-21",
270
- "capabilities": { "streaming": true }
271
- },
272
- {
273
- "id": "NousResearch/Meta-Llama-3-70B-Instruct",
274
- "name": "Meta-Llama-3-70B-Instruct",
275
- "created_by": "nousresearch",
276
- "source": "official",
277
- "last_updated": "2026-03-21",
278
- "capabilities": { "streaming": true }
279
- },
280
- {
281
- "id": "NousResearch/Meta-Llama-3-8B-Instruct",
282
- "name": "Meta-Llama-3-8B-Instruct",
283
- "created_by": "nousresearch",
284
- "source": "official",
285
- "last_updated": "2026-03-21",
286
- "capabilities": { "streaming": true }
287
- },
288
- {
289
- "id": "NousResearch/Meta-Llama-3.1-8B-Instruct",
290
- "name": "Meta-Llama-3.1-8B-Instruct",
291
- "created_by": "nousresearch",
292
- "source": "official",
293
- "last_updated": "2026-03-21",
294
- "capabilities": { "streaming": true }
295
- },
296
- {
297
- "id": "nvidia/Llama-3.1-8B-Instruct-FP8",
298
- "name": "Llama-3.1-8B-Instruct-FP8",
299
- "created_by": "nvidia",
300
- "source": "official",
301
- "last_updated": "2026-03-21",
302
- "capabilities": { "streaming": true }
303
- },
304
- {
305
- "id": "openai/gpt-oss-120b",
306
- "name": "gpt-oss-120b",
307
- "created_by": "openai",
308
- "source": "official",
309
- "last_updated": "2026-03-21",
310
- "family": "gpt-oss",
311
- "capabilities": { "streaming": true }
312
- },
313
- {
314
- "id": "openai/gpt-oss-20b",
315
- "name": "gpt-oss-20b",
316
- "created_by": "openai",
317
- "source": "official",
318
- "last_updated": "2026-03-21",
319
- "family": "gpt-oss",
320
- "capabilities": { "streaming": true }
321
- },
322
- {
323
- "id": "openai/gpt-oss-safeguard-20b",
324
- "name": "gpt-oss-safeguard-20b",
325
- "created_by": "openai",
326
- "source": "official",
327
- "last_updated": "2026-03-21",
328
- "family": "gpt-oss",
329
- "capabilities": { "streaming": true },
330
- "model_type": "moderation"
331
- },
332
- {
333
- "id": "openchat/openchat-3.6-8b-20240522",
334
- "name": "openchat-3.6-8b-20240522",
335
- "created_by": "openchat",
336
- "source": "official",
337
- "last_updated": "2026-03-21",
338
- "alias": "openchat/openchat-3.6-8b",
339
- "capabilities": { "streaming": true }
340
- },
341
- {
342
- "id": "openchat/openchat-3.6-8b",
343
- "name": "openchat-3.6-8b-20240522",
344
- "created_by": "openchat",
345
- "source": "official",
346
- "last_updated": "2026-03-21",
347
- "capabilities": { "streaming": true },
348
- "snapshots": ["openchat/openchat-3.6-8b-20240522"]
349
- },
350
- {
351
- "id": "Orion-zhen/Qwen2.5-7B-Instruct-Uncensored",
352
- "name": "Qwen2.5-7B-Instruct-Uncensored",
353
- "created_by": "orion-zhen",
354
- "source": "official",
355
- "last_updated": "2026-03-21",
356
- "capabilities": { "streaming": true }
357
- },
358
- {
359
- "id": "PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1",
360
- "name": "Llama-3-Patronus-Lynx-8B-Instruct-v1.1",
361
- "created_by": "patronusai",
362
- "source": "official",
363
- "last_updated": "2026-03-21",
364
- "capabilities": { "streaming": true }
365
- },
366
- {
367
- "id": "Qwen/Qwen1.5-1.8B-Chat",
368
- "name": "Qwen1.5-1.8B-Chat",
369
- "created_by": "qwen",
370
- "source": "official",
371
- "last_updated": "2026-03-21",
372
- "capabilities": { "streaming": true }
373
- },
374
- {
375
- "id": "Qwen/Qwen2-1.5B-Instruct",
376
- "name": "Qwen2-1.5B-Instruct",
377
- "created_by": "qwen",
378
- "source": "official",
379
- "last_updated": "2026-03-21",
380
- "capabilities": { "streaming": true }
381
- },
382
- {
383
- "id": "Qwen/Qwen2-72B-Instruct",
384
- "name": "Qwen2-72B-Instruct",
385
- "created_by": "qwen",
386
- "source": "official",
387
- "last_updated": "2026-03-21",
388
- "capabilities": { "streaming": true }
389
- },
390
- {
391
- "id": "Qwen/Qwen2-7B-Instruct",
392
- "name": "Qwen2-7B-Instruct",
393
- "created_by": "qwen",
394
- "source": "official",
395
- "last_updated": "2026-03-21",
396
- "capabilities": { "streaming": true }
397
- },
398
- {
399
- "id": "Qwen/Qwen2.5-1.5B-Instruct",
400
- "name": "Qwen2.5-1.5B-Instruct",
401
- "created_by": "qwen",
402
- "source": "official",
403
- "last_updated": "2026-03-21",
404
- "capabilities": { "streaming": true }
405
- },
406
- {
407
- "id": "Qwen/Qwen2.5-14B-Instruct-1M",
408
- "name": "Qwen2.5-14B-Instruct-1M",
409
- "created_by": "qwen",
410
- "source": "official",
411
- "last_updated": "2026-03-21",
412
- "capabilities": { "streaming": true }
413
- },
414
- {
415
- "id": "Qwen/Qwen2.5-14B-Instruct",
416
- "name": "Qwen2.5-14B-Instruct",
417
- "created_by": "qwen",
418
- "source": "official",
419
- "last_updated": "2026-03-21",
420
- "capabilities": { "streaming": true }
421
- },
422
- {
423
- "id": "Qwen/Qwen2.5-32B-Instruct",
424
- "name": "Qwen2.5-32B-Instruct",
425
- "created_by": "qwen",
426
- "source": "official",
427
- "last_updated": "2026-03-21",
428
- "capabilities": { "streaming": true }
429
- },
430
- {
431
- "id": "Qwen/Qwen2.5-72B-Instruct",
432
- "name": "Qwen2.5-72B-Instruct",
433
- "created_by": "qwen",
434
- "source": "official",
435
- "last_updated": "2026-03-21",
436
- "capabilities": { "streaming": true }
437
- },
438
- {
439
- "id": "Qwen/Qwen2.5-7B-Instruct-1M",
440
- "name": "Qwen2.5-7B-Instruct-1M",
441
- "created_by": "qwen",
442
- "source": "official",
443
- "last_updated": "2026-03-21",
444
- "capabilities": { "streaming": true }
445
- },
446
- {
447
- "id": "Qwen/Qwen2.5-7B-Instruct",
448
- "name": "Qwen2.5-7B-Instruct",
449
- "created_by": "qwen",
450
- "source": "official",
451
- "last_updated": "2026-03-21",
452
- "capabilities": { "streaming": true }
453
- },
454
- {
455
- "id": "Qwen/Qwen2.5-Coder-1.5B-Instruct",
456
- "name": "Qwen2.5-Coder-1.5B-Instruct",
457
- "created_by": "qwen",
458
- "source": "official",
459
- "last_updated": "2026-03-21",
460
- "capabilities": { "streaming": true }
461
- },
462
- {
463
- "id": "Qwen/Qwen2.5-Coder-1.5B",
464
- "name": "Qwen2.5-Coder-1.5B",
465
- "created_by": "qwen",
466
- "source": "official",
467
- "last_updated": "2026-03-21",
468
- "capabilities": { "streaming": true }
469
- },
470
- {
471
- "id": "Qwen/Qwen2.5-Coder-14B-Instruct",
472
- "name": "Qwen2.5-Coder-14B-Instruct",
473
- "created_by": "qwen",
474
- "source": "official",
475
- "last_updated": "2026-03-21",
476
- "capabilities": { "streaming": true }
477
- },
478
- {
479
- "id": "Qwen/Qwen2.5-Coder-14B",
480
- "name": "Qwen2.5-Coder-14B",
481
- "created_by": "qwen",
482
- "source": "official",
483
- "last_updated": "2026-03-21",
484
- "capabilities": { "streaming": true }
485
- },
486
- {
487
- "id": "Qwen/Qwen2.5-Coder-32B-Instruct",
488
- "name": "Qwen2.5-Coder-32B-Instruct",
489
- "created_by": "qwen",
490
- "source": "official",
491
- "last_updated": "2026-03-21",
492
- "capabilities": { "streaming": true }
493
- },
494
- {
495
- "id": "Qwen/Qwen2.5-Coder-32B",
496
- "name": "Qwen2.5-Coder-32B",
497
- "created_by": "qwen",
498
- "source": "official",
499
- "last_updated": "2026-03-21",
500
- "capabilities": { "streaming": true }
501
- },
502
- {
503
- "id": "Qwen/Qwen2.5-Coder-3B-Instruct",
504
- "name": "Qwen2.5-Coder-3B-Instruct",
505
- "created_by": "qwen",
506
- "source": "official",
507
- "last_updated": "2026-03-21",
508
- "capabilities": { "streaming": true }
509
- },
510
- {
511
- "id": "Qwen/Qwen2.5-Coder-7B-Instruct",
512
- "name": "Qwen2.5-Coder-7B-Instruct",
513
- "created_by": "qwen",
514
- "source": "official",
515
- "last_updated": "2026-03-21",
516
- "capabilities": { "streaming": true }
517
- },
518
- {
519
- "id": "Qwen/Qwen2.5-Coder-7B",
520
- "name": "Qwen2.5-Coder-7B",
521
- "created_by": "qwen",
522
- "source": "official",
523
- "last_updated": "2026-03-21",
524
- "capabilities": { "streaming": true }
525
- },
526
- {
527
- "id": "Qwen/Qwen2.5-Math-1.5B-Instruct",
528
- "name": "Qwen2.5-Math-1.5B-Instruct",
529
- "created_by": "qwen",
530
- "source": "official",
531
- "last_updated": "2026-03-21",
532
- "capabilities": { "streaming": true }
533
- },
534
- {
535
- "id": "Qwen/Qwen2.5-Math-7B-Instruct",
536
- "name": "Qwen2.5-Math-7B-Instruct",
537
- "created_by": "qwen",
538
- "source": "official",
539
- "last_updated": "2026-03-21",
540
- "capabilities": { "streaming": true }
541
- },
542
- {
543
- "id": "Qwen/Qwen3-1.7B",
544
- "name": "Qwen3-1.7B",
545
- "created_by": "qwen",
546
- "source": "official",
547
- "last_updated": "2026-03-21",
548
- "capabilities": { "streaming": true }
549
- },
550
- {
551
- "id": "Qwen/Qwen3-14B-Base",
552
- "name": "Qwen3-14B-Base",
553
- "created_by": "qwen",
554
- "source": "official",
555
- "last_updated": "2026-03-21",
556
- "capabilities": { "streaming": true }
557
- },
558
- {
559
- "id": "Qwen/Qwen3-14B",
560
- "name": "Qwen3-14B",
561
- "created_by": "qwen",
562
- "source": "official",
563
- "last_updated": "2026-03-21",
564
- "capabilities": { "streaming": true }
565
- },
566
- {
567
- "id": "Qwen/Qwen3-235B-A22B-Instruct-2507",
568
- "name": "Qwen3-235B-A22B-Instruct-2507",
569
- "created_by": "qwen",
570
- "source": "official",
571
- "last_updated": "2026-03-21",
572
- "alias": "Qwen/Qwen3-235B-A22B-Instruct",
573
- "capabilities": { "streaming": true }
574
- },
575
- {
576
- "id": "Qwen/Qwen3-235B-A22B-Instruct",
577
- "name": "Qwen3-235B-A22B-Instruct-2507",
578
- "created_by": "qwen",
579
- "source": "official",
580
- "last_updated": "2026-03-21",
581
- "capabilities": { "streaming": true },
582
- "snapshots": ["Qwen/Qwen3-235B-A22B-Instruct-2507"]
583
- },
584
- {
585
- "id": "Qwen/Qwen3-235B-A22B-Thinking-2507",
586
- "name": "Qwen3-235B-A22B-Thinking-2507",
587
- "created_by": "qwen",
588
- "source": "official",
589
- "last_updated": "2026-03-21",
590
- "alias": "Qwen/Qwen3-235B-A22B-Thinking",
591
- "capabilities": { "streaming": true }
592
- },
593
- {
594
- "id": "Qwen/Qwen3-235B-A22B-Thinking",
595
- "name": "Qwen3-235B-A22B-Thinking-2507",
596
- "created_by": "qwen",
597
- "source": "official",
598
- "last_updated": "2026-03-21",
599
- "capabilities": { "streaming": true },
600
- "snapshots": ["Qwen/Qwen3-235B-A22B-Thinking-2507"]
601
- },
602
- {
603
- "id": "Qwen/Qwen3-235B-A22B",
604
- "name": "Qwen3-235B-A22B",
605
- "created_by": "qwen",
606
- "source": "official",
607
- "last_updated": "2026-03-21",
608
- "capabilities": { "streaming": true }
609
- },
610
- {
611
- "id": "Qwen/Qwen3-30B-A3B-Instruct-2507",
612
- "name": "Qwen3-30B-A3B-Instruct-2507",
613
- "created_by": "qwen",
614
- "source": "official",
615
- "last_updated": "2026-03-21",
616
- "alias": "Qwen/Qwen3-30B-A3B-Instruct",
617
- "capabilities": { "streaming": true }
618
- },
619
- {
620
- "id": "Qwen/Qwen3-30B-A3B-Instruct",
621
- "name": "Qwen3-30B-A3B-Instruct-2507",
622
- "created_by": "qwen",
623
- "source": "official",
624
- "last_updated": "2026-03-21",
625
- "capabilities": { "streaming": true },
626
- "snapshots": ["Qwen/Qwen3-30B-A3B-Instruct-2507"]
627
- },
628
- {
629
- "id": "Qwen/Qwen3-30B-A3B",
630
- "name": "Qwen3-30B-A3B",
631
- "created_by": "qwen",
632
- "source": "official",
633
- "last_updated": "2026-03-21",
634
- "capabilities": { "streaming": true }
635
- },
636
- {
637
- "id": "Qwen/Qwen3-32B",
638
- "name": "Qwen3-32B",
639
- "created_by": "qwen",
640
- "source": "official",
641
- "last_updated": "2026-03-21",
642
- "capabilities": { "streaming": true }
643
- },
644
- {
645
- "id": "Qwen/Qwen3-4B-Instruct-2507",
646
- "name": "Qwen3-4B-Instruct-2507",
647
- "created_by": "qwen",
648
- "source": "official",
649
- "last_updated": "2026-03-21",
650
- "alias": "Qwen/Qwen3-4B-Instruct",
651
- "capabilities": { "streaming": true }
652
- },
653
- {
654
- "id": "Qwen/Qwen3-4B-Instruct",
655
- "name": "Qwen3-4B-Instruct-2507",
656
- "created_by": "qwen",
657
- "source": "official",
658
- "last_updated": "2026-03-21",
659
- "capabilities": { "streaming": true },
660
- "snapshots": ["Qwen/Qwen3-4B-Instruct-2507"]
661
- },
662
- {
663
- "id": "Qwen/Qwen3-4B-Thinking-2507",
664
- "name": "Qwen3-4B-Thinking-2507",
665
- "created_by": "qwen",
666
- "source": "official",
667
- "last_updated": "2026-03-21",
668
- "alias": "Qwen/Qwen3-4B-Thinking",
669
- "capabilities": { "streaming": true }
670
- },
671
- {
672
- "id": "Qwen/Qwen3-4B-Thinking",
673
- "name": "Qwen3-4B-Thinking-2507",
674
- "created_by": "qwen",
675
- "source": "official",
676
- "last_updated": "2026-03-21",
677
- "capabilities": { "streaming": true },
678
- "snapshots": ["Qwen/Qwen3-4B-Thinking-2507"]
679
- },
680
- {
681
- "id": "Qwen/Qwen3-8B-Base",
682
- "name": "Qwen3-8B-Base",
683
- "created_by": "qwen",
684
- "source": "official",
685
- "last_updated": "2026-03-21",
686
- "capabilities": { "streaming": true }
687
- },
688
- {
689
- "id": "Qwen/Qwen3-8B",
690
- "name": "Qwen3-8B",
691
- "created_by": "qwen",
692
- "source": "official",
693
- "last_updated": "2026-03-21",
694
- "capabilities": { "streaming": true }
695
- },
696
- {
697
- "id": "Qwen/Qwen3-Coder-30B-A3B-Instruct",
698
- "name": "Qwen3-Coder-30B-A3B-Instruct",
699
- "created_by": "qwen",
700
- "source": "official",
701
- "last_updated": "2026-03-21",
702
- "capabilities": { "streaming": true }
703
- },
704
- {
705
- "id": "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8",
706
- "name": "Qwen3-Coder-480B-A35B-Instruct-FP8",
707
- "created_by": "qwen",
708
- "source": "official",
709
- "last_updated": "2026-03-21",
710
- "capabilities": { "streaming": true }
711
- },
712
- {
713
- "id": "Qwen/Qwen3-Coder-480B-A35B-Instruct",
714
- "name": "Qwen3-Coder-480B-A35B-Instruct",
715
- "created_by": "qwen",
716
- "source": "official",
717
- "last_updated": "2026-03-21",
718
- "capabilities": { "streaming": true }
719
- },
720
- {
721
- "id": "Qwen/Qwen3-Coder-Next-FP8",
722
- "name": "Qwen3-Coder-Next-FP8",
723
- "created_by": "qwen",
724
- "source": "official",
725
- "last_updated": "2026-03-21",
726
- "capabilities": { "streaming": true }
727
- },
728
- {
729
- "id": "Qwen/Qwen3-Coder-Next",
730
- "name": "Qwen3-Coder-Next",
731
- "created_by": "qwen",
732
- "source": "official",
733
- "last_updated": "2026-03-21",
734
- "capabilities": { "streaming": true }
735
- },
736
- {
737
- "id": "Qwen/Qwen3-Next-80B-A3B-Instruct",
738
- "name": "Qwen3-Next-80B-A3B-Instruct",
739
- "created_by": "qwen",
740
- "source": "official",
741
- "last_updated": "2026-03-21",
742
- "capabilities": { "streaming": true }
743
- },
744
- {
745
- "id": "Qwen/Qwen3-Next-80B-A3B-Thinking",
746
- "name": "Qwen3-Next-80B-A3B-Thinking",
747
- "created_by": "qwen",
748
- "source": "official",
749
- "last_updated": "2026-03-21",
750
- "capabilities": { "streaming": true }
751
- },
752
- {
753
- "id": "RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV",
754
- "name": "Meta-Llama-3-8B-Instruct-FP8-KV",
755
- "created_by": "redhatai",
756
- "source": "official",
757
- "last_updated": "2026-03-21",
758
- "capabilities": { "streaming": true }
759
- },
760
- {
761
- "id": "SakanaAI/TinySwallow-1.5B-Instruct",
762
- "name": "TinySwallow-1.5B-Instruct",
763
- "created_by": "sakanaai",
764
- "source": "official",
765
- "last_updated": "2026-03-21",
766
- "capabilities": { "streaming": true }
767
- },
768
- {
769
- "id": "speakleash/Bielik-7B-Instruct-v0.1",
770
- "name": "Bielik-7B-Instruct-v0.1",
771
- "created_by": "speakleash",
772
- "source": "official",
773
- "last_updated": "2026-03-21",
774
- "capabilities": { "streaming": true }
775
- },
776
- {
777
- "id": "swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
778
- "name": "LLaMAntino-3-ANITA-8B-Inst-DPO-ITA",
779
- "created_by": "swap-uniba",
780
- "source": "official",
781
- "last_updated": "2026-03-21",
782
- "capabilities": { "streaming": true }
783
- },
784
- {
785
- "id": "swiss-ai/Apertus-8B-Instruct-2509",
786
- "name": "Apertus-8B-Instruct-2509",
787
- "created_by": "swiss-ai",
788
- "source": "official",
789
- "last_updated": "2026-03-21",
790
- "alias": "swiss-ai/Apertus-8B-Instruct",
791
- "capabilities": { "streaming": true }
792
- },
793
- {
794
- "id": "swiss-ai/Apertus-8B-Instruct",
795
- "name": "Apertus-8B-Instruct-2509",
796
- "created_by": "swiss-ai",
797
- "source": "official",
798
- "last_updated": "2026-03-21",
799
- "capabilities": { "streaming": true },
800
- "snapshots": ["swiss-ai/Apertus-8B-Instruct-2509"]
801
- },
802
- {
803
- "id": "tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1",
804
- "name": "Llama-3-Swallow-8B-Instruct-v0.1",
805
- "created_by": "tokyotech-llm",
806
- "source": "official",
807
- "last_updated": "2026-03-21",
808
- "capabilities": { "streaming": true }
809
- },
810
- {
811
- "id": "tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5",
812
- "name": "Llama-3.1-Swallow-8B-Instruct-v0.5",
813
- "created_by": "tokyotech-llm",
814
- "source": "official",
815
- "last_updated": "2026-03-21",
816
- "capabilities": { "streaming": true }
817
- },
818
- {
819
- "id": "unsloth/gemma-2-9b-it",
820
- "name": "gemma-2-9b-it",
821
- "created_by": "unsloth",
822
- "source": "official",
823
- "last_updated": "2026-03-21",
824
- "family": "gemma-2",
825
- "capabilities": { "streaming": true }
826
- },
827
- {
828
- "id": "unsloth/llama-3-8b-Instruct",
829
- "name": "llama-3-8b-Instruct",
830
- "created_by": "unsloth",
831
- "source": "official",
832
- "last_updated": "2026-03-21",
833
- "family": "llama-3",
834
- "capabilities": { "streaming": true }
835
- },
836
- {
837
- "id": "unsloth/Llama-3.1-8B-Instruct",
838
- "name": "Llama-3.1-8B-Instruct",
839
- "created_by": "unsloth",
840
- "source": "official",
841
- "last_updated": "2026-03-21",
842
- "capabilities": { "streaming": true }
843
- },
844
- {
845
- "id": "unsloth/Meta-Llama-3.1-8B-Instruct",
846
- "name": "Meta-Llama-3.1-8B-Instruct",
847
- "created_by": "unsloth",
848
- "source": "official",
849
- "last_updated": "2026-03-21",
850
- "capabilities": { "streaming": true }
851
- },
852
- {
853
- "id": "unsloth/Mistral-Small-24B-Instruct-2501",
854
- "name": "Mistral-Small-24B-Instruct-2501",
855
- "created_by": "unsloth",
856
- "source": "official",
857
- "last_updated": "2026-03-21",
858
- "alias": "unsloth/Mistral-Small-24B-Instruct",
859
- "capabilities": { "streaming": true }
860
- },
861
- {
862
- "id": "unsloth/Mistral-Small-24B-Instruct",
863
- "name": "Mistral-Small-24B-Instruct-2501",
864
- "created_by": "unsloth",
865
- "source": "official",
866
- "last_updated": "2026-03-21",
867
- "capabilities": { "streaming": true },
868
- "snapshots": ["unsloth/Mistral-Small-24B-Instruct-2501"]
869
- },
870
- {
871
- "id": "unsloth/Qwen2.5-14B-Instruct",
872
- "name": "Qwen2.5-14B-Instruct",
873
- "created_by": "unsloth",
874
- "source": "official",
875
- "last_updated": "2026-03-21",
876
- "capabilities": { "streaming": true }
877
- },
878
- {
879
- "id": "unsloth/Qwen2.5-32B-Instruct",
880
- "name": "Qwen2.5-32B-Instruct",
881
- "created_by": "unsloth",
882
- "source": "official",
883
- "last_updated": "2026-03-21",
884
- "capabilities": { "streaming": true }
885
- },
886
- {
887
- "id": "unsloth/Qwen2.5-7B-Instruct",
888
- "name": "Qwen2.5-7B-Instruct",
889
- "created_by": "unsloth",
890
- "source": "official",
891
- "last_updated": "2026-03-21",
892
- "capabilities": { "streaming": true }
893
- },
894
- {
895
- "id": "unsloth/Qwen2.5-Coder-32B-Instruct",
896
- "name": "Qwen2.5-Coder-32B-Instruct",
897
- "created_by": "unsloth",
898
- "source": "official",
899
- "last_updated": "2026-03-21",
900
- "capabilities": { "streaming": true }
901
- },
902
- {
903
- "id": "unsloth/Qwen3-8B",
904
- "name": "Qwen3-8B",
905
- "created_by": "unsloth",
906
- "source": "official",
907
- "last_updated": "2026-03-21",
908
- "capabilities": { "streaming": true }
909
- },
910
- {
911
- "id": "unsloth/tinyllama-chat",
912
- "name": "tinyllama-chat",
913
- "created_by": "unsloth",
914
- "source": "official",
915
- "last_updated": "2026-03-21",
916
- "capabilities": { "streaming": true }
917
- },
918
- {
919
- "id": "VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct",
920
- "name": "Llama-3-SauerkrautLM-8b-Instruct",
921
- "created_by": "vagosolutions",
922
- "source": "official",
923
- "last_updated": "2026-03-21",
924
- "capabilities": { "streaming": true }
925
- },
926
- {
927
- "id": "zai-org/GLM-4-9B-0414",
928
- "name": "GLM-4-9B-0414",
929
- "created_by": "zhipu",
930
- "source": "official",
931
- "last_updated": "2026-03-21",
932
- "capabilities": { "streaming": true }
933
- },
934
- {
935
- "id": "zai-org/GLM-4.5-Air-FP8",
936
- "name": "GLM-4.5-Air-FP8",
937
- "created_by": "zhipu",
938
- "source": "official",
939
- "last_updated": "2026-03-21",
940
- "capabilities": { "streaming": true }
941
- },
942
- {
943
- "id": "zai-org/GLM-4.5-Air",
944
- "name": "GLM-4.5-Air",
945
- "created_by": "zhipu",
946
- "source": "official",
947
- "last_updated": "2026-03-21",
948
- "capabilities": { "streaming": true }
949
- },
950
- {
951
- "id": "zai-org/GLM-4.5",
952
- "name": "GLM-4.5",
953
- "created_by": "zhipu",
954
- "source": "official",
955
- "last_updated": "2026-03-21",
956
- "capabilities": { "streaming": true }
957
- },
958
- {
959
- "id": "zai-org/GLM-4.6-FP8",
960
- "name": "GLM-4.6-FP8",
961
- "created_by": "zhipu",
962
- "source": "official",
963
- "last_updated": "2026-03-21",
964
- "capabilities": { "streaming": true }
965
- },
966
- {
967
- "id": "zai-org/GLM-4.6",
968
- "name": "GLM-4.6",
969
- "created_by": "zhipu",
970
- "source": "official",
971
- "last_updated": "2026-03-21",
972
- "capabilities": { "streaming": true }
973
- },
974
- {
975
- "id": "zai-org/GLM-4.7-Flash",
976
- "name": "GLM-4.7-Flash",
977
- "created_by": "zhipu",
978
- "source": "official",
979
- "last_updated": "2026-03-21",
980
- "capabilities": { "streaming": true }
981
- },
982
- {
983
- "id": "zai-org/GLM-4.7-FP8",
984
- "name": "GLM-4.7-FP8",
985
- "created_by": "zhipu",
986
- "source": "official",
987
- "last_updated": "2026-03-21",
988
- "capabilities": { "streaming": true }
989
- },
990
- {
991
- "id": "zai-org/GLM-4.7",
992
- "name": "GLM-4.7",
993
- "created_by": "zhipu",
994
- "source": "official",
995
- "last_updated": "2026-03-21",
996
- "capabilities": { "streaming": true }
997
- },
998
- {
999
- "id": "zai-org/GLM-5",
1000
- "name": "GLM-5",
1001
- "created_by": "zhipu",
1002
- "source": "official",
1003
- "last_updated": "2026-03-21",
1004
- "capabilities": { "streaming": true }
1005
- },
1006
- {
1007
- "id": "zai-org/GLM-Z1-32B-0414",
1008
- "name": "GLM-Z1-32B-0414",
1009
- "created_by": "zhipu",
1010
- "source": "official",
1011
- "last_updated": "2026-03-21",
1012
- "capabilities": { "streaming": true }
1013
- }
1014
- ]
1015
- };
1016
- const models = provider.models.map((m) => ({
1017
- ...m,
1018
- provider: provider.id
1019
- }));
1020
- //#endregion
1021
- exports.models = models;
1022
- exports.provider = provider;
1
+ Object.defineProperty(exports,Symbol.toStringTag,{value:`Module`});const e={id:`huggingface`,name:`Hugging Face`,region:`US`,headquarters:`New York, NY`,founded:2016,url:`https://huggingface.co`,api_url:`https://router.huggingface.co/v1`,docs_url:`https://huggingface.co/docs/api-inference`,pricing_url:`https://huggingface.co/pricing`,description:`The AI community platform hosting open-source models with serverless inference endpoints.`,type:`aggregator`,sdk:{python:`huggingface_hub`,javascript:`@huggingface/inference`},openai_compatible:!0,free_tier:!0,github_url:`https://github.com/huggingface`,models_url:`https://huggingface.co/models`,twitter_url:`https://x.com/huggingface`,discord_url:`https://discord.com/invite/hugging-face-879548962464493619`,blog_url:`https://huggingface.co/blog`,terms_url:`https://huggingface.co/terms-of-service`,support_url:`https://huggingface.co/support`,icon:`<svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24" fill="currentColor" fill-rule="evenodd">
2
+ <path d="M16.781 3.277c2.997 1.704 4.844 4.851 4.844 8.258 0 .995-.155 1.955-.443 2.857a1.332 1.332 0 011.125.4 1.41 1.41 0 01.2 1.723c.204.165.352.385.428.632l.017.062c.06.222.12.69-.2 1.166.244.37.279.836.093 1.236-.255.57-.893 1.018-2.128 1.5l-.202.078-.131.048c-.478.173-.89.295-1.061.345l-.086.024c-.89.243-1.808.375-2.732.394-1.32 0-2.3-.36-2.923-1.067a9.852 9.852 0 01-3.18.018C9.778 21.647 8.802 22 7.494 22a11.249 11.249 0 01-2.541-.343l-.221-.06-.273-.08a16.574 16.574 0 01-1.175-.405c-1.237-.483-1.875-.93-2.13-1.501-.186-.4-.151-.867.093-1.236a1.42 1.42 0 01-.2-1.166c.069-.273.226-.516.447-.694a1.41 1.41 0 01.2-1.722c.233-.248.557-.391.917-.407l.078-.001a9.385 9.385 0 01-.44-2.85c0-3.407 1.847-6.554 4.844-8.258a9.822 9.822 0 019.687 0zM4.188 14.758c.125.687 2.357 2.35 2.14 2.707-.19.315-.796-.239-.948-.386l-.041-.04-.168-.147c-.561-.479-2.304-1.9-2.74-1.432-.43.46.119.859 1.055 1.42l.784.467.136.083c1.045.643 1.12.84.95 1.113-.188.295-3.07-2.1-3.34-1.083-.27 1.011 2.942 1.304 2.744 2.006-.2.7-2.265-1.324-2.685-.537-.425.79 2.913 1.718 2.94 1.725l.16.04.175.042c1.227.284 3.565.65 4.435-.604.673-.973.64-1.709-.248-2.61l-.057-.057c-.945-.928-1.495-2.288-1.495-2.288l-.017-.058-.025-.072c-.082-.22-.284-.639-.63-.584-.46.073-.798 1.21.12 1.933l.05.038c.977.721-.195 1.21-.573.534l-.058-.104-.143-.25c-.463-.799-1.282-2.111-1.739-2.397-.532-.332-.907-.148-.782.541zm14.842-.541c-.533.335-1.563 2.074-1.94 2.751a.613.613 0 01-.687.302.436.436 0 01-.176-.098.303.303 0 01-.049-.06l-.014-.028-.008-.02-.007-.019-.003-.013-.003-.017a.289.289 0 01-.004-.048c0-.12.071-.266.25-.427.026-.024.054-.047.084-.07l.047-.036c.022-.016.043-.032.063-.049.883-.71.573-1.81.131-1.917l-.031-.006-.056-.004a.368.368 0 00-.062.006l-.028.005-.042.014-.039.017-.028.015-.028.019-.036.027-.023.02c-.173.158-.273.428-.31.542l-.016.054s-.53 1.309-1.439 2.234l-.054.054c-.365.358-.596.69-.702 1.018-.143.437-.066.868.21 1.353.055.097.117.195.187.296.882 1.275 3.282.876 4.494.59l.286-.07.25-.074c.276-.084.736-.233 1.2-.42l.188-.077.065-.028.064-.028.124-.056.081-.038c.529-.252.964-.543.994-.827l.001-.036a.299.299 0 00-.037-.139c-.094-.176-.271-.212-.491-.168l-.045.01c-.044.01-.09.024-.136.04l-.097.035-.054.022c-.559.23-1.238.705-1.607.745h.006a.452.452 0 01-.05.003h-.024l-.024-.003-.023-.005c-.068-.016-.116-.06-.14-.142a.22.22 0 01-.005-.1c.062-.345.958-.595 1.713-.91l.066-.028c.528-.224.97-.483.985-.832v-.04a.47.47 0 00-.016-.098c-.048-.18-.175-.251-.36-.251-.785 0-2.55 1.36-2.92 1.36-.025 0-.048-.007-.058-.024a.6.6 0 01-.046-.088c-.1-.238.068-.462 1.06-1.066l.209-.126c.538-.32 1.01-.588 1.341-.831.29-.212.475-.406.503-.6l.003-.028c.008-.113-.038-.227-.147-.344a.266.266 0 00-.07-.054l-.034-.015-.013-.005a.403.403 0 00-.13-.02c-.162 0-.369.07-.595.18-.637.313-1.431.952-1.826 1.285l-.249.215-.033.033c-.08.078-.288.27-.493.386l-.071.037-.041.019a.535.535 0 01-.122.036h.005a.346.346 0 01-.031.003l.01-.001-.013.001c-.079.005-.145-.021-.19-.095a.113.113 0 01-.014-.065c.027-.465 2.034-1.991 2.152-2.642l.009-.048c.1-.65-.271-.817-.791-.493zM11.938 2.984c-4.798 0-8.688 3.829-8.688 8.55 0 .692.083 1.364.24 2.008l.008-.009c.252-.298.612-.46 1.017-.46.355.008.699.117.993.312.22.14.465.384.715.694.261-.372.69-.598 1.15-.605.852 0 1.367.728 1.562 1.383l.047.105.06.127c.192.396.595 1.139 1.143 1.68 1.06 1.04 1.324 2.115.8 3.266a8.865 8.865 0 002.024-.014c-.505-1.12-.26-2.17.74-3.186l.066-.066c.695-.684 1.157-1.69 1.252-1.912.195-.655.708-1.383 1.56-1.383.46.007.889.233 1.15.605.25-.31.495-.553.718-.694a1.87 1.87 0 01.99-.312c.357 0 .682.126.925.36.14-.61.215-1.245.215-1.898 0-4.722-3.89-8.55-8.687-8.55zm1.857 8.926l.439-.212c.553-.264.89-.383.89.152 0 1.093-.771 3.208-3.155 3.262h-.184c-2.325-.052-3.116-2.06-3.156-3.175l-.001-.087c0-1.107 1.452.586 3.25.586.716 0 1.379-.272 1.917-.526zm4.017-3.143c.45 0 .813.358.813.8 0 .441-.364.8-.813.8a.806.806 0 01-.812-.8c0-.442.364-.8.812-.8zm-11.624 0c.448 0 .812.358.812.8 0 .441-.364.8-.812.8a.806.806 0 01-.813-.8c0-.442.364-.8.813-.8zm7.79-.841c.32-.384.846-.54 1.33-.394.483.146.83.564.878 1.06.048.495-.212.97-.659 1.203-.322.168-.447-.477-.767-.585l.002-.003c-.287-.098-.772.362-.925.079a1.215 1.215 0 01.14-1.36zm-4.323 0c.322.384.377.92.14 1.36-.152.283-.64-.177-.925-.079l.003.003c-.108.036-.194.134-.273.24l-.118.165c-.11.15-.22.262-.377.18a1.226 1.226 0 01-.658-1.204c.048-.495.395-.913.878-1.059a1.262 1.262 0 011.33.394z"/>
3
+ </svg>`,models:[{id:`aaditya/Llama3-OpenBioLLM-8B`,name:`Llama3-OpenBioLLM-8B`,created_by:`aaditya`,source:`official`,last_updated:`2026-03-23`,page_url:`https://huggingface.co/aaditya/Llama3-OpenBioLLM-8B`,model_type:`chat`,capabilities:{streaming:!0},description:`Introducing OpenBioLLM-8B: A State-of-the-Art Open Source Biomedical Large Language Model`,release_date:`2024-04-20`,context_window:8192,license:`llama3`,tagline:`Introducing OpenBioLLM-8B: A State-of-the-Art Open Source Biomedical Large Language Model`,architecture:`transformer`,open_weight:!0},{id:`allenai/Olmo-3-7B-Instruct`,name:`Olmo-3-7B-Instruct`,created_by:`allenai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:0,page_url:`https://huggingface.co/allenai/Olmo-3-7B-Instruct`,release_date:`2025-11-19`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`allenai/Olmo-3.1-32B-Instruct`,name:`Olmo-3.1-32B-Instruct`,created_by:`allenai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:32,page_url:`https://huggingface.co/allenai/Olmo-3.1-32B-Instruct`,release_date:`2025-12-10`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,name:`Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,created_by:`argilla`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1,page_url:`https://huggingface.co/argilla/Llama-3.2-1B-Instruct-APIGen-FC-v0.1`,description:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on [argilla-warehouse/apigen-synth-trl](https://huggingface.co/datasets/argilla-warehouse/apigen-synth-trl) dataset, a version of [argilla/Synth-APIGen-v0.1](https://huggingface.co/datasets/argilla-warehouse/Synth-APIGen-v0.1) ready to do SFT on top of it. It has been trained using [TRL](https://github.com/huggingface/trl).`,release_date:`2024-10-07`,context_window:131072,license:`apache-2.0`,tagline:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on [argilla-warehouse/apigen-synth-trl](https://huggingface.co/datasets/argilla-warehouse/apigen-synth-trl) dataset, a version of [argilla/Synth-APIGen-v0.1](https://huggingface.co/datasets/argilla-warehouse/Synth-APIGen-v0.1) ready to do SFT on top of it.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC`,name:`Llama-3.2-1B-Instruct-v2-FC`,created_by:`argilla-warehouse`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1,page_url:`https://huggingface.co/argilla-warehouse/Llama-3.2-1B-Instruct-v2-FC`,description:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the [argilla-warehouse/apigen-smollm-trl-FC](https://huggingface.co/datasets/argilla-warehouse/apigen-smollm-trl-FC) dataset. It has been trained using [TRL](https://github.com/huggingface/trl).`,release_date:`2024-10-20`,context_window:131072,tagline:`This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the [argilla-warehouse/apigen-smollm-trl-FC](https://huggingface.co/datasets/argilla-warehouse/apigen-smollm-trl-FC) dataset.`,architecture:`transformer`,tools:[`function_calling`]},{id:`DeepMount00/Llama-3-8b-Ita`,name:`Llama-3-8b-Ita`,created_by:`deepmount00`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/DeepMount00/Llama-3-8b-Ita`,description:`**💡 Found this resource helpful?** Creating and maintaining open source AI models and datasets requires significant computational resources. If this work has been valuable to you, consider [supporting my research](https://buymeacoffee.com/michele.montebovi) to help me continue building tools that benefit the entire AI community. Every contribution directly funds more open source innovation! ☕`,release_date:`2024-05-01`,context_window:8192,license:`llama3`,tagline:`**💡 Found this resource helpful?** Creating and maintaining open source AI models and datasets requires significant computational resources.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`deepseek-ai/DeepSeek-R1-0528-Qwen3-8B`,name:`DeepSeek-R1-0528-Qwen3-8B`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`reasoning`,parameters:8.2,page_url:`https://huggingface.co/deepseek-ai/DeepSeek-R1-0528-Qwen3-8B`,release_date:`2025-05-29`,context_window:131072,license:`mit`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Efficient-Large-Model/gemma-2-2b-it`,name:`gemma-2-2b-it`,created_by:`efficient-large-model`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2.6,page_url:`https://huggingface.co/Efficient-Large-Model/gemma-2-2b-it`,description:`**Model Page**: [Gemma](https://ai.google.dev/gemma/docs/base)`,release_date:`2024-12-12`,context_window:8192,license:`gemma`,tagline:`**Model Page**: [Gemma](https://ai.google.dev/gemma/docs/base)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`EssentialAI/rnj-1-instruct`,name:`rnj-1-instruct`,created_by:`essentialai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},page_url:`https://huggingface.co/EssentialAI/rnj-1-instruct`,description:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models. These models perform well across a range of programming languages and boast strong agentic capabilities (e.g., inside agentic frameworks like mini-SWE-agent), while also excelling at tool-calling. They additionally exhibit strong capabilities in math and science.`,release_date:`2025-12-04`,context_window:32768,license:`apache-2.0`,parameters:8.3,tagline:`style="vertical-align: middle;" src="https://img.shields.io/badge/%F0%9F%8C%90%20Website-essential.ai-4b9fe1?color=4b9fe1&logoColor=white"/>`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`failspy/Meta-Llama-3-8B-Instruct-abliterated-v3`,name:`Meta-Llama-3-8B-Instruct-abliterated-v3`,created_by:`failspy`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3`,description:`[My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon](https://huggingface.co/failspy/llama-3-70B-Instruct-abliterated/blob/main/ortho_cookbook.ipynb)`,release_date:`2024-05-20`,context_window:8192,license:`llama3`,tagline:`[My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon](https://huggingface.co/failspy/llama-3-70B-Instruct-abliterated/blob/main/ortho_cookbook.ipynb)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Featherless-Chat-Models/Mistral-7B-Instruct-v0.2`,name:`Mistral-7B-Instruct-v0.2`,created_by:`featherless-chat-models`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.2,page_url:`https://huggingface.co/Featherless-Chat-Models/Mistral-7B-Instruct-v0.2`,release_date:`2025-05-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`FlagAlpha/Llama3-Chinese-8B-Instruct`,name:`Llama3-Chinese-8B-Instruct`,created_by:`flagalpha`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/FlagAlpha/Llama3-Chinese-8B-Instruct`,description:`Llama3-Chinese-8B-Instruct基于Llama3-8B中文微调对话模型,由Llama中文社区和AtomEcho(原子回声)联合研发,我们会持续提供更新的模型参数,模型训练过程见 [https://llama.family](https://llama.family)。`,release_date:`2024-04-23`,context_window:8192,license:`apache-2.0`,tagline:`Llama3-Chinese-8B-Instruct基于Llama3-8B中文微调对话模型,由Llama中文社区和AtomEcho(原子回声)联合研发,我们会持续提供更新的模型参数,模型训练过程见 [https://llama.family](https://llama.family)。`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`google/gemma-2-2b-jpn-it`,name:`gemma-2-2b-jpn-it`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2.6,page_url:`https://huggingface.co/google/gemma-2-2b-jpn-it`,release_date:`2024-09-25`,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`google/gemma-2-9b-it`,name:`gemma-2-9b-it`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:9.2,page_url:`https://huggingface.co/google/gemma-2-9b-it`,release_date:`2024-06-24`,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`GraySwanAI/Llama-3-8B-Instruct-RR`,name:`Llama-3-8B-Instruct-RR`,created_by:`grayswanai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/GraySwanAI/Llama-3-8B-Instruct-RR`,description:`Llama-3-8B-Instruct-RR is a Llama-3 model with circuit breakers inserted using Representation Rerouting (RR).`,release_date:`2024-07-08`,context_window:8192,tagline:`Llama-3-8B-Instruct-RR is a Llama-3 model with circuit breakers inserted using Representation Rerouting (RR).`,architecture:`transformer`,tools:[`function_calling`]},{id:`GritLM/GritLM-7B`,name:`GritLM-7B`,created_by:`gritlm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:7.2,page_url:`https://huggingface.co/GritLM/GritLM-7B`,description:`> GritLM is a generative representational instruction tuned language model. It unifies text representation (embedding) and text generation into a single model achieving state-of-the-art performance on both types of tasks.`,release_date:`2024-02-11`,context_window:32768,license:`apache-2.0`,tagline:`> GritLM is a generative representational instruction tuned language model.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`huihui-ai/Qwen2.5-72B-Instruct-abliterated`,name:`Qwen2.5-72B-Instruct-abliterated`,created_by:`huihui-ai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/huihui-ai/Qwen2.5-72B-Instruct-abliterated`,description:`This is an uncensored version of [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it). This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.`,release_date:`2024-10-26`,context_window:32768,license:`other`,tagline:`This is an uncensored version of [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`instruction-pretrain/finance-Llama3-8B`,name:`finance-Llama3-8B`,created_by:`instruction-pretrain`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},parameters:8,page_url:`https://huggingface.co/instruction-pretrain/finance-Llama3-8B`,description:`This repo contains the **finance model developed from Llama3-8B** in our paper [Instruction Pre-Training: Language Models are Supervised Multitask Learners](https://huggingface.co/papers/2406.14491).`,release_date:`2024-06-18`,context_window:8192,license:`llama3`,tagline:`This repo contains the **finance model developed from Llama3-8B** in our paper [Instruction Pre-Training: Language Models are Supervised Multitask Learners](https://huggingface.co/papers/2406.14491).`,architecture:`transformer`,open_weight:!0},{id:`Intel/neural-chat-7b-v3-3`,name:`neural-chat-7b-v3-3`,created_by:`intel`,source:`official`,last_updated:`2026-03-22`,capabilities:{streaming:!0,tool_call:!0},parameters:7.2,page_url:`https://huggingface.co/Intel/neural-chat-7b-v3-3`,release_date:`2023-12-09`,context_window:32768,license:`apache-2.0`},{id:`invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,name:`Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,created_by:`invalid-coder`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:11,page_url:`https://huggingface.co/invalid-coder/Sakura-SOLAR-Instruct-CarbonVillain-en-10.7B-v2-slerp`,description:`NeuralPipe-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [jeonsworld/CarbonVillain-en-10.7B-v2](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v2) * [kyujinpy/Sakura-SOLAR-Instruct](https://huggingface.co/kyujinpy/Sakura-SOLAR-Instruct)`,release_date:`2024-01-10`,context_window:4096,license:`apache-2.0`,tagline:`NeuralPipe-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [jeonsworld/CarbonVillain-en-10.7B`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix`,name:`Mistral-Nemo-Instruct-2407-lenient-chatfix`,created_by:`m8than`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,page_url:`https://huggingface.co/m8than/Mistral-Nemo-Instruct-2407-lenient-chatfix`,description:`It's just mistral nemo 2407 with a less strict chat format.`,release_date:`2025-05-06`,context_window:131072,license:`apache-2.0`,parameters:12,tagline:`It's just mistral nemo 2407 with a less strict chat format.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`meta-llama/Llama-2-13b-chat-hf`,name:`Llama-2-13b-chat-hf`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-2`,model_type:`chat`,parameters:13,page_url:`https://huggingface.co/meta-llama/Llama-2-13b-chat-hf`,release_date:`2023-07-13`,license:`llama2`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:4096,max_output_tokens:4096,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.1-70B-Instruct`,name:`Llama-3.1-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct`,release_date:`2024-07-16`,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.1-8B-Instruct`,name:`Llama-3.1-8B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct`,release_date:`2024-07-18`,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.2-1B-Instruct`,name:`Llama-3.2-1B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.2`,model_type:`chat`,parameters:1.2,page_url:`https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct`,release_date:`2024-09-18`,license:`llama3.2`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Llama-3.3-70B-Instruct`,name:`Llama-3.3-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct`,release_date:`2024-11-26`,license:`llama3.3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:131072,max_output_tokens:131072,knowledge_cutoff:`2023-12`,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Meta-Llama-3-70B-Instruct`,name:`Meta-Llama-3-70B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct`,release_date:`2024-04-17`,license:`llama3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:8192,max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`meta-llama/Meta-Llama-3-8B-Instruct`,name:`Meta-Llama-3-8B-Instruct`,created_by:`meta`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct`,release_date:`2024-04-17`,license:`llama3`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],context_window:8192,max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`mistralai/Mistral-7B-Instruct-v0.2`,name:`Mistral-7B-Instruct-v0.2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.2,page_url:`https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2`,release_date:`2023-12-11`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated`,name:`Meta-Llama-3.1-8B-Instruct-abliterated`,created_by:`mlabonne`,source:`official`,last_updated:`2026-03-23`,family:`Llama-3.1`,page_url:`https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated`,model_type:`chat`,capabilities:{streaming:!0,tool_call:!0},description:`This is an uncensored version of Llama 3.1 8B Instruct created with abliteration (see [this article](https://huggingface.co/blog/mlabonne/abliteration) to know more about it).`,release_date:`2024-07-24`,context_window:131072,license:`llama3.1`,parameters:8,tagline:`This is an uncensored version of Llama 3.1 8B Instruct created with abliteration (see [this article](https://huggingface.co/blog/mlabonne/abliteration) to know more about it).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`mlabonne/NeuralDaredevil-8B-abliterated`,name:`NeuralDaredevil-8B-abliterated`,created_by:`mlabonne`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated`,description:`This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated), trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k). The DPO fine-tuning successfully recovers the performance loss due to the abliteration process, making it an excellent uncensored model.`,release_date:`2024-05-27`,context_window:8192,license:`llama3`,tagline:`This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated), trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Instruct-0905`,name:`Kimi-K2-Instruct-0905`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Instruct-0905`,release_date:`2025-09-03`,context_window:262144,license:`other`,parameters:1026,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Instruct`,name:`Kimi-K2-Instruct`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Instruct`,release_date:`2025-07-11`,context_window:131072,license:`other`,parameters:1026,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`moonshotai/Kimi-K2-Thinking`,name:`Kimi-K2-Thinking`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},family:`Kimi-K2`,model_type:`chat`,page_url:`https://huggingface.co/moonshotai/Kimi-K2-Thinking`,release_date:`2025-11-04`,context_window:262144,license:`other`,parameters:1058,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],description:`Long-term thinking, multi-step tool usage, complex problem solving.`,tagline:`Long-term thinking, multi-step tool usage, complex problem solving.`},{id:`NousResearch/Meta-Llama-3-70B-Instruct`,name:`Meta-Llama-3-70B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:71,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3-70B-Instruct`,release_date:`2024-04-19`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`NousResearch/Meta-Llama-3-8B-Instruct`,name:`Meta-Llama-3-8B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct`,release_date:`2024-04-18`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`NousResearch/Meta-Llama-3.1-8B-Instruct`,name:`Meta-Llama-3.1-8B-Instruct`,created_by:`nousresearch`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct`,release_date:`2024-07-24`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`nvidia/Llama-3.1-8B-Instruct-FP8`,name:`Llama-3.1-8B-Instruct-FP8`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/nvidia/Llama-3.1-8B-Instruct-FP8`,release_date:`2024-08-29`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`openai/gpt-oss-120b`,name:`gpt-oss-120b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0},model_type:`chat`,parameters:120,page_url:`https://huggingface.co/openai/gpt-oss-120b`,description:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-04`,context_window:131072,license:`apache-2.0`,tagline:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:131072,knowledge_cutoff:`2024-05`,modalities:{input:[`text`],output:[`text`]}},{id:`openai/gpt-oss-20b`,name:`gpt-oss-20b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0},model_type:`chat`,parameters:22,page_url:`https://huggingface.co/openai/gpt-oss-20b`,description:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-04`,context_window:131072,license:`apache-2.0`,tagline:`Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:131072,knowledge_cutoff:`2024-05`,modalities:{input:[`text`],output:[`text`]}},{id:`openai/gpt-oss-safeguard-20b`,name:`gpt-oss-safeguard-20b`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0,tool_call:!0},model_type:`moderation`,parameters:20,page_url:`https://huggingface.co/openai/gpt-oss-safeguard-20b`,description:"`gpt-oss-safeguard-120b` and `gpt-oss-safeguard-20b` are safety reasoning models built-upon gpt-oss. With these models, you can classify text content based on safety policies that you provide and perform a suite of foundational safety tasks. These models are intended for safety use cases. For other applications, we recommend using [gpt-oss models](https://huggingface.co/collections/openai/gpt-oss).",release_date:`2025-09-18`,context_window:131072,license:`apache-2.0`,tagline:"`gpt-oss-safeguard-120b` and `gpt-oss-safeguard-20b` are safety reasoning models built-upon gpt-oss.",architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`openchat/openchat-3.6-8b-20240522`,name:`openchat-3.6-8b-20240522`,created_by:`openchat`,source:`official`,last_updated:`2026-03-23`,alias:`openchat/openchat-3.6-8b`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/openchat/openchat-3.6-8b-20240522`,description:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM.`,release_date:`2024-05-07`,context_window:8192,license:`llama3`,tagline:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`openchat/openchat-3.6-8b`,name:`openchat-3.6-8b-20240522`,created_by:`openchat`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`openchat/openchat-3.6-8b-20240522`],parameters:8,page_url:`https://huggingface.co/openchat/openchat-3.6-8b-20240522`,description:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM.`,release_date:`2024-05-07`,context_window:8192,license:`llama3`,tagline:`To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Orion-zhen/Qwen2.5-7B-Instruct-Uncensored`,name:`Qwen2.5-7B-Instruct-Uncensored`,created_by:`orion-zhen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored`,description:`This model is an uncensored fine-tune version of Qwen2.5-7B-Instruct. However, I can still notice that though uncensored, the model fails to generate detailed descriptions on certain extreme scenarios, which might be associated with deletion on some pretrain datasets in Qwen's pretraining stage.`,release_date:`2024-09-26`,context_window:32768,license:`gpl-3.0`,tagline:`This model is an uncensored fine-tune version of Qwen2.5-7B-Instruct.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,name:`Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,created_by:`patronusai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct-v1.1`,description:`Lynx is an open-source hallucination evaluation model. Patronus-Lynx-8B-Instruct-v1.1 was trained on a mix of datasets including CovidQA, PubmedQA, DROP, RAGTruth. The datasets contain a mix of hand-annotated and synthetic data. The maximum sequence length is 128000 tokens.`,release_date:`2024-07-24`,context_window:131072,license:`cc-by-nc-4.0`,tagline:`Lynx is an open-source hallucination evaluation model.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen1.5-1.8B-Chat`,name:`Qwen1.5-1.8B-Chat`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.8,page_url:`https://huggingface.co/Qwen/Qwen1.5-1.8B-Chat`,release_date:`2024-01-30`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2-1.5B-Instruct`,name:`Qwen2-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2-1.5B-Instruct`,release_date:`2024-06-03`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2-72B-Instruct`,name:`Qwen2-72B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/Qwen/Qwen2-72B-Instruct`,release_date:`2024-05-28`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],max_output_tokens:6144,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2-7B-Instruct`,name:`Qwen2-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2-7B-Instruct`,release_date:`2024-06-04`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-1.5B-Instruct`,name:`Qwen2.5-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-14B-Instruct-1M`,name:`Qwen2.5-14B-Instruct-1M`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-14B-Instruct-1M`,release_date:`2025-01-23`,context_window:101e4,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],max_output_tokens:8192,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-14B-Instruct`,name:`Qwen2.5-14B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:14,page_url:`https://huggingface.co/Qwen/Qwen2.5-14B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-32B-Instruct`,name:`Qwen2.5-32B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:32,page_url:`https://huggingface.co/Qwen/Qwen2.5-32B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-72B-Instruct`,name:`Qwen2.5-72B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:73,page_url:`https://huggingface.co/Qwen/Qwen2.5-72B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-7B-Instruct-1M`,name:`Qwen2.5-7B-Instruct-1M`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-1M`,release_date:`2025-01-23`,context_window:101e4,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-7B-Instruct`,name:`Qwen2.5-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-7B-Instruct`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen2.5-Coder-1.5B-Instruct`,name:`Qwen2.5-Coder-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B-Instruct`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-1.5B`,name:`Qwen2.5-Coder-1.5B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-14B-Instruct`,name:`Qwen2.5-Coder-14B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-14B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-14B`,name:`Qwen2.5-Coder-14B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-14B`,release_date:`2024-11-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-32B-Instruct`,name:`Qwen2.5-Coder-32B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-32B`,name:`Qwen2.5-Coder-32B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-32B`,release_date:`2024-11-08`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-3B-Instruct`,name:`Qwen2.5-Coder-3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:3.1,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-3B-Instruct`,release_date:`2024-11-06`,context_window:32768,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-7B-Instruct`,name:`Qwen2.5-Coder-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct`,release_date:`2024-09-17`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Coder-7B`,name:`Qwen2.5-Coder-7B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Coder-7B`,release_date:`2024-09-16`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Math-1.5B-Instruct`,name:`Qwen2.5-Math-1.5B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:1.5,page_url:`https://huggingface.co/Qwen/Qwen2.5-Math-1.5B-Instruct`,description:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR. We do not recommend using this series of models for other tasks. > </b> > </div>`,release_date:`2024-09-16`,context_window:4096,license:`apache-2.0`,tagline:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen2.5-Math-7B-Instruct`,name:`Qwen2.5-Math-7B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/Qwen/Qwen2.5-Math-7B-Instruct`,description:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR. We do not recommend using this series of models for other tasks. > </b> > </div>`,release_date:`2024-09-19`,context_window:4096,license:`apache-2.0`,tagline:`> [!Warning] > <div align="center"> > <b> > 🚨 Qwen2.5-Math mainly supports solving English and Chinese math problems through CoT and TIR.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen3-1.7B`,name:`Qwen3-1.7B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:2,page_url:`https://huggingface.co/Qwen/Qwen3-1.7B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-14B-Base`,name:`Qwen3-14B-Base`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/Qwen/Qwen3-14B-Base`,release_date:`2025-04-28`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-14B`,name:`Qwen3-14B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:14,page_url:`https://huggingface.co/Qwen/Qwen3-14B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Instruct-2507`,name:`Qwen3-235B-A22B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-235B-A22B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507`,release_date:`2025-07-21`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Instruct`,name:`Qwen3-235B-A22B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-235B-A22B-Instruct-2507`],model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507`,release_date:`2025-07-21`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Thinking-2507`,name:`Qwen3-235B-A22B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-235B-A22B-Thinking`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507`,release_date:`2025-07-25`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B-Thinking`,name:`Qwen3-235B-A22B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-235B-A22B-Thinking-2507`],model_type:`chat`,parameters:235,active_parameters:22,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B-Thinking-2507`,release_date:`2025-07-25`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-235B-A22B`,name:`Qwen3-235B-A22B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:235,active_parameters:15,page_url:`https://huggingface.co/Qwen/Qwen3-235B-A22B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B-Instruct-2507`,name:`Qwen3-30B-A3B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-30B-A3B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507`,release_date:`2025-07-28`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B-Instruct`,name:`Qwen3-30B-A3B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-30B-A3B-Instruct-2507`],model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507`,release_date:`2025-07-28`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-30B-A3B`,name:`Qwen3-30B-A3B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:30,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-30B-A3B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-32B`,name:`Qwen3-32B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:32,page_url:`https://huggingface.co/Qwen/Qwen3-32B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Instruct-2507`,name:`Qwen3-4B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-4B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Instruct`,name:`Qwen3-4B-Instruct-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-4B-Instruct-2507`],model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Thinking-2507`,name:`Qwen3-4B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,alias:`Qwen/Qwen3-4B-Thinking`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-4B-Thinking`,name:`Qwen3-4B-Thinking-2507`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`Qwen/Qwen3-4B-Thinking-2507`],model_type:`chat`,parameters:4,page_url:`https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507`,release_date:`2025-08-05`,context_window:262144,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-8B-Base`,name:`Qwen3-8B-Base`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8.2,page_url:`https://huggingface.co/Qwen/Qwen3-8B-Base`,release_date:`2025-04-28`,context_window:32768,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-8B`,name:`Qwen3-8B`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8.2,page_url:`https://huggingface.co/Qwen/Qwen3-8B`,release_date:`2025-04-27`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-30B-A3B-Instruct`,name:`Qwen3-Coder-30B-A3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:31,active_parameters:1.9,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct`,release_date:`2025-07-31`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8`,name:`Qwen3-Coder-480B-A35B-Instruct-FP8`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:480,active_parameters:35,release_date:`2025-07-22`,context_window:262144,license:`apache-2.0`,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-480B-A35B-Instruct`,name:`Qwen3-Coder-480B-A35B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:480,active_parameters:24,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct`,release_date:`2025-07-22`,context_window:262144,license:`apache-2.0`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-Next-FP8`,name:`Qwen3-Coder-Next-FP8`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-Next-FP8`,release_date:`2026-02-01`,context_window:262144,license:`apache-2.0`,parameters:80,active_parameters:1.6,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Coder-Next`,name:`Qwen3-Coder-Next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,release_date:`2026-01-30`,context_window:262144,license:`apache-2.0`,parameters:80,active_parameters:1.6,page_url:`https://huggingface.co/Qwen/Qwen3-Coder-Next`,architecture:`moe`,open_weight:!0,tools:[`function_calling`],max_output_tokens:65536,modalities:{input:[`text`],output:[`text`]}},{id:`Qwen/Qwen3-Next-80B-A3B-Instruct`,name:`Qwen3-Next-80B-A3B-Instruct`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:81,active_parameters:1.6,page_url:`https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct`,description:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI). We are excited to share our latest advancements in addressing these demands, centered on improving scaling efficiency through innovative model architecture. We call this next-generation foundation models **Qwen3-Next**.`,release_date:`2025-09-09`,context_window:262144,license:`apache-2.0`,tagline:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI).`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`Qwen/Qwen3-Next-80B-A3B-Thinking`,name:`Qwen3-Next-80B-A3B-Thinking`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:80,active_parameters:3,page_url:`https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking`,description:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI). We are excited to share our latest advancements in addressing these demands, centered on improving scaling efficiency through innovative model architecture. We call this next-generation foundation models **Qwen3-Next**.`,release_date:`2025-09-09`,context_window:262144,license:`apache-2.0`,tagline:`Over the past few months, we have observed increasingly clear trends toward scaling both total parameters and context lengths in the pursuit of more powerful and agentic artificial intelligence (AI).`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV`,name:`Meta-Llama-3-8B-Instruct-FP8-KV`,created_by:`redhatai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV`,release_date:`2024-05-20`,context_window:8192,architecture:`transformer`,tools:[`function_calling`]},{id:`SakanaAI/TinySwallow-1.5B-Instruct`,name:`TinySwallow-1.5B-Instruct`,created_by:`sakanaai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},parameters:1.5,page_url:`https://huggingface.co/SakanaAI/TinySwallow-1.5B-Instruct`,description:`🤗 [Models](https://huggingface.co/SakanaAI) | 📚 [Paper](https://arxiv.org/abs/2501.16937) | 📝 [Blog](https://sakana.ai/taid-jp/) | 🐦 [Twitter](https://twitter.com/SakanaAILabs)`,release_date:`2025-01-07`,context_window:32768,license:`apache-2.0`,tagline:`🤗 [Models](https://huggingface.co/SakanaAI) | 📚 [Paper](https://arxiv.org/abs/2501.16937) | 📝 [Blog](https://sakana.ai/taid-jp/) | 🐦 [Twitter](https://twitter.com/SakanaAILabs)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`speakleash/Bielik-7B-Instruct-v0.1`,name:`Bielik-7B-Instruct-v0.1`,created_by:`speakleash`,source:`official`,last_updated:`2026-03-21`,capabilities:{streaming:!0},parameters:7},{id:`swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,name:`LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,created_by:`swap-uniba`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA`,description:`📣 New MODEL FAMILY❗ [https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA](https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA)`,release_date:`2024-04-29`,context_window:8192,license:`llama3`,tagline:`📣 New MODEL FAMILY❗ [https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA](https://huggingface.co/m-polignano/ANITA-NEXT-24B-Magistral-2506-VISION-ITA)`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`swiss-ai/Apertus-8B-Instruct-2509`,name:`Apertus-8B-Instruct-2509`,created_by:`swiss-ai`,source:`official`,last_updated:`2026-03-23`,alias:`swiss-ai/Apertus-8B-Instruct`,capabilities:{streaming:!0,tool_call:!0},parameters:8,page_url:`https://huggingface.co/swiss-ai/Apertus-8B-Instruct-2509`,release_date:`2025-08-13`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`swiss-ai/Apertus-8B-Instruct`,name:`Apertus-8B-Instruct-2509`,created_by:`swiss-ai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`swiss-ai/Apertus-8B-Instruct-2509`],parameters:8,page_url:`https://huggingface.co/swiss-ai/Apertus-8B-Instruct-2509`,release_date:`2025-08-13`,context_window:65536,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1`,name:`Llama-3-Swallow-8B-Instruct-v0.1`,created_by:`tokyotech-llm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1`,description:`Our Swallow model has undergone continual pre-training from the [Llama 3 family](https://huggingface.co/collections/meta-llama/meta-llama-3-66214712577ca38149ebb2b6), primarily with the addition of Japanese language data. The Instruct versions use supervised fine-tuning (SFT) and Chat Vector. Links to other models can be found in the index.`,release_date:`2024-06-26`,context_window:8192,license:`llama3`,tagline:`Our Swallow model has undergone continual pre-training from the [Llama 3 family](https://huggingface.co/collections/meta-llama/meta-llama-3-66214712577ca38149ebb2b6), primarily with the addition of Japanese language data.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5`,name:`Llama-3.1-Swallow-8B-Instruct-v0.5`,created_by:`tokyotech-llm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5`,description:`Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models. Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities.`,release_date:`2025-06-12`,context_window:131072,license:`llama3.3`,tagline:`Llama 3.1 Swallow is a series of large language models (8B, 70B) that were built by continual pre-training on the [Meta Llama 3.1](https://huggingface.co/collections/meta-llama/llama-31-669fc079a0c406a149a5738f) models.`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/gemma-2-9b-it`,name:`gemma-2-9b-it`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,family:`gemma-2`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:9.2,page_url:`https://huggingface.co/unsloth/gemma-2-9b-it`,release_date:`2024-07-03`,context_window:8192,license:`gemma`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/llama-3-8b-Instruct`,name:`llama-3-8b-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,family:`llama-3`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/llama-3-8b-Instruct`,description:"Directly quantized 4bit model with `bitsandbytes`.",release_date:`2024-04-18`,context_window:8192,license:`llama3`,tagline:"Directly quantized 4bit model with `bitsandbytes`.",architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Llama-3.1-8B-Instruct`,name:`Llama-3.1-8B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Llama-3.1-8B-Instruct`,release_date:`2025-02-15`,context_window:131072,license:`llama3.1`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Meta-Llama-3.1-8B-Instruct`,name:`Meta-Llama-3.1-8B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3.1`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Meta-Llama-3.1-8B-Instruct`,description:`We have a free Google Colab Tesla T4 notebook for Llama 3.1 (8B) here: https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRnrIOjZAuxp?usp=sharing`,release_date:`2024-07-23`,context_window:131072,license:`llama3.1`,tagline:`We have a free Google Colab Tesla T4 notebook for Llama 3.1 (8B) here: https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRnrIOjZAuxp?usp=sharing`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Mistral-Small-24B-Instruct-2501`,name:`Mistral-Small-24B-Instruct-2501`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,alias:`unsloth/Mistral-Small-24B-Instruct`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:24,page_url:`https://huggingface.co/unsloth/Mistral-Small-24B-Instruct-2501`,description:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,release_date:`2025-01-30`,context_window:32768,license:`apache-2.0`,tagline:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Mistral-Small-24B-Instruct`,name:`Mistral-Small-24B-Instruct-2501`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},snapshots:[`unsloth/Mistral-Small-24B-Instruct-2501`],model_type:`chat`,parameters:24,page_url:`https://huggingface.co/unsloth/Mistral-Small-24B-Instruct-2501`,description:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,release_date:`2025-01-30`,context_window:32768,license:`apache-2.0`,tagline:`We have a free Google Colab Tesla T4 notebook for Mistral (7B) here: https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Mistral_v0.3_(7B)-Conversational.ipynb`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-14B-Instruct`,name:`Qwen2.5-14B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:15,page_url:`https://huggingface.co/unsloth/Qwen2.5-14B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-23`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-32B-Instruct`,name:`Qwen2.5-32B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:33,page_url:`https://huggingface.co/unsloth/Qwen2.5-32B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-23`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-7B-Instruct`,name:`Qwen2.5-7B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:7.6,page_url:`https://huggingface.co/unsloth/Qwen2.5-7B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-09-18`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen2.5-Coder-32B-Instruct`,name:`Qwen2.5-Coder-32B-Instruct`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,parameters:33,page_url:`https://huggingface.co/unsloth/Qwen2.5-Coder-32B-Instruct`,description:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing). Also a [Qwen 2.5 conversational style notebook](https://colab.research.google.com/drive/1qN1CEalC70EO1wGKhNxs1go1W9So61R5?usp=sharing).`,release_date:`2024-11-12`,context_window:32768,license:`apache-2.0`,tagline:`We have a Qwen 2.5 (all model sizes) [free Google Colab Tesla T4 notebook](https://colab.research.google.com/drive/1Kose-ucXO1IBaZq5BvbwWieuubP7hxvQ?usp=sharing).`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/Qwen3-8B`,name:`Qwen3-8B`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,parameters:8,page_url:`https://huggingface.co/unsloth/Qwen3-8B`,release_date:`2025-04-28`,context_window:40960,license:`apache-2.0`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`unsloth/tinyllama-chat`,name:`tinyllama-chat`,created_by:`unsloth`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},page_url:`https://huggingface.co/unsloth/tinyllama-chat`,description:`[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/Discord%20button.png" width="200"/>](https://discord.gg/u54VK8m8tk) [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/buy%20me%20a%20coffee%20button.png" width="200"/>](https://ko-fi.com/unsloth) [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)`,release_date:`2024-02-14`,context_window:2048,license:`apache-2.0`,parameters:1.1,tagline:`[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/Discord%20button.png" width="200"/>](https://discord.gg/u54VK8m8tk) [<img src="https://raw.githubusercontent.com/unslothai/un`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct`,name:`Llama-3-SauerkrautLM-8b-Instruct`,created_by:`vagosolutions`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`Llama-3`,model_type:`chat`,parameters:8,page_url:`https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct`,release_date:`2024-04-19`,context_window:8192,license:`other`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4-9B-0414`,name:`GLM-4-9B-0414`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4`,model_type:`chat`,parameters:9.4,page_url:`https://huggingface.co/zai-org/GLM-4-9B-0414`,release_date:`2025-04-07`,context_window:32768,license:`mit`,architecture:`transformer`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5-Air-FP8`,name:`GLM-4.5-Air-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5-Air-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5-Air`,name:`GLM-4.5-Air`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5-Air`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,parameters:110,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.5`,name:`GLM-4.5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.5`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-07-20`,context_window:131072,license:`mit`,parameters:358,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.6-FP8`,name:`GLM-4.6-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.6`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.6-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-09-29`,context_window:202752,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.6`,name:`GLM-4.6`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.6`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.6`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-09-29`,context_window:202752,license:`mit`,parameters:357,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7-Flash`,name:`GLM-4.7-Flash`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7-Flash`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2026-01-19`,context_window:202752,license:`mit`,parameters:31,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7-FP8`,name:`GLM-4.7-FP8`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7-FP8`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-12-22`,context_window:202752,license:`mit`,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-4.7`,name:`GLM-4.7`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-4.7`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-4.7`,description:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2025-12-22`,context_window:202752,license:`mit`,parameters:358,tagline:`👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-5`,name:`GLM-5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`GLM-5`,model_type:`chat`,page_url:`https://huggingface.co/zai-org/GLM-5`,description:`👋 Join our <a href="https://raw.githubusercontent.com/zai-org/GLM-5/refs/heads/main/resources/wechat.png" target="_blank">WeChat</a> or <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.`,release_date:`2026-02-11`,context_window:202752,license:`mit`,parameters:754,tagline:`👋 Join our <a href="https://raw.githubusercontent.com/zai-org/GLM-5/refs/heads/main/resources/wechat.png" target="_blank">WeChat</a> or <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord`,architecture:`moe`,open_weight:!0,tools:[`function_calling`]},{id:`zai-org/GLM-Z1-32B-0414`,name:`GLM-Z1-32B-0414`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-21`,capabilities:{streaming:!0},model_type:`chat`,family:`glm`,parameters:32}]},t=e.models.map(t=>({...t,provider:e.id}));exports.models=t,exports.provider=e;