modelpedia 0.0.2 → 0.0.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (136) hide show
  1. package/README.md +141 -0
  2. package/dist/index.cjs +1 -101
  3. package/dist/index.d.cts +10 -2
  4. package/dist/index.d.mts +10 -2
  5. package/dist/index.mjs +1 -92
  6. package/dist/providers/alibaba.cjs +4 -835
  7. package/dist/providers/alibaba.d.cts +1 -1
  8. package/dist/providers/alibaba.d.mts +1 -1
  9. package/dist/providers/alibaba.mjs +4 -833
  10. package/dist/providers/amazon.cjs +4 -1484
  11. package/dist/providers/amazon.d.cts +1 -1
  12. package/dist/providers/amazon.d.mts +1 -1
  13. package/dist/providers/amazon.mjs +4 -1482
  14. package/dist/providers/anthropic.cjs +3 -698
  15. package/dist/providers/anthropic.d.cts +1 -1
  16. package/dist/providers/anthropic.d.mts +1 -1
  17. package/dist/providers/anthropic.mjs +3 -696
  18. package/dist/providers/azure.cjs +6 -1491
  19. package/dist/providers/azure.d.cts +1 -1
  20. package/dist/providers/azure.d.mts +1 -1
  21. package/dist/providers/azure.mjs +6 -1489
  22. package/dist/providers/baseten.cjs +3 -279
  23. package/dist/providers/baseten.d.cts +1 -1
  24. package/dist/providers/baseten.d.mts +1 -1
  25. package/dist/providers/baseten.mjs +3 -277
  26. package/dist/providers/cerebras.cjs +4 -145
  27. package/dist/providers/cerebras.d.cts +1 -1
  28. package/dist/providers/cerebras.d.mts +1 -1
  29. package/dist/providers/cerebras.mjs +4 -143
  30. package/dist/providers/cloudflare-ai-gateway.cjs +4 -0
  31. package/dist/providers/cloudflare-ai-gateway.d.cts +7 -0
  32. package/dist/providers/cloudflare-ai-gateway.d.mts +7 -0
  33. package/dist/providers/cloudflare-ai-gateway.mjs +4 -0
  34. package/dist/providers/cloudflare-workers-ai.cjs +4 -0
  35. package/dist/providers/cloudflare-workers-ai.d.cts +7 -0
  36. package/dist/providers/cloudflare-workers-ai.d.mts +7 -0
  37. package/dist/providers/cloudflare-workers-ai.mjs +4 -0
  38. package/dist/providers/cohere.cjs +5 -525
  39. package/dist/providers/cohere.d.cts +1 -1
  40. package/dist/providers/cohere.d.mts +1 -1
  41. package/dist/providers/cohere.mjs +5 -523
  42. package/dist/providers/cursor.cjs +1 -856
  43. package/dist/providers/cursor.d.cts +1 -1
  44. package/dist/providers/cursor.d.mts +1 -1
  45. package/dist/providers/cursor.mjs +1 -854
  46. package/dist/providers/deepseek.cjs +3 -64
  47. package/dist/providers/deepseek.d.cts +1 -1
  48. package/dist/providers/deepseek.d.mts +1 -1
  49. package/dist/providers/deepseek.mjs +3 -62
  50. package/dist/providers/fireworks.cjs +3 -3794
  51. package/dist/providers/fireworks.d.cts +1 -1
  52. package/dist/providers/fireworks.d.mts +1 -1
  53. package/dist/providers/fireworks.mjs +3 -3792
  54. package/dist/providers/google.cjs +6 -972
  55. package/dist/providers/google.d.cts +1 -1
  56. package/dist/providers/google.d.mts +1 -1
  57. package/dist/providers/google.mjs +6 -970
  58. package/dist/providers/groq.cjs +3 -353
  59. package/dist/providers/groq.d.cts +1 -1
  60. package/dist/providers/groq.d.mts +1 -1
  61. package/dist/providers/groq.mjs +3 -351
  62. package/dist/providers/huggingface.cjs +3 -1022
  63. package/dist/providers/huggingface.d.cts +1 -1
  64. package/dist/providers/huggingface.d.mts +1 -1
  65. package/dist/providers/huggingface.mjs +3 -1020
  66. package/dist/providers/meta.cjs +3 -481
  67. package/dist/providers/meta.d.cts +1 -1
  68. package/dist/providers/meta.d.mts +1 -1
  69. package/dist/providers/meta.mjs +3 -479
  70. package/dist/providers/minimax.cjs +3 -156
  71. package/dist/providers/minimax.d.cts +1 -1
  72. package/dist/providers/minimax.d.mts +1 -1
  73. package/dist/providers/minimax.mjs +3 -154
  74. package/dist/providers/mistral.cjs +3 -1594
  75. package/dist/providers/mistral.d.cts +1 -1
  76. package/dist/providers/mistral.d.mts +1 -1
  77. package/dist/providers/mistral.mjs +3 -1592
  78. package/dist/providers/moonshot.cjs +3 -138
  79. package/dist/providers/moonshot.d.cts +1 -1
  80. package/dist/providers/moonshot.d.mts +1 -1
  81. package/dist/providers/moonshot.mjs +3 -136
  82. package/dist/providers/nvidia.cjs +3 -2040
  83. package/dist/providers/nvidia.d.cts +1 -1
  84. package/dist/providers/nvidia.d.mts +1 -1
  85. package/dist/providers/nvidia.mjs +3 -2038
  86. package/dist/providers/ollama.cjs +3 -326
  87. package/dist/providers/ollama.d.cts +1 -1
  88. package/dist/providers/ollama.d.mts +1 -1
  89. package/dist/providers/ollama.mjs +3 -324
  90. package/dist/providers/openai.cjs +3 -9079
  91. package/dist/providers/openai.d.cts +1 -1
  92. package/dist/providers/openai.d.mts +1 -1
  93. package/dist/providers/openai.mjs +3 -9077
  94. package/dist/providers/opencode.cjs +3 -367
  95. package/dist/providers/opencode.d.cts +1 -1
  96. package/dist/providers/opencode.d.mts +1 -1
  97. package/dist/providers/opencode.mjs +3 -365
  98. package/dist/providers/openrouter.cjs +21 -7835
  99. package/dist/providers/openrouter.d.cts +1 -1
  100. package/dist/providers/openrouter.d.mts +1 -1
  101. package/dist/providers/openrouter.mjs +21 -7833
  102. package/dist/providers/perplexity.cjs +3 -311
  103. package/dist/providers/perplexity.d.cts +1 -1
  104. package/dist/providers/perplexity.d.mts +1 -1
  105. package/dist/providers/perplexity.mjs +3 -309
  106. package/dist/providers/qwen.cjs +3 -835
  107. package/dist/providers/qwen.d.cts +1 -1
  108. package/dist/providers/qwen.d.mts +1 -1
  109. package/dist/providers/qwen.mjs +3 -833
  110. package/dist/providers/together.cjs +4 -360
  111. package/dist/providers/together.d.cts +1 -1
  112. package/dist/providers/together.d.mts +1 -1
  113. package/dist/providers/together.mjs +4 -358
  114. package/dist/providers/vercel.cjs +3 -4037
  115. package/dist/providers/vercel.d.cts +1 -1
  116. package/dist/providers/vercel.d.mts +1 -1
  117. package/dist/providers/vercel.mjs +3 -4035
  118. package/dist/providers/vertex.cjs +10 -1072
  119. package/dist/providers/vertex.d.cts +1 -1
  120. package/dist/providers/vertex.d.mts +1 -1
  121. package/dist/providers/vertex.mjs +10 -1070
  122. package/dist/providers/xai.cjs +3 -301
  123. package/dist/providers/xai.d.cts +1 -1
  124. package/dist/providers/xai.d.mts +1 -1
  125. package/dist/providers/xai.mjs +3 -299
  126. package/dist/providers/zai.cjs +3 -243
  127. package/dist/providers/zai.d.cts +1 -1
  128. package/dist/providers/zai.d.mts +1 -1
  129. package/dist/providers/zai.mjs +3 -241
  130. package/dist/types.d.cts +80 -8
  131. package/dist/types.d.mts +80 -8
  132. package/package.json +3 -3
  133. package/dist/providers/cloudflare.cjs +0 -826
  134. package/dist/providers/cloudflare.d.cts +0 -7
  135. package/dist/providers/cloudflare.d.mts +0 -7
  136. package/dist/providers/cloudflare.mjs +0 -824
@@ -1,326 +1,3 @@
1
- Object.defineProperty(exports, Symbol.toStringTag, { value: "Module" });
2
- //#region src/providers/ollama.ts
3
- const provider = {
4
- "id": "ollama",
5
- "name": "Ollama",
6
- "region": "US",
7
- "url": "https://ollama.com",
8
- "api_url": "http://localhost:11434/api",
9
- "docs_url": "https://github.com/ollama/ollama/blob/main/docs/api.md",
10
- "pricing_url": "https://ollama.com",
11
- "icon": "<svg xmlns=\"http://www.w3.org/2000/svg\" viewBox=\"0 0 24 24\" fill=\"currentColor\" fill-rule=\"evenodd\">\n <path d=\"M7.905 1.09c.216.085.411.225.588.41.295.306.544.744.734 1.263.191.522.315 1.1.362 1.68a5.054 5.054 0 012.049-.636l.051-.004c.87-.07 1.73.087 2.48.474.101.053.2.11.297.17.05-.569.172-1.134.36-1.644.19-.52.439-.957.733-1.264a1.67 1.67 0 01.589-.41c.257-.1.53-.118.796-.042.401.114.745.368 1.016.737.248.337.434.769.561 1.287.23.934.27 2.163.115 3.645l.053.04.026.019c.757.576 1.284 1.397 1.563 2.35.435 1.487.216 3.155-.534 4.088l-.018.021.002.003c.417.762.67 1.567.724 2.4l.002.03c.064 1.065-.2 2.137-.814 3.19l-.007.01.01.024c.472 1.157.62 2.322.438 3.486l-.006.039a.651.651 0 01-.747.536.648.648 0 01-.54-.742c.167-1.033.01-2.069-.48-3.123a.643.643 0 01.04-.617l.004-.006c.604-.924.854-1.83.8-2.72-.046-.779-.325-1.544-.8-2.273a.644.644 0 01.18-.886l.009-.006c.243-.159.467-.565.58-1.12a4.229 4.229 0 00-.095-1.974c-.205-.7-.58-1.284-1.105-1.683-.595-.454-1.383-.673-2.38-.61a.653.653 0 01-.632-.371c-.314-.665-.772-1.141-1.343-1.436a3.288 3.288 0 00-1.772-.332c-1.245.099-2.343.801-2.67 1.686a.652.652 0 01-.61.425c-1.067.002-1.893.252-2.497.703-.522.39-.878.935-1.066 1.588a4.07 4.07 0 00-.068 1.886c.112.558.331 1.02.582 1.269l.008.007c.212.207.257.53.109.785-.36.622-.629 1.549-.673 2.44-.05 1.018.186 1.902.719 2.536l.016.019a.643.643 0 01.095.69c-.576 1.236-.753 2.252-.562 3.052a.652.652 0 01-1.269.298c-.243-1.018-.078-2.184.473-3.498l.014-.035-.008-.012a4.339 4.339 0 01-.598-1.309l-.005-.019a5.764 5.764 0 01-.177-1.785c.044-.91.278-1.842.622-2.59l.012-.026-.002-.002c-.293-.418-.51-.953-.63-1.545l-.005-.024a5.352 5.352 0 01.093-2.49c.262-.915.777-1.701 1.536-2.269.06-.045.123-.09.186-.132-.159-1.493-.119-2.73.112-3.67.127-.518.314-.95.562-1.287.27-.368.614-.622 1.015-.737.266-.076.54-.059.797.042zm4.116 9.09c.936 0 1.8.313 2.446.855.63.527 1.005 1.235 1.005 1.94 0 .888-.406 1.58-1.133 2.022-.62.375-1.451.557-2.403.557-1.009 0-1.871-.259-2.493-.734-.617-.47-.963-1.13-.963-1.845 0-.707.398-1.417 1.056-1.946.668-.537 1.55-.849 2.485-.849zm0 .896a3.07 3.07 0 00-1.916.65c-.461.37-.722.835-.722 1.25 0 .428.21.829.61 1.134.455.347 1.124.548 1.943.548.799 0 1.473-.147 1.932-.426.463-.28.7-.686.7-1.257 0-.423-.246-.89-.683-1.256-.484-.405-1.14-.643-1.864-.643zm.662 1.21l.004.004c.12.151.095.37-.056.49l-.292.23v.446a.375.375 0 01-.376.373.375.375 0 01-.376-.373v-.46l-.271-.218a.347.347 0 01-.052-.49.353.353 0 01.494-.051l.215.172.22-.174a.353.353 0 01.49.051zm-5.04-1.919c.478 0 .867.39.867.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zm8.706 0c.48 0 .868.39.868.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zM7.44 2.3l-.003.002a.659.659 0 00-.285.238l-.005.006c-.138.189-.258.467-.348.832-.17.692-.216 1.631-.124 2.782.43-.128.899-.208 1.404-.237l.01-.001.019-.034c.046-.082.095-.161.148-.239.123-.771.022-1.692-.253-2.444-.134-.364-.297-.65-.453-.813a.628.628 0 00-.107-.09L7.44 2.3zm9.174.04l-.002.001a.628.628 0 00-.107.09c-.156.163-.32.45-.453.814-.29.794-.387 1.776-.23 2.572l.058.097.008.014h.03a5.184 5.184 0 011.466.212c.086-1.124.038-2.043-.128-2.722-.09-.365-.21-.643-.349-.832l-.004-.006a.659.659 0 00-.285-.239h-.004z\"/>\n</svg>",
12
- "models": [
13
- {
14
- "id": "cogito-2.1",
15
- "name": "cogito-2.1",
16
- "created_by": "unknown",
17
- "source": "official",
18
- "last_updated": "2026-03-21",
19
- "capabilities": { "streaming": true }
20
- },
21
- {
22
- "id": "deepseek-v3.1",
23
- "name": "deepseek-v3.1",
24
- "created_by": "deepseek",
25
- "source": "official",
26
- "last_updated": "2026-03-21",
27
- "family": "deepseek",
28
- "capabilities": { "streaming": true }
29
- },
30
- {
31
- "id": "deepseek-v3.2",
32
- "name": "deepseek-v3.2",
33
- "created_by": "deepseek",
34
- "source": "official",
35
- "last_updated": "2026-03-21",
36
- "family": "deepseek",
37
- "capabilities": { "streaming": true }
38
- },
39
- {
40
- "id": "devstral-2",
41
- "name": "devstral-2",
42
- "created_by": "mistral",
43
- "source": "official",
44
- "last_updated": "2026-03-21",
45
- "family": "devstral",
46
- "capabilities": { "streaming": true },
47
- "model_type": "code"
48
- },
49
- {
50
- "id": "devstral-small-2",
51
- "name": "devstral-small-2",
52
- "created_by": "mistral",
53
- "source": "official",
54
- "last_updated": "2026-03-21",
55
- "family": "devstral",
56
- "capabilities": { "streaming": true },
57
- "model_type": "code"
58
- },
59
- {
60
- "id": "gemini-3-flash-preview",
61
- "name": "gemini-3-flash-preview",
62
- "created_by": "unknown",
63
- "source": "official",
64
- "last_updated": "2026-03-21",
65
- "family": "gemini-3",
66
- "capabilities": { "streaming": true }
67
- },
68
- {
69
- "id": "gemma3",
70
- "name": "gemma3",
71
- "created_by": "google",
72
- "source": "official",
73
- "last_updated": "2026-03-21",
74
- "capabilities": { "streaming": true }
75
- },
76
- {
77
- "id": "glm-4.6",
78
- "name": "glm-4.6",
79
- "created_by": "zhipu",
80
- "source": "official",
81
- "last_updated": "2026-03-21",
82
- "capabilities": { "streaming": true }
83
- },
84
- {
85
- "id": "glm-4.7-flash",
86
- "name": "glm-4.7-flash",
87
- "created_by": "zhipu",
88
- "source": "official",
89
- "last_updated": "2026-03-21",
90
- "capabilities": { "streaming": true }
91
- },
92
- {
93
- "id": "glm-4.7",
94
- "name": "glm-4.7",
95
- "created_by": "zhipu",
96
- "source": "official",
97
- "last_updated": "2026-03-21",
98
- "capabilities": { "streaming": true }
99
- },
100
- {
101
- "id": "glm-5",
102
- "name": "glm-5",
103
- "created_by": "zhipu",
104
- "source": "official",
105
- "last_updated": "2026-03-21",
106
- "capabilities": { "streaming": true }
107
- },
108
- {
109
- "id": "glm-ocr",
110
- "name": "glm-ocr",
111
- "created_by": "zhipu",
112
- "source": "official",
113
- "last_updated": "2026-03-21",
114
- "capabilities": { "streaming": true }
115
- },
116
- {
117
- "id": "gpt-oss",
118
- "name": "gpt-oss",
119
- "created_by": "openai",
120
- "source": "official",
121
- "last_updated": "2026-03-21",
122
- "family": "gpt-oss",
123
- "capabilities": { "streaming": true }
124
- },
125
- {
126
- "id": "granite4",
127
- "name": "granite4",
128
- "created_by": "ibm",
129
- "source": "official",
130
- "last_updated": "2026-03-21",
131
- "capabilities": { "streaming": true }
132
- },
133
- {
134
- "id": "kimi-k2-thinking",
135
- "name": "kimi-k2-thinking",
136
- "created_by": "moonshot",
137
- "source": "official",
138
- "last_updated": "2026-03-21",
139
- "capabilities": { "streaming": true }
140
- },
141
- {
142
- "id": "kimi-k2.5",
143
- "name": "kimi-k2.5",
144
- "created_by": "moonshot",
145
- "source": "official",
146
- "last_updated": "2026-03-21",
147
- "capabilities": { "streaming": true }
148
- },
149
- {
150
- "id": "kimi-k2",
151
- "name": "kimi-k2",
152
- "created_by": "moonshot",
153
- "source": "official",
154
- "last_updated": "2026-03-21",
155
- "capabilities": { "streaming": true }
156
- },
157
- {
158
- "id": "lfm2.5-thinking",
159
- "name": "lfm2.5-thinking",
160
- "created_by": "liquid",
161
- "source": "official",
162
- "last_updated": "2026-03-21",
163
- "capabilities": { "streaming": true }
164
- },
165
- {
166
- "id": "lfm2",
167
- "name": "lfm2",
168
- "created_by": "liquid",
169
- "source": "official",
170
- "last_updated": "2026-03-21",
171
- "capabilities": { "streaming": true }
172
- },
173
- {
174
- "id": "minimax-m2.1",
175
- "name": "minimax-m2.1",
176
- "created_by": "minimax",
177
- "source": "official",
178
- "last_updated": "2026-03-21",
179
- "capabilities": { "streaming": true }
180
- },
181
- {
182
- "id": "minimax-m2.5",
183
- "name": "minimax-m2.5",
184
- "created_by": "minimax",
185
- "source": "official",
186
- "last_updated": "2026-03-21",
187
- "capabilities": { "streaming": true }
188
- },
189
- {
190
- "id": "minimax-m2.7",
191
- "name": "minimax-m2.7",
192
- "created_by": "minimax",
193
- "source": "official",
194
- "last_updated": "2026-03-21",
195
- "capabilities": { "streaming": true }
196
- },
197
- {
198
- "id": "minimax-m2",
199
- "name": "minimax-m2",
200
- "created_by": "minimax",
201
- "source": "official",
202
- "last_updated": "2026-03-21",
203
- "capabilities": { "streaming": true }
204
- },
205
- {
206
- "id": "ministral-3",
207
- "name": "ministral-3",
208
- "created_by": "mistral",
209
- "source": "official",
210
- "last_updated": "2026-03-21",
211
- "family": "ministral",
212
- "capabilities": { "streaming": true }
213
- },
214
- {
215
- "id": "mistral-large-3",
216
- "name": "mistral-large-3",
217
- "created_by": "mistral",
218
- "source": "official",
219
- "last_updated": "2026-03-21",
220
- "family": "mistral-large",
221
- "capabilities": { "streaming": true }
222
- },
223
- {
224
- "id": "nemotron-3-nano",
225
- "name": "nemotron-3-nano",
226
- "created_by": "nvidia",
227
- "source": "official",
228
- "last_updated": "2026-03-21",
229
- "capabilities": { "streaming": true }
230
- },
231
- {
232
- "id": "nemotron-3-super",
233
- "name": "nemotron-3-super",
234
- "created_by": "nvidia",
235
- "source": "official",
236
- "last_updated": "2026-03-21",
237
- "capabilities": { "streaming": true }
238
- },
239
- {
240
- "id": "nemotron-cascade-2",
241
- "name": "nemotron-cascade-2",
242
- "created_by": "nvidia",
243
- "source": "official",
244
- "last_updated": "2026-03-21",
245
- "capabilities": { "streaming": true }
246
- },
247
- {
248
- "id": "qwen3-coder-next",
249
- "name": "qwen3-coder-next",
250
- "created_by": "qwen",
251
- "source": "official",
252
- "last_updated": "2026-03-21",
253
- "family": "qwen3",
254
- "capabilities": { "streaming": true }
255
- },
256
- {
257
- "id": "qwen3-coder",
258
- "name": "qwen3-coder",
259
- "created_by": "qwen",
260
- "source": "official",
261
- "last_updated": "2026-03-21",
262
- "family": "qwen3",
263
- "capabilities": { "streaming": true }
264
- },
265
- {
266
- "id": "qwen3-embedding",
267
- "name": "qwen3-embedding",
268
- "created_by": "qwen",
269
- "source": "official",
270
- "last_updated": "2026-03-21",
271
- "family": "qwen3",
272
- "capabilities": { "streaming": true },
273
- "model_type": "embed"
274
- },
275
- {
276
- "id": "qwen3-next",
277
- "name": "qwen3-next",
278
- "created_by": "qwen",
279
- "source": "official",
280
- "last_updated": "2026-03-21",
281
- "family": "qwen3",
282
- "capabilities": { "streaming": true }
283
- },
284
- {
285
- "id": "qwen3-vl",
286
- "name": "qwen3-vl",
287
- "created_by": "qwen",
288
- "source": "official",
289
- "last_updated": "2026-03-21",
290
- "family": "qwen3",
291
- "capabilities": { "streaming": true }
292
- },
293
- {
294
- "id": "qwen3.5",
295
- "name": "qwen3.5",
296
- "created_by": "qwen",
297
- "source": "official",
298
- "last_updated": "2026-03-21",
299
- "family": "qwen3.5",
300
- "capabilities": { "streaming": true }
301
- },
302
- {
303
- "id": "rnj-1",
304
- "name": "rnj-1",
305
- "created_by": "essentialai",
306
- "source": "official",
307
- "last_updated": "2026-03-21",
308
- "capabilities": { "streaming": true }
309
- },
310
- {
311
- "id": "translategemma",
312
- "name": "translategemma",
313
- "created_by": "google",
314
- "source": "official",
315
- "last_updated": "2026-03-21",
316
- "capabilities": { "streaming": true }
317
- }
318
- ]
319
- };
320
- const models = provider.models.map((m) => ({
321
- ...m,
322
- provider: provider.id
323
- }));
324
- //#endregion
325
- exports.models = models;
326
- exports.provider = provider;
1
+ Object.defineProperty(exports,Symbol.toStringTag,{value:`Module`});const e={id:`ollama`,name:`Ollama`,region:`US`,headquarters:`Palo Alto, CA`,founded:2023,url:`https://ollama.com`,api_url:`http://localhost:11434/api`,docs_url:`https://github.com/ollama/ollama/blob/main/docs/api.md`,pricing_url:`https://ollama.com`,description:`Run open-source LLMs locally on your machine with a simple CLI.`,type:`aggregator`,openai_compatible:!0,free_tier:!0,github_url:`https://github.com/ollama/ollama`,models_url:`https://ollama.com/library`,twitter_url:`https://x.com/ollama`,discord_url:`https://discord.com/invite/ollama`,blog_url:`https://ollama.com/blog`,support_url:`https://github.com/ollama/ollama`,icon:`<svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24" fill="currentColor" fill-rule="evenodd">
2
+ <path d="M7.905 1.09c.216.085.411.225.588.41.295.306.544.744.734 1.263.191.522.315 1.1.362 1.68a5.054 5.054 0 012.049-.636l.051-.004c.87-.07 1.73.087 2.48.474.101.053.2.11.297.17.05-.569.172-1.134.36-1.644.19-.52.439-.957.733-1.264a1.67 1.67 0 01.589-.41c.257-.1.53-.118.796-.042.401.114.745.368 1.016.737.248.337.434.769.561 1.287.23.934.27 2.163.115 3.645l.053.04.026.019c.757.576 1.284 1.397 1.563 2.35.435 1.487.216 3.155-.534 4.088l-.018.021.002.003c.417.762.67 1.567.724 2.4l.002.03c.064 1.065-.2 2.137-.814 3.19l-.007.01.01.024c.472 1.157.62 2.322.438 3.486l-.006.039a.651.651 0 01-.747.536.648.648 0 01-.54-.742c.167-1.033.01-2.069-.48-3.123a.643.643 0 01.04-.617l.004-.006c.604-.924.854-1.83.8-2.72-.046-.779-.325-1.544-.8-2.273a.644.644 0 01.18-.886l.009-.006c.243-.159.467-.565.58-1.12a4.229 4.229 0 00-.095-1.974c-.205-.7-.58-1.284-1.105-1.683-.595-.454-1.383-.673-2.38-.61a.653.653 0 01-.632-.371c-.314-.665-.772-1.141-1.343-1.436a3.288 3.288 0 00-1.772-.332c-1.245.099-2.343.801-2.67 1.686a.652.652 0 01-.61.425c-1.067.002-1.893.252-2.497.703-.522.39-.878.935-1.066 1.588a4.07 4.07 0 00-.068 1.886c.112.558.331 1.02.582 1.269l.008.007c.212.207.257.53.109.785-.36.622-.629 1.549-.673 2.44-.05 1.018.186 1.902.719 2.536l.016.019a.643.643 0 01.095.69c-.576 1.236-.753 2.252-.562 3.052a.652.652 0 01-1.269.298c-.243-1.018-.078-2.184.473-3.498l.014-.035-.008-.012a4.339 4.339 0 01-.598-1.309l-.005-.019a5.764 5.764 0 01-.177-1.785c.044-.91.278-1.842.622-2.59l.012-.026-.002-.002c-.293-.418-.51-.953-.63-1.545l-.005-.024a5.352 5.352 0 01.093-2.49c.262-.915.777-1.701 1.536-2.269.06-.045.123-.09.186-.132-.159-1.493-.119-2.73.112-3.67.127-.518.314-.95.562-1.287.27-.368.614-.622 1.015-.737.266-.076.54-.059.797.042zm4.116 9.09c.936 0 1.8.313 2.446.855.63.527 1.005 1.235 1.005 1.94 0 .888-.406 1.58-1.133 2.022-.62.375-1.451.557-2.403.557-1.009 0-1.871-.259-2.493-.734-.617-.47-.963-1.13-.963-1.845 0-.707.398-1.417 1.056-1.946.668-.537 1.55-.849 2.485-.849zm0 .896a3.07 3.07 0 00-1.916.65c-.461.37-.722.835-.722 1.25 0 .428.21.829.61 1.134.455.347 1.124.548 1.943.548.799 0 1.473-.147 1.932-.426.463-.28.7-.686.7-1.257 0-.423-.246-.89-.683-1.256-.484-.405-1.14-.643-1.864-.643zm.662 1.21l.004.004c.12.151.095.37-.056.49l-.292.23v.446a.375.375 0 01-.376.373.375.375 0 01-.376-.373v-.46l-.271-.218a.347.347 0 01-.052-.49.353.353 0 01.494-.051l.215.172.22-.174a.353.353 0 01.49.051zm-5.04-1.919c.478 0 .867.39.867.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zm8.706 0c.48 0 .868.39.868.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zM7.44 2.3l-.003.002a.659.659 0 00-.285.238l-.005.006c-.138.189-.258.467-.348.832-.17.692-.216 1.631-.124 2.782.43-.128.899-.208 1.404-.237l.01-.001.019-.034c.046-.082.095-.161.148-.239.123-.771.022-1.692-.253-2.444-.134-.364-.297-.65-.453-.813a.628.628 0 00-.107-.09L7.44 2.3zm9.174.04l-.002.001a.628.628 0 00-.107.09c-.156.163-.32.45-.453.814-.29.794-.387 1.776-.23 2.572l.058.097.008.014h.03a5.184 5.184 0 011.466.212c.086-1.124.038-2.043-.128-2.722-.09-.365-.21-.643-.349-.832l-.004-.006a.659.659 0 00-.285-.239h-.004z"/>
3
+ </svg>`,models:[{id:`cogito-2.1`,name:`cogito-2.1`,created_by:`deepcogito`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},description:`The Cogito v2.1 LLMs are instruction tuned generative models. All models are released under MIT license for commercial use.`,release_date:`2025-11-19`,parameters:671,page_url:`https://ollama.com/library/cogito-2.1`,tagline:`The Cogito v2.1 LLMs are instruction tuned generative models.`,context_window:163840,license:`mit`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`deepseek-v3.1`,name:`deepseek-v3.1`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,family:`deepseek`,capabilities:{streaming:!0},description:`DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.`,release_date:`2025-11-20`,parameters:671,page_url:`https://ollama.com/library/deepseek-v3.1`,tagline:`DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.`,context_window:163840,license:`open-weight`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`deepseek-v3.2`,name:`deepseek-v3.2`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,family:`deepseek`,capabilities:{streaming:!0},description:`DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.`,release_date:`2025-12-02`,page_url:`https://ollama.com/library/deepseek-v3.2`,tagline:`DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.`,context_window:163840,license:`open-weight`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`devstral-2`,name:`devstral-2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`devstral`,capabilities:{streaming:!0},model_type:`code`,description:`123B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,release_date:`2025-12-08`,parameters:123,page_url:`https://ollama.com/library/devstral-2`,tagline:`123B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,context_window:262144,license:`mit`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`devstral-small-2`,name:`devstral-small-2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`devstral`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`code`,description:`24B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,release_date:`2025-12-09`,parameters:24,page_url:`https://ollama.com/library/devstral-small-2`,tagline:`24B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,context_window:393216,license:`apache-2.0`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`gemini-3-flash-preview`,name:`gemini-3-flash-preview`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemini-3`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0,batch:!0,vision:!0},model_type:`chat`,description:`Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost.`,release_date:`2025-12-17`,page_url:`https://ollama.com/library/gemini-3-flash-preview`,tagline:`Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost.`,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]},context_window:1048576,max_output_tokens:65536,knowledge_cutoff:`2025-01`},{id:`gemma3`,name:`gemma3`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,description:`The current, most capable model that runs on a single GPU.`,release_date:`2025-03-12`,page_url:`https://ollama.com/library/gemma3`,tagline:`The current, most capable model that runs on a single GPU.`,context_window:131072,license:`gemma`,parameters:4,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.6`,name:`glm-4.6`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},family:`glm-4.6`,model_type:`chat`,description:`Advanced agentic, reasoning and coding capabilities.`,release_date:`2025-09-29`,page_url:`https://ollama.com/library/glm-4.6`,tagline:`Advanced agentic, reasoning and coding capabilities.`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.7-flash`,name:`glm-4.7-flash`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},family:`glm-4.7`,model_type:`chat`,description:`As the strongest model in the 30B class, GLM-4.7-Flash offers a new option for lightweight deployment that balances performance and efficiency.`,page_url:`https://ollama.com/library/glm-4.7-flash`,tagline:`As the strongest model in the 30B class, GLM-4.7-Flash offers a new option for lightweight deployment that balances performance and efficiency.`,context_window:202752,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.7`,name:`glm-4.7`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},family:`glm-4.7`,model_type:`chat`,description:`Advancing the Coding Capability`,release_date:`2025-12-22`,page_url:`https://ollama.com/library/glm-4.7`,tagline:`Advancing the Coding Capability`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-5`,name:`glm-5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,family:`glm-5`,description:`A strong reasoning and agentic model from Z.ai with 744B total parameters (40B active), built for complex systems engineering and long-horizon tasks.`,release_date:`2026-02-11`,page_url:`https://ollama.com/library/glm-5`,tagline:`A strong reasoning and agentic model from Z.ai with 744B total parameters (40B active), built for complex systems engineering and long-horizon tasks.`,context_window:202752,parameters:744,license:`open-weight`,active_parameters:40,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-ocr`,name:`glm-ocr`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`chat`,family:`glm`,description:`GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.`,page_url:`https://ollama.com/library/glm-ocr`,tagline:`GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.`,context_window:131072,parameters:.9,license:`open-weight`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`gpt-oss`,name:`gpt-oss`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0},model_type:`chat`,description:`OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-05`,page_url:`https://ollama.com/library/gpt-oss`,tagline:`OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,context_window:131072,license:`apache-2.0`,parameters:20,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`granite4`,name:`granite4`,created_by:`ibm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`Granite 4 features improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.`,page_url:`https://ollama.com/library/granite4`,tagline:`Granite 4 features improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.`,context_window:131072,license:`apache-2.0`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2-thinking`,name:`kimi-k2-thinking`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,family:`kimi-k2`,description:`Kimi K2 Thinking, Moonshot AI&#39;s best open-source thinking model.`,release_date:`2025-11-06`,page_url:`https://ollama.com/library/kimi-k2-thinking`,tagline:`Kimi K2 Thinking, Moonshot AI&#39;s best open-source thinking model.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2.5`,name:`kimi-k2.5`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},family:`kimi-k2.5`,model_type:`chat`,description:`Kimi K2.5 is an open-source, native multimodal agentic model that seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as conversational and agentic paradigms.`,release_date:`2026-01-26`,page_url:`https://ollama.com/library/kimi-k2.5`,tagline:`Kimi K2.5 is an open-source, native multimodal agentic model that seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as con`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2`,name:`kimi-k2`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,family:`kimi-k2`,description:`A state-of-the-art mixture-of-experts (MoE) language model. Kimi K2-Instruct-0905 demonstrates significant improvements in performance on public benchmarks and real-world coding agent tasks.`,release_date:`2025-09-05`,parameters:1e3,page_url:`https://ollama.com/library/kimi-k2`,tagline:`A state-of-the-art mixture-of-experts (MoE) language model.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`lfm2.5-thinking`,name:`lfm2.5-thinking`,created_by:`liquid`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`LFM2.5 is a new family of hybrid models designed for on-device deployment.`,page_url:`https://ollama.com/library/lfm2.5-thinking`,tagline:`LFM2.5 is a new family of hybrid models designed for on-device deployment.`,context_window:128e3,parameters:1.2,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`lfm2`,name:`lfm2`,created_by:`liquid`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`LFM2 is a family of hybrid models designed for on-device deployment. LFM2-24B-A2B is the largest model in the family, scaling the architecture to 24 billion parameters while keeping inference efficient.`,page_url:`https://ollama.com/library/lfm2`,tagline:`LFM2 is a family of hybrid models designed for on-device deployment.`,context_window:32768,parameters:24,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.1`,name:`minimax-m2.1`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`Exceptional multilingual capabilities to elevate code engineering`,release_date:`2025-12-20`,page_url:`https://ollama.com/library/minimax-m2.1`,tagline:`Exceptional multilingual capabilities to elevate code engineering`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.5`,name:`minimax-m2.5`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`MiniMax-M2.5 is a state-of-the-art large language model designed for real-world productivity and coding tasks.`,release_date:`2026-02-12`,page_url:`https://ollama.com/library/minimax-m2.5`,tagline:`MiniMax-M2.5 is a state-of-the-art large language model designed for real-world productivity and coding tasks.`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.7`,name:`minimax-m2.7`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`MiniMax&#39;s M2-series model for coding, agentic workflows, and professional productivity.`,release_date:`2026-03-18`,page_url:`https://ollama.com/library/minimax-m2.7`,tagline:`MiniMax&#39;s M2-series model for coding, agentic workflows, and professional productivity.`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2`,name:`minimax-m2`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,family:`minimax`,description:`MiniMax M2 is a high-efficiency large language model built for coding and agentic workflows.`,release_date:`2025-10-27`,page_url:`https://ollama.com/library/minimax-m2`,tagline:`MiniMax M2 is a high-efficiency large language model built for coding and agentic workflows.`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`ministral-3`,name:`ministral-3`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`ministral`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`chat`,description:`The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.`,release_date:`2025-12-02`,page_url:`https://ollama.com/library/ministral-3`,tagline:`The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.`,context_window:262144,license:`apache-2.0`,parameters:8,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`mistral-large-3`,name:`mistral-large-3`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`mistral-large`,capabilities:{streaming:!0},model_type:`chat`,description:`A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.`,release_date:`2025-12-02`,parameters:675,page_url:`https://ollama.com/library/mistral-large-3`,tagline:`A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.`,context_window:262144,license:`apache-2.0`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-3-nano`,name:`nemotron-3-nano`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`Nemotron-3-Nano is a new Standard for Efficient, Open, and Intelligent Agentic Models, now updated with a 4B parameter count model.`,release_date:`2025-12-15`,parameters:30,page_url:`https://ollama.com/library/nemotron-3-nano`,tagline:`Nemotron-3-Nano is a new Standard for Efficient, Open, and Intelligent Agentic Models, now updated with a 4B parameter count model.`,context_window:262144,license:`nvidia-open`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-3-super`,name:`nemotron-3-super`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.`,release_date:`2026-03-11`,page_url:`https://ollama.com/library/nemotron-3-super`,tagline:`NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.`,context_window:262144,parameters:120,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-cascade-2`,name:`nemotron-cascade-2`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`An open 30B MoE model from NVIDIA with 3B activated parameters that delivers strong reasoning and agentic capabilities.`,page_url:`https://ollama.com/library/nemotron-cascade-2`,tagline:`An open 30B MoE model from NVIDIA with 3B activated parameters that delivers strong reasoning and agentic capabilities.`,context_window:262144,parameters:30,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-coder-next`,name:`qwen3-coder-next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,description:`Qwen3-Coder-Next is a coding-focused language model from Alibaba&#39;s Qwen team, optimized for agentic coding workflows and local development.`,release_date:`2025-02-04`,page_url:`https://ollama.com/library/qwen3-coder-next`,tagline:`Qwen3-Coder-Next is a coding-focused language model from Alibaba&#39;s Qwen team, optimized for agentic coding workflows and local development.`,context_window:262144,license:`open-weight`,parameters:80,active_parameters:3,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]},max_output_tokens:65536},{id:`qwen3-coder`,name:`qwen3-coder`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0},model_type:`code`,description:`Alibaba&#39;s performant long context models for agentic and coding tasks.`,release_date:`2025-07-22`,parameters:480,page_url:`https://ollama.com/library/qwen3-coder`,tagline:`Alibaba&#39;s performant long context models for agentic and coding tasks.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-embedding`,name:`qwen3-embedding`,created_by:`qwen`,source:`official`,last_updated:`2026-03-21`,family:`qwen3`,capabilities:{streaming:!0},model_type:`embed`},{id:`qwen3-next`,name:`qwen3-next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`The first installment in the Qwen3-Next series with strong performance in terms of both parameter efficiency and inference speed.`,release_date:`2025-09-10`,parameters:80,page_url:`https://ollama.com/library/qwen3-next`,tagline:`The first installment in the Qwen3-Next series with strong performance in terms of both parameter efficiency and inference speed.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-vl`,name:`qwen3-vl`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`The most powerful vision-language model in the Qwen model family to date.`,release_date:`2025-09-22`,parameters:235,page_url:`https://ollama.com/library/qwen3-vl`,tagline:`The most powerful vision-language model in the Qwen model family to date.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`,`image`],output:[`text`]}},{id:`qwen3.5`,name:`qwen3.5`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3.5`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`Qwen 3.5 is a family of open-source multimodal models that delivers exceptional utility and performance.`,release_date:`2026-02-16`,parameters:397,page_url:`https://ollama.com/library/qwen3.5`,tagline:`Qwen 3.5 is a family of open-source multimodal models that delivers exceptional utility and performance.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`,`image`],output:[`text`]}},{id:`rnj-1`,name:`rnj-1`,created_by:`essentialai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models.`,release_date:`2025-12-09`,parameters:8,page_url:`https://ollama.com/library/rnj-1`,tagline:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models.`,context_window:32768,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`translategemma`,name:`translategemma`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0},model_type:`translation`,description:`A new collection of open translation models built on Gemma 3, helping people communicate across 55 languages.`,page_url:`https://ollama.com/library/translategemma`,tagline:`A new collection of open translation models built on Gemma 3, helping people communicate across 55 languages.`,context_window:131072,parameters:4,license:`open-weight`,open_weight:!0,modalities:{input:[`text`,`image`],output:[`text`]}}]},t=e.models.map(t=>({...t,provider:e.id}));exports.models=t,exports.provider=e;
@@ -1,4 +1,4 @@
1
- import { f as ProviderWithModels, n as Model } from "../types.cjs";
1
+ import { m as ProviderWithModels, r as Model } from "../types.cjs";
2
2
 
3
3
  //#region src/providers/ollama.d.ts
4
4
  declare const provider: ProviderWithModels;
@@ -1,4 +1,4 @@
1
- import { f as ProviderWithModels, n as Model } from "../types.mjs";
1
+ import { m as ProviderWithModels, r as Model } from "../types.mjs";
2
2
 
3
3
  //#region src/providers/ollama.d.ts
4
4
  declare const provider: ProviderWithModels;