modelpedia 0.0.2 → 0.0.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +141 -0
- package/dist/index.cjs +1 -101
- package/dist/index.d.cts +10 -2
- package/dist/index.d.mts +10 -2
- package/dist/index.mjs +1 -92
- package/dist/providers/alibaba.cjs +4 -835
- package/dist/providers/alibaba.d.cts +1 -1
- package/dist/providers/alibaba.d.mts +1 -1
- package/dist/providers/alibaba.mjs +4 -833
- package/dist/providers/amazon.cjs +4 -1484
- package/dist/providers/amazon.d.cts +1 -1
- package/dist/providers/amazon.d.mts +1 -1
- package/dist/providers/amazon.mjs +4 -1482
- package/dist/providers/anthropic.cjs +3 -698
- package/dist/providers/anthropic.d.cts +1 -1
- package/dist/providers/anthropic.d.mts +1 -1
- package/dist/providers/anthropic.mjs +3 -696
- package/dist/providers/azure.cjs +6 -1491
- package/dist/providers/azure.d.cts +1 -1
- package/dist/providers/azure.d.mts +1 -1
- package/dist/providers/azure.mjs +6 -1489
- package/dist/providers/baseten.cjs +3 -279
- package/dist/providers/baseten.d.cts +1 -1
- package/dist/providers/baseten.d.mts +1 -1
- package/dist/providers/baseten.mjs +3 -277
- package/dist/providers/cerebras.cjs +4 -145
- package/dist/providers/cerebras.d.cts +1 -1
- package/dist/providers/cerebras.d.mts +1 -1
- package/dist/providers/cerebras.mjs +4 -143
- package/dist/providers/cloudflare-ai-gateway.cjs +4 -0
- package/dist/providers/cloudflare-ai-gateway.d.cts +7 -0
- package/dist/providers/cloudflare-ai-gateway.d.mts +7 -0
- package/dist/providers/cloudflare-ai-gateway.mjs +4 -0
- package/dist/providers/cloudflare-workers-ai.cjs +4 -0
- package/dist/providers/cloudflare-workers-ai.d.cts +7 -0
- package/dist/providers/cloudflare-workers-ai.d.mts +7 -0
- package/dist/providers/cloudflare-workers-ai.mjs +4 -0
- package/dist/providers/cohere.cjs +5 -525
- package/dist/providers/cohere.d.cts +1 -1
- package/dist/providers/cohere.d.mts +1 -1
- package/dist/providers/cohere.mjs +5 -523
- package/dist/providers/cursor.cjs +1 -856
- package/dist/providers/cursor.d.cts +1 -1
- package/dist/providers/cursor.d.mts +1 -1
- package/dist/providers/cursor.mjs +1 -854
- package/dist/providers/deepseek.cjs +3 -64
- package/dist/providers/deepseek.d.cts +1 -1
- package/dist/providers/deepseek.d.mts +1 -1
- package/dist/providers/deepseek.mjs +3 -62
- package/dist/providers/fireworks.cjs +3 -3794
- package/dist/providers/fireworks.d.cts +1 -1
- package/dist/providers/fireworks.d.mts +1 -1
- package/dist/providers/fireworks.mjs +3 -3792
- package/dist/providers/google.cjs +6 -972
- package/dist/providers/google.d.cts +1 -1
- package/dist/providers/google.d.mts +1 -1
- package/dist/providers/google.mjs +6 -970
- package/dist/providers/groq.cjs +3 -353
- package/dist/providers/groq.d.cts +1 -1
- package/dist/providers/groq.d.mts +1 -1
- package/dist/providers/groq.mjs +3 -351
- package/dist/providers/huggingface.cjs +3 -1022
- package/dist/providers/huggingface.d.cts +1 -1
- package/dist/providers/huggingface.d.mts +1 -1
- package/dist/providers/huggingface.mjs +3 -1020
- package/dist/providers/meta.cjs +3 -481
- package/dist/providers/meta.d.cts +1 -1
- package/dist/providers/meta.d.mts +1 -1
- package/dist/providers/meta.mjs +3 -479
- package/dist/providers/minimax.cjs +3 -156
- package/dist/providers/minimax.d.cts +1 -1
- package/dist/providers/minimax.d.mts +1 -1
- package/dist/providers/minimax.mjs +3 -154
- package/dist/providers/mistral.cjs +3 -1594
- package/dist/providers/mistral.d.cts +1 -1
- package/dist/providers/mistral.d.mts +1 -1
- package/dist/providers/mistral.mjs +3 -1592
- package/dist/providers/moonshot.cjs +3 -138
- package/dist/providers/moonshot.d.cts +1 -1
- package/dist/providers/moonshot.d.mts +1 -1
- package/dist/providers/moonshot.mjs +3 -136
- package/dist/providers/nvidia.cjs +3 -2040
- package/dist/providers/nvidia.d.cts +1 -1
- package/dist/providers/nvidia.d.mts +1 -1
- package/dist/providers/nvidia.mjs +3 -2038
- package/dist/providers/ollama.cjs +3 -326
- package/dist/providers/ollama.d.cts +1 -1
- package/dist/providers/ollama.d.mts +1 -1
- package/dist/providers/ollama.mjs +3 -324
- package/dist/providers/openai.cjs +3 -9079
- package/dist/providers/openai.d.cts +1 -1
- package/dist/providers/openai.d.mts +1 -1
- package/dist/providers/openai.mjs +3 -9077
- package/dist/providers/opencode.cjs +3 -367
- package/dist/providers/opencode.d.cts +1 -1
- package/dist/providers/opencode.d.mts +1 -1
- package/dist/providers/opencode.mjs +3 -365
- package/dist/providers/openrouter.cjs +21 -7835
- package/dist/providers/openrouter.d.cts +1 -1
- package/dist/providers/openrouter.d.mts +1 -1
- package/dist/providers/openrouter.mjs +21 -7833
- package/dist/providers/perplexity.cjs +3 -311
- package/dist/providers/perplexity.d.cts +1 -1
- package/dist/providers/perplexity.d.mts +1 -1
- package/dist/providers/perplexity.mjs +3 -309
- package/dist/providers/qwen.cjs +3 -835
- package/dist/providers/qwen.d.cts +1 -1
- package/dist/providers/qwen.d.mts +1 -1
- package/dist/providers/qwen.mjs +3 -833
- package/dist/providers/together.cjs +4 -360
- package/dist/providers/together.d.cts +1 -1
- package/dist/providers/together.d.mts +1 -1
- package/dist/providers/together.mjs +4 -358
- package/dist/providers/vercel.cjs +3 -4037
- package/dist/providers/vercel.d.cts +1 -1
- package/dist/providers/vercel.d.mts +1 -1
- package/dist/providers/vercel.mjs +3 -4035
- package/dist/providers/vertex.cjs +10 -1072
- package/dist/providers/vertex.d.cts +1 -1
- package/dist/providers/vertex.d.mts +1 -1
- package/dist/providers/vertex.mjs +10 -1070
- package/dist/providers/xai.cjs +3 -301
- package/dist/providers/xai.d.cts +1 -1
- package/dist/providers/xai.d.mts +1 -1
- package/dist/providers/xai.mjs +3 -299
- package/dist/providers/zai.cjs +3 -243
- package/dist/providers/zai.d.cts +1 -1
- package/dist/providers/zai.d.mts +1 -1
- package/dist/providers/zai.mjs +3 -241
- package/dist/types.d.cts +80 -8
- package/dist/types.d.mts +80 -8
- package/package.json +3 -3
- package/dist/providers/cloudflare.cjs +0 -826
- package/dist/providers/cloudflare.d.cts +0 -7
- package/dist/providers/cloudflare.d.mts +0 -7
- package/dist/providers/cloudflare.mjs +0 -824
|
@@ -1,324 +1,3 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
"id": "ollama",
|
|
4
|
-
"name": "Ollama",
|
|
5
|
-
"region": "US",
|
|
6
|
-
"url": "https://ollama.com",
|
|
7
|
-
"api_url": "http://localhost:11434/api",
|
|
8
|
-
"docs_url": "https://github.com/ollama/ollama/blob/main/docs/api.md",
|
|
9
|
-
"pricing_url": "https://ollama.com",
|
|
10
|
-
"icon": "<svg xmlns=\"http://www.w3.org/2000/svg\" viewBox=\"0 0 24 24\" fill=\"currentColor\" fill-rule=\"evenodd\">\n <path d=\"M7.905 1.09c.216.085.411.225.588.41.295.306.544.744.734 1.263.191.522.315 1.1.362 1.68a5.054 5.054 0 012.049-.636l.051-.004c.87-.07 1.73.087 2.48.474.101.053.2.11.297.17.05-.569.172-1.134.36-1.644.19-.52.439-.957.733-1.264a1.67 1.67 0 01.589-.41c.257-.1.53-.118.796-.042.401.114.745.368 1.016.737.248.337.434.769.561 1.287.23.934.27 2.163.115 3.645l.053.04.026.019c.757.576 1.284 1.397 1.563 2.35.435 1.487.216 3.155-.534 4.088l-.018.021.002.003c.417.762.67 1.567.724 2.4l.002.03c.064 1.065-.2 2.137-.814 3.19l-.007.01.01.024c.472 1.157.62 2.322.438 3.486l-.006.039a.651.651 0 01-.747.536.648.648 0 01-.54-.742c.167-1.033.01-2.069-.48-3.123a.643.643 0 01.04-.617l.004-.006c.604-.924.854-1.83.8-2.72-.046-.779-.325-1.544-.8-2.273a.644.644 0 01.18-.886l.009-.006c.243-.159.467-.565.58-1.12a4.229 4.229 0 00-.095-1.974c-.205-.7-.58-1.284-1.105-1.683-.595-.454-1.383-.673-2.38-.61a.653.653 0 01-.632-.371c-.314-.665-.772-1.141-1.343-1.436a3.288 3.288 0 00-1.772-.332c-1.245.099-2.343.801-2.67 1.686a.652.652 0 01-.61.425c-1.067.002-1.893.252-2.497.703-.522.39-.878.935-1.066 1.588a4.07 4.07 0 00-.068 1.886c.112.558.331 1.02.582 1.269l.008.007c.212.207.257.53.109.785-.36.622-.629 1.549-.673 2.44-.05 1.018.186 1.902.719 2.536l.016.019a.643.643 0 01.095.69c-.576 1.236-.753 2.252-.562 3.052a.652.652 0 01-1.269.298c-.243-1.018-.078-2.184.473-3.498l.014-.035-.008-.012a4.339 4.339 0 01-.598-1.309l-.005-.019a5.764 5.764 0 01-.177-1.785c.044-.91.278-1.842.622-2.59l.012-.026-.002-.002c-.293-.418-.51-.953-.63-1.545l-.005-.024a5.352 5.352 0 01.093-2.49c.262-.915.777-1.701 1.536-2.269.06-.045.123-.09.186-.132-.159-1.493-.119-2.73.112-3.67.127-.518.314-.95.562-1.287.27-.368.614-.622 1.015-.737.266-.076.54-.059.797.042zm4.116 9.09c.936 0 1.8.313 2.446.855.63.527 1.005 1.235 1.005 1.94 0 .888-.406 1.58-1.133 2.022-.62.375-1.451.557-2.403.557-1.009 0-1.871-.259-2.493-.734-.617-.47-.963-1.13-.963-1.845 0-.707.398-1.417 1.056-1.946.668-.537 1.55-.849 2.485-.849zm0 .896a3.07 3.07 0 00-1.916.65c-.461.37-.722.835-.722 1.25 0 .428.21.829.61 1.134.455.347 1.124.548 1.943.548.799 0 1.473-.147 1.932-.426.463-.28.7-.686.7-1.257 0-.423-.246-.89-.683-1.256-.484-.405-1.14-.643-1.864-.643zm.662 1.21l.004.004c.12.151.095.37-.056.49l-.292.23v.446a.375.375 0 01-.376.373.375.375 0 01-.376-.373v-.46l-.271-.218a.347.347 0 01-.052-.49.353.353 0 01.494-.051l.215.172.22-.174a.353.353 0 01.49.051zm-5.04-1.919c.478 0 .867.39.867.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zm8.706 0c.48 0 .868.39.868.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zM7.44 2.3l-.003.002a.659.659 0 00-.285.238l-.005.006c-.138.189-.258.467-.348.832-.17.692-.216 1.631-.124 2.782.43-.128.899-.208 1.404-.237l.01-.001.019-.034c.046-.082.095-.161.148-.239.123-.771.022-1.692-.253-2.444-.134-.364-.297-.65-.453-.813a.628.628 0 00-.107-.09L7.44 2.3zm9.174.04l-.002.001a.628.628 0 00-.107.09c-.156.163-.32.45-.453.814-.29.794-.387 1.776-.23 2.572l.058.097.008.014h.03a5.184 5.184 0 011.466.212c.086-1.124.038-2.043-.128-2.722-.09-.365-.21-.643-.349-.832l-.004-.006a.659.659 0 00-.285-.239h-.004z\"/>\n</svg>",
|
|
11
|
-
"models": [
|
|
12
|
-
{
|
|
13
|
-
"id": "cogito-2.1",
|
|
14
|
-
"name": "cogito-2.1",
|
|
15
|
-
"created_by": "unknown",
|
|
16
|
-
"source": "official",
|
|
17
|
-
"last_updated": "2026-03-21",
|
|
18
|
-
"capabilities": { "streaming": true }
|
|
19
|
-
},
|
|
20
|
-
{
|
|
21
|
-
"id": "deepseek-v3.1",
|
|
22
|
-
"name": "deepseek-v3.1",
|
|
23
|
-
"created_by": "deepseek",
|
|
24
|
-
"source": "official",
|
|
25
|
-
"last_updated": "2026-03-21",
|
|
26
|
-
"family": "deepseek",
|
|
27
|
-
"capabilities": { "streaming": true }
|
|
28
|
-
},
|
|
29
|
-
{
|
|
30
|
-
"id": "deepseek-v3.2",
|
|
31
|
-
"name": "deepseek-v3.2",
|
|
32
|
-
"created_by": "deepseek",
|
|
33
|
-
"source": "official",
|
|
34
|
-
"last_updated": "2026-03-21",
|
|
35
|
-
"family": "deepseek",
|
|
36
|
-
"capabilities": { "streaming": true }
|
|
37
|
-
},
|
|
38
|
-
{
|
|
39
|
-
"id": "devstral-2",
|
|
40
|
-
"name": "devstral-2",
|
|
41
|
-
"created_by": "mistral",
|
|
42
|
-
"source": "official",
|
|
43
|
-
"last_updated": "2026-03-21",
|
|
44
|
-
"family": "devstral",
|
|
45
|
-
"capabilities": { "streaming": true },
|
|
46
|
-
"model_type": "code"
|
|
47
|
-
},
|
|
48
|
-
{
|
|
49
|
-
"id": "devstral-small-2",
|
|
50
|
-
"name": "devstral-small-2",
|
|
51
|
-
"created_by": "mistral",
|
|
52
|
-
"source": "official",
|
|
53
|
-
"last_updated": "2026-03-21",
|
|
54
|
-
"family": "devstral",
|
|
55
|
-
"capabilities": { "streaming": true },
|
|
56
|
-
"model_type": "code"
|
|
57
|
-
},
|
|
58
|
-
{
|
|
59
|
-
"id": "gemini-3-flash-preview",
|
|
60
|
-
"name": "gemini-3-flash-preview",
|
|
61
|
-
"created_by": "unknown",
|
|
62
|
-
"source": "official",
|
|
63
|
-
"last_updated": "2026-03-21",
|
|
64
|
-
"family": "gemini-3",
|
|
65
|
-
"capabilities": { "streaming": true }
|
|
66
|
-
},
|
|
67
|
-
{
|
|
68
|
-
"id": "gemma3",
|
|
69
|
-
"name": "gemma3",
|
|
70
|
-
"created_by": "google",
|
|
71
|
-
"source": "official",
|
|
72
|
-
"last_updated": "2026-03-21",
|
|
73
|
-
"capabilities": { "streaming": true }
|
|
74
|
-
},
|
|
75
|
-
{
|
|
76
|
-
"id": "glm-4.6",
|
|
77
|
-
"name": "glm-4.6",
|
|
78
|
-
"created_by": "zhipu",
|
|
79
|
-
"source": "official",
|
|
80
|
-
"last_updated": "2026-03-21",
|
|
81
|
-
"capabilities": { "streaming": true }
|
|
82
|
-
},
|
|
83
|
-
{
|
|
84
|
-
"id": "glm-4.7-flash",
|
|
85
|
-
"name": "glm-4.7-flash",
|
|
86
|
-
"created_by": "zhipu",
|
|
87
|
-
"source": "official",
|
|
88
|
-
"last_updated": "2026-03-21",
|
|
89
|
-
"capabilities": { "streaming": true }
|
|
90
|
-
},
|
|
91
|
-
{
|
|
92
|
-
"id": "glm-4.7",
|
|
93
|
-
"name": "glm-4.7",
|
|
94
|
-
"created_by": "zhipu",
|
|
95
|
-
"source": "official",
|
|
96
|
-
"last_updated": "2026-03-21",
|
|
97
|
-
"capabilities": { "streaming": true }
|
|
98
|
-
},
|
|
99
|
-
{
|
|
100
|
-
"id": "glm-5",
|
|
101
|
-
"name": "glm-5",
|
|
102
|
-
"created_by": "zhipu",
|
|
103
|
-
"source": "official",
|
|
104
|
-
"last_updated": "2026-03-21",
|
|
105
|
-
"capabilities": { "streaming": true }
|
|
106
|
-
},
|
|
107
|
-
{
|
|
108
|
-
"id": "glm-ocr",
|
|
109
|
-
"name": "glm-ocr",
|
|
110
|
-
"created_by": "zhipu",
|
|
111
|
-
"source": "official",
|
|
112
|
-
"last_updated": "2026-03-21",
|
|
113
|
-
"capabilities": { "streaming": true }
|
|
114
|
-
},
|
|
115
|
-
{
|
|
116
|
-
"id": "gpt-oss",
|
|
117
|
-
"name": "gpt-oss",
|
|
118
|
-
"created_by": "openai",
|
|
119
|
-
"source": "official",
|
|
120
|
-
"last_updated": "2026-03-21",
|
|
121
|
-
"family": "gpt-oss",
|
|
122
|
-
"capabilities": { "streaming": true }
|
|
123
|
-
},
|
|
124
|
-
{
|
|
125
|
-
"id": "granite4",
|
|
126
|
-
"name": "granite4",
|
|
127
|
-
"created_by": "ibm",
|
|
128
|
-
"source": "official",
|
|
129
|
-
"last_updated": "2026-03-21",
|
|
130
|
-
"capabilities": { "streaming": true }
|
|
131
|
-
},
|
|
132
|
-
{
|
|
133
|
-
"id": "kimi-k2-thinking",
|
|
134
|
-
"name": "kimi-k2-thinking",
|
|
135
|
-
"created_by": "moonshot",
|
|
136
|
-
"source": "official",
|
|
137
|
-
"last_updated": "2026-03-21",
|
|
138
|
-
"capabilities": { "streaming": true }
|
|
139
|
-
},
|
|
140
|
-
{
|
|
141
|
-
"id": "kimi-k2.5",
|
|
142
|
-
"name": "kimi-k2.5",
|
|
143
|
-
"created_by": "moonshot",
|
|
144
|
-
"source": "official",
|
|
145
|
-
"last_updated": "2026-03-21",
|
|
146
|
-
"capabilities": { "streaming": true }
|
|
147
|
-
},
|
|
148
|
-
{
|
|
149
|
-
"id": "kimi-k2",
|
|
150
|
-
"name": "kimi-k2",
|
|
151
|
-
"created_by": "moonshot",
|
|
152
|
-
"source": "official",
|
|
153
|
-
"last_updated": "2026-03-21",
|
|
154
|
-
"capabilities": { "streaming": true }
|
|
155
|
-
},
|
|
156
|
-
{
|
|
157
|
-
"id": "lfm2.5-thinking",
|
|
158
|
-
"name": "lfm2.5-thinking",
|
|
159
|
-
"created_by": "liquid",
|
|
160
|
-
"source": "official",
|
|
161
|
-
"last_updated": "2026-03-21",
|
|
162
|
-
"capabilities": { "streaming": true }
|
|
163
|
-
},
|
|
164
|
-
{
|
|
165
|
-
"id": "lfm2",
|
|
166
|
-
"name": "lfm2",
|
|
167
|
-
"created_by": "liquid",
|
|
168
|
-
"source": "official",
|
|
169
|
-
"last_updated": "2026-03-21",
|
|
170
|
-
"capabilities": { "streaming": true }
|
|
171
|
-
},
|
|
172
|
-
{
|
|
173
|
-
"id": "minimax-m2.1",
|
|
174
|
-
"name": "minimax-m2.1",
|
|
175
|
-
"created_by": "minimax",
|
|
176
|
-
"source": "official",
|
|
177
|
-
"last_updated": "2026-03-21",
|
|
178
|
-
"capabilities": { "streaming": true }
|
|
179
|
-
},
|
|
180
|
-
{
|
|
181
|
-
"id": "minimax-m2.5",
|
|
182
|
-
"name": "minimax-m2.5",
|
|
183
|
-
"created_by": "minimax",
|
|
184
|
-
"source": "official",
|
|
185
|
-
"last_updated": "2026-03-21",
|
|
186
|
-
"capabilities": { "streaming": true }
|
|
187
|
-
},
|
|
188
|
-
{
|
|
189
|
-
"id": "minimax-m2.7",
|
|
190
|
-
"name": "minimax-m2.7",
|
|
191
|
-
"created_by": "minimax",
|
|
192
|
-
"source": "official",
|
|
193
|
-
"last_updated": "2026-03-21",
|
|
194
|
-
"capabilities": { "streaming": true }
|
|
195
|
-
},
|
|
196
|
-
{
|
|
197
|
-
"id": "minimax-m2",
|
|
198
|
-
"name": "minimax-m2",
|
|
199
|
-
"created_by": "minimax",
|
|
200
|
-
"source": "official",
|
|
201
|
-
"last_updated": "2026-03-21",
|
|
202
|
-
"capabilities": { "streaming": true }
|
|
203
|
-
},
|
|
204
|
-
{
|
|
205
|
-
"id": "ministral-3",
|
|
206
|
-
"name": "ministral-3",
|
|
207
|
-
"created_by": "mistral",
|
|
208
|
-
"source": "official",
|
|
209
|
-
"last_updated": "2026-03-21",
|
|
210
|
-
"family": "ministral",
|
|
211
|
-
"capabilities": { "streaming": true }
|
|
212
|
-
},
|
|
213
|
-
{
|
|
214
|
-
"id": "mistral-large-3",
|
|
215
|
-
"name": "mistral-large-3",
|
|
216
|
-
"created_by": "mistral",
|
|
217
|
-
"source": "official",
|
|
218
|
-
"last_updated": "2026-03-21",
|
|
219
|
-
"family": "mistral-large",
|
|
220
|
-
"capabilities": { "streaming": true }
|
|
221
|
-
},
|
|
222
|
-
{
|
|
223
|
-
"id": "nemotron-3-nano",
|
|
224
|
-
"name": "nemotron-3-nano",
|
|
225
|
-
"created_by": "nvidia",
|
|
226
|
-
"source": "official",
|
|
227
|
-
"last_updated": "2026-03-21",
|
|
228
|
-
"capabilities": { "streaming": true }
|
|
229
|
-
},
|
|
230
|
-
{
|
|
231
|
-
"id": "nemotron-3-super",
|
|
232
|
-
"name": "nemotron-3-super",
|
|
233
|
-
"created_by": "nvidia",
|
|
234
|
-
"source": "official",
|
|
235
|
-
"last_updated": "2026-03-21",
|
|
236
|
-
"capabilities": { "streaming": true }
|
|
237
|
-
},
|
|
238
|
-
{
|
|
239
|
-
"id": "nemotron-cascade-2",
|
|
240
|
-
"name": "nemotron-cascade-2",
|
|
241
|
-
"created_by": "nvidia",
|
|
242
|
-
"source": "official",
|
|
243
|
-
"last_updated": "2026-03-21",
|
|
244
|
-
"capabilities": { "streaming": true }
|
|
245
|
-
},
|
|
246
|
-
{
|
|
247
|
-
"id": "qwen3-coder-next",
|
|
248
|
-
"name": "qwen3-coder-next",
|
|
249
|
-
"created_by": "qwen",
|
|
250
|
-
"source": "official",
|
|
251
|
-
"last_updated": "2026-03-21",
|
|
252
|
-
"family": "qwen3",
|
|
253
|
-
"capabilities": { "streaming": true }
|
|
254
|
-
},
|
|
255
|
-
{
|
|
256
|
-
"id": "qwen3-coder",
|
|
257
|
-
"name": "qwen3-coder",
|
|
258
|
-
"created_by": "qwen",
|
|
259
|
-
"source": "official",
|
|
260
|
-
"last_updated": "2026-03-21",
|
|
261
|
-
"family": "qwen3",
|
|
262
|
-
"capabilities": { "streaming": true }
|
|
263
|
-
},
|
|
264
|
-
{
|
|
265
|
-
"id": "qwen3-embedding",
|
|
266
|
-
"name": "qwen3-embedding",
|
|
267
|
-
"created_by": "qwen",
|
|
268
|
-
"source": "official",
|
|
269
|
-
"last_updated": "2026-03-21",
|
|
270
|
-
"family": "qwen3",
|
|
271
|
-
"capabilities": { "streaming": true },
|
|
272
|
-
"model_type": "embed"
|
|
273
|
-
},
|
|
274
|
-
{
|
|
275
|
-
"id": "qwen3-next",
|
|
276
|
-
"name": "qwen3-next",
|
|
277
|
-
"created_by": "qwen",
|
|
278
|
-
"source": "official",
|
|
279
|
-
"last_updated": "2026-03-21",
|
|
280
|
-
"family": "qwen3",
|
|
281
|
-
"capabilities": { "streaming": true }
|
|
282
|
-
},
|
|
283
|
-
{
|
|
284
|
-
"id": "qwen3-vl",
|
|
285
|
-
"name": "qwen3-vl",
|
|
286
|
-
"created_by": "qwen",
|
|
287
|
-
"source": "official",
|
|
288
|
-
"last_updated": "2026-03-21",
|
|
289
|
-
"family": "qwen3",
|
|
290
|
-
"capabilities": { "streaming": true }
|
|
291
|
-
},
|
|
292
|
-
{
|
|
293
|
-
"id": "qwen3.5",
|
|
294
|
-
"name": "qwen3.5",
|
|
295
|
-
"created_by": "qwen",
|
|
296
|
-
"source": "official",
|
|
297
|
-
"last_updated": "2026-03-21",
|
|
298
|
-
"family": "qwen3.5",
|
|
299
|
-
"capabilities": { "streaming": true }
|
|
300
|
-
},
|
|
301
|
-
{
|
|
302
|
-
"id": "rnj-1",
|
|
303
|
-
"name": "rnj-1",
|
|
304
|
-
"created_by": "essentialai",
|
|
305
|
-
"source": "official",
|
|
306
|
-
"last_updated": "2026-03-21",
|
|
307
|
-
"capabilities": { "streaming": true }
|
|
308
|
-
},
|
|
309
|
-
{
|
|
310
|
-
"id": "translategemma",
|
|
311
|
-
"name": "translategemma",
|
|
312
|
-
"created_by": "google",
|
|
313
|
-
"source": "official",
|
|
314
|
-
"last_updated": "2026-03-21",
|
|
315
|
-
"capabilities": { "streaming": true }
|
|
316
|
-
}
|
|
317
|
-
]
|
|
318
|
-
};
|
|
319
|
-
const models = provider.models.map((m) => ({
|
|
320
|
-
...m,
|
|
321
|
-
provider: provider.id
|
|
322
|
-
}));
|
|
323
|
-
//#endregion
|
|
324
|
-
export { models, provider };
|
|
1
|
+
const e={id:`ollama`,name:`Ollama`,region:`US`,headquarters:`Palo Alto, CA`,founded:2023,url:`https://ollama.com`,api_url:`http://localhost:11434/api`,docs_url:`https://github.com/ollama/ollama/blob/main/docs/api.md`,pricing_url:`https://ollama.com`,description:`Run open-source LLMs locally on your machine with a simple CLI.`,type:`aggregator`,openai_compatible:!0,free_tier:!0,github_url:`https://github.com/ollama/ollama`,models_url:`https://ollama.com/library`,twitter_url:`https://x.com/ollama`,discord_url:`https://discord.com/invite/ollama`,blog_url:`https://ollama.com/blog`,support_url:`https://github.com/ollama/ollama`,icon:`<svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24" fill="currentColor" fill-rule="evenodd">
|
|
2
|
+
<path d="M7.905 1.09c.216.085.411.225.588.41.295.306.544.744.734 1.263.191.522.315 1.1.362 1.68a5.054 5.054 0 012.049-.636l.051-.004c.87-.07 1.73.087 2.48.474.101.053.2.11.297.17.05-.569.172-1.134.36-1.644.19-.52.439-.957.733-1.264a1.67 1.67 0 01.589-.41c.257-.1.53-.118.796-.042.401.114.745.368 1.016.737.248.337.434.769.561 1.287.23.934.27 2.163.115 3.645l.053.04.026.019c.757.576 1.284 1.397 1.563 2.35.435 1.487.216 3.155-.534 4.088l-.018.021.002.003c.417.762.67 1.567.724 2.4l.002.03c.064 1.065-.2 2.137-.814 3.19l-.007.01.01.024c.472 1.157.62 2.322.438 3.486l-.006.039a.651.651 0 01-.747.536.648.648 0 01-.54-.742c.167-1.033.01-2.069-.48-3.123a.643.643 0 01.04-.617l.004-.006c.604-.924.854-1.83.8-2.72-.046-.779-.325-1.544-.8-2.273a.644.644 0 01.18-.886l.009-.006c.243-.159.467-.565.58-1.12a4.229 4.229 0 00-.095-1.974c-.205-.7-.58-1.284-1.105-1.683-.595-.454-1.383-.673-2.38-.61a.653.653 0 01-.632-.371c-.314-.665-.772-1.141-1.343-1.436a3.288 3.288 0 00-1.772-.332c-1.245.099-2.343.801-2.67 1.686a.652.652 0 01-.61.425c-1.067.002-1.893.252-2.497.703-.522.39-.878.935-1.066 1.588a4.07 4.07 0 00-.068 1.886c.112.558.331 1.02.582 1.269l.008.007c.212.207.257.53.109.785-.36.622-.629 1.549-.673 2.44-.05 1.018.186 1.902.719 2.536l.016.019a.643.643 0 01.095.69c-.576 1.236-.753 2.252-.562 3.052a.652.652 0 01-1.269.298c-.243-1.018-.078-2.184.473-3.498l.014-.035-.008-.012a4.339 4.339 0 01-.598-1.309l-.005-.019a5.764 5.764 0 01-.177-1.785c.044-.91.278-1.842.622-2.59l.012-.026-.002-.002c-.293-.418-.51-.953-.63-1.545l-.005-.024a5.352 5.352 0 01.093-2.49c.262-.915.777-1.701 1.536-2.269.06-.045.123-.09.186-.132-.159-1.493-.119-2.73.112-3.67.127-.518.314-.95.562-1.287.27-.368.614-.622 1.015-.737.266-.076.54-.059.797.042zm4.116 9.09c.936 0 1.8.313 2.446.855.63.527 1.005 1.235 1.005 1.94 0 .888-.406 1.58-1.133 2.022-.62.375-1.451.557-2.403.557-1.009 0-1.871-.259-2.493-.734-.617-.47-.963-1.13-.963-1.845 0-.707.398-1.417 1.056-1.946.668-.537 1.55-.849 2.485-.849zm0 .896a3.07 3.07 0 00-1.916.65c-.461.37-.722.835-.722 1.25 0 .428.21.829.61 1.134.455.347 1.124.548 1.943.548.799 0 1.473-.147 1.932-.426.463-.28.7-.686.7-1.257 0-.423-.246-.89-.683-1.256-.484-.405-1.14-.643-1.864-.643zm.662 1.21l.004.004c.12.151.095.37-.056.49l-.292.23v.446a.375.375 0 01-.376.373.375.375 0 01-.376-.373v-.46l-.271-.218a.347.347 0 01-.052-.49.353.353 0 01.494-.051l.215.172.22-.174a.353.353 0 01.49.051zm-5.04-1.919c.478 0 .867.39.867.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zm8.706 0c.48 0 .868.39.868.871a.87.87 0 01-.868.871.87.87 0 01-.867-.87.87.87 0 01.867-.872zM7.44 2.3l-.003.002a.659.659 0 00-.285.238l-.005.006c-.138.189-.258.467-.348.832-.17.692-.216 1.631-.124 2.782.43-.128.899-.208 1.404-.237l.01-.001.019-.034c.046-.082.095-.161.148-.239.123-.771.022-1.692-.253-2.444-.134-.364-.297-.65-.453-.813a.628.628 0 00-.107-.09L7.44 2.3zm9.174.04l-.002.001a.628.628 0 00-.107.09c-.156.163-.32.45-.453.814-.29.794-.387 1.776-.23 2.572l.058.097.008.014h.03a5.184 5.184 0 011.466.212c.086-1.124.038-2.043-.128-2.722-.09-.365-.21-.643-.349-.832l-.004-.006a.659.659 0 00-.285-.239h-.004z"/>
|
|
3
|
+
</svg>`,models:[{id:`cogito-2.1`,name:`cogito-2.1`,created_by:`deepcogito`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},description:`The Cogito v2.1 LLMs are instruction tuned generative models. All models are released under MIT license for commercial use.`,release_date:`2025-11-19`,parameters:671,page_url:`https://ollama.com/library/cogito-2.1`,tagline:`The Cogito v2.1 LLMs are instruction tuned generative models.`,context_window:163840,license:`mit`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`deepseek-v3.1`,name:`deepseek-v3.1`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,family:`deepseek`,capabilities:{streaming:!0},description:`DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.`,release_date:`2025-11-20`,parameters:671,page_url:`https://ollama.com/library/deepseek-v3.1`,tagline:`DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.`,context_window:163840,license:`open-weight`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`deepseek-v3.2`,name:`deepseek-v3.2`,created_by:`deepseek`,source:`official`,last_updated:`2026-03-23`,family:`deepseek`,capabilities:{streaming:!0},description:`DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.`,release_date:`2025-12-02`,page_url:`https://ollama.com/library/deepseek-v3.2`,tagline:`DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.`,context_window:163840,license:`open-weight`,open_weight:!0,model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`devstral-2`,name:`devstral-2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`devstral`,capabilities:{streaming:!0},model_type:`code`,description:`123B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,release_date:`2025-12-08`,parameters:123,page_url:`https://ollama.com/library/devstral-2`,tagline:`123B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,context_window:262144,license:`mit`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`devstral-small-2`,name:`devstral-small-2`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`devstral`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`code`,description:`24B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,release_date:`2025-12-09`,parameters:24,page_url:`https://ollama.com/library/devstral-small-2`,tagline:`24B model that excels at using tools to explore codebases, editing multiple files and power software engineering agents.`,context_window:393216,license:`apache-2.0`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`gemini-3-flash-preview`,name:`gemini-3-flash-preview`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,family:`gemini-3`,capabilities:{streaming:!0,tool_call:!0,structured_output:!0,reasoning:!0,batch:!0,vision:!0},model_type:`chat`,description:`Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost.`,release_date:`2025-12-17`,page_url:`https://ollama.com/library/gemini-3-flash-preview`,tagline:`Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost.`,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]},context_window:1048576,max_output_tokens:65536,knowledge_cutoff:`2025-01`},{id:`gemma3`,name:`gemma3`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,description:`The current, most capable model that runs on a single GPU.`,release_date:`2025-03-12`,page_url:`https://ollama.com/library/gemma3`,tagline:`The current, most capable model that runs on a single GPU.`,context_window:131072,license:`gemma`,parameters:4,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.6`,name:`glm-4.6`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},family:`glm-4.6`,model_type:`chat`,description:`Advanced agentic, reasoning and coding capabilities.`,release_date:`2025-09-29`,page_url:`https://ollama.com/library/glm-4.6`,tagline:`Advanced agentic, reasoning and coding capabilities.`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.7-flash`,name:`glm-4.7-flash`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},family:`glm-4.7`,model_type:`chat`,description:`As the strongest model in the 30B class, GLM-4.7-Flash offers a new option for lightweight deployment that balances performance and efficiency.`,page_url:`https://ollama.com/library/glm-4.7-flash`,tagline:`As the strongest model in the 30B class, GLM-4.7-Flash offers a new option for lightweight deployment that balances performance and efficiency.`,context_window:202752,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-4.7`,name:`glm-4.7`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},family:`glm-4.7`,model_type:`chat`,description:`Advancing the Coding Capability`,release_date:`2025-12-22`,page_url:`https://ollama.com/library/glm-4.7`,tagline:`Advancing the Coding Capability`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-5`,name:`glm-5`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,family:`glm-5`,description:`A strong reasoning and agentic model from Z.ai with 744B total parameters (40B active), built for complex systems engineering and long-horizon tasks.`,release_date:`2026-02-11`,page_url:`https://ollama.com/library/glm-5`,tagline:`A strong reasoning and agentic model from Z.ai with 744B total parameters (40B active), built for complex systems engineering and long-horizon tasks.`,context_window:202752,parameters:744,license:`open-weight`,active_parameters:40,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`glm-ocr`,name:`glm-ocr`,created_by:`zhipu`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`chat`,family:`glm`,description:`GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.`,page_url:`https://ollama.com/library/glm-ocr`,tagline:`GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.`,context_window:131072,parameters:.9,license:`open-weight`,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`gpt-oss`,name:`gpt-oss`,created_by:`openai`,source:`official`,last_updated:`2026-03-23`,family:`gpt-oss`,capabilities:{streaming:!0},model_type:`chat`,description:`OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,release_date:`2025-08-05`,page_url:`https://ollama.com/library/gpt-oss`,tagline:`OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases.`,context_window:131072,license:`apache-2.0`,parameters:20,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`granite4`,name:`granite4`,created_by:`ibm`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`Granite 4 features improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.`,page_url:`https://ollama.com/library/granite4`,tagline:`Granite 4 features improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.`,context_window:131072,license:`apache-2.0`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2-thinking`,name:`kimi-k2-thinking`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,family:`kimi-k2`,description:`Kimi K2 Thinking, Moonshot AI's best open-source thinking model.`,release_date:`2025-11-06`,page_url:`https://ollama.com/library/kimi-k2-thinking`,tagline:`Kimi K2 Thinking, Moonshot AI's best open-source thinking model.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2.5`,name:`kimi-k2.5`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},family:`kimi-k2.5`,model_type:`chat`,description:`Kimi K2.5 is an open-source, native multimodal agentic model that seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as conversational and agentic paradigms.`,release_date:`2026-01-26`,page_url:`https://ollama.com/library/kimi-k2.5`,tagline:`Kimi K2.5 is an open-source, native multimodal agentic model that seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as con`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`kimi-k2`,name:`kimi-k2`,created_by:`moonshot`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0},model_type:`chat`,family:`kimi-k2`,description:`A state-of-the-art mixture-of-experts (MoE) language model. Kimi K2-Instruct-0905 demonstrates significant improvements in performance on public benchmarks and real-world coding agent tasks.`,release_date:`2025-09-05`,parameters:1e3,page_url:`https://ollama.com/library/kimi-k2`,tagline:`A state-of-the-art mixture-of-experts (MoE) language model.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`lfm2.5-thinking`,name:`lfm2.5-thinking`,created_by:`liquid`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`LFM2.5 is a new family of hybrid models designed for on-device deployment.`,page_url:`https://ollama.com/library/lfm2.5-thinking`,tagline:`LFM2.5 is a new family of hybrid models designed for on-device deployment.`,context_window:128e3,parameters:1.2,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`lfm2`,name:`lfm2`,created_by:`liquid`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`LFM2 is a family of hybrid models designed for on-device deployment. LFM2-24B-A2B is the largest model in the family, scaling the architecture to 24 billion parameters while keeping inference efficient.`,page_url:`https://ollama.com/library/lfm2`,tagline:`LFM2 is a family of hybrid models designed for on-device deployment.`,context_window:32768,parameters:24,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.1`,name:`minimax-m2.1`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`Exceptional multilingual capabilities to elevate code engineering`,release_date:`2025-12-20`,page_url:`https://ollama.com/library/minimax-m2.1`,tagline:`Exceptional multilingual capabilities to elevate code engineering`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.5`,name:`minimax-m2.5`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`MiniMax-M2.5 is a state-of-the-art large language model designed for real-world productivity and coding tasks.`,release_date:`2026-02-12`,page_url:`https://ollama.com/library/minimax-m2.5`,tagline:`MiniMax-M2.5 is a state-of-the-art large language model designed for real-world productivity and coding tasks.`,context_window:202752,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2.7`,name:`minimax-m2.7`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},family:`minimax`,model_type:`chat`,description:`MiniMax's M2-series model for coding, agentic workflows, and professional productivity.`,release_date:`2026-03-18`,page_url:`https://ollama.com/library/minimax-m2.7`,tagline:`MiniMax's M2-series model for coding, agentic workflows, and professional productivity.`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`minimax-m2`,name:`minimax-m2`,created_by:`minimax`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},model_type:`chat`,family:`minimax`,description:`MiniMax M2 is a high-efficiency large language model built for coding and agentic workflows.`,release_date:`2025-10-27`,page_url:`https://ollama.com/library/minimax-m2`,tagline:`MiniMax M2 is a high-efficiency large language model built for coding and agentic workflows.`,context_window:204800,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`ministral-3`,name:`ministral-3`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`ministral`,capabilities:{streaming:!0,vision:!0,tool_call:!0},model_type:`chat`,description:`The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.`,release_date:`2025-12-02`,page_url:`https://ollama.com/library/ministral-3`,tagline:`The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.`,context_window:262144,license:`apache-2.0`,parameters:8,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`,`image`],output:[`text`]}},{id:`mistral-large-3`,name:`mistral-large-3`,created_by:`mistral`,source:`official`,last_updated:`2026-03-23`,family:`mistral-large`,capabilities:{streaming:!0},model_type:`chat`,description:`A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.`,release_date:`2025-12-02`,parameters:675,page_url:`https://ollama.com/library/mistral-large-3`,tagline:`A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.`,context_window:262144,license:`apache-2.0`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-3-nano`,name:`nemotron-3-nano`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`Nemotron-3-Nano is a new Standard for Efficient, Open, and Intelligent Agentic Models, now updated with a 4B parameter count model.`,release_date:`2025-12-15`,parameters:30,page_url:`https://ollama.com/library/nemotron-3-nano`,tagline:`Nemotron-3-Nano is a new Standard for Efficient, Open, and Intelligent Agentic Models, now updated with a 4B parameter count model.`,context_window:262144,license:`nvidia-open`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-3-super`,name:`nemotron-3-super`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.`,release_date:`2026-03-11`,page_url:`https://ollama.com/library/nemotron-3-super`,tagline:`NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.`,context_window:262144,parameters:120,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`nemotron-cascade-2`,name:`nemotron-cascade-2`,created_by:`nvidia`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`An open 30B MoE model from NVIDIA with 3B activated parameters that delivers strong reasoning and agentic capabilities.`,page_url:`https://ollama.com/library/nemotron-cascade-2`,tagline:`An open 30B MoE model from NVIDIA with 3B activated parameters that delivers strong reasoning and agentic capabilities.`,context_window:262144,parameters:30,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-coder-next`,name:`qwen3-coder-next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,tool_call:!0},model_type:`code`,description:`Qwen3-Coder-Next is a coding-focused language model from Alibaba's Qwen team, optimized for agentic coding workflows and local development.`,release_date:`2025-02-04`,page_url:`https://ollama.com/library/qwen3-coder-next`,tagline:`Qwen3-Coder-Next is a coding-focused language model from Alibaba's Qwen team, optimized for agentic coding workflows and local development.`,context_window:262144,license:`open-weight`,parameters:80,active_parameters:3,open_weight:!0,tools:[`function_calling`],modalities:{input:[`text`],output:[`text`]},max_output_tokens:65536},{id:`qwen3-coder`,name:`qwen3-coder`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0},model_type:`code`,description:`Alibaba's performant long context models for agentic and coding tasks.`,release_date:`2025-07-22`,parameters:480,page_url:`https://ollama.com/library/qwen3-coder`,tagline:`Alibaba's performant long context models for agentic and coding tasks.`,context_window:262144,license:`open-weight`,open_weight:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-embedding`,name:`qwen3-embedding`,created_by:`qwen`,source:`official`,last_updated:`2026-03-21`,family:`qwen3`,capabilities:{streaming:!0},model_type:`embed`},{id:`qwen3-next`,name:`qwen3-next`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`The first installment in the Qwen3-Next series with strong performance in terms of both parameter efficiency and inference speed.`,release_date:`2025-09-10`,parameters:80,page_url:`https://ollama.com/library/qwen3-next`,tagline:`The first installment in the Qwen3-Next series with strong performance in terms of both parameter efficiency and inference speed.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`],output:[`text`]}},{id:`qwen3-vl`,name:`qwen3-vl`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`The most powerful vision-language model in the Qwen model family to date.`,release_date:`2025-09-22`,parameters:235,page_url:`https://ollama.com/library/qwen3-vl`,tagline:`The most powerful vision-language model in the Qwen model family to date.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`,`image`],output:[`text`]}},{id:`qwen3.5`,name:`qwen3.5`,created_by:`qwen`,source:`official`,last_updated:`2026-03-23`,family:`qwen3.5`,capabilities:{streaming:!0,vision:!0,tool_call:!0,reasoning:!0},model_type:`chat`,description:`Qwen 3.5 is a family of open-source multimodal models that delivers exceptional utility and performance.`,release_date:`2026-02-16`,parameters:397,page_url:`https://ollama.com/library/qwen3.5`,tagline:`Qwen 3.5 is a family of open-source multimodal models that delivers exceptional utility and performance.`,context_window:262144,license:`open-weight`,open_weight:!0,tools:[`function_calling`],reasoning_tokens:!0,modalities:{input:[`text`,`image`],output:[`text`]}},{id:`rnj-1`,name:`rnj-1`,created_by:`essentialai`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,tool_call:!0},description:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models.`,release_date:`2025-12-09`,parameters:8,page_url:`https://ollama.com/library/rnj-1`,tagline:`Rnj-1 is a family of 8B parameter open-weight, dense models trained from scratch by Essential AI, optimized for code and STEM with capabilities on par with SOTA open-weight models.`,context_window:32768,license:`open-weight`,open_weight:!0,tools:[`function_calling`],model_type:`chat`,modalities:{input:[`text`],output:[`text`]}},{id:`translategemma`,name:`translategemma`,created_by:`google`,source:`official`,last_updated:`2026-03-23`,capabilities:{streaming:!0,vision:!0},model_type:`translation`,description:`A new collection of open translation models built on Gemma 3, helping people communicate across 55 languages.`,page_url:`https://ollama.com/library/translategemma`,tagline:`A new collection of open translation models built on Gemma 3, helping people communicate across 55 languages.`,context_window:131072,parameters:4,license:`open-weight`,open_weight:!0,modalities:{input:[`text`,`image`],output:[`text`]}}]},t=e.models.map(t=>({...t,provider:e.id}));export{t as models,e as provider};
|