@anolilab/ai-model-registry 1.4.0 → 2.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (147) hide show
  1. package/CHANGELOG.md +41 -0
  2. package/LICENSE.md +8 -13
  3. package/MIGRATION.md +474 -0
  4. package/README.md +94 -49
  5. package/dist/index.d.ts +16 -16
  6. package/dist/index.js +1 -3
  7. package/dist/packem_chunks/alibaba.js +1 -0
  8. package/dist/packem_chunks/amazon-bedrock.js +1 -0
  9. package/dist/packem_chunks/anthropic.js +1 -0
  10. package/dist/packem_chunks/api.js +75 -0
  11. package/dist/packem_chunks/azure-open-ai.js +1 -0
  12. package/dist/packem_chunks/cerebras.js +1 -0
  13. package/dist/packem_chunks/chutes.js +1 -0
  14. package/dist/packem_chunks/cloudflare.js +3 -0
  15. package/dist/packem_chunks/deep-infra.js +1 -0
  16. package/dist/packem_chunks/deep-seek.js +1 -0
  17. package/dist/packem_chunks/fireworks-ai.js +1 -0
  18. package/dist/packem_chunks/git-hub-copilot.js +1 -0
  19. package/dist/packem_chunks/git-hub-models.js +1 -0
  20. package/dist/packem_chunks/google-partner.js +1 -0
  21. package/dist/packem_chunks/google-vertex.js +1 -0
  22. package/dist/packem_chunks/google.js +1 -0
  23. package/dist/packem_chunks/groq.js +1 -0
  24. package/dist/packem_chunks/hugging-face.js +1 -0
  25. package/dist/packem_chunks/inception.js +1 -0
  26. package/dist/packem_chunks/inference.js +1 -0
  27. package/dist/packem_chunks/meta.js +1 -0
  28. package/dist/packem_chunks/mistral.js +1 -0
  29. package/dist/packem_chunks/model-scope.js +1 -0
  30. package/dist/packem_chunks/morph.js +1 -0
  31. package/dist/packem_chunks/open-ai.js +1 -0
  32. package/dist/packem_chunks/open-router.js +1 -0
  33. package/dist/packem_chunks/providers.js +1 -0
  34. package/dist/packem_chunks/requesty.js +73 -0
  35. package/dist/packem_chunks/together-ai.js +1 -0
  36. package/dist/packem_chunks/upstage.js +1 -0
  37. package/dist/packem_chunks/v0.js +1 -0
  38. package/dist/packem_chunks/venice.js +1 -0
  39. package/dist/packem_chunks/vercel.js +1 -0
  40. package/dist/packem_chunks/weights-_-biases.js +1 -0
  41. package/dist/packem_chunks/xai.js +1 -0
  42. package/dist/schema.d.ts +1 -1
  43. package/dist/schema.js +1 -1
  44. package/dist/types/alibaba.d.ts +6 -0
  45. package/dist/types/alibaba.js +0 -0
  46. package/dist/types/amazon-bedrock.d.ts +6 -0
  47. package/dist/types/amazon-bedrock.js +0 -0
  48. package/dist/types/anthropic.d.ts +6 -0
  49. package/dist/types/anthropic.js +0 -0
  50. package/dist/types/azure-open-ai.d.ts +6 -0
  51. package/dist/types/azure-open-ai.js +0 -0
  52. package/dist/types/cerebras.d.ts +6 -0
  53. package/dist/types/cerebras.js +0 -0
  54. package/dist/types/chutes.d.ts +6 -0
  55. package/dist/types/chutes.js +0 -0
  56. package/dist/types/cloudflare.d.ts +6 -0
  57. package/dist/types/cloudflare.js +0 -0
  58. package/dist/types/deep-infra.d.ts +6 -0
  59. package/dist/types/deep-infra.js +0 -0
  60. package/dist/types/deep-seek.d.ts +6 -0
  61. package/dist/types/deep-seek.js +0 -0
  62. package/dist/types/fireworks-ai.d.ts +6 -0
  63. package/dist/types/fireworks-ai.js +0 -0
  64. package/dist/types/git-hub-copilot.d.ts +6 -0
  65. package/dist/types/git-hub-copilot.js +0 -0
  66. package/dist/types/git-hub-models.d.ts +6 -0
  67. package/dist/types/git-hub-models.js +0 -0
  68. package/dist/types/google-partner.d.ts +6 -0
  69. package/dist/types/google-partner.js +0 -0
  70. package/dist/types/google-vertex.d.ts +6 -0
  71. package/dist/types/google-vertex.js +0 -0
  72. package/dist/types/google.d.ts +6 -0
  73. package/dist/types/google.js +0 -0
  74. package/dist/types/groq.d.ts +6 -0
  75. package/dist/types/groq.js +0 -0
  76. package/dist/types/hugging-face.d.ts +6 -0
  77. package/dist/types/hugging-face.js +0 -0
  78. package/dist/types/inception.d.ts +6 -0
  79. package/dist/types/inception.js +0 -0
  80. package/dist/types/inference.d.ts +6 -0
  81. package/dist/types/inference.js +0 -0
  82. package/dist/types/meta.d.ts +6 -0
  83. package/dist/types/meta.js +0 -0
  84. package/dist/types/mistral.d.ts +6 -0
  85. package/dist/types/mistral.js +0 -0
  86. package/dist/types/model-scope.d.ts +6 -0
  87. package/dist/types/model-scope.js +0 -0
  88. package/dist/types/morph.d.ts +6 -0
  89. package/dist/types/morph.js +0 -0
  90. package/dist/types/open-ai.d.ts +6 -0
  91. package/dist/types/open-ai.js +0 -0
  92. package/dist/types/open-router.d.ts +6 -0
  93. package/dist/types/open-router.js +0 -0
  94. package/dist/types/providers.d.ts +11 -0
  95. package/dist/types/providers.js +1 -0
  96. package/dist/types/requesty.d.ts +6 -0
  97. package/dist/types/requesty.js +0 -0
  98. package/dist/types/together-ai.d.ts +6 -0
  99. package/dist/types/together-ai.js +0 -0
  100. package/dist/types/upstage.d.ts +6 -0
  101. package/dist/types/upstage.js +0 -0
  102. package/dist/types/v0.d.ts +6 -0
  103. package/dist/types/v0.js +0 -0
  104. package/dist/types/venice.d.ts +6 -0
  105. package/dist/types/venice.js +0 -0
  106. package/dist/types/vercel.d.ts +6 -0
  107. package/dist/types/vercel.js +0 -0
  108. package/dist/types/weights-_-biases.d.ts +6 -0
  109. package/dist/types/weights-_-biases.js +0 -0
  110. package/dist/types/xai.d.ts +6 -0
  111. package/dist/types/xai.js +0 -0
  112. package/package.json +6 -1
  113. package/public/alibaba.json +326 -0
  114. package/public/amazon-bedrock.json +2141 -0
  115. package/public/anthropic.json +396 -0
  116. package/public/api.json +13504 -3
  117. package/public/azure-open-ai.json +332 -0
  118. package/public/cerebras.json +288 -0
  119. package/public/chutes.json +3220 -0
  120. package/public/cloudflare.json +3094 -0
  121. package/public/deep-infra.json +956 -0
  122. package/public/deep-seek.json +46 -0
  123. package/public/fireworks-ai.json +14486 -0
  124. package/public/git-hub-copilot.json +676 -0
  125. package/public/git-hub-models.json +256 -0
  126. package/public/google-partner.json +536 -0
  127. package/public/google-vertex.json +1376 -0
  128. package/public/google.json +536 -0
  129. package/public/groq.json +882 -0
  130. package/public/hugging-face.json +1096 -0
  131. package/public/inception.json +81 -0
  132. package/public/inference.json +291 -0
  133. package/public/meta.json +151 -0
  134. package/public/mistral.json +2181 -0
  135. package/public/model-scope.json +37839 -0
  136. package/public/morph.json +46 -0
  137. package/public/open-ai.json +606 -0
  138. package/public/open-router.json +15341 -0
  139. package/public/providers.json +43 -0
  140. package/public/requesty.json +13757 -0
  141. package/public/together-ai.json +641 -0
  142. package/public/upstage.json +606 -0
  143. package/public/v0.json +431 -0
  144. package/public/venice.json +501 -0
  145. package/public/vercel.json +4071 -0
  146. package/public/weights-&-biases.json +551 -0
  147. package/public/xai.json +256 -0
@@ -0,0 +1,3094 @@
1
+ {
2
+ "metadata": {
3
+ "description": "AI Models API - Models from Cloudflare",
4
+ "lastUpdated": "2026-01-10T23:24:47.798Z",
5
+ "provider": "Cloudflare",
6
+ "totalModels": 79,
7
+ "version": "0.0.0-development"
8
+ },
9
+ "models": [
10
+ {
11
+ "attachment": false,
12
+ "audioGeneration": true,
13
+ "cost": {
14
+ "input": null,
15
+ "inputCacheHit": null,
16
+ "output": null
17
+ },
18
+ "description": "Aura is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",
19
+ "id": "aura-1",
20
+ "knowledge": null,
21
+ "lastUpdated": null,
22
+ "limit": {
23
+ "context": null,
24
+ "output": null
25
+ },
26
+ "modalities": {
27
+ "input": ["audio"],
28
+ "output": ["text"]
29
+ },
30
+ "name": "aura-1",
31
+ "openWeights": false,
32
+ "provider": "Cloudflare",
33
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
34
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
35
+ "providerId": "cloudflare",
36
+ "providerNpm": "@ai-sdk/cloudflare",
37
+ "reasoning": false,
38
+ "releaseDate": null,
39
+ "streamingSupported": true,
40
+ "structuredOutputs": true,
41
+ "supportsStructuredOutput": true,
42
+ "supportsTools": true,
43
+ "temperature": true,
44
+ "toolCall": true,
45
+ "version": null,
46
+ "vision": true,
47
+ "icon": "cloudflare"
48
+ },
49
+ {
50
+ "attachment": false,
51
+ "audioGeneration": true,
52
+ "cost": {
53
+ "input": null,
54
+ "inputCacheHit": null,
55
+ "output": null
56
+ },
57
+ "description": "Aura-2 is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",
58
+ "id": "aura-2-en",
59
+ "knowledge": null,
60
+ "lastUpdated": null,
61
+ "limit": {
62
+ "context": null,
63
+ "output": null
64
+ },
65
+ "modalities": {
66
+ "input": ["audio"],
67
+ "output": ["text"]
68
+ },
69
+ "name": "aura-2-en",
70
+ "openWeights": false,
71
+ "provider": "Cloudflare",
72
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
73
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
74
+ "providerId": "cloudflare",
75
+ "providerNpm": "@ai-sdk/cloudflare",
76
+ "reasoning": false,
77
+ "releaseDate": null,
78
+ "streamingSupported": true,
79
+ "structuredOutputs": true,
80
+ "supportsStructuredOutput": true,
81
+ "supportsTools": true,
82
+ "temperature": true,
83
+ "toolCall": true,
84
+ "version": null,
85
+ "vision": true,
86
+ "icon": "cloudflare"
87
+ },
88
+ {
89
+ "attachment": false,
90
+ "audioGeneration": true,
91
+ "cost": {
92
+ "input": null,
93
+ "inputCacheHit": null,
94
+ "output": null
95
+ },
96
+ "description": "Aura-2 is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",
97
+ "id": "aura-2-es",
98
+ "knowledge": null,
99
+ "lastUpdated": null,
100
+ "limit": {
101
+ "context": null,
102
+ "output": null
103
+ },
104
+ "modalities": {
105
+ "input": ["audio"],
106
+ "output": ["text"]
107
+ },
108
+ "name": "aura-2-es",
109
+ "openWeights": false,
110
+ "provider": "Cloudflare",
111
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
112
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
113
+ "providerId": "cloudflare",
114
+ "providerNpm": "@ai-sdk/cloudflare",
115
+ "reasoning": false,
116
+ "releaseDate": null,
117
+ "streamingSupported": true,
118
+ "structuredOutputs": true,
119
+ "supportsStructuredOutput": true,
120
+ "supportsTools": true,
121
+ "temperature": true,
122
+ "toolCall": true,
123
+ "version": null,
124
+ "vision": true,
125
+ "icon": "cloudflare"
126
+ },
127
+ {
128
+ "attachment": false,
129
+ "audioGeneration": false,
130
+ "cost": {
131
+ "input": null,
132
+ "inputCacheHit": null,
133
+ "output": null
134
+ },
135
+ "description": "BART is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. You can use this model for text summarization.",
136
+ "id": "bart-large-cnn",
137
+ "knowledge": null,
138
+ "lastUpdated": null,
139
+ "limit": {
140
+ "context": null,
141
+ "output": null
142
+ },
143
+ "modalities": {
144
+ "input": ["text"],
145
+ "output": ["text"]
146
+ },
147
+ "name": "bart-large-cnn",
148
+ "openWeights": false,
149
+ "provider": "Cloudflare",
150
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
151
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
152
+ "providerId": "cloudflare",
153
+ "providerNpm": "@ai-sdk/cloudflare",
154
+ "reasoning": false,
155
+ "releaseDate": null,
156
+ "streamingSupported": true,
157
+ "structuredOutputs": true,
158
+ "supportsStructuredOutput": true,
159
+ "supportsTools": true,
160
+ "temperature": true,
161
+ "toolCall": true,
162
+ "version": null,
163
+ "vision": false,
164
+ "icon": "cloudflare"
165
+ },
166
+ {
167
+ "attachment": false,
168
+ "audioGeneration": false,
169
+ "cost": {
170
+ "input": null,
171
+ "inputCacheHit": null,
172
+ "output": null
173
+ },
174
+ "description": "BAAI general embedding (Base) model that transforms any given text into a 768-dimensional vector",
175
+ "id": "bge-base-en-v1.5",
176
+ "knowledge": null,
177
+ "lastUpdated": null,
178
+ "limit": {
179
+ "context": null,
180
+ "output": null
181
+ },
182
+ "modalities": {
183
+ "input": ["text"],
184
+ "output": ["embedding"]
185
+ },
186
+ "name": "bge-base-en-v1.5",
187
+ "openWeights": false,
188
+ "provider": "Cloudflare",
189
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
190
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
191
+ "providerId": "cloudflare",
192
+ "providerNpm": "@ai-sdk/cloudflare",
193
+ "reasoning": false,
194
+ "releaseDate": null,
195
+ "streamingSupported": true,
196
+ "structuredOutputs": true,
197
+ "supportsStructuredOutput": true,
198
+ "supportsTools": true,
199
+ "temperature": true,
200
+ "toolCall": true,
201
+ "version": null,
202
+ "vision": false,
203
+ "icon": "cloudflare"
204
+ },
205
+ {
206
+ "attachment": false,
207
+ "audioGeneration": false,
208
+ "cost": {
209
+ "input": null,
210
+ "inputCacheHit": null,
211
+ "output": null
212
+ },
213
+ "description": "BAAI general embedding (Large) model that transforms any given text into a 1024-dimensional vector",
214
+ "id": "bge-large-en-v1.5",
215
+ "knowledge": null,
216
+ "lastUpdated": null,
217
+ "limit": {
218
+ "context": null,
219
+ "output": null
220
+ },
221
+ "modalities": {
222
+ "input": ["text"],
223
+ "output": ["embedding"]
224
+ },
225
+ "name": "bge-large-en-v1.5",
226
+ "openWeights": false,
227
+ "provider": "Cloudflare",
228
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
229
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
230
+ "providerId": "cloudflare",
231
+ "providerNpm": "@ai-sdk/cloudflare",
232
+ "reasoning": false,
233
+ "releaseDate": null,
234
+ "streamingSupported": true,
235
+ "structuredOutputs": true,
236
+ "supportsStructuredOutput": true,
237
+ "supportsTools": true,
238
+ "temperature": true,
239
+ "toolCall": true,
240
+ "version": null,
241
+ "vision": false,
242
+ "icon": "cloudflare"
243
+ },
244
+ {
245
+ "attachment": false,
246
+ "audioGeneration": false,
247
+ "cost": {
248
+ "input": null,
249
+ "inputCacheHit": null,
250
+ "output": null
251
+ },
252
+ "description": "Multi-Functionality, Multi-Linguality, and Multi-Granularity embeddings model.",
253
+ "id": "bge-m3",
254
+ "knowledge": null,
255
+ "lastUpdated": null,
256
+ "limit": {
257
+ "context": 60000000,
258
+ "output": null
259
+ },
260
+ "modalities": {
261
+ "input": ["text"],
262
+ "output": ["embedding"]
263
+ },
264
+ "name": "bge-m3",
265
+ "openWeights": false,
266
+ "provider": "Cloudflare",
267
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
268
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
269
+ "providerId": "cloudflare",
270
+ "providerNpm": "@ai-sdk/cloudflare",
271
+ "reasoning": false,
272
+ "releaseDate": null,
273
+ "streamingSupported": true,
274
+ "structuredOutputs": true,
275
+ "supportsStructuredOutput": true,
276
+ "supportsTools": true,
277
+ "temperature": true,
278
+ "toolCall": true,
279
+ "version": null,
280
+ "vision": false,
281
+ "icon": "cloudflare"
282
+ },
283
+ {
284
+ "attachment": false,
285
+ "audioGeneration": false,
286
+ "cost": {
287
+ "input": null,
288
+ "inputCacheHit": null,
289
+ "output": null
290
+ },
291
+ "description": "BAAI general embedding (Small) model that transforms any given text into a 384-dimensional vector",
292
+ "id": "bge-small-en-v1.5",
293
+ "knowledge": null,
294
+ "lastUpdated": null,
295
+ "limit": {
296
+ "context": null,
297
+ "output": null
298
+ },
299
+ "modalities": {
300
+ "input": ["text"],
301
+ "output": ["embedding"]
302
+ },
303
+ "name": "bge-small-en-v1.5",
304
+ "openWeights": false,
305
+ "provider": "Cloudflare",
306
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
307
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
308
+ "providerId": "cloudflare",
309
+ "providerNpm": "@ai-sdk/cloudflare",
310
+ "reasoning": false,
311
+ "releaseDate": null,
312
+ "streamingSupported": true,
313
+ "structuredOutputs": true,
314
+ "supportsStructuredOutput": true,
315
+ "supportsTools": true,
316
+ "temperature": true,
317
+ "toolCall": true,
318
+ "version": null,
319
+ "vision": false,
320
+ "icon": "cloudflare"
321
+ },
322
+ {
323
+ "attachment": false,
324
+ "audioGeneration": false,
325
+ "cost": {
326
+ "input": null,
327
+ "inputCacheHit": null,
328
+ "output": null
329
+ },
330
+ "description": "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",
331
+ "id": "deepseek-coder-6.7b-base-awq",
332
+ "knowledge": null,
333
+ "lastUpdated": null,
334
+ "limit": {
335
+ "context": 4096000,
336
+ "output": null
337
+ },
338
+ "modalities": {
339
+ "input": ["text"],
340
+ "output": ["text"]
341
+ },
342
+ "name": "deepseek-coder-6.7b-base-awq",
343
+ "openWeights": false,
344
+ "provider": "Cloudflare",
345
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
346
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
347
+ "providerId": "cloudflare",
348
+ "providerNpm": "@ai-sdk/cloudflare",
349
+ "reasoning": false,
350
+ "releaseDate": null,
351
+ "streamingSupported": true,
352
+ "structuredOutputs": true,
353
+ "supportsStructuredOutput": true,
354
+ "supportsTools": true,
355
+ "temperature": true,
356
+ "toolCall": true,
357
+ "version": null,
358
+ "vision": false,
359
+ "icon": "cloudflare"
360
+ },
361
+ {
362
+ "attachment": false,
363
+ "audioGeneration": false,
364
+ "cost": {
365
+ "input": null,
366
+ "inputCacheHit": null,
367
+ "output": null
368
+ },
369
+ "description": "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",
370
+ "id": "deepseek-coder-6.7b-instruct-awq",
371
+ "knowledge": null,
372
+ "lastUpdated": null,
373
+ "limit": {
374
+ "context": 4096000,
375
+ "output": null
376
+ },
377
+ "modalities": {
378
+ "input": ["text"],
379
+ "output": ["text"]
380
+ },
381
+ "name": "deepseek-coder-6.7b-instruct-awq",
382
+ "openWeights": false,
383
+ "provider": "Cloudflare",
384
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
385
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
386
+ "providerId": "cloudflare",
387
+ "providerNpm": "@ai-sdk/cloudflare",
388
+ "reasoning": false,
389
+ "releaseDate": null,
390
+ "streamingSupported": true,
391
+ "structuredOutputs": true,
392
+ "supportsStructuredOutput": true,
393
+ "supportsTools": true,
394
+ "temperature": true,
395
+ "toolCall": true,
396
+ "version": null,
397
+ "vision": false,
398
+ "icon": "cloudflare"
399
+ },
400
+ {
401
+ "attachment": false,
402
+ "audioGeneration": false,
403
+ "cost": {
404
+ "input": null,
405
+ "inputCacheHit": null,
406
+ "output": null
407
+ },
408
+ "description": "DeepSeekMath-Instruct 7B is a mathematically instructed tuning model derived from DeepSeekMath-Base 7B. DeepSeekMath is initialized with DeepSeek-Coder-v1.5 7B and continues pre-training on math-related tokens sourced from Common Crawl, together with natural language and code data for 500B tokens.",
409
+ "id": "deepseek-math-7b-instruct",
410
+ "knowledge": null,
411
+ "lastUpdated": null,
412
+ "limit": {
413
+ "context": 4096000,
414
+ "output": null
415
+ },
416
+ "modalities": {
417
+ "input": ["text"],
418
+ "output": ["text"]
419
+ },
420
+ "name": "deepseek-math-7b-instruct",
421
+ "openWeights": false,
422
+ "provider": "Cloudflare",
423
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
424
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
425
+ "providerId": "cloudflare",
426
+ "providerNpm": "@ai-sdk/cloudflare",
427
+ "reasoning": false,
428
+ "releaseDate": null,
429
+ "streamingSupported": true,
430
+ "structuredOutputs": true,
431
+ "supportsStructuredOutput": true,
432
+ "supportsTools": true,
433
+ "temperature": true,
434
+ "toolCall": true,
435
+ "version": null,
436
+ "vision": false,
437
+ "icon": "cloudflare"
438
+ },
439
+ {
440
+ "attachment": false,
441
+ "audioGeneration": false,
442
+ "cost": {
443
+ "input": 0.0003,
444
+ "inputCacheHit": null,
445
+ "output": 0.0003
446
+ },
447
+ "description": "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.",
448
+ "extendedThinking": false,
449
+ "id": "deepseek-r1-distill-qwen-32b",
450
+ "knowledge": null,
451
+ "lastUpdated": null,
452
+ "limit": {
453
+ "context": 80000000,
454
+ "output": null
455
+ },
456
+ "modalities": {
457
+ "input": ["text"],
458
+ "output": ["text"]
459
+ },
460
+ "name": "deepseek-r1-distill-qwen-32b",
461
+ "openWeights": false,
462
+ "provider": "Cloudflare",
463
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
464
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
465
+ "providerId": "cloudflare",
466
+ "providerNpm": "@ai-sdk/cloudflare",
467
+ "reasoning": false,
468
+ "releaseDate": "2025-01-29",
469
+ "streamingSupported": true,
470
+ "structuredOutputs": true,
471
+ "supportsStructuredOutput": true,
472
+ "supportsTools": true,
473
+ "temperature": true,
474
+ "toolCall": true,
475
+ "version": null,
476
+ "vision": false,
477
+ "icon": "cloudflare"
478
+ },
479
+ {
480
+ "attachment": false,
481
+ "audioGeneration": false,
482
+ "cost": {
483
+ "input": null,
484
+ "inputCacheHit": null,
485
+ "output": null
486
+ },
487
+ "description": "DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).",
488
+ "id": "detr-resnet-50",
489
+ "knowledge": null,
490
+ "lastUpdated": null,
491
+ "limit": {
492
+ "context": null,
493
+ "output": null
494
+ },
495
+ "modalities": {
496
+ "input": ["text", "image"],
497
+ "output": ["text"]
498
+ },
499
+ "name": "detr-resnet-50",
500
+ "openWeights": false,
501
+ "provider": "Cloudflare",
502
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
503
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
504
+ "providerId": "cloudflare",
505
+ "providerNpm": "@ai-sdk/cloudflare",
506
+ "reasoning": false,
507
+ "releaseDate": null,
508
+ "streamingSupported": true,
509
+ "structuredOutputs": true,
510
+ "supportsStructuredOutput": true,
511
+ "supportsTools": true,
512
+ "temperature": true,
513
+ "toolCall": true,
514
+ "version": null,
515
+ "vision": true,
516
+ "icon": "cloudflare"
517
+ },
518
+ {
519
+ "attachment": false,
520
+ "audioGeneration": false,
521
+ "cost": {
522
+ "input": null,
523
+ "inputCacheHit": null,
524
+ "output": null
525
+ },
526
+ "description": "DiscoLM German 7b is a Mistral-based large language model with a focus on German-language applications. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",
527
+ "id": "discolm-german-7b-v1-awq",
528
+ "knowledge": null,
529
+ "lastUpdated": null,
530
+ "limit": {
531
+ "context": 4096000,
532
+ "output": null
533
+ },
534
+ "modalities": {
535
+ "input": ["text"],
536
+ "output": ["text"]
537
+ },
538
+ "name": "discolm-german-7b-v1-awq",
539
+ "openWeights": false,
540
+ "provider": "Cloudflare",
541
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
542
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
543
+ "providerId": "cloudflare",
544
+ "providerNpm": "@ai-sdk/cloudflare",
545
+ "reasoning": false,
546
+ "releaseDate": null,
547
+ "streamingSupported": true,
548
+ "structuredOutputs": true,
549
+ "supportsStructuredOutput": true,
550
+ "supportsTools": true,
551
+ "temperature": true,
552
+ "toolCall": true,
553
+ "version": null,
554
+ "vision": false,
555
+ "icon": "cloudflare"
556
+ },
557
+ {
558
+ "attachment": false,
559
+ "audioGeneration": false,
560
+ "cost": {
561
+ "input": null,
562
+ "inputCacheHit": null,
563
+ "output": null
564
+ },
565
+ "description": "Distilled BERT model that was finetuned on SST-2 for sentiment classification",
566
+ "id": "distilbert-sst-2-int8",
567
+ "knowledge": null,
568
+ "lastUpdated": null,
569
+ "limit": {
570
+ "context": null,
571
+ "output": null
572
+ },
573
+ "modalities": {
574
+ "input": ["text"],
575
+ "output": ["text"]
576
+ },
577
+ "name": "distilbert-sst-2-int8",
578
+ "openWeights": false,
579
+ "provider": "Cloudflare",
580
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
581
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
582
+ "providerId": "cloudflare",
583
+ "providerNpm": "@ai-sdk/cloudflare",
584
+ "reasoning": false,
585
+ "releaseDate": null,
586
+ "streamingSupported": true,
587
+ "structuredOutputs": true,
588
+ "supportsStructuredOutput": true,
589
+ "supportsTools": true,
590
+ "temperature": true,
591
+ "toolCall": true,
592
+ "version": null,
593
+ "vision": false,
594
+ "icon": "cloudflare"
595
+ },
596
+ {
597
+ "attachment": false,
598
+ "audioGeneration": false,
599
+ "cost": {
600
+ "input": null,
601
+ "inputCacheHit": null,
602
+ "output": null
603
+ },
604
+ "description": "Stable Diffusion model that has been fine-tuned to be better at photorealism without sacrificing range.",
605
+ "id": "dreamshaper-8-lcm",
606
+ "knowledge": null,
607
+ "lastUpdated": null,
608
+ "limit": {
609
+ "context": null,
610
+ "output": null
611
+ },
612
+ "modalities": {
613
+ "input": ["text", "image"],
614
+ "output": ["text"]
615
+ },
616
+ "name": "dreamshaper-8-lcm",
617
+ "openWeights": false,
618
+ "provider": "Cloudflare",
619
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
620
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
621
+ "providerId": "cloudflare",
622
+ "providerNpm": "@ai-sdk/cloudflare",
623
+ "reasoning": false,
624
+ "releaseDate": null,
625
+ "streamingSupported": true,
626
+ "structuredOutputs": true,
627
+ "supportsStructuredOutput": true,
628
+ "supportsTools": true,
629
+ "temperature": true,
630
+ "toolCall": true,
631
+ "version": null,
632
+ "vision": true,
633
+ "icon": "cloudflare"
634
+ },
635
+ {
636
+ "attachment": false,
637
+ "audioGeneration": false,
638
+ "cost": {
639
+ "input": null,
640
+ "inputCacheHit": null,
641
+ "output": null
642
+ },
643
+ "description": "EmbeddingGemma is a 300M parameter, state-of-the-art for its size, open embedding model from Google, built from Gemma 3 (with T5Gemma initialization) and the same research and technology used to create Gemini models. EmbeddingGemma produces vector representations of text, making it well-suited for search and retrieval tasks, including classification, clustering, and semantic similarity search. This model was trained with data in 100+ spoken languages.",
644
+ "id": "embeddinggemma-300m",
645
+ "knowledge": null,
646
+ "lastUpdated": null,
647
+ "limit": {
648
+ "context": null,
649
+ "output": null
650
+ },
651
+ "modalities": {
652
+ "input": ["text"],
653
+ "output": ["embedding"]
654
+ },
655
+ "name": "embeddinggemma-300m",
656
+ "openWeights": false,
657
+ "provider": "Cloudflare",
658
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
659
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
660
+ "providerId": "cloudflare",
661
+ "providerNpm": "@ai-sdk/cloudflare",
662
+ "reasoning": false,
663
+ "releaseDate": null,
664
+ "streamingSupported": true,
665
+ "structuredOutputs": true,
666
+ "supportsStructuredOutput": true,
667
+ "supportsTools": true,
668
+ "temperature": true,
669
+ "toolCall": true,
670
+ "version": null,
671
+ "vision": false,
672
+ "icon": "cloudflare"
673
+ },
674
+ {
675
+ "attachment": false,
676
+ "audioGeneration": false,
677
+ "cost": {
678
+ "input": null,
679
+ "inputCacheHit": null,
680
+ "output": null
681
+ },
682
+ "description": "Falcon-7B-Instruct is a 7B parameters causal decoder-only model built by TII based on Falcon-7B and finetuned on a mixture of chat/instruct datasets.",
683
+ "id": "falcon-7b-instruct",
684
+ "knowledge": null,
685
+ "lastUpdated": null,
686
+ "limit": {
687
+ "context": 4096000,
688
+ "output": null
689
+ },
690
+ "modalities": {
691
+ "input": ["text"],
692
+ "output": ["text"]
693
+ },
694
+ "name": "falcon-7b-instruct",
695
+ "openWeights": false,
696
+ "provider": "Cloudflare",
697
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
698
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
699
+ "providerId": "cloudflare",
700
+ "providerNpm": "@ai-sdk/cloudflare",
701
+ "reasoning": false,
702
+ "releaseDate": null,
703
+ "streamingSupported": true,
704
+ "structuredOutputs": true,
705
+ "supportsStructuredOutput": true,
706
+ "supportsTools": true,
707
+ "temperature": true,
708
+ "toolCall": true,
709
+ "version": null,
710
+ "vision": false,
711
+ "icon": "cloudflare"
712
+ },
713
+ {
714
+ "attachment": false,
715
+ "audioGeneration": true,
716
+ "cost": {
717
+ "input": 0,
718
+ "inputCacheHit": null,
719
+ "output": 0
720
+ },
721
+ "description": "Flux is the first conversational speech recognition model built specifically for voice agents.",
722
+ "id": "flux",
723
+ "knowledge": null,
724
+ "lastUpdated": null,
725
+ "limit": {
726
+ "context": null,
727
+ "output": null
728
+ },
729
+ "modalities": {
730
+ "input": ["audio"],
731
+ "output": ["text"]
732
+ },
733
+ "name": "flux",
734
+ "openWeights": false,
735
+ "provider": "Cloudflare",
736
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
737
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
738
+ "providerId": "cloudflare",
739
+ "providerNpm": "@ai-sdk/cloudflare",
740
+ "reasoning": false,
741
+ "releaseDate": null,
742
+ "streamingSupported": true,
743
+ "structuredOutputs": true,
744
+ "supportsStructuredOutput": true,
745
+ "supportsTools": true,
746
+ "temperature": true,
747
+ "toolCall": true,
748
+ "version": null,
749
+ "vision": false,
750
+ "icon": "cloudflare"
751
+ },
752
+ {
753
+ "attachment": false,
754
+ "audioGeneration": false,
755
+ "cost": {
756
+ "input": null,
757
+ "inputCacheHit": null,
758
+ "output": null
759
+ },
760
+ "description": "This is a Gemma-2B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",
761
+ "id": "gemma-2b-it-lora",
762
+ "knowledge": null,
763
+ "lastUpdated": null,
764
+ "limit": {
765
+ "context": 8192000,
766
+ "output": null
767
+ },
768
+ "modalities": {
769
+ "input": ["text"],
770
+ "output": ["text"]
771
+ },
772
+ "name": "gemma-2b-it-lora",
773
+ "openWeights": false,
774
+ "provider": "Cloudflare",
775
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
776
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
777
+ "providerId": "cloudflare",
778
+ "providerNpm": "@ai-sdk/cloudflare",
779
+ "reasoning": false,
780
+ "releaseDate": null,
781
+ "streamingSupported": true,
782
+ "structuredOutputs": true,
783
+ "supportsStructuredOutput": true,
784
+ "supportsTools": true,
785
+ "temperature": true,
786
+ "toolCall": true,
787
+ "version": null,
788
+ "vision": false,
789
+ "icon": "cloudflare"
790
+ },
791
+ {
792
+ "attachment": false,
793
+ "audioGeneration": false,
794
+ "cost": {
795
+ "input": 0.00003,
796
+ "inputCacheHit": null,
797
+ "output": 0.0001
798
+ },
799
+ "description": "Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Gemma 3 models are multimodal, handling text and image input and generating text output, with a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions.",
800
+ "id": "gemma-3-12b-it",
801
+ "knowledge": null,
802
+ "lastUpdated": null,
803
+ "limit": {
804
+ "context": 80000000,
805
+ "output": null
806
+ },
807
+ "modalities": {
808
+ "input": ["text", "image"],
809
+ "output": ["text"]
810
+ },
811
+ "name": "gemma-3-12b-it",
812
+ "openWeights": false,
813
+ "provider": "Cloudflare",
814
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
815
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
816
+ "providerId": "cloudflare",
817
+ "providerNpm": "@ai-sdk/cloudflare",
818
+ "reasoning": false,
819
+ "releaseDate": null,
820
+ "streamingSupported": true,
821
+ "structuredOutputs": true,
822
+ "supportsStructuredOutput": true,
823
+ "supportsTools": true,
824
+ "temperature": true,
825
+ "toolCall": true,
826
+ "version": null,
827
+ "vision": true,
828
+ "icon": "cloudflare"
829
+ },
830
+ {
831
+ "attachment": false,
832
+ "audioGeneration": false,
833
+ "cost": {
834
+ "input": 0.00007,
835
+ "inputCacheHit": null,
836
+ "output": 0.00007
837
+ },
838
+ "description": "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.",
839
+ "id": "gemma-7b-it",
840
+ "knowledge": null,
841
+ "lastUpdated": null,
842
+ "limit": {
843
+ "context": 8192000,
844
+ "output": null
845
+ },
846
+ "modalities": {
847
+ "input": ["text"],
848
+ "output": ["text"]
849
+ },
850
+ "name": "gemma-7b-it",
851
+ "openWeights": false,
852
+ "provider": "Cloudflare",
853
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
854
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
855
+ "providerId": "cloudflare",
856
+ "providerNpm": "@ai-sdk/cloudflare",
857
+ "reasoning": false,
858
+ "releaseDate": null,
859
+ "streamingSupported": true,
860
+ "structuredOutputs": true,
861
+ "supportsStructuredOutput": true,
862
+ "supportsTools": true,
863
+ "temperature": true,
864
+ "toolCall": true,
865
+ "version": null,
866
+ "vision": false,
867
+ "icon": "cloudflare"
868
+ },
869
+ {
870
+ "attachment": false,
871
+ "audioGeneration": false,
872
+ "cost": {
873
+ "input": 0.00007,
874
+ "inputCacheHit": null,
875
+ "output": 0.00007
876
+ },
877
+ "description": "This is a Gemma-7B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",
878
+ "id": "gemma-7b-it-lora",
879
+ "knowledge": null,
880
+ "lastUpdated": null,
881
+ "limit": {
882
+ "context": 3500000,
883
+ "output": null
884
+ },
885
+ "modalities": {
886
+ "input": ["text"],
887
+ "output": ["text"]
888
+ },
889
+ "name": "gemma-7b-it-lora",
890
+ "openWeights": false,
891
+ "provider": "Cloudflare",
892
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
893
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
894
+ "providerId": "cloudflare",
895
+ "providerNpm": "@ai-sdk/cloudflare",
896
+ "reasoning": false,
897
+ "releaseDate": null,
898
+ "streamingSupported": true,
899
+ "structuredOutputs": true,
900
+ "supportsStructuredOutput": true,
901
+ "supportsTools": true,
902
+ "temperature": true,
903
+ "toolCall": true,
904
+ "version": null,
905
+ "vision": false,
906
+ "icon": "cloudflare"
907
+ },
908
+ {
909
+ "attachment": false,
910
+ "audioGeneration": false,
911
+ "cost": {
912
+ "input": null,
913
+ "inputCacheHit": null,
914
+ "output": null
915
+ },
916
+ "description": "SEA-LION stands for Southeast Asian Languages In One Network, which is a collection of Large Language Models (LLMs) which have been pretrained and instruct-tuned for the Southeast Asia (SEA) region.",
917
+ "id": "gemma-sea-lion-v4-27b-it",
918
+ "knowledge": null,
919
+ "lastUpdated": null,
920
+ "limit": {
921
+ "context": 128000000,
922
+ "output": null
923
+ },
924
+ "modalities": {
925
+ "input": ["text"],
926
+ "output": ["text"]
927
+ },
928
+ "name": "gemma-sea-lion-v4-27b-it",
929
+ "openWeights": false,
930
+ "provider": "Cloudflare",
931
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
932
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
933
+ "providerId": "cloudflare",
934
+ "providerNpm": "@ai-sdk/cloudflare",
935
+ "reasoning": false,
936
+ "releaseDate": null,
937
+ "streamingSupported": true,
938
+ "structuredOutputs": true,
939
+ "supportsStructuredOutput": true,
940
+ "supportsTools": true,
941
+ "temperature": true,
942
+ "toolCall": true,
943
+ "version": null,
944
+ "vision": false,
945
+ "icon": "cloudflare"
946
+ },
947
+ {
948
+ "attachment": false,
949
+ "audioGeneration": false,
950
+ "cost": {
951
+ "input": 0.0001,
952
+ "inputCacheHit": null,
953
+ "output": 0.0001
954
+ },
955
+ "description": "OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases – gpt-oss-20b is for lower latency, and local or specialized use-cases.",
956
+ "id": "gpt-oss-20b",
957
+ "knowledge": null,
958
+ "lastUpdated": null,
959
+ "limit": {
960
+ "context": 128000000,
961
+ "output": null
962
+ },
963
+ "modalities": {
964
+ "input": ["text"],
965
+ "output": ["text"]
966
+ },
967
+ "name": "gpt-oss-20b",
968
+ "openWeights": false,
969
+ "provider": "Cloudflare",
970
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
971
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
972
+ "providerId": "cloudflare",
973
+ "providerNpm": "@ai-sdk/cloudflare",
974
+ "reasoning": false,
975
+ "releaseDate": null,
976
+ "streamingSupported": true,
977
+ "structuredOutputs": true,
978
+ "supportsStructuredOutput": true,
979
+ "supportsTools": true,
980
+ "temperature": true,
981
+ "toolCall": true,
982
+ "version": null,
983
+ "vision": false,
984
+ "icon": "cloudflare"
985
+ },
986
+ {
987
+ "attachment": false,
988
+ "audioGeneration": false,
989
+ "cost": {
990
+ "input": 0.000017,
991
+ "inputCacheHit": null,
992
+ "output": 0.00011
993
+ },
994
+ "description": "Granite 4.0 instruct models deliver strong performance across benchmarks, achieving industry-leading results in key agentic tasks like instruction following and function calling. These efficiencies make the models well-suited for a wide range of use cases like retrieval-augmented generation (RAG), multi-agent workflows, and edge deployments.",
995
+ "id": "granite-4.0-h-micro",
996
+ "knowledge": null,
997
+ "lastUpdated": null,
998
+ "limit": {
999
+ "context": 131000000,
1000
+ "output": null
1001
+ },
1002
+ "modalities": {
1003
+ "input": ["text"],
1004
+ "output": ["text"]
1005
+ },
1006
+ "name": "granite-4.0-h-micro",
1007
+ "openWeights": false,
1008
+ "provider": "Cloudflare",
1009
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1010
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1011
+ "providerId": "cloudflare",
1012
+ "providerNpm": "@ai-sdk/cloudflare",
1013
+ "reasoning": false,
1014
+ "releaseDate": null,
1015
+ "streamingSupported": true,
1016
+ "structuredOutputs": true,
1017
+ "supportsStructuredOutput": true,
1018
+ "supportsTools": true,
1019
+ "temperature": true,
1020
+ "toolCall": true,
1021
+ "version": null,
1022
+ "vision": false,
1023
+ "icon": "cloudflare"
1024
+ },
1025
+ {
1026
+ "attachment": false,
1027
+ "audioGeneration": false,
1028
+ "cost": {
1029
+ "input": null,
1030
+ "inputCacheHit": null,
1031
+ "output": null
1032
+ },
1033
+ "description": "IndicTrans2 is the first open-source transformer-based multilingual NMT model that supports high-quality translations across all the 22 scheduled Indic languages",
1034
+ "id": "indictrans2-en-indic-1B",
1035
+ "knowledge": null,
1036
+ "lastUpdated": null,
1037
+ "limit": {
1038
+ "context": null,
1039
+ "output": null
1040
+ },
1041
+ "modalities": {
1042
+ "input": ["text"],
1043
+ "output": ["text"]
1044
+ },
1045
+ "name": "indictrans2-en-indic-1B",
1046
+ "openWeights": false,
1047
+ "provider": "Cloudflare",
1048
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1049
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1050
+ "providerId": "cloudflare",
1051
+ "providerNpm": "@ai-sdk/cloudflare",
1052
+ "reasoning": false,
1053
+ "releaseDate": null,
1054
+ "streamingSupported": true,
1055
+ "structuredOutputs": true,
1056
+ "supportsStructuredOutput": true,
1057
+ "supportsTools": true,
1058
+ "temperature": true,
1059
+ "toolCall": true,
1060
+ "version": null,
1061
+ "vision": false,
1062
+ "icon": "cloudflare"
1063
+ },
1064
+ {
1065
+ "attachment": false,
1066
+ "audioGeneration": false,
1067
+ "cost": {
1068
+ "input": 0.0003,
1069
+ "inputCacheHit": null,
1070
+ "output": 0.0003
1071
+ },
1072
+ "description": "Llama 2 13B Chat AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Llama 2 variant.",
1073
+ "id": "llama-2-13b-chat-awq",
1074
+ "knowledge": null,
1075
+ "lastUpdated": null,
1076
+ "limit": {
1077
+ "context": 4096000,
1078
+ "output": null
1079
+ },
1080
+ "modalities": {
1081
+ "input": ["text"],
1082
+ "output": ["text"]
1083
+ },
1084
+ "name": "llama-2-13b-chat-awq",
1085
+ "openWeights": false,
1086
+ "provider": "Cloudflare",
1087
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1088
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1089
+ "providerId": "cloudflare",
1090
+ "providerNpm": "@ai-sdk/cloudflare",
1091
+ "reasoning": false,
1092
+ "releaseDate": null,
1093
+ "streamingSupported": true,
1094
+ "structuredOutputs": true,
1095
+ "supportsStructuredOutput": true,
1096
+ "supportsTools": true,
1097
+ "temperature": true,
1098
+ "toolCall": true,
1099
+ "version": null,
1100
+ "vision": false,
1101
+ "icon": "cloudflare"
1102
+ },
1103
+ {
1104
+ "attachment": false,
1105
+ "audioGeneration": false,
1106
+ "cost": {
1107
+ "input": 0.0003,
1108
+ "inputCacheHit": null,
1109
+ "output": 0.0003
1110
+ },
1111
+ "description": "Full precision (fp16) generative text model with 7 billion parameters from Meta",
1112
+ "id": "llama-2-7b-chat-fp16",
1113
+ "knowledge": null,
1114
+ "lastUpdated": null,
1115
+ "limit": {
1116
+ "context": 4096000,
1117
+ "output": null
1118
+ },
1119
+ "modalities": {
1120
+ "input": ["text"],
1121
+ "output": ["text"]
1122
+ },
1123
+ "name": "llama-2-7b-chat-fp16",
1124
+ "openWeights": false,
1125
+ "provider": "Cloudflare",
1126
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1127
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1128
+ "providerId": "cloudflare",
1129
+ "providerNpm": "@ai-sdk/cloudflare",
1130
+ "reasoning": false,
1131
+ "releaseDate": null,
1132
+ "streamingSupported": true,
1133
+ "structuredOutputs": true,
1134
+ "supportsStructuredOutput": true,
1135
+ "supportsTools": true,
1136
+ "temperature": true,
1137
+ "toolCall": true,
1138
+ "version": null,
1139
+ "vision": false,
1140
+ "icon": "cloudflare"
1141
+ },
1142
+ {
1143
+ "attachment": false,
1144
+ "audioGeneration": false,
1145
+ "cost": {
1146
+ "input": 0.0003,
1147
+ "inputCacheHit": null,
1148
+ "output": 0.0003
1149
+ },
1150
+ "description": "This is a Llama2 base model that Cloudflare dedicated for inference with LoRA adapters. Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format.",
1151
+ "id": "llama-2-7b-chat-hf-lora",
1152
+ "knowledge": null,
1153
+ "lastUpdated": null,
1154
+ "limit": {
1155
+ "context": 8192000,
1156
+ "output": null
1157
+ },
1158
+ "modalities": {
1159
+ "input": ["text"],
1160
+ "output": ["text"]
1161
+ },
1162
+ "name": "llama-2-7b-chat-hf-lora",
1163
+ "openWeights": false,
1164
+ "provider": "Cloudflare",
1165
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1166
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1167
+ "providerId": "cloudflare",
1168
+ "providerNpm": "@ai-sdk/cloudflare",
1169
+ "reasoning": false,
1170
+ "releaseDate": null,
1171
+ "streamingSupported": true,
1172
+ "structuredOutputs": true,
1173
+ "supportsStructuredOutput": true,
1174
+ "supportsTools": true,
1175
+ "temperature": true,
1176
+ "toolCall": true,
1177
+ "version": null,
1178
+ "vision": false,
1179
+ "icon": "cloudflare"
1180
+ },
1181
+ {
1182
+ "attachment": false,
1183
+ "audioGeneration": false,
1184
+ "cost": {
1185
+ "input": 0.0003,
1186
+ "inputCacheHit": null,
1187
+ "output": 0.0003
1188
+ },
1189
+ "description": "Quantized (int8) generative text model with 7 billion parameters from Meta",
1190
+ "id": "llama-2-7b-chat-int8",
1191
+ "knowledge": null,
1192
+ "lastUpdated": null,
1193
+ "limit": {
1194
+ "context": 8192000,
1195
+ "output": null
1196
+ },
1197
+ "modalities": {
1198
+ "input": ["text"],
1199
+ "output": ["text"]
1200
+ },
1201
+ "name": "llama-2-7b-chat-int8",
1202
+ "openWeights": false,
1203
+ "provider": "Cloudflare",
1204
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1205
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1206
+ "providerId": "cloudflare",
1207
+ "providerNpm": "@ai-sdk/cloudflare",
1208
+ "reasoning": false,
1209
+ "releaseDate": null,
1210
+ "streamingSupported": true,
1211
+ "structuredOutputs": true,
1212
+ "supportsStructuredOutput": true,
1213
+ "supportsTools": true,
1214
+ "temperature": true,
1215
+ "toolCall": true,
1216
+ "version": null,
1217
+ "vision": false,
1218
+ "icon": "cloudflare"
1219
+ },
1220
+ {
1221
+ "attachment": false,
1222
+ "audioGeneration": false,
1223
+ "cost": {
1224
+ "input": 0.00045,
1225
+ "inputCacheHit": null,
1226
+ "output": 0.00045
1227
+ },
1228
+ "description": "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1229
+ "id": "llama-3.1-70b-instruct",
1230
+ "knowledge": null,
1231
+ "lastUpdated": null,
1232
+ "limit": {
1233
+ "context": 24000000,
1234
+ "output": null
1235
+ },
1236
+ "modalities": {
1237
+ "input": ["text", "image"],
1238
+ "output": ["text"]
1239
+ },
1240
+ "name": "llama-3.1-70b-instruct",
1241
+ "openWeights": false,
1242
+ "provider": "Cloudflare",
1243
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1244
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1245
+ "providerId": "cloudflare",
1246
+ "providerNpm": "@ai-sdk/cloudflare",
1247
+ "reasoning": false,
1248
+ "releaseDate": null,
1249
+ "streamingSupported": true,
1250
+ "structuredOutputs": true,
1251
+ "supportsStructuredOutput": true,
1252
+ "supportsTools": true,
1253
+ "temperature": true,
1254
+ "toolCall": true,
1255
+ "version": null,
1256
+ "vision": true,
1257
+ "icon": "cloudflare"
1258
+ },
1259
+ {
1260
+ "attachment": false,
1261
+ "audioGeneration": false,
1262
+ "cost": {
1263
+ "input": 0.0001,
1264
+ "inputCacheHit": null,
1265
+ "output": 0.0001
1266
+ },
1267
+ "description": "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1268
+ "id": "llama-3.1-8b-instruct",
1269
+ "knowledge": null,
1270
+ "lastUpdated": null,
1271
+ "limit": {
1272
+ "context": 7968000,
1273
+ "output": null
1274
+ },
1275
+ "modalities": {
1276
+ "input": ["text"],
1277
+ "output": ["text"]
1278
+ },
1279
+ "name": "llama-3.1-8b-instruct",
1280
+ "openWeights": false,
1281
+ "provider": "Cloudflare",
1282
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1283
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1284
+ "providerId": "cloudflare",
1285
+ "providerNpm": "@ai-sdk/cloudflare",
1286
+ "reasoning": false,
1287
+ "releaseDate": null,
1288
+ "streamingSupported": true,
1289
+ "structuredOutputs": true,
1290
+ "supportsStructuredOutput": true,
1291
+ "supportsTools": true,
1292
+ "temperature": true,
1293
+ "toolCall": true,
1294
+ "version": null,
1295
+ "vision": false,
1296
+ "icon": "cloudflare"
1297
+ },
1298
+ {
1299
+ "attachment": false,
1300
+ "audioGeneration": false,
1301
+ "cost": {
1302
+ "input": 0.0003,
1303
+ "inputCacheHit": null,
1304
+ "output": 0.0003
1305
+ },
1306
+ "description": "Quantized (int4) generative text model with 8 billion parameters from Meta.",
1307
+ "id": "llama-3.1-8b-instruct-awq",
1308
+ "knowledge": null,
1309
+ "lastUpdated": null,
1310
+ "limit": {
1311
+ "context": 8192000,
1312
+ "output": null
1313
+ },
1314
+ "modalities": {
1315
+ "input": ["text"],
1316
+ "output": ["text"]
1317
+ },
1318
+ "name": "llama-3.1-8b-instruct-awq",
1319
+ "openWeights": false,
1320
+ "provider": "Cloudflare",
1321
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1322
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1323
+ "providerId": "cloudflare",
1324
+ "providerNpm": "@ai-sdk/cloudflare",
1325
+ "reasoning": false,
1326
+ "releaseDate": null,
1327
+ "streamingSupported": true,
1328
+ "structuredOutputs": true,
1329
+ "supportsStructuredOutput": true,
1330
+ "supportsTools": true,
1331
+ "temperature": true,
1332
+ "toolCall": true,
1333
+ "version": null,
1334
+ "vision": false,
1335
+ "icon": "cloudflare"
1336
+ },
1337
+ {
1338
+ "attachment": false,
1339
+ "audioGeneration": false,
1340
+ "cost": {
1341
+ "input": 0.0003,
1342
+ "inputCacheHit": null,
1343
+ "output": 0.0003
1344
+ },
1345
+ "description": "[Fast version] The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1346
+ "id": "llama-3.1-8b-instruct-fast",
1347
+ "knowledge": null,
1348
+ "lastUpdated": null,
1349
+ "limit": {
1350
+ "context": 128000000,
1351
+ "output": null
1352
+ },
1353
+ "modalities": {
1354
+ "input": ["text", "image"],
1355
+ "output": ["text"]
1356
+ },
1357
+ "name": "llama-3.1-8b-instruct-fast",
1358
+ "openWeights": false,
1359
+ "provider": "Cloudflare",
1360
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1361
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1362
+ "providerId": "cloudflare",
1363
+ "providerNpm": "@ai-sdk/cloudflare",
1364
+ "reasoning": false,
1365
+ "releaseDate": null,
1366
+ "streamingSupported": true,
1367
+ "structuredOutputs": true,
1368
+ "supportsStructuredOutput": true,
1369
+ "supportsTools": true,
1370
+ "temperature": true,
1371
+ "toolCall": true,
1372
+ "version": null,
1373
+ "vision": true,
1374
+ "icon": "cloudflare"
1375
+ },
1376
+ {
1377
+ "attachment": false,
1378
+ "audioGeneration": false,
1379
+ "cost": {
1380
+ "input": 0.0003,
1381
+ "inputCacheHit": null,
1382
+ "output": 0.0003
1383
+ },
1384
+ "description": "Llama 3.1 8B quantized to FP8 precision",
1385
+ "id": "llama-3.1-8b-instruct-fp8",
1386
+ "knowledge": null,
1387
+ "lastUpdated": null,
1388
+ "limit": {
1389
+ "context": 32000000,
1390
+ "output": null
1391
+ },
1392
+ "modalities": {
1393
+ "input": ["text"],
1394
+ "output": ["text"]
1395
+ },
1396
+ "name": "llama-3.1-8b-instruct-fp8",
1397
+ "openWeights": false,
1398
+ "provider": "Cloudflare",
1399
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1400
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1401
+ "providerId": "cloudflare",
1402
+ "providerNpm": "@ai-sdk/cloudflare",
1403
+ "reasoning": false,
1404
+ "releaseDate": null,
1405
+ "streamingSupported": true,
1406
+ "structuredOutputs": true,
1407
+ "supportsStructuredOutput": true,
1408
+ "supportsTools": true,
1409
+ "temperature": true,
1410
+ "toolCall": true,
1411
+ "version": null,
1412
+ "vision": false,
1413
+ "icon": "cloudflare"
1414
+ },
1415
+ {
1416
+ "attachment": false,
1417
+ "audioGeneration": false,
1418
+ "cost": {
1419
+ "input": 0.00006,
1420
+ "inputCacheHit": null,
1421
+ "output": 0.00006
1422
+ },
1423
+ "description": "The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.",
1424
+ "id": "llama-3.2-11b-vision-instruct",
1425
+ "knowledge": null,
1426
+ "lastUpdated": null,
1427
+ "limit": {
1428
+ "context": 128000000,
1429
+ "output": null
1430
+ },
1431
+ "modalities": {
1432
+ "input": ["text", "image"],
1433
+ "output": ["text"]
1434
+ },
1435
+ "name": "llama-3.2-11b-vision-instruct",
1436
+ "openWeights": false,
1437
+ "provider": "Cloudflare",
1438
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1439
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1440
+ "providerId": "cloudflare",
1441
+ "providerNpm": "@ai-sdk/cloudflare",
1442
+ "reasoning": false,
1443
+ "releaseDate": null,
1444
+ "streamingSupported": true,
1445
+ "structuredOutputs": true,
1446
+ "supportsStructuredOutput": true,
1447
+ "supportsTools": true,
1448
+ "temperature": true,
1449
+ "toolCall": true,
1450
+ "version": null,
1451
+ "vision": true,
1452
+ "icon": "cloudflare"
1453
+ },
1454
+ {
1455
+ "attachment": false,
1456
+ "audioGeneration": false,
1457
+ "cost": {
1458
+ "input": 0.00002,
1459
+ "inputCacheHit": null,
1460
+ "output": 0.00002
1461
+ },
1462
+ "description": "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",
1463
+ "id": "llama-3.2-1b-instruct",
1464
+ "knowledge": null,
1465
+ "lastUpdated": null,
1466
+ "limit": {
1467
+ "context": 60000000,
1468
+ "output": null
1469
+ },
1470
+ "modalities": {
1471
+ "input": ["text"],
1472
+ "output": ["text"]
1473
+ },
1474
+ "name": "llama-3.2-1b-instruct",
1475
+ "openWeights": false,
1476
+ "provider": "Cloudflare",
1477
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1478
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1479
+ "providerId": "cloudflare",
1480
+ "providerNpm": "@ai-sdk/cloudflare",
1481
+ "reasoning": false,
1482
+ "releaseDate": null,
1483
+ "streamingSupported": true,
1484
+ "structuredOutputs": true,
1485
+ "supportsStructuredOutput": true,
1486
+ "supportsTools": true,
1487
+ "temperature": true,
1488
+ "toolCall": true,
1489
+ "version": null,
1490
+ "vision": false,
1491
+ "icon": "cloudflare"
1492
+ },
1493
+ {
1494
+ "attachment": false,
1495
+ "audioGeneration": false,
1496
+ "cost": {
1497
+ "input": 0.00003,
1498
+ "inputCacheHit": null,
1499
+ "output": 0.00005
1500
+ },
1501
+ "description": "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",
1502
+ "id": "llama-3.2-3b-instruct",
1503
+ "knowledge": null,
1504
+ "lastUpdated": null,
1505
+ "limit": {
1506
+ "context": 128000000,
1507
+ "output": null
1508
+ },
1509
+ "modalities": {
1510
+ "input": ["text"],
1511
+ "output": ["text"]
1512
+ },
1513
+ "name": "llama-3.2-3b-instruct",
1514
+ "openWeights": false,
1515
+ "provider": "Cloudflare",
1516
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1517
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1518
+ "providerId": "cloudflare",
1519
+ "providerNpm": "@ai-sdk/cloudflare",
1520
+ "reasoning": false,
1521
+ "releaseDate": null,
1522
+ "streamingSupported": true,
1523
+ "structuredOutputs": true,
1524
+ "supportsStructuredOutput": true,
1525
+ "supportsTools": true,
1526
+ "temperature": true,
1527
+ "toolCall": true,
1528
+ "version": null,
1529
+ "vision": false,
1530
+ "icon": "cloudflare"
1531
+ },
1532
+ {
1533
+ "attachment": false,
1534
+ "audioGeneration": false,
1535
+ "cost": {
1536
+ "input": 0,
1537
+ "inputCacheHit": null,
1538
+ "output": 0
1539
+ },
1540
+ "description": "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.",
1541
+ "id": "llama-3.3-70b-instruct-fp8-fast",
1542
+ "knowledge": null,
1543
+ "lastUpdated": null,
1544
+ "limit": {
1545
+ "context": 24000000,
1546
+ "output": null
1547
+ },
1548
+ "modalities": {
1549
+ "input": ["text"],
1550
+ "output": ["text"]
1551
+ },
1552
+ "name": "llama-3.3-70b-instruct-fp8-fast",
1553
+ "openWeights": false,
1554
+ "provider": "Cloudflare",
1555
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1556
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1557
+ "providerId": "cloudflare",
1558
+ "providerNpm": "@ai-sdk/cloudflare",
1559
+ "reasoning": false,
1560
+ "releaseDate": null,
1561
+ "streamingSupported": true,
1562
+ "structuredOutputs": true,
1563
+ "supportsStructuredOutput": true,
1564
+ "supportsTools": true,
1565
+ "temperature": true,
1566
+ "toolCall": true,
1567
+ "version": null,
1568
+ "vision": false,
1569
+ "icon": "cloudflare"
1570
+ },
1571
+ {
1572
+ "attachment": false,
1573
+ "audioGeneration": false,
1574
+ "cost": {
1575
+ "input": 0.0003,
1576
+ "inputCacheHit": null,
1577
+ "output": 0.0003
1578
+ },
1579
+ "description": "Quantized (int4) generative text model with 8 billion parameters from Meta.",
1580
+ "id": "llama-3-8b-instruct-awq",
1581
+ "knowledge": null,
1582
+ "lastUpdated": null,
1583
+ "limit": {
1584
+ "context": 8192000,
1585
+ "output": null
1586
+ },
1587
+ "modalities": {
1588
+ "input": ["text"],
1589
+ "output": ["text"]
1590
+ },
1591
+ "name": "llama-3-8b-instruct-awq",
1592
+ "openWeights": false,
1593
+ "provider": "Cloudflare",
1594
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1595
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1596
+ "providerId": "cloudflare",
1597
+ "providerNpm": "@ai-sdk/cloudflare",
1598
+ "reasoning": false,
1599
+ "releaseDate": null,
1600
+ "streamingSupported": true,
1601
+ "structuredOutputs": true,
1602
+ "supportsStructuredOutput": true,
1603
+ "supportsTools": true,
1604
+ "temperature": true,
1605
+ "toolCall": true,
1606
+ "version": null,
1607
+ "vision": false,
1608
+ "icon": "cloudflare"
1609
+ },
1610
+ {
1611
+ "attachment": false,
1612
+ "audioGeneration": false,
1613
+ "cost": {
1614
+ "input": 0.0003,
1615
+ "inputCacheHit": null,
1616
+ "output": 0.0003
1617
+ },
1618
+ "description": "Llama Guard is a model for classifying the safety of LLM prompts and responses, using a taxonomy of safety risks.",
1619
+ "id": "llamaguard-7b-awq",
1620
+ "knowledge": null,
1621
+ "lastUpdated": null,
1622
+ "limit": {
1623
+ "context": 4096000,
1624
+ "output": null
1625
+ },
1626
+ "modalities": {
1627
+ "input": ["text"],
1628
+ "output": ["text"]
1629
+ },
1630
+ "name": "llamaguard-7b-awq",
1631
+ "openWeights": false,
1632
+ "provider": "Cloudflare",
1633
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1634
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1635
+ "providerId": "cloudflare",
1636
+ "providerNpm": "@ai-sdk/cloudflare",
1637
+ "reasoning": false,
1638
+ "releaseDate": null,
1639
+ "streamingSupported": true,
1640
+ "structuredOutputs": true,
1641
+ "supportsStructuredOutput": true,
1642
+ "supportsTools": true,
1643
+ "temperature": true,
1644
+ "toolCall": true,
1645
+ "version": null,
1646
+ "vision": false,
1647
+ "icon": "cloudflare"
1648
+ },
1649
+ {
1650
+ "attachment": false,
1651
+ "audioGeneration": false,
1652
+ "cost": {
1653
+ "input": null,
1654
+ "inputCacheHit": null,
1655
+ "output": null
1656
+ },
1657
+ "description": "LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.",
1658
+ "id": "llava-1.5-7b-hf",
1659
+ "knowledge": null,
1660
+ "lastUpdated": null,
1661
+ "limit": {
1662
+ "context": null,
1663
+ "output": null
1664
+ },
1665
+ "modalities": {
1666
+ "input": ["text", "image"],
1667
+ "output": ["text"]
1668
+ },
1669
+ "name": "llava-1.5-7b-hf",
1670
+ "openWeights": false,
1671
+ "provider": "Cloudflare",
1672
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1673
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1674
+ "providerId": "cloudflare",
1675
+ "providerNpm": "@ai-sdk/cloudflare",
1676
+ "reasoning": false,
1677
+ "releaseDate": null,
1678
+ "streamingSupported": true,
1679
+ "structuredOutputs": true,
1680
+ "supportsStructuredOutput": true,
1681
+ "supportsTools": true,
1682
+ "temperature": true,
1683
+ "toolCall": true,
1684
+ "version": null,
1685
+ "vision": true,
1686
+ "icon": "cloudflare"
1687
+ },
1688
+ {
1689
+ "attachment": false,
1690
+ "audioGeneration": false,
1691
+ "cost": {
1692
+ "input": null,
1693
+ "inputCacheHit": null,
1694
+ "output": null
1695
+ },
1696
+ "description": "Lucid Origin from Leonardo.AI is their most adaptable and prompt-responsive model to date. Whether you're generating images with sharp graphic design, stunning full-HD renders, or highly specific creative direction, it adheres closely to your prompts, renders text with accuracy, and supports a wide array of visual styles and aesthetics – from stylized concept art to crisp product mockups.",
1697
+ "id": "lucid-origin",
1698
+ "knowledge": null,
1699
+ "lastUpdated": null,
1700
+ "limit": {
1701
+ "context": null,
1702
+ "output": null
1703
+ },
1704
+ "modalities": {
1705
+ "input": ["text", "image"],
1706
+ "output": ["text"]
1707
+ },
1708
+ "name": "lucid-origin",
1709
+ "openWeights": false,
1710
+ "provider": "Cloudflare",
1711
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1712
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1713
+ "providerId": "cloudflare",
1714
+ "providerNpm": "@ai-sdk/cloudflare",
1715
+ "reasoning": false,
1716
+ "releaseDate": null,
1717
+ "streamingSupported": true,
1718
+ "structuredOutputs": true,
1719
+ "supportsStructuredOutput": true,
1720
+ "supportsTools": true,
1721
+ "temperature": true,
1722
+ "toolCall": true,
1723
+ "version": null,
1724
+ "vision": true,
1725
+ "icon": "cloudflare"
1726
+ },
1727
+ {
1728
+ "attachment": false,
1729
+ "audioGeneration": false,
1730
+ "cost": {
1731
+ "input": null,
1732
+ "inputCacheHit": null,
1733
+ "output": null
1734
+ },
1735
+ "description": "Multilingual encoder-decoder (seq-to-seq) model trained for Many-to-Many multilingual translation",
1736
+ "id": "m2m100-1.2b",
1737
+ "knowledge": null,
1738
+ "lastUpdated": null,
1739
+ "limit": {
1740
+ "context": null,
1741
+ "output": null
1742
+ },
1743
+ "modalities": {
1744
+ "input": ["text"],
1745
+ "output": ["text"]
1746
+ },
1747
+ "name": "m2m100-1.2b",
1748
+ "openWeights": false,
1749
+ "provider": "Cloudflare",
1750
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1751
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1752
+ "providerId": "cloudflare",
1753
+ "providerNpm": "@ai-sdk/cloudflare",
1754
+ "reasoning": false,
1755
+ "releaseDate": null,
1756
+ "streamingSupported": true,
1757
+ "structuredOutputs": true,
1758
+ "supportsStructuredOutput": true,
1759
+ "supportsTools": true,
1760
+ "temperature": true,
1761
+ "toolCall": true,
1762
+ "version": null,
1763
+ "vision": false,
1764
+ "icon": "cloudflare"
1765
+ },
1766
+ {
1767
+ "attachment": false,
1768
+ "audioGeneration": true,
1769
+ "cost": {
1770
+ "input": null,
1771
+ "inputCacheHit": null,
1772
+ "output": null
1773
+ },
1774
+ "description": "MeloTTS is a high-quality multi-lingual text-to-speech library by MyShell.ai.",
1775
+ "id": "melotts",
1776
+ "knowledge": null,
1777
+ "lastUpdated": null,
1778
+ "limit": {
1779
+ "context": null,
1780
+ "output": null
1781
+ },
1782
+ "modalities": {
1783
+ "input": ["audio"],
1784
+ "output": ["text"]
1785
+ },
1786
+ "name": "melotts",
1787
+ "openWeights": false,
1788
+ "provider": "Cloudflare",
1789
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1790
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1791
+ "providerId": "cloudflare",
1792
+ "providerNpm": "@ai-sdk/cloudflare",
1793
+ "reasoning": false,
1794
+ "releaseDate": null,
1795
+ "streamingSupported": true,
1796
+ "structuredOutputs": true,
1797
+ "supportsStructuredOutput": true,
1798
+ "supportsTools": true,
1799
+ "temperature": true,
1800
+ "toolCall": true,
1801
+ "version": null,
1802
+ "vision": false,
1803
+ "icon": "cloudflare"
1804
+ },
1805
+ {
1806
+ "attachment": false,
1807
+ "audioGeneration": false,
1808
+ "cost": {
1809
+ "input": 0.00014,
1810
+ "inputCacheHit": null,
1811
+ "output": 0.00042
1812
+ },
1813
+ "description": "Instruct fine-tuned version of the Mistral-7b generative text model with 7 billion parameters",
1814
+ "id": "mistral-7b-instruct-v0.1",
1815
+ "knowledge": null,
1816
+ "lastUpdated": null,
1817
+ "limit": {
1818
+ "context": 2824000,
1819
+ "output": null
1820
+ },
1821
+ "modalities": {
1822
+ "input": ["text"],
1823
+ "output": ["text"]
1824
+ },
1825
+ "name": "mistral-7b-instruct-v0.1",
1826
+ "openWeights": false,
1827
+ "provider": "Cloudflare",
1828
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1829
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1830
+ "providerId": "cloudflare",
1831
+ "providerNpm": "@ai-sdk/cloudflare",
1832
+ "reasoning": false,
1833
+ "releaseDate": null,
1834
+ "streamingSupported": true,
1835
+ "structuredOutputs": true,
1836
+ "supportsStructuredOutput": true,
1837
+ "supportsTools": true,
1838
+ "temperature": true,
1839
+ "toolCall": true,
1840
+ "version": null,
1841
+ "vision": false,
1842
+ "icon": "cloudflare"
1843
+ },
1844
+ {
1845
+ "attachment": false,
1846
+ "audioGeneration": false,
1847
+ "cost": {
1848
+ "input": 0.00014,
1849
+ "inputCacheHit": null,
1850
+ "output": 0.00042
1851
+ },
1852
+ "description": "Mistral 7B Instruct v0.1 AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Mistral variant.",
1853
+ "id": "mistral-7b-instruct-v0.1-awq",
1854
+ "knowledge": null,
1855
+ "lastUpdated": null,
1856
+ "limit": {
1857
+ "context": 4096000,
1858
+ "output": null
1859
+ },
1860
+ "modalities": {
1861
+ "input": ["text"],
1862
+ "output": ["text"]
1863
+ },
1864
+ "name": "mistral-7b-instruct-v0.1-awq",
1865
+ "openWeights": false,
1866
+ "provider": "Cloudflare",
1867
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1868
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1869
+ "providerId": "cloudflare",
1870
+ "providerNpm": "@ai-sdk/cloudflare",
1871
+ "reasoning": false,
1872
+ "releaseDate": null,
1873
+ "streamingSupported": true,
1874
+ "structuredOutputs": true,
1875
+ "supportsStructuredOutput": true,
1876
+ "supportsTools": true,
1877
+ "temperature": true,
1878
+ "toolCall": true,
1879
+ "version": null,
1880
+ "vision": false,
1881
+ "icon": "cloudflare"
1882
+ },
1883
+ {
1884
+ "attachment": false,
1885
+ "audioGeneration": false,
1886
+ "cost": {
1887
+ "input": 0.00014,
1888
+ "inputCacheHit": null,
1889
+ "output": 0.00042
1890
+ },
1891
+ "description": "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2. Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1: 32k context window (vs 8k context in v0.1), rope-theta = 1e6, and no Sliding-Window Attention.",
1892
+ "id": "mistral-7b-instruct-v0.2",
1893
+ "knowledge": null,
1894
+ "lastUpdated": null,
1895
+ "limit": {
1896
+ "context": 3072000,
1897
+ "output": null
1898
+ },
1899
+ "modalities": {
1900
+ "input": ["text"],
1901
+ "output": ["text"]
1902
+ },
1903
+ "name": "mistral-7b-instruct-v0.2",
1904
+ "openWeights": false,
1905
+ "provider": "Cloudflare",
1906
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1907
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1908
+ "providerId": "cloudflare",
1909
+ "providerNpm": "@ai-sdk/cloudflare",
1910
+ "reasoning": false,
1911
+ "releaseDate": null,
1912
+ "streamingSupported": true,
1913
+ "structuredOutputs": true,
1914
+ "supportsStructuredOutput": true,
1915
+ "supportsTools": true,
1916
+ "temperature": true,
1917
+ "toolCall": true,
1918
+ "version": null,
1919
+ "vision": false,
1920
+ "icon": "cloudflare"
1921
+ },
1922
+ {
1923
+ "attachment": false,
1924
+ "audioGeneration": false,
1925
+ "cost": {
1926
+ "input": 0.00014,
1927
+ "inputCacheHit": null,
1928
+ "output": 0.00042
1929
+ },
1930
+ "description": "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.",
1931
+ "id": "mistral-7b-instruct-v0.2-lora",
1932
+ "knowledge": null,
1933
+ "lastUpdated": null,
1934
+ "limit": {
1935
+ "context": 15000000,
1936
+ "output": null
1937
+ },
1938
+ "modalities": {
1939
+ "input": ["text"],
1940
+ "output": ["text"]
1941
+ },
1942
+ "name": "mistral-7b-instruct-v0.2-lora",
1943
+ "openWeights": false,
1944
+ "provider": "Cloudflare",
1945
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1946
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1947
+ "providerId": "cloudflare",
1948
+ "providerNpm": "@ai-sdk/cloudflare",
1949
+ "reasoning": false,
1950
+ "releaseDate": null,
1951
+ "streamingSupported": true,
1952
+ "structuredOutputs": true,
1953
+ "supportsStructuredOutput": true,
1954
+ "supportsTools": true,
1955
+ "temperature": true,
1956
+ "toolCall": true,
1957
+ "version": null,
1958
+ "vision": false,
1959
+ "icon": "cloudflare"
1960
+ },
1961
+ {
1962
+ "attachment": false,
1963
+ "audioGeneration": false,
1964
+ "cost": {
1965
+ "input": 0.00014,
1966
+ "inputCacheHit": null,
1967
+ "output": 0.00042
1968
+ },
1969
+ "description": "Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance. With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.",
1970
+ "id": "mistral-small-3.1-24b-instruct",
1971
+ "knowledge": null,
1972
+ "lastUpdated": null,
1973
+ "limit": {
1974
+ "context": 128000,
1975
+ "output": null
1976
+ },
1977
+ "modalities": {
1978
+ "input": ["text", "image"],
1979
+ "output": ["text"]
1980
+ },
1981
+ "name": "mistral-small-3.1-24b-instruct",
1982
+ "openWeights": false,
1983
+ "provider": "Cloudflare",
1984
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1985
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1986
+ "providerId": "cloudflare",
1987
+ "providerNpm": "@ai-sdk/cloudflare",
1988
+ "reasoning": false,
1989
+ "releaseDate": null,
1990
+ "streamingSupported": true,
1991
+ "structuredOutputs": true,
1992
+ "supportsStructuredOutput": true,
1993
+ "supportsTools": true,
1994
+ "temperature": true,
1995
+ "toolCall": true,
1996
+ "version": null,
1997
+ "vision": true,
1998
+ "icon": "cloudflare"
1999
+ },
2000
+ {
2001
+ "attachment": false,
2002
+ "audioGeneration": false,
2003
+ "cost": {
2004
+ "input": null,
2005
+ "inputCacheHit": null,
2006
+ "output": null
2007
+ },
2008
+ "description": "This model is a fine-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the mistralai/Mistral-7B-v0.1 on the open source dataset Open-Orca/SlimOrca.",
2009
+ "id": "neural-chat-7b-v3-1-awq",
2010
+ "knowledge": null,
2011
+ "lastUpdated": null,
2012
+ "limit": {
2013
+ "context": 4096000,
2014
+ "output": null
2015
+ },
2016
+ "modalities": {
2017
+ "input": ["text"],
2018
+ "output": ["text"]
2019
+ },
2020
+ "name": "neural-chat-7b-v3-1-awq",
2021
+ "openWeights": false,
2022
+ "provider": "Cloudflare",
2023
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2024
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2025
+ "providerId": "cloudflare",
2026
+ "providerNpm": "@ai-sdk/cloudflare",
2027
+ "reasoning": false,
2028
+ "releaseDate": null,
2029
+ "streamingSupported": true,
2030
+ "structuredOutputs": true,
2031
+ "supportsStructuredOutput": true,
2032
+ "supportsTools": true,
2033
+ "temperature": true,
2034
+ "toolCall": true,
2035
+ "version": null,
2036
+ "vision": false,
2037
+ "icon": "cloudflare"
2038
+ },
2039
+ {
2040
+ "attachment": false,
2041
+ "audioGeneration": true,
2042
+ "cost": {
2043
+ "input": 5.2e-9,
2044
+ "inputCacheHit": null,
2045
+ "output": 9.2e-9
2046
+ },
2047
+ "description": "Transcribe audio using Deepgram’s speech-to-text model",
2048
+ "id": "nova-3",
2049
+ "knowledge": null,
2050
+ "lastUpdated": null,
2051
+ "limit": {
2052
+ "context": null,
2053
+ "output": null
2054
+ },
2055
+ "modalities": {
2056
+ "input": ["audio"],
2057
+ "output": ["text"]
2058
+ },
2059
+ "name": "nova-3",
2060
+ "openWeights": false,
2061
+ "provider": "Cloudflare",
2062
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2063
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2064
+ "providerId": "cloudflare",
2065
+ "providerNpm": "@ai-sdk/cloudflare",
2066
+ "reasoning": false,
2067
+ "releaseDate": null,
2068
+ "streamingSupported": true,
2069
+ "structuredOutputs": true,
2070
+ "supportsStructuredOutput": true,
2071
+ "supportsTools": true,
2072
+ "temperature": true,
2073
+ "toolCall": true,
2074
+ "version": null,
2075
+ "vision": false,
2076
+ "icon": "cloudflare"
2077
+ },
2078
+ {
2079
+ "attachment": false,
2080
+ "audioGeneration": false,
2081
+ "cost": {
2082
+ "input": null,
2083
+ "inputCacheHit": null,
2084
+ "output": null
2085
+ },
2086
+ "description": "OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning.",
2087
+ "id": "openchat-3.5-0106",
2088
+ "knowledge": null,
2089
+ "lastUpdated": null,
2090
+ "limit": {
2091
+ "context": 8192000,
2092
+ "output": null
2093
+ },
2094
+ "modalities": {
2095
+ "input": ["text"],
2096
+ "output": ["text"]
2097
+ },
2098
+ "name": "openchat-3.5-0106",
2099
+ "openWeights": false,
2100
+ "provider": "Cloudflare",
2101
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2102
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2103
+ "providerId": "cloudflare",
2104
+ "providerNpm": "@ai-sdk/cloudflare",
2105
+ "reasoning": false,
2106
+ "releaseDate": null,
2107
+ "streamingSupported": true,
2108
+ "structuredOutputs": true,
2109
+ "supportsStructuredOutput": true,
2110
+ "supportsTools": true,
2111
+ "temperature": true,
2112
+ "toolCall": true,
2113
+ "version": null,
2114
+ "vision": false,
2115
+ "icon": "cloudflare"
2116
+ },
2117
+ {
2118
+ "attachment": false,
2119
+ "audioGeneration": false,
2120
+ "cost": {
2121
+ "input": 0.0003,
2122
+ "inputCacheHit": null,
2123
+ "output": 0.0003
2124
+ },
2125
+ "description": "OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets.",
2126
+ "id": "openhermes-2.5-mistral-7b-awq",
2127
+ "knowledge": null,
2128
+ "lastUpdated": null,
2129
+ "limit": {
2130
+ "context": 4096000,
2131
+ "output": null
2132
+ },
2133
+ "modalities": {
2134
+ "input": ["text"],
2135
+ "output": ["text"]
2136
+ },
2137
+ "name": "openhermes-2.5-mistral-7b-awq",
2138
+ "openWeights": false,
2139
+ "provider": "Cloudflare",
2140
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2141
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2142
+ "providerId": "cloudflare",
2143
+ "providerNpm": "@ai-sdk/cloudflare",
2144
+ "reasoning": false,
2145
+ "releaseDate": null,
2146
+ "streamingSupported": true,
2147
+ "structuredOutputs": true,
2148
+ "supportsStructuredOutput": true,
2149
+ "supportsTools": true,
2150
+ "temperature": true,
2151
+ "toolCall": true,
2152
+ "version": null,
2153
+ "vision": false,
2154
+ "icon": "cloudflare"
2155
+ },
2156
+ {
2157
+ "attachment": false,
2158
+ "audioGeneration": false,
2159
+ "cost": {
2160
+ "input": 0.0001,
2161
+ "inputCacheHit": null,
2162
+ "output": 0.0001
2163
+ },
2164
+ "description": "Phi-2 is a Transformer-based model with a next-word prediction objective, trained on 1.4T tokens from multiple passes on a mixture of Synthetic and Web datasets for NLP and coding.",
2165
+ "id": "phi-2",
2166
+ "knowledge": null,
2167
+ "lastUpdated": null,
2168
+ "limit": {
2169
+ "context": 2048000,
2170
+ "output": null
2171
+ },
2172
+ "modalities": {
2173
+ "input": ["text"],
2174
+ "output": ["text"]
2175
+ },
2176
+ "name": "phi-2",
2177
+ "openWeights": false,
2178
+ "provider": "Cloudflare",
2179
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2180
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2181
+ "providerId": "cloudflare",
2182
+ "providerNpm": "@ai-sdk/cloudflare",
2183
+ "reasoning": false,
2184
+ "releaseDate": null,
2185
+ "streamingSupported": true,
2186
+ "structuredOutputs": true,
2187
+ "supportsStructuredOutput": true,
2188
+ "supportsTools": true,
2189
+ "temperature": true,
2190
+ "toolCall": true,
2191
+ "version": null,
2192
+ "vision": false,
2193
+ "icon": "cloudflare"
2194
+ },
2195
+ {
2196
+ "attachment": false,
2197
+ "audioGeneration": false,
2198
+ "cost": {
2199
+ "input": null,
2200
+ "inputCacheHit": null,
2201
+ "output": null
2202
+ },
2203
+ "description": "Phoenix 1.0 is a model by Leonardo.Ai that generates images with exceptional prompt adherence and coherent text.",
2204
+ "id": "phoenix-1.0",
2205
+ "knowledge": null,
2206
+ "lastUpdated": null,
2207
+ "limit": {
2208
+ "context": null,
2209
+ "output": null
2210
+ },
2211
+ "modalities": {
2212
+ "input": ["text", "image"],
2213
+ "output": ["text"]
2214
+ },
2215
+ "name": "phoenix-1.0",
2216
+ "openWeights": false,
2217
+ "provider": "Cloudflare",
2218
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2219
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2220
+ "providerId": "cloudflare",
2221
+ "providerNpm": "@ai-sdk/cloudflare",
2222
+ "reasoning": false,
2223
+ "releaseDate": null,
2224
+ "streamingSupported": true,
2225
+ "structuredOutputs": true,
2226
+ "supportsStructuredOutput": true,
2227
+ "supportsTools": true,
2228
+ "temperature": true,
2229
+ "toolCall": true,
2230
+ "version": null,
2231
+ "vision": true,
2232
+ "icon": "cloudflare"
2233
+ },
2234
+ {
2235
+ "attachment": false,
2236
+ "audioGeneration": false,
2237
+ "cost": {
2238
+ "input": null,
2239
+ "inputCacheHit": null,
2240
+ "output": null
2241
+ },
2242
+ "description": "PLaMo-Embedding-1B is a Japanese text embedding model developed by Preferred Networks, Inc.\n\nIt can convert Japanese text input into numerical vectors and can be used for a wide range of applications, including information retrieval, text classification, and clustering.",
2243
+ "id": "plamo-embedding-1b",
2244
+ "knowledge": null,
2245
+ "lastUpdated": null,
2246
+ "limit": {
2247
+ "context": null,
2248
+ "output": null
2249
+ },
2250
+ "modalities": {
2251
+ "input": ["text"],
2252
+ "output": ["embedding"]
2253
+ },
2254
+ "name": "plamo-embedding-1b",
2255
+ "openWeights": false,
2256
+ "provider": "Cloudflare",
2257
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2258
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2259
+ "providerId": "cloudflare",
2260
+ "providerNpm": "@ai-sdk/cloudflare",
2261
+ "reasoning": false,
2262
+ "releaseDate": null,
2263
+ "streamingSupported": true,
2264
+ "structuredOutputs": true,
2265
+ "supportsStructuredOutput": true,
2266
+ "supportsTools": true,
2267
+ "temperature": true,
2268
+ "toolCall": true,
2269
+ "version": null,
2270
+ "vision": false,
2271
+ "icon": "cloudflare"
2272
+ },
2273
+ {
2274
+ "attachment": false,
2275
+ "audioGeneration": false,
2276
+ "cost": {
2277
+ "input": 0.0001,
2278
+ "inputCacheHit": null,
2279
+ "output": 0.0001
2280
+ },
2281
+ "description": "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.",
2282
+ "id": "qwen1.5-0.5b-chat",
2283
+ "knowledge": null,
2284
+ "lastUpdated": null,
2285
+ "limit": {
2286
+ "context": 32000000,
2287
+ "output": null
2288
+ },
2289
+ "modalities": {
2290
+ "input": ["text"],
2291
+ "output": ["text"]
2292
+ },
2293
+ "name": "qwen1.5-0.5b-chat",
2294
+ "openWeights": false,
2295
+ "provider": "Cloudflare",
2296
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2297
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2298
+ "providerId": "cloudflare",
2299
+ "providerNpm": "@ai-sdk/cloudflare",
2300
+ "reasoning": false,
2301
+ "releaseDate": null,
2302
+ "streamingSupported": true,
2303
+ "structuredOutputs": true,
2304
+ "supportsStructuredOutput": true,
2305
+ "supportsTools": true,
2306
+ "temperature": true,
2307
+ "toolCall": true,
2308
+ "version": null,
2309
+ "vision": false,
2310
+ "icon": "cloudflare"
2311
+ },
2312
+ {
2313
+ "attachment": false,
2314
+ "audioGeneration": false,
2315
+ "cost": {
2316
+ "input": null,
2317
+ "inputCacheHit": null,
2318
+ "output": null
2319
+ },
2320
+ "description": "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",
2321
+ "id": "qwen1.5-14b-chat-awq",
2322
+ "knowledge": null,
2323
+ "lastUpdated": null,
2324
+ "limit": {
2325
+ "context": 7500000,
2326
+ "output": null
2327
+ },
2328
+ "modalities": {
2329
+ "input": ["text"],
2330
+ "output": ["text"]
2331
+ },
2332
+ "name": "qwen1.5-14b-chat-awq",
2333
+ "openWeights": false,
2334
+ "provider": "Cloudflare",
2335
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2336
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2337
+ "providerId": "cloudflare",
2338
+ "providerNpm": "@ai-sdk/cloudflare",
2339
+ "reasoning": false,
2340
+ "releaseDate": null,
2341
+ "streamingSupported": true,
2342
+ "structuredOutputs": true,
2343
+ "supportsStructuredOutput": true,
2344
+ "supportsTools": true,
2345
+ "temperature": true,
2346
+ "toolCall": true,
2347
+ "version": null,
2348
+ "vision": false,
2349
+ "icon": "cloudflare"
2350
+ },
2351
+ {
2352
+ "attachment": false,
2353
+ "audioGeneration": false,
2354
+ "cost": {
2355
+ "input": 0.0001,
2356
+ "inputCacheHit": null,
2357
+ "output": 0.0001
2358
+ },
2359
+ "description": "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.",
2360
+ "id": "qwen1.5-1.8b-chat",
2361
+ "knowledge": null,
2362
+ "lastUpdated": null,
2363
+ "limit": {
2364
+ "context": 32000000,
2365
+ "output": null
2366
+ },
2367
+ "modalities": {
2368
+ "input": ["text"],
2369
+ "output": ["text"]
2370
+ },
2371
+ "name": "qwen1.5-1.8b-chat",
2372
+ "openWeights": false,
2373
+ "provider": "Cloudflare",
2374
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2375
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2376
+ "providerId": "cloudflare",
2377
+ "providerNpm": "@ai-sdk/cloudflare",
2378
+ "reasoning": false,
2379
+ "releaseDate": null,
2380
+ "streamingSupported": true,
2381
+ "structuredOutputs": true,
2382
+ "supportsStructuredOutput": true,
2383
+ "supportsTools": true,
2384
+ "temperature": true,
2385
+ "toolCall": true,
2386
+ "version": null,
2387
+ "vision": false,
2388
+ "icon": "cloudflare"
2389
+ },
2390
+ {
2391
+ "attachment": false,
2392
+ "audioGeneration": false,
2393
+ "cost": {
2394
+ "input": null,
2395
+ "inputCacheHit": null,
2396
+ "output": null
2397
+ },
2398
+ "description": "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",
2399
+ "id": "qwen1.5-7b-chat-awq",
2400
+ "knowledge": null,
2401
+ "lastUpdated": null,
2402
+ "limit": {
2403
+ "context": 20000000,
2404
+ "output": null
2405
+ },
2406
+ "modalities": {
2407
+ "input": ["text"],
2408
+ "output": ["text"]
2409
+ },
2410
+ "name": "qwen1.5-7b-chat-awq",
2411
+ "openWeights": false,
2412
+ "provider": "Cloudflare",
2413
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2414
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2415
+ "providerId": "cloudflare",
2416
+ "providerNpm": "@ai-sdk/cloudflare",
2417
+ "reasoning": false,
2418
+ "releaseDate": null,
2419
+ "streamingSupported": true,
2420
+ "structuredOutputs": true,
2421
+ "supportsStructuredOutput": true,
2422
+ "supportsTools": true,
2423
+ "temperature": true,
2424
+ "toolCall": true,
2425
+ "version": null,
2426
+ "vision": false,
2427
+ "icon": "cloudflare"
2428
+ },
2429
+ {
2430
+ "attachment": false,
2431
+ "audioGeneration": false,
2432
+ "cost": {
2433
+ "input": 0.0008,
2434
+ "inputCacheHit": null,
2435
+ "output": 0.0008
2436
+ },
2437
+ "description": "Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:",
2438
+ "extendedThinking": false,
2439
+ "id": "qwen2.5-coder-32b-instruct",
2440
+ "knowledge": null,
2441
+ "lastUpdated": null,
2442
+ "limit": {
2443
+ "context": 32768000,
2444
+ "output": 32768
2445
+ },
2446
+ "modalities": {
2447
+ "input": ["text"],
2448
+ "output": ["text"]
2449
+ },
2450
+ "name": "qwen2.5-coder-32b-instruct",
2451
+ "openWeights": false,
2452
+ "provider": "Cloudflare",
2453
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2454
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2455
+ "providerId": "cloudflare",
2456
+ "providerNpm": "@ai-sdk/cloudflare",
2457
+ "reasoning": false,
2458
+ "releaseDate": "2024-11-11",
2459
+ "streamingSupported": true,
2460
+ "structuredOutputs": true,
2461
+ "supportsStructuredOutput": true,
2462
+ "supportsTools": true,
2463
+ "temperature": true,
2464
+ "toolCall": true,
2465
+ "version": null,
2466
+ "vision": false,
2467
+ "icon": "cloudflare"
2468
+ },
2469
+ {
2470
+ "attachment": false,
2471
+ "audioGeneration": false,
2472
+ "cost": {
2473
+ "input": null,
2474
+ "inputCacheHit": null,
2475
+ "output": null
2476
+ },
2477
+ "description": "Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support.",
2478
+ "id": "qwen3-30b-a3b-fp8",
2479
+ "knowledge": null,
2480
+ "lastUpdated": null,
2481
+ "limit": {
2482
+ "context": 32768000,
2483
+ "output": null
2484
+ },
2485
+ "modalities": {
2486
+ "input": ["text"],
2487
+ "output": ["text"]
2488
+ },
2489
+ "name": "qwen3-30b-a3b-fp8",
2490
+ "openWeights": false,
2491
+ "provider": "Cloudflare",
2492
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2493
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2494
+ "providerId": "cloudflare",
2495
+ "providerNpm": "@ai-sdk/cloudflare",
2496
+ "reasoning": false,
2497
+ "releaseDate": null,
2498
+ "streamingSupported": true,
2499
+ "structuredOutputs": true,
2500
+ "supportsStructuredOutput": true,
2501
+ "supportsTools": true,
2502
+ "temperature": true,
2503
+ "toolCall": true,
2504
+ "version": null,
2505
+ "vision": false,
2506
+ "icon": "cloudflare"
2507
+ },
2508
+ {
2509
+ "attachment": false,
2510
+ "audioGeneration": false,
2511
+ "cost": {
2512
+ "input": null,
2513
+ "inputCacheHit": null,
2514
+ "output": null
2515
+ },
2516
+ "description": "The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks.",
2517
+ "id": "qwen3-embedding-0.6b",
2518
+ "knowledge": null,
2519
+ "lastUpdated": null,
2520
+ "limit": {
2521
+ "context": null,
2522
+ "output": null
2523
+ },
2524
+ "modalities": {
2525
+ "input": ["text"],
2526
+ "output": ["embedding"]
2527
+ },
2528
+ "name": "qwen3-embedding-0.6b",
2529
+ "openWeights": false,
2530
+ "provider": "Cloudflare",
2531
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2532
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2533
+ "providerId": "cloudflare",
2534
+ "providerNpm": "@ai-sdk/cloudflare",
2535
+ "reasoning": false,
2536
+ "releaseDate": null,
2537
+ "streamingSupported": true,
2538
+ "structuredOutputs": true,
2539
+ "supportsStructuredOutput": true,
2540
+ "supportsTools": true,
2541
+ "temperature": true,
2542
+ "toolCall": true,
2543
+ "version": null,
2544
+ "vision": false,
2545
+ "icon": "cloudflare"
2546
+ },
2547
+ {
2548
+ "attachment": false,
2549
+ "audioGeneration": false,
2550
+ "cost": {
2551
+ "input": null,
2552
+ "inputCacheHit": null,
2553
+ "output": null
2554
+ },
2555
+ "description": "50 layers deep image classification CNN trained on more than 1M images from ImageNet",
2556
+ "id": "resnet-50",
2557
+ "knowledge": null,
2558
+ "lastUpdated": null,
2559
+ "limit": {
2560
+ "context": null,
2561
+ "output": null
2562
+ },
2563
+ "modalities": {
2564
+ "input": ["text", "image"],
2565
+ "output": ["text"]
2566
+ },
2567
+ "name": "resnet-50",
2568
+ "openWeights": false,
2569
+ "provider": "Cloudflare",
2570
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2571
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2572
+ "providerId": "cloudflare",
2573
+ "providerNpm": "@ai-sdk/cloudflare",
2574
+ "reasoning": false,
2575
+ "releaseDate": null,
2576
+ "streamingSupported": true,
2577
+ "structuredOutputs": true,
2578
+ "supportsStructuredOutput": true,
2579
+ "supportsTools": true,
2580
+ "temperature": true,
2581
+ "toolCall": true,
2582
+ "version": null,
2583
+ "vision": true,
2584
+ "icon": "cloudflare"
2585
+ },
2586
+ {
2587
+ "attachment": false,
2588
+ "audioGeneration": true,
2589
+ "cost": {
2590
+ "input": null,
2591
+ "inputCacheHit": null,
2592
+ "output": null
2593
+ },
2594
+ "description": "An open source, community-driven, native audio turn detection model in 2nd version",
2595
+ "id": "smart-turn-v2",
2596
+ "knowledge": null,
2597
+ "lastUpdated": null,
2598
+ "limit": {
2599
+ "context": null,
2600
+ "output": null
2601
+ },
2602
+ "modalities": {
2603
+ "input": ["text"],
2604
+ "output": ["audio"]
2605
+ },
2606
+ "name": "smart-turn-v2",
2607
+ "openWeights": false,
2608
+ "provider": "Cloudflare",
2609
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2610
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2611
+ "providerId": "cloudflare",
2612
+ "providerNpm": "@ai-sdk/cloudflare",
2613
+ "reasoning": false,
2614
+ "releaseDate": null,
2615
+ "streamingSupported": true,
2616
+ "structuredOutputs": true,
2617
+ "supportsStructuredOutput": true,
2618
+ "supportsTools": true,
2619
+ "temperature": true,
2620
+ "toolCall": true,
2621
+ "version": null,
2622
+ "vision": false,
2623
+ "icon": "cloudflare"
2624
+ },
2625
+ {
2626
+ "attachment": false,
2627
+ "audioGeneration": false,
2628
+ "cost": {
2629
+ "input": null,
2630
+ "inputCacheHit": null,
2631
+ "output": null
2632
+ },
2633
+ "description": "This model is intended to be used by non-technical users to understand data inside their SQL databases.",
2634
+ "id": "sqlcoder-7b-2",
2635
+ "knowledge": null,
2636
+ "lastUpdated": null,
2637
+ "limit": {
2638
+ "context": 10000000,
2639
+ "output": null
2640
+ },
2641
+ "modalities": {
2642
+ "input": ["text"],
2643
+ "output": ["text"]
2644
+ },
2645
+ "name": "sqlcoder-7b-2",
2646
+ "openWeights": false,
2647
+ "provider": "Cloudflare",
2648
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2649
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2650
+ "providerId": "cloudflare",
2651
+ "providerNpm": "@ai-sdk/cloudflare",
2652
+ "reasoning": false,
2653
+ "releaseDate": null,
2654
+ "streamingSupported": true,
2655
+ "structuredOutputs": true,
2656
+ "supportsStructuredOutput": true,
2657
+ "supportsTools": true,
2658
+ "temperature": true,
2659
+ "toolCall": true,
2660
+ "version": null,
2661
+ "vision": false,
2662
+ "icon": "cloudflare"
2663
+ },
2664
+ {
2665
+ "attachment": false,
2666
+ "audioGeneration": false,
2667
+ "cost": {
2668
+ "input": null,
2669
+ "inputCacheHit": null,
2670
+ "output": null
2671
+ },
2672
+ "description": "Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images. Img2img generate a new image from an input image with Stable Diffusion.",
2673
+ "id": "stable-diffusion-v1-5-img2img",
2674
+ "knowledge": null,
2675
+ "lastUpdated": null,
2676
+ "limit": {
2677
+ "context": null,
2678
+ "output": null
2679
+ },
2680
+ "modalities": {
2681
+ "input": ["text", "image"],
2682
+ "output": ["text"]
2683
+ },
2684
+ "name": "stable-diffusion-v1-5-img2img",
2685
+ "openWeights": false,
2686
+ "provider": "Cloudflare",
2687
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2688
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2689
+ "providerId": "cloudflare",
2690
+ "providerNpm": "@ai-sdk/cloudflare",
2691
+ "reasoning": false,
2692
+ "releaseDate": null,
2693
+ "streamingSupported": true,
2694
+ "structuredOutputs": true,
2695
+ "supportsStructuredOutput": true,
2696
+ "supportsTools": true,
2697
+ "temperature": true,
2698
+ "toolCall": true,
2699
+ "version": null,
2700
+ "vision": true,
2701
+ "icon": "cloudflare"
2702
+ },
2703
+ {
2704
+ "attachment": false,
2705
+ "audioGeneration": false,
2706
+ "cost": {
2707
+ "input": null,
2708
+ "inputCacheHit": null,
2709
+ "output": null
2710
+ },
2711
+ "description": "Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.",
2712
+ "id": "stable-diffusion-v1-5-inpainting",
2713
+ "knowledge": null,
2714
+ "lastUpdated": null,
2715
+ "limit": {
2716
+ "context": null,
2717
+ "output": null
2718
+ },
2719
+ "modalities": {
2720
+ "input": ["text", "image"],
2721
+ "output": ["text"]
2722
+ },
2723
+ "name": "stable-diffusion-v1-5-inpainting",
2724
+ "openWeights": false,
2725
+ "provider": "Cloudflare",
2726
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2727
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2728
+ "providerId": "cloudflare",
2729
+ "providerNpm": "@ai-sdk/cloudflare",
2730
+ "reasoning": false,
2731
+ "releaseDate": null,
2732
+ "streamingSupported": true,
2733
+ "structuredOutputs": true,
2734
+ "supportsStructuredOutput": true,
2735
+ "supportsTools": true,
2736
+ "temperature": true,
2737
+ "toolCall": true,
2738
+ "version": null,
2739
+ "vision": true,
2740
+ "icon": "cloudflare"
2741
+ },
2742
+ {
2743
+ "attachment": false,
2744
+ "audioGeneration": false,
2745
+ "cost": {
2746
+ "input": null,
2747
+ "inputCacheHit": null,
2748
+ "output": null
2749
+ },
2750
+ "description": "Diffusion-based text-to-image generative model by Stability AI. Generates and modify images based on text prompts.",
2751
+ "id": "stable-diffusion-xl-base-1.0",
2752
+ "knowledge": null,
2753
+ "lastUpdated": null,
2754
+ "limit": {
2755
+ "context": null,
2756
+ "output": null
2757
+ },
2758
+ "modalities": {
2759
+ "input": ["text", "image"],
2760
+ "output": ["text"]
2761
+ },
2762
+ "name": "stable-diffusion-xl-base-1.0",
2763
+ "openWeights": false,
2764
+ "provider": "Cloudflare",
2765
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2766
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2767
+ "providerId": "cloudflare",
2768
+ "providerNpm": "@ai-sdk/cloudflare",
2769
+ "reasoning": false,
2770
+ "releaseDate": null,
2771
+ "streamingSupported": true,
2772
+ "structuredOutputs": true,
2773
+ "supportsStructuredOutput": true,
2774
+ "supportsTools": true,
2775
+ "temperature": true,
2776
+ "toolCall": true,
2777
+ "version": null,
2778
+ "vision": true,
2779
+ "icon": "cloudflare"
2780
+ },
2781
+ {
2782
+ "attachment": false,
2783
+ "audioGeneration": false,
2784
+ "cost": {
2785
+ "input": null,
2786
+ "inputCacheHit": null,
2787
+ "output": null
2788
+ },
2789
+ "description": "SDXL-Lightning is a lightning-fast text-to-image generation model. It can generate high-quality 1024px images in a few steps.",
2790
+ "id": "stable-diffusion-xl-lightning",
2791
+ "knowledge": null,
2792
+ "lastUpdated": null,
2793
+ "limit": {
2794
+ "context": null,
2795
+ "output": null
2796
+ },
2797
+ "modalities": {
2798
+ "input": ["text", "image"],
2799
+ "output": ["text"]
2800
+ },
2801
+ "name": "stable-diffusion-xl-lightning",
2802
+ "openWeights": false,
2803
+ "provider": "Cloudflare",
2804
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2805
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2806
+ "providerId": "cloudflare",
2807
+ "providerNpm": "@ai-sdk/cloudflare",
2808
+ "reasoning": false,
2809
+ "releaseDate": null,
2810
+ "streamingSupported": true,
2811
+ "structuredOutputs": true,
2812
+ "supportsStructuredOutput": true,
2813
+ "supportsTools": true,
2814
+ "temperature": true,
2815
+ "toolCall": true,
2816
+ "version": null,
2817
+ "vision": true,
2818
+ "icon": "cloudflare"
2819
+ },
2820
+ {
2821
+ "attachment": false,
2822
+ "audioGeneration": false,
2823
+ "cost": {
2824
+ "input": null,
2825
+ "inputCacheHit": null,
2826
+ "output": null
2827
+ },
2828
+ "description": "We introduce Starling-LM-7B-beta, an open large language model (LLM) trained by Reinforcement Learning from AI Feedback (RLAIF). Starling-LM-7B-beta is trained from Openchat-3.5-0106 with our new reward model Nexusflow/Starling-RM-34B and policy optimization method Fine-Tuning Language Models from Human Preferences (PPO).",
2829
+ "id": "starling-lm-7b-beta",
2830
+ "knowledge": null,
2831
+ "lastUpdated": null,
2832
+ "limit": {
2833
+ "context": 4096000,
2834
+ "output": null
2835
+ },
2836
+ "modalities": {
2837
+ "input": ["text"],
2838
+ "output": ["text"]
2839
+ },
2840
+ "name": "starling-lm-7b-beta",
2841
+ "openWeights": false,
2842
+ "provider": "Cloudflare",
2843
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2844
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2845
+ "providerId": "cloudflare",
2846
+ "providerNpm": "@ai-sdk/cloudflare",
2847
+ "reasoning": false,
2848
+ "releaseDate": null,
2849
+ "streamingSupported": true,
2850
+ "structuredOutputs": true,
2851
+ "supportsStructuredOutput": true,
2852
+ "supportsTools": true,
2853
+ "temperature": true,
2854
+ "toolCall": true,
2855
+ "version": null,
2856
+ "vision": false,
2857
+ "icon": "cloudflare"
2858
+ },
2859
+ {
2860
+ "attachment": false,
2861
+ "audioGeneration": false,
2862
+ "cost": {
2863
+ "input": 0.0003,
2864
+ "inputCacheHit": null,
2865
+ "output": 0.0003
2866
+ },
2867
+ "description": "The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T.",
2868
+ "id": "tinyllama-1.1b-chat-v1.0",
2869
+ "knowledge": null,
2870
+ "lastUpdated": null,
2871
+ "limit": {
2872
+ "context": 2048000,
2873
+ "output": null
2874
+ },
2875
+ "modalities": {
2876
+ "input": ["text"],
2877
+ "output": ["text"]
2878
+ },
2879
+ "name": "tinyllama-1.1b-chat-v1.0",
2880
+ "openWeights": false,
2881
+ "provider": "Cloudflare",
2882
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2883
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2884
+ "providerId": "cloudflare",
2885
+ "providerNpm": "@ai-sdk/cloudflare",
2886
+ "reasoning": false,
2887
+ "releaseDate": null,
2888
+ "streamingSupported": true,
2889
+ "structuredOutputs": true,
2890
+ "supportsStructuredOutput": true,
2891
+ "supportsTools": true,
2892
+ "temperature": true,
2893
+ "toolCall": true,
2894
+ "version": null,
2895
+ "vision": false,
2896
+ "icon": "cloudflare"
2897
+ },
2898
+ {
2899
+ "attachment": false,
2900
+ "audioGeneration": false,
2901
+ "cost": {
2902
+ "input": null,
2903
+ "inputCacheHit": null,
2904
+ "output": null
2905
+ },
2906
+ "description": "UForm-Gen is a small generative vision-language model primarily designed for Image Captioning and Visual Question Answering. The model was pre-trained on the internal image captioning dataset and fine-tuned on public instructions datasets: SVIT, LVIS, VQAs datasets.",
2907
+ "id": "uform-gen2-qwen-500m",
2908
+ "knowledge": null,
2909
+ "lastUpdated": null,
2910
+ "limit": {
2911
+ "context": null,
2912
+ "output": null
2913
+ },
2914
+ "modalities": {
2915
+ "input": ["text", "image"],
2916
+ "output": ["text"]
2917
+ },
2918
+ "name": "uform-gen2-qwen-500m",
2919
+ "openWeights": false,
2920
+ "provider": "Cloudflare",
2921
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2922
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2923
+ "providerId": "cloudflare",
2924
+ "providerNpm": "@ai-sdk/cloudflare",
2925
+ "reasoning": false,
2926
+ "releaseDate": null,
2927
+ "streamingSupported": true,
2928
+ "structuredOutputs": true,
2929
+ "supportsStructuredOutput": true,
2930
+ "supportsTools": true,
2931
+ "temperature": true,
2932
+ "toolCall": true,
2933
+ "version": null,
2934
+ "vision": true,
2935
+ "icon": "cloudflare"
2936
+ },
2937
+ {
2938
+ "attachment": false,
2939
+ "audioGeneration": false,
2940
+ "cost": {
2941
+ "input": null,
2942
+ "inputCacheHit": null,
2943
+ "output": null
2944
+ },
2945
+ "description": "Cybertron 7B v2 is a 7B MistralAI based model, best on it's series. It was trained with SFT, DPO and UNA (Unified Neural Alignment) on multiple datasets.",
2946
+ "id": "una-cybertron-7b-v2-bf16",
2947
+ "knowledge": null,
2948
+ "lastUpdated": null,
2949
+ "limit": {
2950
+ "context": 15000000,
2951
+ "output": null
2952
+ },
2953
+ "modalities": {
2954
+ "input": ["text"],
2955
+ "output": ["text"]
2956
+ },
2957
+ "name": "una-cybertron-7b-v2-bf16",
2958
+ "openWeights": false,
2959
+ "provider": "Cloudflare",
2960
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2961
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2962
+ "providerId": "cloudflare",
2963
+ "providerNpm": "@ai-sdk/cloudflare",
2964
+ "reasoning": false,
2965
+ "releaseDate": null,
2966
+ "streamingSupported": true,
2967
+ "structuredOutputs": true,
2968
+ "supportsStructuredOutput": true,
2969
+ "supportsTools": true,
2970
+ "temperature": true,
2971
+ "toolCall": true,
2972
+ "version": null,
2973
+ "vision": false,
2974
+ "icon": "cloudflare"
2975
+ },
2976
+ {
2977
+ "attachment": false,
2978
+ "audioGeneration": true,
2979
+ "cost": {
2980
+ "input": null,
2981
+ "inputCacheHit": null,
2982
+ "output": null
2983
+ },
2984
+ "description": "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation.",
2985
+ "id": "whisper-large-v3-turbo",
2986
+ "knowledge": null,
2987
+ "lastUpdated": null,
2988
+ "limit": {
2989
+ "context": null,
2990
+ "output": null
2991
+ },
2992
+ "modalities": {
2993
+ "input": ["audio"],
2994
+ "output": ["text"]
2995
+ },
2996
+ "name": "whisper-large-v3-turbo",
2997
+ "openWeights": false,
2998
+ "provider": "Cloudflare",
2999
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
3000
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
3001
+ "providerId": "cloudflare",
3002
+ "providerNpm": "@ai-sdk/cloudflare",
3003
+ "reasoning": false,
3004
+ "releaseDate": null,
3005
+ "streamingSupported": true,
3006
+ "structuredOutputs": true,
3007
+ "supportsStructuredOutput": true,
3008
+ "supportsTools": true,
3009
+ "temperature": true,
3010
+ "toolCall": true,
3011
+ "version": null,
3012
+ "vision": false,
3013
+ "icon": "cloudflare"
3014
+ },
3015
+ {
3016
+ "attachment": false,
3017
+ "audioGeneration": true,
3018
+ "cost": {
3019
+ "input": null,
3020
+ "inputCacheHit": null,
3021
+ "output": null
3022
+ },
3023
+ "description": "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalize to many datasets and domains without the need for fine-tuning. This is the English-only version of the Whisper Tiny model which was trained on the task of speech recognition.",
3024
+ "id": "whisper-tiny-en",
3025
+ "knowledge": null,
3026
+ "lastUpdated": null,
3027
+ "limit": {
3028
+ "context": null,
3029
+ "output": null
3030
+ },
3031
+ "modalities": {
3032
+ "input": ["audio"],
3033
+ "output": ["text"]
3034
+ },
3035
+ "name": "whisper-tiny-en",
3036
+ "openWeights": false,
3037
+ "provider": "Cloudflare",
3038
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
3039
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
3040
+ "providerId": "cloudflare",
3041
+ "providerNpm": "@ai-sdk/cloudflare",
3042
+ "reasoning": false,
3043
+ "releaseDate": null,
3044
+ "streamingSupported": true,
3045
+ "structuredOutputs": true,
3046
+ "supportsStructuredOutput": true,
3047
+ "supportsTools": true,
3048
+ "temperature": true,
3049
+ "toolCall": true,
3050
+ "version": null,
3051
+ "vision": false,
3052
+ "icon": "cloudflare"
3053
+ },
3054
+ {
3055
+ "attachment": false,
3056
+ "audioGeneration": false,
3057
+ "cost": {
3058
+ "input": null,
3059
+ "inputCacheHit": null,
3060
+ "output": null
3061
+ },
3062
+ "description": "Zephyr 7B Beta AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Zephyr model variant.",
3063
+ "id": "zephyr-7b-beta-awq",
3064
+ "knowledge": null,
3065
+ "lastUpdated": null,
3066
+ "limit": {
3067
+ "context": 4096000,
3068
+ "output": null
3069
+ },
3070
+ "modalities": {
3071
+ "input": ["text"],
3072
+ "output": ["text"]
3073
+ },
3074
+ "name": "zephyr-7b-beta-awq",
3075
+ "openWeights": false,
3076
+ "provider": "Cloudflare",
3077
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
3078
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
3079
+ "providerId": "cloudflare",
3080
+ "providerNpm": "@ai-sdk/cloudflare",
3081
+ "reasoning": false,
3082
+ "releaseDate": null,
3083
+ "streamingSupported": true,
3084
+ "structuredOutputs": true,
3085
+ "supportsStructuredOutput": true,
3086
+ "supportsTools": true,
3087
+ "temperature": true,
3088
+ "toolCall": true,
3089
+ "version": null,
3090
+ "vision": false,
3091
+ "icon": "cloudflare"
3092
+ }
3093
+ ]
3094
+ }