@anolilab/ai-model-registry 1.3.0 → 2.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (149) hide show
  1. package/CHANGELOG.md +62 -0
  2. package/LICENSE.md +8 -13
  3. package/MIGRATION.md +474 -0
  4. package/README.md +275 -200
  5. package/dist/icons-sprite.d.ts +1 -1
  6. package/dist/icons-sprite.js +11 -11
  7. package/dist/index.d.ts +16 -16
  8. package/dist/index.js +1 -16
  9. package/dist/packem_chunks/alibaba.js +1 -0
  10. package/dist/packem_chunks/amazon-bedrock.js +1 -0
  11. package/dist/packem_chunks/anthropic.js +1 -0
  12. package/dist/packem_chunks/api.js +75 -0
  13. package/dist/packem_chunks/azure-open-ai.js +1 -0
  14. package/dist/packem_chunks/cerebras.js +1 -0
  15. package/dist/packem_chunks/chutes.js +1 -0
  16. package/dist/packem_chunks/cloudflare.js +3 -0
  17. package/dist/packem_chunks/deep-infra.js +1 -0
  18. package/dist/packem_chunks/deep-seek.js +1 -0
  19. package/dist/packem_chunks/fireworks-ai.js +1 -0
  20. package/dist/packem_chunks/git-hub-copilot.js +1 -0
  21. package/dist/packem_chunks/git-hub-models.js +1 -0
  22. package/dist/packem_chunks/google-partner.js +1 -0
  23. package/dist/packem_chunks/google-vertex.js +1 -0
  24. package/dist/packem_chunks/google.js +1 -0
  25. package/dist/packem_chunks/groq.js +1 -0
  26. package/dist/packem_chunks/hugging-face.js +1 -0
  27. package/dist/packem_chunks/inception.js +1 -0
  28. package/dist/packem_chunks/inference.js +1 -0
  29. package/dist/packem_chunks/meta.js +1 -0
  30. package/dist/packem_chunks/mistral.js +1 -0
  31. package/dist/packem_chunks/model-scope.js +1 -0
  32. package/dist/packem_chunks/morph.js +1 -0
  33. package/dist/packem_chunks/open-ai.js +1 -0
  34. package/dist/packem_chunks/open-router.js +1 -0
  35. package/dist/packem_chunks/providers.js +1 -0
  36. package/dist/packem_chunks/requesty.js +73 -0
  37. package/dist/packem_chunks/together-ai.js +1 -0
  38. package/dist/packem_chunks/upstage.js +1 -0
  39. package/dist/packem_chunks/v0.js +1 -0
  40. package/dist/packem_chunks/venice.js +1 -0
  41. package/dist/packem_chunks/vercel.js +1 -0
  42. package/dist/packem_chunks/weights-_-biases.js +1 -0
  43. package/dist/packem_chunks/xai.js +1 -0
  44. package/dist/schema.d.ts +1 -1
  45. package/dist/schema.js +1 -1
  46. package/dist/types/alibaba.d.ts +6 -0
  47. package/dist/types/alibaba.js +0 -0
  48. package/dist/types/amazon-bedrock.d.ts +6 -0
  49. package/dist/types/amazon-bedrock.js +0 -0
  50. package/dist/types/anthropic.d.ts +6 -0
  51. package/dist/types/anthropic.js +0 -0
  52. package/dist/types/azure-open-ai.d.ts +6 -0
  53. package/dist/types/azure-open-ai.js +0 -0
  54. package/dist/types/cerebras.d.ts +6 -0
  55. package/dist/types/cerebras.js +0 -0
  56. package/dist/types/chutes.d.ts +6 -0
  57. package/dist/types/chutes.js +0 -0
  58. package/dist/types/cloudflare.d.ts +6 -0
  59. package/dist/types/cloudflare.js +0 -0
  60. package/dist/types/deep-infra.d.ts +6 -0
  61. package/dist/types/deep-infra.js +0 -0
  62. package/dist/types/deep-seek.d.ts +6 -0
  63. package/dist/types/deep-seek.js +0 -0
  64. package/dist/types/fireworks-ai.d.ts +6 -0
  65. package/dist/types/fireworks-ai.js +0 -0
  66. package/dist/types/git-hub-copilot.d.ts +6 -0
  67. package/dist/types/git-hub-copilot.js +0 -0
  68. package/dist/types/git-hub-models.d.ts +6 -0
  69. package/dist/types/git-hub-models.js +0 -0
  70. package/dist/types/google-partner.d.ts +6 -0
  71. package/dist/types/google-partner.js +0 -0
  72. package/dist/types/google-vertex.d.ts +6 -0
  73. package/dist/types/google-vertex.js +0 -0
  74. package/dist/types/google.d.ts +6 -0
  75. package/dist/types/google.js +0 -0
  76. package/dist/types/groq.d.ts +6 -0
  77. package/dist/types/groq.js +0 -0
  78. package/dist/types/hugging-face.d.ts +6 -0
  79. package/dist/types/hugging-face.js +0 -0
  80. package/dist/types/inception.d.ts +6 -0
  81. package/dist/types/inception.js +0 -0
  82. package/dist/types/inference.d.ts +6 -0
  83. package/dist/types/inference.js +0 -0
  84. package/dist/types/meta.d.ts +6 -0
  85. package/dist/types/meta.js +0 -0
  86. package/dist/types/mistral.d.ts +6 -0
  87. package/dist/types/mistral.js +0 -0
  88. package/dist/types/model-scope.d.ts +6 -0
  89. package/dist/types/model-scope.js +0 -0
  90. package/dist/types/morph.d.ts +6 -0
  91. package/dist/types/morph.js +0 -0
  92. package/dist/types/open-ai.d.ts +6 -0
  93. package/dist/types/open-ai.js +0 -0
  94. package/dist/types/open-router.d.ts +6 -0
  95. package/dist/types/open-router.js +0 -0
  96. package/dist/types/providers.d.ts +11 -0
  97. package/dist/types/providers.js +1 -0
  98. package/dist/types/requesty.d.ts +6 -0
  99. package/dist/types/requesty.js +0 -0
  100. package/dist/types/together-ai.d.ts +6 -0
  101. package/dist/types/together-ai.js +0 -0
  102. package/dist/types/upstage.d.ts +6 -0
  103. package/dist/types/upstage.js +0 -0
  104. package/dist/types/v0.d.ts +6 -0
  105. package/dist/types/v0.js +0 -0
  106. package/dist/types/venice.d.ts +6 -0
  107. package/dist/types/venice.js +0 -0
  108. package/dist/types/vercel.d.ts +6 -0
  109. package/dist/types/vercel.js +0 -0
  110. package/dist/types/weights-_-biases.d.ts +6 -0
  111. package/dist/types/weights-_-biases.js +0 -0
  112. package/dist/types/xai.d.ts +6 -0
  113. package/dist/types/xai.js +0 -0
  114. package/package.json +7 -2
  115. package/public/alibaba.json +326 -0
  116. package/public/amazon-bedrock.json +2141 -0
  117. package/public/anthropic.json +396 -0
  118. package/public/api.json +66961 -19318
  119. package/public/azure-open-ai.json +332 -0
  120. package/public/cerebras.json +288 -0
  121. package/public/chutes.json +3220 -0
  122. package/public/cloudflare.json +3094 -0
  123. package/public/deep-infra.json +956 -0
  124. package/public/deep-seek.json +46 -0
  125. package/public/fireworks-ai.json +14486 -0
  126. package/public/git-hub-copilot.json +676 -0
  127. package/public/git-hub-models.json +256 -0
  128. package/public/google-partner.json +536 -0
  129. package/public/google-vertex.json +1376 -0
  130. package/public/google.json +536 -0
  131. package/public/groq.json +882 -0
  132. package/public/hugging-face.json +1096 -0
  133. package/public/inception.json +81 -0
  134. package/public/inference.json +291 -0
  135. package/public/meta.json +151 -0
  136. package/public/mistral.json +2181 -0
  137. package/public/model-scope.json +37839 -0
  138. package/public/morph.json +46 -0
  139. package/public/open-ai.json +606 -0
  140. package/public/open-router.json +15341 -0
  141. package/public/providers.json +43 -0
  142. package/public/requesty.json +13757 -0
  143. package/public/together-ai.json +641 -0
  144. package/public/upstage.json +606 -0
  145. package/public/v0.json +431 -0
  146. package/public/venice.json +501 -0
  147. package/public/vercel.json +4071 -0
  148. package/public/weights-&-biases.json +551 -0
  149. package/public/xai.json +256 -0
@@ -0,0 +1 @@
1
+ var e={description:"AI Models API - Models from Azure OpenAI",lastUpdated:"2026-01-10T23:24:47.786Z",provider:"Azure OpenAI",totalModels:9,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:.0015,inputCacheHit:null,output:.002},extendedThinking:!1,id:"gpt-35-turbo-(0125)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"gpt-35-turbo (0125)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"azure"},{attachment:!1,cost:{input:.0015,inputCacheHit:null,output:.002},extendedThinking:!1,id:"gpt-35-turbo-(1106)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"gpt-35-turbo (1106)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"azure"},{attachment:!1,cost:{input:.03,inputCacheHit:null,output:.06},extendedThinking:!1,id:"gpt-4-(turbo-2024-04-09)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"gpt-4 (turbo-2024-04-09)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"azure"},{attachment:!1,cost:{input:.03,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-11-20 | Retirement: 2026-06-03 | Replacement: gpt-5.1",extendedThinking:!1,id:"gpt-4-o-(2024-05-13)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"gpt-4o (2024-05-13)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"azure"},{attachment:!1,cost:{input:.03,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-11-20 | Retirement: 2026-06-03 | Replacement: gpt-5.1",extendedThinking:!1,id:"gpt-4-o-(2024-08-06)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"gpt-4o (2024-08-06)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"azure"},{attachment:!1,cost:{input:.03,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-11-20 | Retirement: 2026-06-03 | Replacement: gpt-5.1",extendedThinking:!1,id:"gpt-4-o-(2024-11-20)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"gpt-4o (2024-11-20)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"azure"},{attachment:!1,cost:{input:.03,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-11-20 | Retirement: 2026-06-03 | Replacement: gpt-5.1",extendedThinking:!1,id:"gpt-4-o-mini-(2024-07-18)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"gpt-4o-mini (2024-07-18)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"azure"},{attachment:!1,cost:{input:.015,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-12-17 | Retirement: 2026-06-18 | Replacement: o3",extendedThinking:!1,id:"o-1-mini-(2024-09-12)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"o1-mini (2024-09-12)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"azure"},{attachment:!1,cost:{input:.015,inputCacheHit:null,output:.06},description:"Lifecycle: Generally Available | Deprecation: 2025-12-17 | Retirement: 2026-06-18 | Replacement: o3",extendedThinking:!1,id:"o-1-preview-(2024-09-12)",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"o1-preview (2024-09-12)",openWeights:!1,provider:"Azure OpenAI",providerDoc:"https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models",providerEnv:["AZURE_OPENAI_API_KEY","AZURE_OPENAI_ENDPOINT"],providerId:"azure-open-ai",providerModelsDevId:"azure-openai",providerNpm:"@ai-sdk/azure-openai",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"azure"}];const i={metadata:e,models:t};export{i as default,e as metadata,t as models};
@@ -0,0 +1 @@
1
+ var e={description:"AI Models API - Models from cerebras",lastUpdated:"2026-01-10T23:24:47.786Z",provider:"cerebras",totalModels:8,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:1e-4,inputCacheHit:null,output:3e-4},extendedThinking:!0,id:"FP16",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"qwen-3-32b",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"cerebras"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!0,id:"FP16/FP8 (weights only)1",knowledge:null,lastUpdated:null,limit:{context:202752,output:65535},modalities:{input:["text"],output:["text"]},name:"zai-glm-4.7",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:"2025-12-22",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"},{attachment:!1,cost:{input:.35,inputCacheHit:null,output:.75},extendedThinking:!0,id:"gpt-oss-120b",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"OpenAI GPT OSS",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras",audioGeneration:!1,description:"OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases – gpt-oss-120b is for production, general purpose, high reasoning use-cases.",structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0},{attachment:!1,cost:{input:.1,inputCacheHit:null,output:.1},extendedThinking:!0,id:"llama3.1-8b",knowledge:null,lastUpdated:null,limit:{context:4096,output:null},modalities:{input:["text"],output:["text"]},name:"Llama 3.1 8B",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"},{attachment:!1,cost:{input:.85,inputCacheHit:null,output:1.2},extendedThinking:!0,id:"llama-3.3-70b",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"Llama 3.3 70B",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"},{attachment:!1,cost:{input:2,inputCacheHit:null,output:2},extendedThinking:!0,id:"qwen-3-coder-480b",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen 3 480B Coder",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"},{attachment:!1,cost:{input:2.25,inputCacheHit:null,output:2.75},extendedThinking:!0,id:"zai-glm-4.6",knowledge:null,lastUpdated:null,limit:{context:32768,output:65536},modalities:{input:["text"],output:["text"]},name:"Z.ai GLM 4.6",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:"2025-09-30",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"},{attachment:!1,cost:{input:2.25,inputCacheHit:null,output:2.75},extendedThinking:!0,id:"zai-glm-4.7",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"Z.ai GLM 4.7 1",openWeights:!1,provider:"cerebras",providerDoc:"https://inference-docs.cerebras.ai/models/overview",providerEnv:["CEREBRAS_API_KEY"],providerId:"cerebras",providerNpm:"@ai-sdk/cerebras",reasoning:!0,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"cerebras"}];const r={metadata:e,models:t};export{r as default,e as metadata,t as models};
@@ -0,0 +1 @@
1
+ var e={description:"AI Models API - Models from chutes",lastUpdated:"2026-01-10T23:24:47.794Z",provider:"chutes",totalModels:94,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:15e-6,inputCacheHit:null,output:15e-6},extendedThinking:!1,id:"agentica-org/DeepCoder-14B-Preview",knowledge:null,lastUpdated:null,limit:{context:384e3,output:null},modalities:{input:["text"],output:["text"]},name:"DeepCoder-14B-Preview",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"all-hands/openhands-lm-32b-v0.1-ep3",knowledge:null,lastUpdated:null,limit:{context:65536,output:null},modalities:{input:["text"],output:["text"]},name:"openhands-lm-32b-v0.1-ep3",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:11e-5},extendedThinking:!1,id:"ArliAI/QwQ-32B-ArliAI-RpR-v1",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"QwQ-32B-ArliAI-RpR-v1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:5e-5,inputCacheHit:null,output:22e-5},extendedThinking:!1,id:"chutesai/Devstral-Small-2505",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Devstral-Small-2505",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:0,inputCacheHit:null,output:0},extendedThinking:!1,id:"chutesai/Llama-4-Maverick-17B-128E-Instruct-FP8",knowledge:null,lastUpdated:null,limit:{context:512e3,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-4-Maverick-17B-128E-Instruct-FP8",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},extendedThinking:!1,id:"chutesai/Mistral-Small-3.1-24B-Instruct-2503",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Mistral-Small-3.1-24B-Instruct-2503",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},extendedThinking:!1,id:"chutesai/Mistral-Small-3.2-24B-Instruct-2506",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Mistral-Small-3.2-24B-Instruct-2506",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:4e-5,inputCacheHit:null,output:17e-5},extendedThinking:!1,id:"cognitivecomputations/Dolphin3.0-Mistral-24B",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Dolphin3.0-Mistral-24B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-02-13",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},extendedThinking:!1,id:"cognitivecomputations/Dolphin3.0-R1-Mistral-24B",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Dolphin3.0-R1-Mistral-24B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-02-13",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:75e-5,inputCacheHit:null,output:99e-5},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1",knowledge:null,lastUpdated:null,limit:{context:655360,output:163840},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-01-20",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:4e-4,inputCacheHit:null,output:.00175},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-0528",knowledge:null,lastUpdated:null,limit:{context:655360,output:65536},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-0528",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-05-28",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:11e-5},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-0528-Qwen3-8B",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-0528-Qwen3-8B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-0528-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-0528-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:75e-5,inputCacheHit:null,output:99e-5},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-Distill-Llama-70B",knowledge:null,lastUpdated:null,limit:{context:524288,output:131072},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-Distill-Llama-70B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-01-23",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:9e-4,inputCacheHit:null,output:9e-4},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3-0324",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3-0324",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3-0324-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3-0324-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.1-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.1-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.1-Terminus-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.1-Terminus-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.2-Speciale-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.2-Speciale-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.2-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.2-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3-Base",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3-Base",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"internlm/Intern-S1",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"Intern-S1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"microsoft/MAI-DS-R1-FP8",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"MAI-DS-R1-FP8",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"MiniMaxAI/MiniMax-M2.1-TEE",knowledge:null,lastUpdated:null,limit:{context:786432,output:null},modalities:{input:["text"],output:["text"]},name:"MiniMax-M2.1-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"miromind-ai/MiroThinker-v1.5-235B",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"MiroThinker-v1.5-235B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"mistralai/Devstral-2-123B-Instruct-2512",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Devstral-2-123B-Instruct-2512",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"mistralai/Devstral-2-123B-Instruct-2512-TEE",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Devstral-2-123B-Instruct-2512-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:29e-5,inputCacheHit:null,output:.00115},extendedThinking:!1,id:"moonshotai/Kimi-Dev-72B",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-Dev-72B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:.001,inputCacheHit:null,output:.003},extendedThinking:!1,id:"moonshotai/Kimi-K2-Instruct",knowledge:null,lastUpdated:null,limit:{context:3e5,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-K2-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"moonshotai/Kimi-K2-Instruct-0905",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-K2-Instruct-0905",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"moonshotai/Kimi-K2-Thinking-TEE",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-K2-Thinking-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"moonshotai/Kimi-VL-A3B-Thinking",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-VL-A3B-Thinking",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:11e-5},extendedThinking:!1,id:"NousResearch/DeepHermes-3-Llama-3-8B-Preview",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"DeepHermes-3-Llama-3-8B-Preview",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:15e-5,inputCacheHit:null,output:59e-5},extendedThinking:!1,id:"NousResearch/DeepHermes-3-Mistral-24B-Preview",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"DeepHermes-3-Mistral-24B-Preview",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"NousResearch/Hermes-4-14B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Hermes-4-14B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"NousResearch/Hermes-4.3-36B",knowledge:null,lastUpdated:null,limit:{context:2097152,output:null},modalities:{input:["text"],output:["text"]},name:"Hermes-4.3-36B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"NousResearch/Hermes-4-405B-FP8-TEE",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Hermes-4-405B-FP8-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:11e-5,inputCacheHit:null,output:38e-5},extendedThinking:!1,id:"NousResearch/Hermes-4-70B",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Hermes-4-70B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},extendedThinking:!1,id:"nvidia/Llama-3_3-Nemotron-Super-49B-v1_5",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-3_3-Nemotron-Super-49B-v1_5",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"NVIDIA-Nemotron-3-Nano-30B-A3B-BF16",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"OpenGVLab/InternVL3-78B-TEE",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"InternVL3-78B-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"openai/gpt-oss-120b-TEE",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"gpt-oss-120b-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:.0012,inputCacheHit:null,output:.0012},extendedThinking:!1,id:"Qwen/Qwen2.5-72B-Instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:16384},modalities:{input:["text"],output:["text"]},name:"Qwen2.5-72B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2024-09-19",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:8e-4,inputCacheHit:null,output:8e-4},extendedThinking:!1,id:"Qwen/Qwen2.5-Coder-32B-Instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:32768},modalities:{input:["text"],output:["text"]},name:"Qwen2.5-Coder-32B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2024-11-11",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:5e-5,inputCacheHit:null,output:22e-5},extendedThinking:!1,id:"Qwen/Qwen2.5-VL-32B-Instruct",knowledge:null,lastUpdated:null,limit:{context:65536,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen2.5-VL-32B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:8e-5,inputCacheHit:null,output:33e-5},extendedThinking:!1,id:"Qwen/Qwen2.5-VL-72B-Instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen2.5-VL-72B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen2.5-VL-72B-Instruct-TEE",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen2.5-VL-72B-Instruct-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:5e-5,inputCacheHit:null,output:22e-5},extendedThinking:!1,id:"Qwen/Qwen3-14B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-14B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:18e-5,inputCacheHit:null,output:54e-5},extendedThinking:!1,id:"Qwen/Qwen3-235B-A22B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-235B-A22B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen3-235B-A22B-Instruct-2507",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-235B-A22B-Instruct-2507",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen3-235B-A22B-Instruct-2507-TEE",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-235B-A22B-Instruct-2507-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:11e-5,inputCacheHit:null,output:6e-4},extendedThinking:!1,id:"Qwen/Qwen3-235B-A22B-Thinking-2507",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-235B-A22B-Thinking-2507",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:6e-5,inputCacheHit:null,output:22e-5},extendedThinking:!1,id:"Qwen/Qwen3-30B-A3B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-30B-A3B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:8e-5,inputCacheHit:null,output:33e-5},extendedThinking:!1,id:"Qwen/Qwen3-30B-A3B-Instruct-2507",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-30B-A3B-Instruct-2507",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:29e-5,inputCacheHit:null,output:59e-5},extendedThinking:!1,id:"Qwen/Qwen3-32B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-32B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:11e-5},extendedThinking:!1,id:"Qwen/Qwen3-8B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-8B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:6e-5,inputCacheHit:null,output:25e-5},extendedThinking:!1,id:"Qwen/Qwen3-Coder-30B-A3B-Instruct",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Coder-30B-A3B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Coder-480B-A35B-Instruct-FP8",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8-TEE",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Coder-480B-A35B-Instruct-FP8-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Qwen/Qwen3Guard-Gen-0.6B",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3Guard-Gen-0.6B",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:1e-4,inputCacheHit:null,output:8e-4},extendedThinking:!1,id:"Qwen/Qwen3-Next-80B-A3B-Instruct",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Next-80B-A3B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:.0012},extendedThinking:!1,id:"Qwen/Qwen3-VL-235B-A22B-Instruct",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-VL-235B-A22B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"rednote-hilab/dots.ocr",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"dots.ocr",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Salesforce/xgen-small-9B-instruct-r",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"xgen-small-9B-instruct-r",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:3e-5},extendedThinking:!1,id:"tencent/Hunyuan-A13B-Instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Hunyuan-A13B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"Tesslate/UIGEN-X-32B-0727",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"UIGEN-X-32B-0727",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"TheDrummer/Cydonia-24B-v2.1",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Cydonia-24B-v2.1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"TheDrummer/Gemmasutra-Pro-27B-v1.1",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"Gemmasutra-Pro-27B-v1.1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:8e-5,inputCacheHit:null,output:33e-5},extendedThinking:!1,id:"TheDrummer/Skyfall-36B-v2",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Skyfall-36B-v2",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"TheDrummer/Tunguska-39B-v1",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Tunguska-39B-v1",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:.0012},extendedThinking:!1,id:"tngtech/DeepSeek-R1T-Chimera",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1T-Chimera",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"tngtech/DeepSeek-TNG-R1T2-Chimera",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-TNG-R1T2-Chimera",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"tngtech/TNG-R1T-Chimera-TEE",knowledge:null,lastUpdated:null,limit:{context:655360,output:null},modalities:{input:["text"],output:["text"]},name:"TNG-R1T-Chimera-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"tplr/TEMPLAR-I",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"TEMPLAR-I",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:8e-5,inputCacheHit:null,output:8e-5},extendedThinking:!1,id:"unsloth/gemma-2-9b-it",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-2-9b-it",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:1e-4},extendedThinking:!1,id:"unsloth/gemma-3-12b-it",knowledge:null,lastUpdated:null,limit:{context:4194304,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-3-12b-it",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:9e-5,inputCacheHit:null,output:16e-5},extendedThinking:!1,id:"unsloth/gemma-3-27b-it",knowledge:null,lastUpdated:null,limit:{context:4194304,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-3-27b-it",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:170301e-10,inputCacheHit:null,output:681536e-10},extendedThinking:!1,id:"unsloth/gemma-3-4b-it",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-3-4b-it",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:2e-5,inputCacheHit:null,output:2e-5},extendedThinking:!1,id:"unsloth/Llama-3.2-1B-Instruct",knowledge:null,lastUpdated:null,limit:{context:65536,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-3.2-1B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:5e-5},extendedThinking:!1,id:"unsloth/Llama-3.2-3B-Instruct",knowledge:null,lastUpdated:null,limit:{context:65536,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-3.2-3B-Instruct",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},extendedThinking:!1,id:"unsloth/Mistral-Nemo-Instruct-2407",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"Mistral-Nemo-Instruct-2407",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},extendedThinking:!1,id:"unsloth/Mistral-Small-24B-Instruct-2501",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Mistral-Small-24B-Instruct-2501",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"XiaomiMiMo/MiMo-V2-Flash",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"MiMo-V2-Flash",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4-32B-0414",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4-32B-0414",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:13e-5,inputCacheHit:null,output:85e-5},extendedThinking:!1,id:"zai-org/GLM-4.5-Air",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5-Air",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes",providerModelsDevId:"deepinfra"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.5-FP8",knowledge:null,lastUpdated:null,limit:{context:393216,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5-FP8",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.5-TEE",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.5V-FP8",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5V-FP8",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.6-TEE",knowledge:null,lastUpdated:null,limit:{context:811008,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.6-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.6V",knowledge:null,lastUpdated:null,limit:{context:524288,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.6V",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-4.7-TEE",knowledge:null,lastUpdated:null,limit:{context:811008,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.7-TEE",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"zai-org/GLM-Z1-32B-0414",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-Z1-32B-0414",openWeights:!1,provider:"chutes",providerDoc:"https://llm.chutes.ai/v1/models",providerEnv:["CHUTES_API_KEY"],providerId:"chutes",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"chutes"}];const l={metadata:e,models:t};export{l as default,e as metadata,t as models};
@@ -0,0 +1,3 @@
1
+ var e={description:"AI Models API - Models from Cloudflare",lastUpdated:"2026-01-10T23:24:47.798Z",provider:"Cloudflare",totalModels:79,version:"0.0.0-development"},t=[{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"Aura is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",id:"aura-1",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"aura-1",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"Aura-2 is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",id:"aura-2-en",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"aura-2-en",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"Aura-2 is a context-aware text-to-speech (TTS) model that applies natural pacing, expressiveness, and fillers based on the context of the provided text. The quality of your text input directly impacts the naturalness of the audio output.",id:"aura-2-es",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"aura-2-es",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"BART is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. You can use this model for text summarization.",id:"bart-large-cnn",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"bart-large-cnn",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"BAAI general embedding (Base) model that transforms any given text into a 768-dimensional vector",id:"bge-base-en-v1.5",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"bge-base-en-v1.5",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"BAAI general embedding (Large) model that transforms any given text into a 1024-dimensional vector",id:"bge-large-en-v1.5",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"bge-large-en-v1.5",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Multi-Functionality, Multi-Linguality, and Multi-Granularity embeddings model.",id:"bge-m3",knowledge:null,lastUpdated:null,limit:{context:6e7,output:null},modalities:{input:["text"],output:["embedding"]},name:"bge-m3",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"BAAI general embedding (Small) model that transforms any given text into a 384-dimensional vector",id:"bge-small-en-v1.5",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"bge-small-en-v1.5",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",id:"deepseek-coder-6.7b-base-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"deepseek-coder-6.7b-base-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.",id:"deepseek-coder-6.7b-instruct-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"deepseek-coder-6.7b-instruct-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"DeepSeekMath-Instruct 7B is a mathematically instructed tuning model derived from DeepSeekMath-Base 7B. DeepSeekMath is initialized with DeepSeek-Coder-v1.5 7B and continues pre-training on math-related tokens sourced from Common Crawl, together with natural language and code data for 500B tokens.",id:"deepseek-math-7b-instruct",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"deepseek-math-7b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.",extendedThinking:!1,id:"deepseek-r1-distill-qwen-32b",knowledge:null,lastUpdated:null,limit:{context:8e7,output:null},modalities:{input:["text"],output:["text"]},name:"deepseek-r1-distill-qwen-32b",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:"2025-01-29",streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).",id:"detr-resnet-50",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"detr-resnet-50",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"DiscoLM German 7b is a Mistral-based large language model with a focus on German-language applications. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",id:"discolm-german-7b-v1-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"discolm-german-7b-v1-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Distilled BERT model that was finetuned on SST-2 for sentiment classification",id:"distilbert-sst-2-int8",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"distilbert-sst-2-int8",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Stable Diffusion model that has been fine-tuned to be better at photorealism without sacrificing range.",id:"dreamshaper-8-lcm",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"dreamshaper-8-lcm",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"EmbeddingGemma is a 300M parameter, state-of-the-art for its size, open embedding model from Google, built from Gemma 3 (with T5Gemma initialization) and the same research and technology used to create Gemini models. EmbeddingGemma produces vector representations of text, making it well-suited for search and retrieval tasks, including classification, clustering, and semantic similarity search. This model was trained with data in 100+ spoken languages.",id:"embeddinggemma-300m",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"embeddinggemma-300m",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Falcon-7B-Instruct is a 7B parameters causal decoder-only model built by TII based on Falcon-7B and finetuned on a mixture of chat/instruct datasets.",id:"falcon-7b-instruct",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"falcon-7b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:0,inputCacheHit:null,output:0},description:"Flux is the first conversational speech recognition model built specifically for voice agents.",id:"flux",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"flux",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"This is a Gemma-2B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",id:"gemma-2b-it-lora",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-2b-it-lora",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-5,inputCacheHit:null,output:1e-4},description:"Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Gemma 3 models are multimodal, handling text and image input and generating text output, with a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions.",id:"gemma-3-12b-it",knowledge:null,lastUpdated:null,limit:{context:8e7,output:null},modalities:{input:["text","image"],output:["text"]},name:"gemma-3-12b-it",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:7e-5,inputCacheHit:null,output:7e-5},description:"Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.",id:"gemma-7b-it",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-7b-it",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:7e-5,inputCacheHit:null,output:7e-5},description:"This is a Gemma-7B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.",id:"gemma-7b-it-lora",knowledge:null,lastUpdated:null,limit:{context:35e5,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-7b-it-lora",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"SEA-LION stands for Southeast Asian Languages In One Network, which is a collection of Large Language Models (LLMs) which have been pretrained and instruct-tuned for the Southeast Asia (SEA) region.",id:"gemma-sea-lion-v4-27b-it",knowledge:null,lastUpdated:null,limit:{context:128e6,output:null},modalities:{input:["text"],output:["text"]},name:"gemma-sea-lion-v4-27b-it",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:1e-4,inputCacheHit:null,output:1e-4},description:"OpenAI’s open-weight models designed for powerful reasoning, agentic tasks, and versatile developer use cases – gpt-oss-20b is for lower latency, and local or specialized use-cases.",id:"gpt-oss-20b",knowledge:null,lastUpdated:null,limit:{context:128e6,output:null},modalities:{input:["text"],output:["text"]},name:"gpt-oss-20b",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:17e-6,inputCacheHit:null,output:11e-5},description:"Granite 4.0 instruct models deliver strong performance across benchmarks, achieving industry-leading results in key agentic tasks like instruction following and function calling. These efficiencies make the models well-suited for a wide range of use cases like retrieval-augmented generation (RAG), multi-agent workflows, and edge deployments.",id:"granite-4.0-h-micro",knowledge:null,lastUpdated:null,limit:{context:131e6,output:null},modalities:{input:["text"],output:["text"]},name:"granite-4.0-h-micro",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"IndicTrans2 is the first open-source transformer-based multilingual NMT model that supports high-quality translations across all the 22 scheduled Indic languages",id:"indictrans2-en-indic-1B",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"indictrans2-en-indic-1B",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Llama 2 13B Chat AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Llama 2 variant.",id:"llama-2-13b-chat-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-2-13b-chat-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Full precision (fp16) generative text model with 7 billion parameters from Meta",id:"llama-2-7b-chat-fp16",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-2-7b-chat-fp16",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"This is a Llama2 base model that Cloudflare dedicated for inference with LoRA adapters. Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format.",id:"llama-2-7b-chat-hf-lora",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-2-7b-chat-hf-lora",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Quantized (int8) generative text model with 7 billion parameters from Meta",id:"llama-2-7b-chat-int8",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-2-7b-chat-int8",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:45e-5,inputCacheHit:null,output:45e-5},description:"The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",id:"llama-3.1-70b-instruct",knowledge:null,lastUpdated:null,limit:{context:24e6,output:null},modalities:{input:["text","image"],output:["text"]},name:"llama-3.1-70b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:1e-4,inputCacheHit:null,output:1e-4},description:"The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",id:"llama-3.1-8b-instruct",knowledge:null,lastUpdated:null,limit:{context:7968e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.1-8b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Quantized (int4) generative text model with 8 billion parameters from Meta.",id:"llama-3.1-8b-instruct-awq",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.1-8b-instruct-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"[Fast version] The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",id:"llama-3.1-8b-instruct-fast",knowledge:null,lastUpdated:null,limit:{context:128e6,output:null},modalities:{input:["text","image"],output:["text"]},name:"llama-3.1-8b-instruct-fast",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Llama 3.1 8B quantized to FP8 precision",id:"llama-3.1-8b-instruct-fp8",knowledge:null,lastUpdated:null,limit:{context:32e6,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.1-8b-instruct-fp8",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:6e-5,inputCacheHit:null,output:6e-5},description:"The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.",id:"llama-3.2-11b-vision-instruct",knowledge:null,lastUpdated:null,limit:{context:128e6,output:null},modalities:{input:["text","image"],output:["text"]},name:"llama-3.2-11b-vision-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:2e-5,inputCacheHit:null,output:2e-5},description:"The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",id:"llama-3.2-1b-instruct",knowledge:null,lastUpdated:null,limit:{context:6e7,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.2-1b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-5,inputCacheHit:null,output:5e-5},description:"The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",id:"llama-3.2-3b-instruct",knowledge:null,lastUpdated:null,limit:{context:128e6,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.2-3b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:0,inputCacheHit:null,output:0},description:"Llama 3.3 70B quantized to fp8 precision, optimized to be faster.",id:"llama-3.3-70b-instruct-fp8-fast",knowledge:null,lastUpdated:null,limit:{context:24e6,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3.3-70b-instruct-fp8-fast",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Quantized (int4) generative text model with 8 billion parameters from Meta.",id:"llama-3-8b-instruct-awq",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"llama-3-8b-instruct-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"Llama Guard is a model for classifying the safety of LLM prompts and responses, using a taxonomy of safety risks.",id:"llamaguard-7b-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"llamaguard-7b-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.",id:"llava-1.5-7b-hf",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"llava-1.5-7b-hf",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Lucid Origin from Leonardo.AI is their most adaptable and prompt-responsive model to date. Whether you're generating images with sharp graphic design, stunning full-HD renders, or highly specific creative direction, it adheres closely to your prompts, renders text with accuracy, and supports a wide array of visual styles and aesthetics – from stylized concept art to crisp product mockups.",id:"lucid-origin",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"lucid-origin",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Multilingual encoder-decoder (seq-to-seq) model trained for Many-to-Many multilingual translation",id:"m2m100-1.2b",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"m2m100-1.2b",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"MeloTTS is a high-quality multi-lingual text-to-speech library by MyShell.ai.",id:"melotts",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"melotts",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},description:"Instruct fine-tuned version of the Mistral-7b generative text model with 7 billion parameters",id:"mistral-7b-instruct-v0.1",knowledge:null,lastUpdated:null,limit:{context:2824e3,output:null},modalities:{input:["text"],output:["text"]},name:"mistral-7b-instruct-v0.1",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},description:"Mistral 7B Instruct v0.1 AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Mistral variant.",id:"mistral-7b-instruct-v0.1-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"mistral-7b-instruct-v0.1-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},description:"The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2. Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1: 32k context window (vs 8k context in v0.1), rope-theta = 1e6, and no Sliding-Window Attention.",id:"mistral-7b-instruct-v0.2",knowledge:null,lastUpdated:null,limit:{context:3072e3,output:null},modalities:{input:["text"],output:["text"]},name:"mistral-7b-instruct-v0.2",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},description:"The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.",id:"mistral-7b-instruct-v0.2-lora",knowledge:null,lastUpdated:null,limit:{context:15e6,output:null},modalities:{input:["text"],output:["text"]},name:"mistral-7b-instruct-v0.2-lora",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:14e-5,inputCacheHit:null,output:42e-5},description:"Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance. With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.",id:"mistral-small-3.1-24b-instruct",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text","image"],output:["text"]},name:"mistral-small-3.1-24b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"This model is a fine-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the mistralai/Mistral-7B-v0.1 on the open source dataset Open-Orca/SlimOrca.",id:"neural-chat-7b-v3-1-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"neural-chat-7b-v3-1-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:52e-10,inputCacheHit:null,output:92e-10},description:"Transcribe audio using Deepgram’s speech-to-text model",id:"nova-3",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"nova-3",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning.",id:"openchat-3.5-0106",knowledge:null,lastUpdated:null,limit:{context:8192e3,output:null},modalities:{input:["text"],output:["text"]},name:"openchat-3.5-0106",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets.",id:"openhermes-2.5-mistral-7b-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"openhermes-2.5-mistral-7b-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:1e-4,inputCacheHit:null,output:1e-4},description:"Phi-2 is a Transformer-based model with a next-word prediction objective, trained on 1.4T tokens from multiple passes on a mixture of Synthetic and Web datasets for NLP and coding.",id:"phi-2",knowledge:null,lastUpdated:null,limit:{context:2048e3,output:null},modalities:{input:["text"],output:["text"]},name:"phi-2",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Phoenix 1.0 is a model by Leonardo.Ai that generates images with exceptional prompt adherence and coherent text.",id:"phoenix-1.0",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"phoenix-1.0",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:`PLaMo-Embedding-1B is a Japanese text embedding model developed by Preferred Networks, Inc.
2
+
3
+ It can convert Japanese text input into numerical vectors and can be used for a wide range of applications, including information retrieval, text classification, and clustering.`,id:"plamo-embedding-1b",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"plamo-embedding-1b",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:1e-4,inputCacheHit:null,output:1e-4},description:"Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.",id:"qwen1.5-0.5b-chat",knowledge:null,lastUpdated:null,limit:{context:32e6,output:null},modalities:{input:["text"],output:["text"]},name:"qwen1.5-0.5b-chat",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",id:"qwen1.5-14b-chat-awq",knowledge:null,lastUpdated:null,limit:{context:75e5,output:null},modalities:{input:["text"],output:["text"]},name:"qwen1.5-14b-chat-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:1e-4,inputCacheHit:null,output:1e-4},description:"Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.",id:"qwen1.5-1.8b-chat",knowledge:null,lastUpdated:null,limit:{context:32e6,output:null},modalities:{input:["text"],output:["text"]},name:"qwen1.5-1.8b-chat",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.",id:"qwen1.5-7b-chat-awq",knowledge:null,lastUpdated:null,limit:{context:2e7,output:null},modalities:{input:["text"],output:["text"]},name:"qwen1.5-7b-chat-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:8e-4,inputCacheHit:null,output:8e-4},description:"Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:",extendedThinking:!1,id:"qwen2.5-coder-32b-instruct",knowledge:null,lastUpdated:null,limit:{context:32768e3,output:32768},modalities:{input:["text"],output:["text"]},name:"qwen2.5-coder-32b-instruct",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:"2024-11-11",streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support.",id:"qwen3-30b-a3b-fp8",knowledge:null,lastUpdated:null,limit:{context:32768e3,output:null},modalities:{input:["text"],output:["text"]},name:"qwen3-30b-a3b-fp8",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks.",id:"qwen3-embedding-0.6b",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["embedding"]},name:"qwen3-embedding-0.6b",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"50 layers deep image classification CNN trained on more than 1M images from ImageNet",id:"resnet-50",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"resnet-50",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"An open source, community-driven, native audio turn detection model in 2nd version",id:"smart-turn-v2",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["audio"]},name:"smart-turn-v2",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"This model is intended to be used by non-technical users to understand data inside their SQL databases.",id:"sqlcoder-7b-2",knowledge:null,lastUpdated:null,limit:{context:1e7,output:null},modalities:{input:["text"],output:["text"]},name:"sqlcoder-7b-2",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images. Img2img generate a new image from an input image with Stable Diffusion.",id:"stable-diffusion-v1-5-img2img",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"stable-diffusion-v1-5-img2img",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.",id:"stable-diffusion-v1-5-inpainting",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"stable-diffusion-v1-5-inpainting",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Diffusion-based text-to-image generative model by Stability AI. Generates and modify images based on text prompts.",id:"stable-diffusion-xl-base-1.0",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"stable-diffusion-xl-base-1.0",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"SDXL-Lightning is a lightning-fast text-to-image generation model. It can generate high-quality 1024px images in a few steps.",id:"stable-diffusion-xl-lightning",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"stable-diffusion-xl-lightning",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"We introduce Starling-LM-7B-beta, an open large language model (LLM) trained by Reinforcement Learning from AI Feedback (RLAIF). Starling-LM-7B-beta is trained from Openchat-3.5-0106 with our new reward model Nexusflow/Starling-RM-34B and policy optimization method Fine-Tuning Language Models from Human Preferences (PPO).",id:"starling-lm-7b-beta",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"starling-lm-7b-beta",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T.",id:"tinyllama-1.1b-chat-v1.0",knowledge:null,lastUpdated:null,limit:{context:2048e3,output:null},modalities:{input:["text"],output:["text"]},name:"tinyllama-1.1b-chat-v1.0",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"UForm-Gen is a small generative vision-language model primarily designed for Image Captioning and Visual Question Answering. The model was pre-trained on the internal image captioning dataset and fine-tuned on public instructions datasets: SVIT, LVIS, VQAs datasets.",id:"uform-gen2-qwen-500m",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"uform-gen2-qwen-500m",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!0,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Cybertron 7B v2 is a 7B MistralAI based model, best on it's series. It was trained with SFT, DPO and UNA (Unified Neural Alignment) on multiple datasets.",id:"una-cybertron-7b-v2-bf16",knowledge:null,lastUpdated:null,limit:{context:15e6,output:null},modalities:{input:["text"],output:["text"]},name:"una-cybertron-7b-v2-bf16",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation.",id:"whisper-large-v3-turbo",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"whisper-large-v3-turbo",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!0,cost:{input:null,inputCacheHit:null,output:null},description:"Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalize to many datasets and domains without the need for fine-tuning. This is the English-only version of the Whisper Tiny model which was trained on the task of speech recognition.",id:"whisper-tiny-en",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["audio"],output:["text"]},name:"whisper-tiny-en",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"},{attachment:!1,audioGeneration:!1,cost:{input:null,inputCacheHit:null,output:null},description:"Zephyr 7B Beta AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Zephyr model variant.",id:"zephyr-7b-beta-awq",knowledge:null,lastUpdated:null,limit:{context:4096e3,output:null},modalities:{input:["text"],output:["text"]},name:"zephyr-7b-beta-awq",openWeights:!1,provider:"Cloudflare",providerDoc:"https://developers.cloudflare.com/workers-ai/models",providerEnv:["CLOUDFLARE_API_TOKEN"],providerId:"cloudflare",providerNpm:"@ai-sdk/cloudflare",reasoning:!1,releaseDate:null,streamingSupported:!0,structuredOutputs:!0,supportsStructuredOutput:!0,supportsTools:!0,temperature:!0,toolCall:!0,version:null,vision:!1,icon:"cloudflare"}];const l={metadata:e,models:t};export{l as default,e as metadata,t as models};
@@ -0,0 +1 @@
1
+ var e={description:"AI Models API - Models from Deep Infra",lastUpdated:"2026-01-10T23:24:47.802Z",provider:"Deep Infra",totalModels:27,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:9e-6,inputCacheHit:null,output:19e-6},extendedThinking:!1,id:"allenai/olmOCR-2-7B-1025",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"olmOCR-2-7B-1025",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:3e-5,inputCacheHit:null,output:18e-5},extendedThinking:!1,id:"allenai/olmOCR-7B-0725-FP8",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"olmOCR-7B-0725-FP8",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:.00165,inputCacheHit:null,output:.00825},extendedThinking:!1,id:"anthropic/claude-4-opus",knowledge:null,lastUpdated:null,limit:{context:2e5,output:null},modalities:{input:["text"],output:["text"]},name:"claude-4-opus",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:33e-5,inputCacheHit:null,output:.00165},extendedThinking:!1,id:"anthropic/claude-4-sonnet",knowledge:null,lastUpdated:null,limit:{context:2e5,output:null},modalities:{input:["text"],output:["text"]},name:"claude-4-sonnet",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:3e-6,inputCacheHit:null,output:1e-5},extendedThinking:!1,id:"deepseek-ai/DeepSeek-OCR",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-OCR",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:1e-4,inputCacheHit:null,output:3e-4},extendedThinking:!1,id:"deepseek-ai/DeepSeek-R1-0528-Turbo",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-R1-0528-Turbo",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:1e-4,inputCacheHit:null,output:3e-4},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3-0324-Turbo",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3-0324-Turbo",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:21e-6,inputCacheHit:null,output:79e-6},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.1",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.1",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:21e-6,inputCacheHit:null,output:79e-6},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.1-Terminus",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.1-Terminus",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:26e-6,inputCacheHit:null,output:39e-6},extendedThinking:!1,id:"deepseek-ai/DeepSeek-V3.2",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3.2",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:23e-6,inputCacheHit:null,output:4e-5},extendedThinking:!1,id:"meta-llama/Llama-3.3-70B-Instruct",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-3.3-70B-Instruct",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:1e-5,inputCacheHit:null,output:32e-6},extendedThinking:!1,id:"meta-llama/Llama-3.3-70B-Instruct-Turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-3.3-70B-Instruct-Turbo",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:15e-6,inputCacheHit:null,output:6e-5},extendedThinking:!1,id:"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-4-Maverick-17B-128E-Instruct-FP8",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:5e-5,inputCacheHit:null,output:5e-5},extendedThinking:!1,id:"meta-llama/Llama-4-Maverick-17B-128E-Instruct-Turbo",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-4-Maverick-17B-128E-Instruct-Turbo",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:8e-6,inputCacheHit:null,output:3e-5},extendedThinking:!1,id:"meta-llama/Llama-4-Scout-17B-16E-Instruct",knowledge:null,lastUpdated:null,limit:{context:327680,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-4-Scout-17B-16E-Instruct",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:18e-6,inputCacheHit:null,output:18e-6},extendedThinking:!1,id:"meta-llama/Llama-Guard-4-12B",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"Llama-Guard-4-12B",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:254e-7,inputCacheHit:null,output:102e-6},extendedThinking:!1,id:"MiniMaxAI/MiniMax-M2",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"MiniMax-M2",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:7e-6,inputCacheHit:null,output:28e-6},extendedThinking:!1,id:"mistralai/Devstral-Small-2507",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"Devstral-Small-2507",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:75e-7,inputCacheHit:null,output:2e-5},extendedThinking:!1,id:"mistralai/Mistral-Small-3.2-24B-Instruct-2506",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"Mistral-Small-3.2-24B-Instruct-2506",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:47e-6,inputCacheHit:null,output:2e-4},extendedThinking:!1,id:"moonshotai/Kimi-K2-Thinking",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"Kimi-K2-Thinking",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:6e-6,inputCacheHit:null,output:24e-6},extendedThinking:!1,id:"nvidia/Nemotron-3-Nano-30B-A3B",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"Nemotron-3-Nano-30B-A3B",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:14e-6,inputCacheHit:null,output:8e-5},extendedThinking:!1,id:"PaddlePaddle/PaddleOCR-VL-0.9B",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"PaddleOCR-VL-0.9B",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:75e-7,inputCacheHit:null,output:15e-6},extendedThinking:!1,id:"Qwen/QwQ-32B",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"QwQ-32B",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:4e-5,inputCacheHit:null,output:16e-5},extendedThinking:!1,id:"Qwen/Qwen3-Coder-480B-A35B-Instruct",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Coder-480B-A35B-Instruct",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:28e-6,inputCacheHit:null,output:12e-5},extendedThinking:!1,id:"Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"Qwen3-Coder-480B-A35B-Instruct-Turbo",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:55e-6,inputCacheHit:null,output:2e-4},extendedThinking:!1,id:"zai-org/GLM-4.5",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"},{attachment:!1,cost:{input:5e-5,inputCacheHit:null,output:17e-5},extendedThinking:!1,id:"zai-org/GLM-4.5V",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"GLM-4.5V",openWeights:!1,provider:"Deep Infra",providerDoc:"https://deepinfra.com/docs",providerEnv:[],providerId:"deep-infra",providerModelsDevId:"deepinfra",providerNpm:"@deepinfra/sdk",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"deepinfra"}];const n={metadata:e,models:t};export{n as default,e as metadata,t as models};
@@ -0,0 +1 @@
1
+ var e={description:"AI Models API - Models from DeepSeek",lastUpdated:"2026-01-10T23:24:47.803Z",provider:"DeepSeek",totalModels:1,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"deep-seek-v3-0324",knowledge:null,lastUpdated:"2025-03-25",limit:{context:null,output:8e3},modalities:{input:["text"],output:["text"]},name:"DeepSeek-V3-0324",openWeights:!1,provider:"DeepSeek",providerDoc:"https://api-docs.deepseek.com/quick_start/pricing",providerEnv:["DEEPSEEK_API_KEY"],providerId:"deep-seek",providerModelsDevId:"deepseek",providerNpm:"@ai-sdk/deepseek",reasoning:!1,releaseDate:"2024-03-24",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"deepseek"}];const d={metadata:e,models:t};export{d as default,e as metadata,t as models};