@anolilab/ai-model-registry 1.4.0 → 2.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +41 -0
- package/LICENSE.md +8 -13
- package/MIGRATION.md +474 -0
- package/README.md +94 -49
- package/dist/index.d.ts +16 -16
- package/dist/index.js +1 -3
- package/dist/packem_chunks/alibaba.js +1 -0
- package/dist/packem_chunks/amazon-bedrock.js +1 -0
- package/dist/packem_chunks/anthropic.js +1 -0
- package/dist/packem_chunks/api.js +75 -0
- package/dist/packem_chunks/azure-open-ai.js +1 -0
- package/dist/packem_chunks/cerebras.js +1 -0
- package/dist/packem_chunks/chutes.js +1 -0
- package/dist/packem_chunks/cloudflare.js +3 -0
- package/dist/packem_chunks/deep-infra.js +1 -0
- package/dist/packem_chunks/deep-seek.js +1 -0
- package/dist/packem_chunks/fireworks-ai.js +1 -0
- package/dist/packem_chunks/git-hub-copilot.js +1 -0
- package/dist/packem_chunks/git-hub-models.js +1 -0
- package/dist/packem_chunks/google-partner.js +1 -0
- package/dist/packem_chunks/google-vertex.js +1 -0
- package/dist/packem_chunks/google.js +1 -0
- package/dist/packem_chunks/groq.js +1 -0
- package/dist/packem_chunks/hugging-face.js +1 -0
- package/dist/packem_chunks/inception.js +1 -0
- package/dist/packem_chunks/inference.js +1 -0
- package/dist/packem_chunks/meta.js +1 -0
- package/dist/packem_chunks/mistral.js +1 -0
- package/dist/packem_chunks/model-scope.js +1 -0
- package/dist/packem_chunks/morph.js +1 -0
- package/dist/packem_chunks/open-ai.js +1 -0
- package/dist/packem_chunks/open-router.js +1 -0
- package/dist/packem_chunks/providers.js +1 -0
- package/dist/packem_chunks/requesty.js +73 -0
- package/dist/packem_chunks/together-ai.js +1 -0
- package/dist/packem_chunks/upstage.js +1 -0
- package/dist/packem_chunks/v0.js +1 -0
- package/dist/packem_chunks/venice.js +1 -0
- package/dist/packem_chunks/vercel.js +1 -0
- package/dist/packem_chunks/weights-_-biases.js +1 -0
- package/dist/packem_chunks/xai.js +1 -0
- package/dist/schema.d.ts +1 -1
- package/dist/schema.js +1 -1
- package/dist/types/alibaba.d.ts +6 -0
- package/dist/types/alibaba.js +0 -0
- package/dist/types/amazon-bedrock.d.ts +6 -0
- package/dist/types/amazon-bedrock.js +0 -0
- package/dist/types/anthropic.d.ts +6 -0
- package/dist/types/anthropic.js +0 -0
- package/dist/types/azure-open-ai.d.ts +6 -0
- package/dist/types/azure-open-ai.js +0 -0
- package/dist/types/cerebras.d.ts +6 -0
- package/dist/types/cerebras.js +0 -0
- package/dist/types/chutes.d.ts +6 -0
- package/dist/types/chutes.js +0 -0
- package/dist/types/cloudflare.d.ts +6 -0
- package/dist/types/cloudflare.js +0 -0
- package/dist/types/deep-infra.d.ts +6 -0
- package/dist/types/deep-infra.js +0 -0
- package/dist/types/deep-seek.d.ts +6 -0
- package/dist/types/deep-seek.js +0 -0
- package/dist/types/fireworks-ai.d.ts +6 -0
- package/dist/types/fireworks-ai.js +0 -0
- package/dist/types/git-hub-copilot.d.ts +6 -0
- package/dist/types/git-hub-copilot.js +0 -0
- package/dist/types/git-hub-models.d.ts +6 -0
- package/dist/types/git-hub-models.js +0 -0
- package/dist/types/google-partner.d.ts +6 -0
- package/dist/types/google-partner.js +0 -0
- package/dist/types/google-vertex.d.ts +6 -0
- package/dist/types/google-vertex.js +0 -0
- package/dist/types/google.d.ts +6 -0
- package/dist/types/google.js +0 -0
- package/dist/types/groq.d.ts +6 -0
- package/dist/types/groq.js +0 -0
- package/dist/types/hugging-face.d.ts +6 -0
- package/dist/types/hugging-face.js +0 -0
- package/dist/types/inception.d.ts +6 -0
- package/dist/types/inception.js +0 -0
- package/dist/types/inference.d.ts +6 -0
- package/dist/types/inference.js +0 -0
- package/dist/types/meta.d.ts +6 -0
- package/dist/types/meta.js +0 -0
- package/dist/types/mistral.d.ts +6 -0
- package/dist/types/mistral.js +0 -0
- package/dist/types/model-scope.d.ts +6 -0
- package/dist/types/model-scope.js +0 -0
- package/dist/types/morph.d.ts +6 -0
- package/dist/types/morph.js +0 -0
- package/dist/types/open-ai.d.ts +6 -0
- package/dist/types/open-ai.js +0 -0
- package/dist/types/open-router.d.ts +6 -0
- package/dist/types/open-router.js +0 -0
- package/dist/types/providers.d.ts +11 -0
- package/dist/types/providers.js +1 -0
- package/dist/types/requesty.d.ts +6 -0
- package/dist/types/requesty.js +0 -0
- package/dist/types/together-ai.d.ts +6 -0
- package/dist/types/together-ai.js +0 -0
- package/dist/types/upstage.d.ts +6 -0
- package/dist/types/upstage.js +0 -0
- package/dist/types/v0.d.ts +6 -0
- package/dist/types/v0.js +0 -0
- package/dist/types/venice.d.ts +6 -0
- package/dist/types/venice.js +0 -0
- package/dist/types/vercel.d.ts +6 -0
- package/dist/types/vercel.js +0 -0
- package/dist/types/weights-_-biases.d.ts +6 -0
- package/dist/types/weights-_-biases.js +0 -0
- package/dist/types/xai.d.ts +6 -0
- package/dist/types/xai.js +0 -0
- package/package.json +6 -1
- package/public/alibaba.json +326 -0
- package/public/amazon-bedrock.json +2141 -0
- package/public/anthropic.json +396 -0
- package/public/api.json +13504 -3
- package/public/azure-open-ai.json +332 -0
- package/public/cerebras.json +288 -0
- package/public/chutes.json +3220 -0
- package/public/cloudflare.json +3094 -0
- package/public/deep-infra.json +956 -0
- package/public/deep-seek.json +46 -0
- package/public/fireworks-ai.json +14486 -0
- package/public/git-hub-copilot.json +676 -0
- package/public/git-hub-models.json +256 -0
- package/public/google-partner.json +536 -0
- package/public/google-vertex.json +1376 -0
- package/public/google.json +536 -0
- package/public/groq.json +882 -0
- package/public/hugging-face.json +1096 -0
- package/public/inception.json +81 -0
- package/public/inference.json +291 -0
- package/public/meta.json +151 -0
- package/public/mistral.json +2181 -0
- package/public/model-scope.json +37839 -0
- package/public/morph.json +46 -0
- package/public/open-ai.json +606 -0
- package/public/open-router.json +15341 -0
- package/public/providers.json +43 -0
- package/public/requesty.json +13757 -0
- package/public/together-ai.json +641 -0
- package/public/upstage.json +606 -0
- package/public/v0.json +431 -0
- package/public/venice.json +501 -0
- package/public/vercel.json +4071 -0
- package/public/weights-&-biases.json +551 -0
- package/public/xai.json +256 -0
|
@@ -0,0 +1,73 @@
|
|
|
1
|
+
var e={description:"AI Models API - Models from Requesty",lastUpdated:"2026-01-10T23:24:47.806Z",provider:"Requesty",totalModels:373,version:"0.0.0-development"},t=[{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"252-frequesty",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"252Frequesty",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"2-b-requesty",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"2BRequesty",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"content-requesty",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"contentRequesty",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"page-requesty",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"pageRequesty",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"u-003-c-requesty-api-key",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"u003cREQUESTY_API_KEY",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"why-choose-requesty",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"why-choose-requesty",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"your-requesty-api-key",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"YOUR_REQUESTY_API_KEY",openWeights:!1,provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:null,streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:8e-7},description:"Qwen3-30B-A3B-Instruct-2507 is a 30.5B-parameter mixture-of-experts language model from Qwen, with 3.3B active parameters per inference. It operates in non-thinking mode and is designed for high-quality instruction following, multilingual understanding, and agentic tool use. Post-trained on instruction data, it demonstrates competitive performance across reasoning (AIME, ZebraLogic), coding (MultiPL-E, LiveCodeBench), and alignment (IFEval, WritingBench) benchmarks. It outperforms its non-instruct variant on subjective and open-ended tasks while retaining strong factual and coding performance.",extendedThinking:!1,id:"alibaba-qwen-3-30-b-a-3-b-instruct-2507",knowledge:null,lastUpdated:null,limit:{context:131072,output:65536},modalities:{input:["text","image"],output:["text"]},name:"alibaba/qwen3-30b-a3b-instruct-2507",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-31T09:17:43.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:8e-8,output:15e-7},extendedThinking:!1,id:"alibaba-qwen-3-coder-flash",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65536},modalities:{input:["text","image"],output:["text"]},name:"alibaba/qwen3-coder-flash",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-24T09:17:43.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:null,output:5e-6},extendedThinking:!1,id:"alibaba-qwen-3-coder-plus",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65536},modalities:{input:["text","image"],output:["text"]},name:"alibaba/qwen3-coder-plus",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-24T09:17:43.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:861e-9,inputCacheHit:null,output:3441e-9},description:"This is the best-performing model in the Qwen series. It is ideal for complex, multi-step tasks.",extendedThinking:!1,id:"alibaba-qwen-3-max",knowledge:null,lastUpdated:null,limit:{context:262144,output:65536},modalities:{input:["text","image"],output:["text"]},name:"alibaba/qwen3-max",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-09-10T17:22:09.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:16e-7,inputCacheHit:16e-7,output:64e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
2
|
+
|
|
3
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"alibaba-qwen-max",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"alibaba/qwen-max",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:12e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
4
|
+
|
|
5
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"alibaba-qwen-plus",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"alibaba/qwen-plus",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-8,output:2e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
6
|
+
|
|
7
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"alibaba-qwen-turbo",knowledge:null,lastUpdated:null,limit:{context:1e6,output:null},modalities:{input:["text"],output:["text"]},name:"alibaba/qwen-turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:8e-7,inputCacheHit:8e-8,output:4e-6},description:"Anthropic's fastest model. Intelligence at blazing speeds.",extendedThinking:!1,id:"anthropic-claude-3-5-haiku",knowledge:null,lastUpdated:null,limit:{context:2e5,output:8192},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-3-5-haiku",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-04T00:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"anthropic-claude-3-7-sonnet",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-3-7-sonnet",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:3e-8,output:125e-8},description:"Fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance.",extendedThinking:!1,id:"anthropic-claude-3-haiku",knowledge:null,lastUpdated:null,limit:{context:2e5,output:4096},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-3-haiku",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-04T00:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-7,output:5e-6},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"anthropic-claude-haiku-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-haiku-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-11-24T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"anthropic-claude-opus-4",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-opus-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"anthropic-claude-opus-4-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-opus-4-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-05T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-7,output:25e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"anthropic-claude-opus-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-opus-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-05T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Claude Sonnet 4 significantly improves on Sonnet 3.7's industry-leading capabilities, excelling in coding with a state-of-the-art 72.7% on SWE-bench. The model balances performance and efficiency for internal and external use cases, with enhanced steerability for greater control over implementations.",extendedThinking:!0,id:"anthropic-claude-sonnet-4",knowledge:null,lastUpdated:null,limit:{context:1e6,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-sonnet-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Claude Sonnet 4.5 is Anthropic's best coding model in the world, leading on SWE-bench Verified (77.2%) and OSWorld (61.4%). It delivers sustained autonomous performance on complex tasks for over 30 hours—up from seven hours for Opus 4—maintaining focus and reliability throughout the entire software development lifecycle, with enhanced capabilities in tool handling, memory management, and context processing that make it the strongest model for building complex agents.",extendedThinking:!0,id:"anthropic-claude-sonnet-4-5",knowledge:null,lastUpdated:null,limit:{context:1e6,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"anthropic/claude-sonnet-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-09-29T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1-@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1-@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1-@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1-@uksouth",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-gpt-4-1-@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini@uksouth",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-gpt-4-1-mini@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-mini@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano@uksouth",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-gpt-4-1-nano@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/gpt-4.1-nano@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-11-26T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5-1-@swedencentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5.1@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-11-26T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5-@eastus-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5-@swedencentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},extendedThinking:!0,id:"azure-gpt-5-@uksouth",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},extendedThinking:!0,id:"azure-gpt-5-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},extendedThinking:!0,id:"azure-gpt-5-mini@eastus-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-mini@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},extendedThinking:!0,id:"azure-gpt-5-mini@swedencentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-mini@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},extendedThinking:!0,id:"azure-gpt-5-mini@uksouth",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-mini@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-9,output:4e-7},extendedThinking:!0,id:"azure-gpt-5-nano",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-9,output:4e-7},extendedThinking:!0,id:"azure-gpt-5-nano@eastus-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-nano@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-9,output:4e-7},extendedThinking:!0,id:"azure-gpt-5-nano@swedencentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/gpt-5-nano@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini@eastus-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini@francecentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini@swedencentral",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini@uksouth",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini@uksouth",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"azure-o-4-mini@westus-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"azure/o4-mini@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-mini",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-mini@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-mini@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-mini@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-mini@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-mini@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-mini@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-mini@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-mini@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-nano",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-nano@eastus-2",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-nano@eastus2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-nano@francecentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-nano@francecentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-nano@swedencentral",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-nano@swedencentral",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"azure-openai-responses-gpt-4-1-nano@westus-3",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"azure/openai-responses/gpt-4.1-nano@westus3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@eu-central-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@eu-central-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@eu-north-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@eu-north-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@eu-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@eu-west-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@eu-west-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@eu-west-3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@us-east-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@us-east-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@us-east-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@us-east-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-3-7-sonnet@us-west-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-3-7-sonnet@us-west-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-7,output:5e-6},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@eu-central-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@eu-central-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@eu-north-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@eu-north-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@eu-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@eu-west-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@eu-west-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@eu-west-3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@us-east-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@us-east-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@us-east-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@us-east-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:11e-8,output:55e-7},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"bedrock-claude-haiku-4-5-@us-west-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-haiku-4-5@us-west-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-23T10:45:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"bedrock-claude-opus-4",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-opus-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-7,output:25e-6},description:"Premium model combining maximum intelligence with practical performance",extendedThinking:!0,id:"bedrock-claude-opus-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-opus-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-01T22:17:16.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"bedrock-claude-opus-4-@us-east-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-opus-4@us-east-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"bedrock-claude-opus-4-@us-east-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-opus-4@us-east-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"bedrock-claude-opus-4-@us-west-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-opus-4@us-west-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@eu-central-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@eu-central-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@eu-north-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@eu-north-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@eu-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@eu-west-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@eu-west-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@eu-west-3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@us-east-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@us-east-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@us-east-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@us-east-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:33e-7,inputCacheHit:3e-7,output:165e-7},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-5-@us-west-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4-5@us-west-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@eu-central-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@eu-central-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@eu-north-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@eu-north-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@eu-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@eu-west-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@eu-west-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@eu-west-3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@us-east-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@us-east-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@us-east-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@us-east-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"bedrock-claude-sonnet-4-@us-west-2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"bedrock/claude-sonnet-4@us-west-2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219:1024",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:1024",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219:16384",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:16384",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219:64000",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:64000",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219:8192",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:8192",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219-:high",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:high",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219-:low",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:low",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219-:max",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:max",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"coding-claude-3-7-sonnet-20250219-:medium",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-3-7-sonnet-20250219:medium",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-24T18:35:10.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Claude Opus 4 is Anthropic's most powerful model yet and the best coding model in the world, leading on SWE-bench (72.5%) and Terminal-bench (43.2%). It delivers sustained performance on long-running tasks that require focused effort and thousands of steps, with the ability to work continuously for several hours—dramatically outperforming all Sonnet models and significantly expanding what AI agents can accomplish.",extendedThinking:!0,id:"coding-claude-opus-4-20250514",knowledge:null,lastUpdated:null,limit:{context:2e5,output:32e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-opus-4-20250514",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:27:25.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Claude Sonnet 4 significantly improves on Sonnet 3.7's industry-leading capabilities, excelling in coding with a state-of-the-art 72.7% on SWE-bench. The model balances performance and efficiency for internal and external use cases, with enhanced steerability for greater control over implementations.",extendedThinking:!0,id:"coding-claude-sonnet-4-20250514",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"coding/claude-sonnet-4-20250514",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@europe-central-2",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@europe-central2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@europe-north-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@europe-north1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@europe-west-4",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@europe-west4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@europe-west-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@europe-west8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@us-central-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@us-central1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@us-east-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@us-east1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@us-east-5",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@us-south-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@us-south1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"coding-gemini-2-5-flash@us-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-flash@us-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@europe-central-2",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@europe-central2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@europe-north-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@europe-north1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@europe-west-4",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@europe-west4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@europe-west-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@europe-west8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@us-central-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@us-central1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@us-east-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@us-east1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@us-east-5",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@us-south-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@us-south1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"coding-gemini-2-5-pro@us-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"coding/gemini-2.5-pro@us-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:85e-8,inputCacheHit:85e-8,output:25e-7},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"deepinfra-deepseek-ai-deep-seek-r1",knowledge:null,lastUpdated:null,limit:{context:64e3,output:8192},modalities:{input:["text"],output:["text"]},name:"deepinfra/deepseek-ai/DeepSeek-R1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:23e-8,inputCacheHit:23e-8,output:69e-8},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"deepinfra-deepseek-ai-deep-seek-r1-distill-llama-70b",knowledge:null,lastUpdated:null,limit:{context:64e3,output:8192},modalities:{input:["text"],output:["text"]},name:"deepinfra/deepseek-ai/DeepSeek-R1-Distill-Llama-70B",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:85e-8,inputCacheHit:85e-8,output:9e-7},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"deepinfra-deepseek-ai-deep-seek-v3",knowledge:null,lastUpdated:null,limit:{context:128e3,output:8192},modalities:{input:["text"],output:["text"]},name:"deepinfra/deepseek-ai/DeepSeek-V3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:1e-6},extendedThinking:!1,id:"deepinfra-deepseek-ai-deep-seek-v3-1",knowledge:null,lastUpdated:null,limit:{context:163840,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/deepseek-ai/DeepSeek-V3.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-26T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:35e-8,inputCacheHit:35e-8,output:4e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-llama-3-2-90b-vision-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:4096},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Llama-3.2-90B-Vision-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:23e-8,inputCacheHit:23e-8,output:4e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-llama-3-3-70b-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Llama-3.3-70B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:12e-8,inputCacheHit:12e-8,output:3e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-llama-3-3-70b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Llama-3.3-70B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:8e-7,inputCacheHit:8e-7,output:8e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-meta-llama-3-1-405b-instruct",knowledge:null,lastUpdated:null,limit:{context:130815,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Meta-Llama-3.1-405B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:23e-8,inputCacheHit:23e-8,output:4e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-meta-llama-3-1-70b-instruct",knowledge:null,lastUpdated:null,limit:{context:130815,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Meta-Llama-3.1-70B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-8,inputCacheHit:2e-8,output:5e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"deepinfra-meta-llama-meta-llama-3-1-8b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-05-14T14:42:34.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:7e-8,inputCacheHit:7e-8,output:14e-8},description:`Phi-4-reasoning-plus is an enhanced 14B parameter model from Microsoft, fine-tuned from Phi-4 with additional reinforcement learning to boost accuracy on math, science, and code reasoning tasks. It uses the same dense decoder-only transformer architecture as Phi-4, but generates longer, more comprehensive outputs structured into a step-by-step reasoning trace and final answer.
|
|
8
|
+
|
|
9
|
+
While it offers improved benchmark scores over Phi-4-reasoning across tasks like AIME, OmniMath, and HumanEvalPlus, its responses are typically ~50% longer, resulting in higher latency. Designed for English-only applications, it is well-suited for structured reasoning workflows where output quality takes priority over response speed.`,extendedThinking:!1,id:"deepinfra-microsoft-phi-4",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/microsoft/phi-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-01T21:22:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:23e-8,inputCacheHit:23e-8,output:4e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
10
|
+
|
|
11
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"deepinfra-qwen-qwen-2-5-72b-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/Qwen/Qwen2.5-72B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:7e-8,inputCacheHit:7e-8,output:16e-8},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
12
|
+
|
|
13
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"deepinfra-qwen-qwen-2-5-coder-32b-instruct",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/Qwen/Qwen2.5-Coder-32B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:6e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
14
|
+
|
|
15
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"deepinfra-qwen-qwen-3-235b-a22b",knowledge:null,lastUpdated:null,limit:{context:4e4,output:4096},modalities:{input:["text"],output:["text"]},name:"deepinfra/Qwen/Qwen3-235B-A22B",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:3e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
16
|
+
|
|
17
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"deepinfra-qwen-qwen-3-32b",knowledge:null,lastUpdated:null,limit:{context:4e4,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/Qwen/Qwen3-32B",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:16e-7},extendedThinking:!1,id:"deepinfra-qwen-qwen-3-coder-480b-a35b-instruct",knowledge:null,lastUpdated:null,limit:{context:262144,output:null},modalities:{input:["text"],output:["text"]},name:"deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-25T14:38:14.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:6e-7,output:22e-7},description:"The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities to meet the complex demands of intelligent agent applications.",extendedThinking:!1,id:"deepinfra-zai-org-glm-4-5",knowledge:null,lastUpdated:null,limit:{context:131072,output:4096},modalities:{input:["text"],output:["text"]},name:"deepinfra/zai-org/GLM-4.5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-30T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:11e-7},description:"The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities to meet the complex demands of intelligent agent applications.",extendedThinking:!1,id:"deepinfra-zai-org-glm-4-5-air",knowledge:null,lastUpdated:null,limit:{context:131072,output:4096},modalities:{input:["text"],output:["text"]},name:"deepinfra/zai-org/GLM-4.5-Air",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-30T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:28e-8,inputCacheHit:28e-9,output:42e-8},description:`DeepSeek-V3 achieves a significant breakthrough in inference speed over previous models.
|
|
18
|
+
|
|
19
|
+
It tops the leaderboard among open-source models and rivals the most advanced closed-source models globally.`,extendedThinking:!1,id:"deepseek-deepseek-chat",knowledge:null,lastUpdated:null,limit:{context:128e3,output:8e3},modalities:{input:["text"],output:["text"]},name:"deepseek/deepseek-chat",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:28e-8,inputCacheHit:28e-9,output:42e-8},description:"Fully open-source model & technical report. Performance on par with OpenAI-o1.",extendedThinking:!1,id:"deepseek-deepseek-reasoner",knowledge:null,lastUpdated:null,limit:{context:128e3,output:64e3},modalities:{input:["text"],output:["text"]},name:"deepseek/deepseek-reasoner",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:4e-7},description:"Gemini Flash 2.0 offers a significantly faster time to first token (TTFT) compared to [Gemini Flash 1.5](/google/gemini-flash-1.5), while maintaining quality on par with larger models like [Gemini Pro 1.5](/google/gemini-pro-1.5). It introduces notable enhancements in multimodal understanding, coding capabilities, complex instruction following, and function calling. These advancements come together to deliver more seamless and robust agentic experiences.",extendedThinking:!1,id:"google-gemini-2-0-flash-001",knowledge:null,lastUpdated:null,limit:{context:1048576,output:8192},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-2.0-flash-001",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-05T15:30:13.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"google-gemini-2-5-flash",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-2.5-flash",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"google-gemini-2-5-flash-lite",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-2.5-flash-lite",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"google-gemini-2-5-pro",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-2.5-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-7,inputCacheHit:5e-8,output:3e-6},description:"Gemini 3 Flash Preview is designed to deliver strong agentic capabilities (near-Pro level) at substantial speed and value. Making it perfect for engaging multi-turn chats, and collaborating back and forth with your coding agent without getting out of flow. Compared to 2.5 Flash it delivers significant improvements across the board.",extendedThinking:!0,id:"google-gemini-3-flash-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-3-flash-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-18T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-7,output:12e-6},description:"Gemini 3 Pro is designed to tackle the most challenging agentic problems with strong coding and state-of-the-art reasoning capabilities. It is the best model for complex multimodal understanding. Compared to Gemini 2.5 Pro, it improves significantly on complex instruction following and delivers outcomes with better output efficiency.",extendedThinking:!0,id:"google-gemini-3-pro-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"google/gemini-3-pro-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-6,output:3e-6},extendedThinking:!1,id:"groq-moonshotai-kimi-k-2-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:16384},modalities:{input:["text"],output:["text"]},name:"groq/moonshotai/kimi-k2-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-14T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-6,output:3e-6},description:"Moonshot AI’s cutting‑edge model, moonshotai/Kimi-K2-Instruct-0905, is now live on GroqCloud.",extendedThinking:!1,id:"groq-moonshotai-kimi-k2-instruct-0905",knowledge:null,lastUpdated:null,limit:{context:256e3,output:16384},modalities:{input:["text"],output:["text"]},name:"groq/moonshotai/Kimi-K2-Instruct-0905",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-09-10T17:16:37.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:15e-8,inputCacheHit:15e-8,output:75e-8},extendedThinking:!1,id:"groq-openai-gpt-oss-120-b",knowledge:null,lastUpdated:null,limit:{context:131072,output:32768},modalities:{input:["text"],output:["text"]},name:"groq/openai/gpt-oss-120b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-05T17:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:5e-7},extendedThinking:!1,id:"groq-openai-gpt-oss-20-b",knowledge:null,lastUpdated:null,limit:{context:131072,output:32768},modalities:{input:["text"],output:["text"]},name:"groq/openai/gpt-oss-20b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-05T17:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:12e-7},description:"MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency.",extendedThinking:!1,id:"minimaxi-mini-max-m2",knowledge:null,lastUpdated:null,limit:{context:2e5,output:128e3},modalities:{input:["text"],output:["text"]},name:"minimaxi/MiniMax-M2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-28T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:9e-7},extendedThinking:!1,id:"mistral-codestral-latest",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/codestral-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-12T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-4,inputCacheHit:null,output:3e-4},description:"An enterprise grade text model, that excels at using tools to explore codebases, editing multiple files and power software engineering agents.",extendedThinking:!1,id:"mistral-devstral-latest",knowledge:null,lastUpdated:null,limit:{context:256e3,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/devstral-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-12-10T00:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:2e-6},extendedThinking:!1,id:"mistral-devstral-medium-2507",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/devstral-medium-2507",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-12T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:3e-7},extendedThinking:!1,id:"mistral-devstral-small-2507",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/devstral-small-2507",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-12T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:3e-7},extendedThinking:!1,id:"mistral-devstral-small-latest",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/devstral-small-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-15T18:03:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-6,output:6e-6},extendedThinking:!1,id:"mistral-mistral-large-latest",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/mistral-large-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-02-24T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:2e-6},extendedThinking:!1,id:"mistral-mistral-medium-latest",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/mistral-medium-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-12T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:3e-7},extendedThinking:!1,id:"mistral-mistral-small-latest",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/mistral-small-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-02-24T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-8,output:25e-8},extendedThinking:!1,id:"mistral-open-mistral-7-b",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/open-mistral-7b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-02-24T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-6,output:5e-6},extendedThinking:!1,id:"mistral-pixtral-large-latest",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"mistral/pixtral-large-latest",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-12T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:15e-8,output:25e-7},description:"A Mixture-of-Experts (MoE) foundation model with exceptional coding and agent capabilities, featuring 1 trillion total parameters and 32 billion activated parameters. In benchmark evaluations covering general knowledge reasoning, programming, mathematics, and agent-related tasks, the K2 model outperforms other leading open-source models.",extendedThinking:!1,id:"moonshot-kimi-k-2-0711-preview",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"moonshot/kimi-k2-0711-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-14T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:15e-8,output:25e-7},description:"Based on kimi-k2-0711-preview, with enhanced agentic coding abilities, improved frontend code quality and practicality, and better context understanding",extendedThinking:!1,id:"moonshot-kimi-k-2-0905-preview",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"moonshot/kimi-k2-0905-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-06T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:15e-8,output:25e-7},description:"A thinking model with general agentic and reasoning capabilities, specializing in deep reasoning tasks.",extendedThinking:!1,id:"moonshot-kimi-k-2-thinking",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"moonshot/kimi-k2-thinking",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-06T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:15e-8,output:25e-7},description:"High-speed version of kimi-k2-thinking, suitable for scenarios requiring both deep reasoning and extremely fast responses",extendedThinking:!1,id:"moonshot-kimi-k-2-thinking-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"moonshot/kimi-k2-thinking-turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-06T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:12e-7,inputCacheHit:3e-7,output:5e-6},description:"A Mixture-of-Experts (MoE) foundation model with exceptional coding and agent capabilities, featuring 1 trillion total parameters and 32 billion activated parameters. In benchmark evaluations covering general knowledge reasoning, programming, mathematics, and agent-related tasks, the K2 model outperforms other leading open-source models.",extendedThinking:!1,id:"moonshot-kimi-k-2-turbo-preview",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"moonshot/kimi-k2-turbo-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-06T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:8e-7,inputCacheHit:8e-7,output:24e-7},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"nebius-deepseek-ai-deep-seek-r1-0528",knowledge:null,lastUpdated:null,limit:{context:164e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/deepseek-ai/DeepSeek-R1-0528",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-7,inputCacheHit:5e-7,output:15e-7},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"nebius-deepseek-ai-deep-seek-v3-0324",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/deepseek-ai/DeepSeek-V3-0324",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-6,output:6e-6},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"nebius-deepseek-ai-deep-seek-v3-0324-fast",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/deepseek-ai/DeepSeek-V3-0324-fast",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:13e-8,inputCacheHit:13e-8,output:4e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"nebius-meta-llama-llama-3-3-70b-instruct",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/meta-llama/Llama-3.3-70B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-05-14T14:42:34.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-7,inputCacheHit:5e-7,output:24e-7},extendedThinking:!1,id:"nebius-moonshotai-kimi-k2-instruct",knowledge:null,lastUpdated:null,limit:{context:131e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/moonshotai/Kimi-K2-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-10T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:18e-7},extendedThinking:!1,id:"nebius-qwen-qwen-3-coder-480b-a35b-instruct",knowledge:null,lastUpdated:null,limit:{context:262e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/Qwen/Qwen3-Coder-480B-A35B-Instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-10T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:6e-7,output:22e-7},extendedThinking:!1,id:"nebius-zai-org-glm-4-5",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"nebius/zai-org/GLM-4.5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-10T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:7e-7,inputCacheHit:7e-7,output:25e-7},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"novita-deepseek-deepseek-prover-v-2-671-b",knowledge:null,lastUpdated:null,limit:{context:16e4,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-prover-v2-671b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-6,inputCacheHit:4e-6,output:4e-6},description:"DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.",extendedThinking:!1,id:"novita-deepseek-deepseek-r-1",knowledge:null,lastUpdated:null,limit:{context:64e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-r1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:8e-7,inputCacheHit:8e-7,output:8e-7},description:"DeepSeek R1 Distill LLama 70B",extendedThinking:!1,id:"novita-deepseek-deepseek-r-1-distill-llama-70-b",knowledge:null,lastUpdated:null,limit:{context:32e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-r1-distill-llama-70b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:15e-8,inputCacheHit:15e-8,output:15e-8},description:`DeepSeek R1 Distill Qwen 14B is a distilled large language model based on Qwen 2.5 14B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.
|
|
20
|
+
|
|
21
|
+
Other benchmark results include:
|
|
22
|
+
|
|
23
|
+
AIME 2024 pass@1: 69.7
|
|
24
|
+
MATH-500 pass@1: 93.9
|
|
25
|
+
CodeForces Rating: 1481
|
|
26
|
+
The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.`,extendedThinking:!1,id:"novita-deepseek-deepseek-r-1-distill-qwen-14-b",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-r1-distill-qwen-14b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:3e-7},description:`DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.
|
|
27
|
+
|
|
28
|
+
Other benchmark results include:
|
|
29
|
+
AIME 2024 pass@1: 72.6
|
|
30
|
+
MATH-500 pass@1: 94.3
|
|
31
|
+
CodeForces Rating: 1691
|
|
32
|
+
The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.`,extendedThinking:!1,id:"novita-deepseek-deepseek-r-1-distill-qwen-32-b",knowledge:null,lastUpdated:null,limit:{context:12800,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-r1-distill-qwen-32b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:7e-7,inputCacheHit:7e-7,output:25e-7},description:"DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.",extendedThinking:!1,id:"novita-deepseek-deepseek-r-1-turbo",knowledge:null,lastUpdated:null,limit:{context:64e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-r1-turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:89e-8,inputCacheHit:89e-8,output:89e-8},description:"DeepSeek-V3 is the latest model from the DeepSeek team, building upon the instruction following and coding abilities of the previous versions. Pre-trained on nearly 15 trillion tokens, the reported evaluations reveal that the model outperforms other open-source models and rivals leading closed-source models.",extendedThinking:!1,id:"novita-deepseek-deepseek-v-3",knowledge:null,lastUpdated:null,limit:{context:64e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek_v3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:13e-7},description:"DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.",extendedThinking:!1,id:"novita-deepseek-deepseek-v-3-0324",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-v3-0324",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:4e-7,output:13e-7},description:"DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.",extendedThinking:!1,id:"novita-deepseek-deepseek-v-3-turbo",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/deepseek/deepseek-v3-turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:9e-8,inputCacheHit:9e-8,output:9e-8},description:"The idea behind this merge is that each layer is composed of several tensors, which are in turn responsible for specific functions. Using MythoLogic-L2's robust understanding as its input and Huginn's extensive writing capability as its output seems to have resulted in a model that exceeds at both, confirming my theory. (More details to be released at a later time).",extendedThinking:!1,id:"novita-gryphe-mythomax-l-2-13-b",knowledge:null,lastUpdated:null,limit:{context:4096,output:null},modalities:{input:["text"],output:["text"]},name:"novita/gryphe/mythomax-l2-13b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2023-07-02T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-8,output:5e-8},description:"Meta's latest class of models, Llama 3.1, launched with a variety of sizes and configurations. The 8B instruct-tuned version is particularly fast and efficient. It has demonstrated strong performance in human evaluations, outperforming several leading closed-source models.",extendedThinking:!1,id:"novita-meta-llama-llama-3-1-8-b-instruct",knowledge:null,lastUpdated:null,limit:{context:16384,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3.1-8b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-8,inputCacheHit:2e-8,output:2e-8},description:"The Meta Llama 3.2 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 1B and 3B sizes (text in/text out).",extendedThinking:!1,id:"novita-meta-llama-llama-3-2-1-b-instruct",knowledge:null,lastUpdated:null,limit:{context:131e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3.2-1b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-8,inputCacheHit:3e-8,output:5e-8},description:"The Meta Llama 3.2 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 1B and 3B sizes (text in/text out)",extendedThinking:!1,id:"novita-meta-llama-llama-3-2-3-b-instruct",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3.2-3b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:39e-8,inputCacheHit:39e-8,output:39e-8},description:`The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model is optimized for multilingual dialogue use cases and outperforms many of the available open source and closed chat models on common industry benchmarks.
|
|
33
|
+
|
|
34
|
+
Supported languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.`,extendedThinking:!1,id:"novita-meta-llama-llama-3-3-70-b-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3.3-70b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:51e-8,inputCacheHit:51e-8,output:74e-8},description:"Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations.",extendedThinking:!1,id:"novita-meta-llama-llama-3-70-b-instruct",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3-70b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:4e-8,inputCacheHit:4e-8,output:4e-8},description:"Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations.",extendedThinking:!1,id:"novita-meta-llama-llama-3-8-b-instruct",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-3-8b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:85e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"novita-meta-llama-llama-4-maverick-17-b-128-e-instruct-fp-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:1048576},modalities:{input:["text"],output:["text"]},name:"novita/meta-llama/llama-4-maverick-17b-128e-instruct-fp8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:62e-8,inputCacheHit:62e-8,output:62e-8},description:"WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing state-of-the-art opensource models.",extendedThinking:!1,id:"novita-microsoft-wizardlm-2-8-x-22-b",knowledge:null,lastUpdated:null,limit:{context:65535,output:null},modalities:{input:["text"],output:["text"]},name:"novita/microsoft/wizardlm-2-8x22b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:17e-8,inputCacheHit:17e-8,output:17e-8},description:"A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It supports function calling and is released under the Apache 2.0 license.",extendedThinking:!1,id:"novita-mistralai-mistral-nemo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"novita/mistralai/mistral-nemo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-06-10T16:32:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:57e-8,inputCacheHit:57e-8,output:23e-7},extendedThinking:!1,id:"novita-moonshotai-kimi-k-2-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"novita/moonshotai/kimi-k2-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-11T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:14e-8,inputCacheHit:14e-8,output:14e-8},description:"Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",extendedThinking:!1,id:"novita-nousresearch-hermes-2-pro-llama-3-8-b",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"novita/nousresearch/hermes-2-pro-llama-3-8b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:38e-8,inputCacheHit:38e-8,output:4e-7},description:"Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters.",extendedThinking:!1,id:"novita-qwen-qwen-2-5-72-b-instruct",knowledge:null,lastUpdated:null,limit:{context:32e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/qwen/qwen-2.5-72b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:8e-7,inputCacheHit:8e-7,output:8e-7},description:`Qwen2 VL 72B is a multimodal LLM from the Qwen Team with the following key enhancements:
|
|
35
|
+
|
|
36
|
+
SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc.
|
|
37
|
+
|
|
38
|
+
Understanding videos of 20min+: Qwen2-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc.
|
|
39
|
+
|
|
40
|
+
Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions.
|
|
41
|
+
|
|
42
|
+
Multilingual Support: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside images, including most European languages, Japanese, Korean, Arabic, Vietnamese, etc.`,extendedThinking:!1,id:"novita-qwen-qwen-2-5-vl-72-b-instruct",knowledge:null,lastUpdated:null,limit:{context:96e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/qwen/qwen2.5-vl-72b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:8e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
43
|
+
|
|
44
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"novita-qwen-qwen-3-235-b-a-22-b-fp-8",knowledge:null,lastUpdated:null,limit:{context:128e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/qwen/qwen3-235b-a22b-fp8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:148e-8,inputCacheHit:148e-8,output:148e-8},description:"Euryale L3.1 70B v2.2 is a model focused on creative roleplay from Sao10k. It is the successor of Euryale L3 70B v2.1.",extendedThinking:!1,id:"novita-sao-10-k-l-31-70-b-euryale-v-2-2",knowledge:null,lastUpdated:null,limit:{context:16e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/sao10k/l31-70b-euryale-v2.2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-12-18T15:32:08.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:148e-8,inputCacheHit:148e-8,output:148e-8},description:"The uncensored llama3 model is a powerhouse of creativity, excelling in both roleplay and story writing. It offers a liberating experience during roleplays, free from any restrictions. This model stands out for its immense creativity, boasting a vast array of unique ideas and plots, truly a treasure trove for those seeking originality. Its unrestricted nature during roleplays allows for the full breadth of imagination to unfold, akin to an enhanced, big-brained version of Stheno. Perfect for creative minds seeking a boundless platform for their imaginative expressions, the uncensored llama3 model is an ideal choice",extendedThinking:!1,id:"novita-sao-10-k-l-3-70-b-euryale-v-2-1",knowledge:null,lastUpdated:null,limit:{context:16e3,output:null},modalities:{input:["text"],output:["text"]},name:"novita/sao10k/l3-70b-euryale-v2.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-12-18T15:32:08.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-8,output:5e-8},description:"A generalist / roleplaying model merge based on Llama 3.",extendedThinking:!1,id:"novita-sao-10-k-l-3-8-b-lunaris",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"novita/sao10k/l3-8b-lunaris",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-12-18T15:32:08.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-8,output:5e-8},description:"Sao10K/L3-8B-Stheno-v3.2 is a highly skilled actor that excels at fully immersing itself in any role assigned.",extendedThinking:!1,id:"novita-sao-10-k-l3-8b-stheno-v-3-2",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"novita/Sao10K/L3-8B-Stheno-v3.2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-12-18T15:32:08.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:6e-7,output:22e-7},extendedThinking:!1,id:"novita-zai-org-glm-4-5",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"novita/zai-org/glm-4.5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-30T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:6e-7,output:22e-7},description:`GLM-4.6 is Z AI’s latest flagship model, designed to push agentic and coding performance further. It expands the context window from 128K to 200K tokens, improves reasoning and tool-use capabilities, and delivers stronger results in coding benchmarks and real-world development workflows. GLM-4.6 demonstrates refined writing quality, more capable agent behavior, and higher token efficiency (≈15% fewer tokens vs. GLM-4.5).
|
|
45
|
+
|
|
46
|
+
Evaluations show clear gains over GLM-4.5 across reasoning, agents, and coding, reaching near parity with Claude Sonnet 4 in practical tasks while outperforming other open-source baselines. GLM-4.6 is available through the Z.ai API platform, OpenRouter, coding agents (Claude Code, Roo Code, Cline, Kilo Code), and soon as downloadable weights on HuggingFace and ModelScope.`,extendedThinking:!1,id:"novita-zai-org-glm-4-6",knowledge:null,lastUpdated:null,limit:{context:204800,output:131072},modalities:{input:["text"],output:["text"]},name:"novita/zai-org/glm-4.6",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-30T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-6,output:15e-6},description:`OpenAI ChatGPT 4o is continually updated by OpenAI to point to the current version of GPT-4o used by ChatGPT. It therefore differs slightly from the API version of [GPT-4o](/models/openai/gpt-4o) in that it has additional RLHF. It is intended for research and evaluation.
|
|
47
|
+
|
|
48
|
+
OpenAI notes that this model is not suited for production use-cases as it may be removed or redirected to another model in the future.`,extendedThinking:!1,id:"openai-chatgpt-4-o",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16e3},modalities:{input:["text","image"],output:["text"]},name:"openai/chatgpt-4o",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-08-14T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"openai-gpt-4-1",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"openai-gpt-4-1-mini",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4.1-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"openai-gpt-4-1-nano",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4.1-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-7,inputCacheHit:125e-8,output:1e-5},description:`The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses.
|
|
49
|
+
|
|
50
|
+
GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.`,extendedThinking:!1,id:"openai-gpt-4-o",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4o",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-20T18:33:14.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-7,inputCacheHit:25e-7,output:1e-5},description:`The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses.
|
|
51
|
+
|
|
52
|
+
GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.`,extendedThinking:!1,id:"openai-gpt-4-o-2024-05-13",knowledge:null,lastUpdated:null,limit:{context:128e3,output:4096},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4o-2024-05-13",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-20T18:33:14.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-7,inputCacheHit:125e-8,output:1e-5},description:`The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses.
|
|
53
|
+
|
|
54
|
+
GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.`,extendedThinking:!1,id:"openai-gpt-4-o-2024-08-06",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4o-2024-08-06",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-20T18:33:14.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-7,inputCacheHit:125e-8,output:1e-5},description:`The 2024-11-20 version of GPT-4o offers a leveled-up creative writing ability with more natural, engaging, and tailored writing to improve relevance & readability. It’s also better at working with uploaded files, providing deeper insights & more thorough responses.
|
|
55
|
+
|
|
56
|
+
GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.`,extendedThinking:!1,id:"openai-gpt-4-o-2024-11-20",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4o-2024-11-20",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-11-20T18:33:14.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-8,inputCacheHit:75e-9,output:6e-7},description:`GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs.
|
|
57
|
+
|
|
58
|
+
As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective.
|
|
59
|
+
|
|
60
|
+
GPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/).
|
|
61
|
+
|
|
62
|
+
Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more.
|
|
63
|
+
|
|
64
|
+
#multimodal`,extendedThinking:!1,id:"openai-gpt-4-o-mini",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-4o-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-07-18T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5-1",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5.1 Chat points to the GPT-5.1 snapshot currently used in ChatGPT. We recommend GPT-5.1 for most API usage, but feel free to use this GPT-5.1 Chat model to test our latest improvements for chat use cases.",extendedThinking:!0,id:"openai-gpt-5-1-chat",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5.1-chat",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-01T17:28:11.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:175e-8,inputCacheHit:175e-9,output:14e-6},description:"The best model for coding and agentic tasks across industries",extendedThinking:!0,id:"openai-gpt-5-2",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5.2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-11T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:175e-8,inputCacheHit:175e-9,output:14e-6},description:"GPT‑5.2 sets a new state of the art across many benchmarks, including GDPval, where it outperforms industry professionals at well-specified knowledge work tasks spanning 44 occupations.",extendedThinking:!0,id:"openai-gpt-5-2-chat",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5.2-chat",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-19T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:625e-9,inputCacheHit:625e-10,output:5e-6},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5-:flex",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5:flex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-7,inputCacheHit:25e-8,output:2e-5},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5-:priority",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5:priority",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5-chat",knowledge:null,lastUpdated:null,limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-chat",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},description:"GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.",extendedThinking:!0,id:"openai-gpt-5-mini",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:05:19.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-9,inputCacheHit:125e-10,output:1e-6},description:"GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.",extendedThinking:!0,id:"openai-gpt-5-mini:flex",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-mini:flex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:05:19.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:45e-8,inputCacheHit:45e-9,output:36e-7},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-gpt-5-mini:priority",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-mini:priority",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-9,output:4e-7},description:"GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.",extendedThinking:!0,id:"openai-gpt-5-nano",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:07:35.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-9,inputCacheHit:25e-10,output:2e-7},description:"GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.",extendedThinking:!0,id:"openai-gpt-5-nano:flex",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai/gpt-5-nano:flex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:07:35.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:75e-7,output:6e-5},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-12-17T18:26:39.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:75e-7,output:6e-5},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-1-:high",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o1:high",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-12-17T18:26:39.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:75e-7,output:6e-5},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-1-:low",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o1:low",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-12-17T18:26:39.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:75e-7,output:6e-5},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-1-:medium",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o1:medium",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-12-17T18:26:39.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-5,inputCacheHit:15e-5,output:6e-4},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-1-pro",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o1-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"The o1 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-o-3",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T18:10:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:25e-8,output:4e-6},description:"O3 Flex is a cheaper version of the o3 model",extendedThinking:!0,id:"openai-o-3-:flex",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o3:flex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T18:10:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-5,inputCacheHit:25e-7,output:4e-5},description:"O3 Deep Research is a premium OpenAI model tuned for long-context research and high-recall reasoning tasks, optimized for analytical depth.",extendedThinking:!0,id:"openai-o-3-deep-research",knowledge:null,lastUpdated:null,limit:{context:1e5,output:2e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o3-deep-research",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-07-17T17:01:06.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:55e-8,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-3-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o3-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-01-31T19:28:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:55e-8,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-3-mini:high",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o3-mini:high",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-01-31T19:28:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:55e-8,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-3-mini:low",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o3-mini:low",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-01-31T19:28:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:55e-8,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-3-mini:medium",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o3-mini:medium",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-01-31T19:28:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-4-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o4-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:55e-8,inputCacheHit:138e-9,output:22e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-4-mini:flex",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o4-mini:flex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-4-mini:high",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o4-mini:high",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-4-mini:low",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o4-mini:low",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-o-4-mini:medium",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai/o4-mini:medium",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"Optimized for fast reasoning and minimal latency, O4 Mini Deep Research supports caching and high-speed inference. Ideal for lightweight agent use.",extendedThinking:!0,id:"openai-o-4-mini-deep-research",knowledge:null,lastUpdated:null,limit:{context:1e5,output:2e5},modalities:{input:["text","image"],output:["text"]},name:"openai/o4-mini-deep-research",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-07-17T17:00:50.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:5e-7,output:8e-6},description:"GPT-4.1 is a flagship large language model optimized for advanced instruction following, real-world software engineering, and long-context reasoning. It supports a 1 million token context window and outperforms GPT-4o and GPT-4.5 across coding (54.6% SWE-bench Verified), instruction compliance (87.4% IFEval), and multimodal understanding benchmarks. It is tuned for precise code diffs, agent reliability, and high recall in large document contexts, making it ideal for agents, IDE tooling, and enterprise knowledge retrieval.",extendedThinking:!1,id:"openai-responses-gpt-4-1",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-4.1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:05.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:4e-7,inputCacheHit:1e-7,output:16e-7},description:"GPT-4.1 Mini is a mid-sized model delivering performance competitive with GPT-4o at substantially lower latency and cost. It retains a 1 million token context window and scores 45.1% on hard instruction evals, 35.8% on MultiChallenge, and 84.1% on IFEval. Mini also shows strong coding ability (e.g., 31.6% on Aider’s polyglot diff benchmark) and vision understanding, making it suitable for interactive applications with tight performance constraints.",extendedThinking:!1,id:"openai-responses-gpt-4-1-mini",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-4.1-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:23:01.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:25e-9,output:4e-7},description:"For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.",extendedThinking:!1,id:"openai-responses-gpt-4-1-nano",knowledge:null,lastUpdated:null,limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-4.1-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-14T18:22:49.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5 is OpenAI's flagship model for coding, reasoning, and agentic tasks across domains.",extendedThinking:!0,id:"openai-responses-gpt-5",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:08:26.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5.1-Codex is a version of GPT-5 optimized for agentic coding tasks in Codex or similar environments",extendedThinking:!0,id:"openai-responses-gpt-5-1-codex",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5.1-codex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-02T23:43:36.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-9,output:1e-5},description:"GPT-5-Codex is a version of GPT-5 optimized for agentic coding tasks in Codex or similar environments",extendedThinking:!0,id:"openai-responses-gpt-5-codex",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5-codex",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-09-24T11:05:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:25e-8,inputCacheHit:25e-9,output:2e-6},description:"GPT-5 mini is a faster, more cost-efficient version of GPT-5. It's great for well-defined tasks and precise prompts.",extendedThinking:!0,id:"openai-responses-gpt-5-mini",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:05:19.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-8,inputCacheHit:5e-9,output:4e-7},description:"GPT-5 nano is OpenAI's fastest, cheapest version of GPT-5. It's great for summarization and classification tasks.",extendedThinking:!0,id:"openai-responses-gpt-5-nano",knowledge:null,lastUpdated:null,limit:{context:4e5,output:128e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5-nano",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-08-07T17:07:35.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-6,output:12e-5},description:"GPT-5 Pro is OpenAI’s extended-reasoning tier of GPT-5, built to push reliability on hard problems, long tool chains, and agentic workflows. It keeps GPT-5’s multimodal skills and very large context (API page lists up to 400K tokens) while allocating more compute to think longer and plan better, improving code generation, math, and complex writing beyond standard GPT-5/“Thinking.” OpenAI positions Pro as the version that “uses extended reasoning for even more comprehensive and accurate answers,” targeting high-stakes tasks and enterprise use.",extendedThinking:!0,id:"openai-responses-gpt-5-pro",knowledge:null,lastUpdated:null,limit:{context:4e5,output:272e3},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/gpt-5-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-06T11:05:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:55e-8,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-responses-o-3-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai-responses/o3-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-01-31T19:28:41.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-5,inputCacheHit:2e-5,output:8e-5},description:"The o3 series of models are trained with reinforcement learning to perform complex reasoning. o1 models think before they answer, producing a long internal chain of thought before responding to the user. The o1 reasoning model is designed to solve hard problems across domains. The knowledge cutoff for o1 and o1-mini models is October, 2023.",extendedThinking:!0,id:"openai-responses-o-3-pro",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"openai-responses/o3-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-06-11T00:32:32.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:11e-7,inputCacheHit:275e-9,output:44e-7},description:"o3-mini is OpenAI's most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks.",extendedThinking:!0,id:"openai-responses-o-4-mini",knowledge:null,lastUpdated:null,limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"openai-responses/o4-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-04-16T17:29:02.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:5e-7},description:"Gemma 3 1B is the smallest of the new Gemma 3 family. It handles context windows up to 32k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Note: Gemma 3 1B is not multimodal. For the smallest multimodal Gemma 3 model, please see [Gemma 3 4B](google/gemma-3-4b-it)",extendedThinking:!1,id:"parasail-parasail-gemma-3-27-b-it",knowledge:null,lastUpdated:null,limit:{context:128e3,output:8192},modalities:{input:["text"],output:["text"]},name:"parasail/parasail-gemma3-27b-it",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-03-14T14:45:56.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:99e-8,inputCacheHit:99e-8,output:299e-8},extendedThinking:!1,id:"parasail-parasail-kimi-k-2-instruct",knowledge:null,lastUpdated:null,limit:{context:131072,output:16384},modalities:{input:["text"],output:["text"]},name:"parasail/parasail-kimi-k2-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-08-18T16:23:40.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:7e-7,inputCacheHit:7e-7,output:7e-7},extendedThinking:!1,id:"parasail-parasail-qwen-25-vl-72-b-instruct",knowledge:null,lastUpdated:null,limit:{context:32768,output:8192},modalities:{input:["text","image"],output:["text"]},name:"parasail/parasail-qwen25-vl-72b-instruct",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-8,inputCacheHit:15e-8,output:85e-8},extendedThinking:!1,id:"parasail-parasail-qwen-3-235-b-a-22-b-instruct-2507",knowledge:null,lastUpdated:null,limit:{context:262144,output:8192},modalities:{input:["text"],output:["text"]},name:"parasail/parasail-qwen3-235b-a22b-instruct-2507",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-07-21T16:23:40.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-6,output:1e-6},description:"Lightweight offering with search grounding, quicker and cheaper than Sonar Pro.",extendedThinking:!1,id:"perplexity-sonar",knowledge:null,lastUpdated:null,limit:{context:131072,output:8192},modalities:{input:["text"],output:["text"]},name:"perplexity/sonar",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-03-07T02:08:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-6,output:15e-6},description:"Premier search offering with search grounding, supporting advanced queries and follow-ups.",extendedThinking:!1,id:"perplexity-sonar-pro",knowledge:null,lastUpdated:null,limit:{context:204800,output:8192},modalities:{input:["text"],output:["text"]},name:"perplexity/sonar-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-03-07T02:08:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-6,output:8e-6},description:"Premier reasoning offering powered by DeepSeek R1 with Chain of Thought (CoT).",extendedThinking:!1,id:"perplexity-sonar-reasoning-pro",knowledge:null,lastUpdated:null,limit:{context:131072,output:8192},modalities:{input:["text"],output:["text"]},name:"perplexity/sonar-reasoning-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-03-07T02:08:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!0,id:"smart-task",knowledge:null,lastUpdated:null,limit:{context:null,output:null},modalities:{input:["text","image"],output:["text"]},name:"smart/task",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-02-06T18:18:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-6,output:7e-6},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"together-deepseek-ai-deep-seek-r1",knowledge:null,lastUpdated:null,limit:{context:64e3,output:8192},modalities:{input:["text"],output:["text"]},name:"together/deepseek-ai/DeepSeek-R1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:125e-8,output:125e-8},description:"DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.",extendedThinking:!1,id:"together-deepseek-ai-deep-seek-v3",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"together/deepseek-ai/DeepSeek-V3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-8,inputCacheHit:6e-8,output:6e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-llama-3-2-3b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/Llama-3.2-3B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:88e-8,inputCacheHit:88e-8,output:88e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-llama-3-3-70b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/Llama-3.3-70B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-7,output:2e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-llama-guard-2-8-b",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/LlamaGuard-2-8b",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:88e-8,inputCacheHit:88e-8,output:88e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-meta-llama-3-1-70b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:18e-8,inputCacheHit:18e-8,output:18e-8},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-meta-llama-3-1-8b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-7,output:1e-7},description:"A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",extendedThinking:!1,id:"together-meta-llama-meta-llama-3-8b-instruct-lite",knowledge:null,lastUpdated:null,limit:{context:8192,output:null},modalities:{input:["text"],output:["text"]},name:"together/meta-llama/Meta-Llama-3-8B-Instruct-Lite",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-01-30T19:03:57.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:12e-7,inputCacheHit:12e-7,output:12e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
65
|
+
|
|
66
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"together-qwen-qwen-2-5-72b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"together/Qwen/Qwen2.5-72B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:3e-7},description:`Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique ability to switch seamlessly between a thinking mode for complex reasoning and a non-thinking mode for efficient dialogue ensures versatile, high-quality performance.
|
|
67
|
+
|
|
68
|
+
Significantly outperforming prior models like QwQ and Qwen2.5, Qwen3 delivers superior mathematics, coding, commonsense reasoning, creative writing, and interactive dialogue capabilities. The Qwen3-30B-A3B variant includes 30.5 billion parameters (3.3 billion activated), 48 layers, 128 experts (8 activated per task), and supports up to 131K token contexts with YaRN, setting a new standard among open-source models.`,extendedThinking:!1,id:"together-qwen-qwen-2-5-7b-instruct-turbo",knowledge:null,lastUpdated:null,limit:{context:32768,output:null},modalities:{input:["text"],output:["text"]},name:"together/Qwen/Qwen2.5-7B-Instruct-Turbo",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-04-28T23:16:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's previous most intelligent model. High level of intelligence and capability. Excells in coding.",extendedThinking:!1,id:"vertex-claude-3-5-sonnet",knowledge:null,lastUpdated:null,limit:{context:2e5,output:8192},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-5-sonnet",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's previous most intelligent model. High level of intelligence and capability. Excells in coding.",extendedThinking:!1,id:"vertex-claude-3-5-sonnet@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:8192},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-5-sonnet@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's previous most intelligent model. High level of intelligence and capability. Excells in coding.",extendedThinking:!1,id:"vertex-claude-3-5-sonnet@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:8192},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-5-sonnet@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-3-7-sonnet",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-7-sonnet",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-3-7-sonnet@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-7-sonnet@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-3-7-sonnet@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-3-7-sonnet@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2024-06-20T01:00:00.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-7,output:5e-6},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"vertex-claude-haiku-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-haiku-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-7,output:5e-6},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"vertex-claude-haiku-4-5-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-haiku-4-5@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-6,inputCacheHit:1e-7,output:5e-6},description:"Anthropic Haiku 4.5",extendedThinking:!1,id:"vertex-claude-haiku-4-5-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-haiku-4-5@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4-1-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-1@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4-1-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-1@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-7,output:25e-6},extendedThinking:!0,id:"vertex-claude-opus-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-11-24T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-7,output:25e-6},extendedThinking:!0,id:"vertex-claude-opus-4-5-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-5@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-11-24T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-7,output:25e-6},extendedThinking:!0,id:"vertex-claude-opus-4-5-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4-5@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-11-24T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:15e-6,inputCacheHit:15e-7,output:75e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-opus-4-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-opus-4@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4-5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4-5-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4-5@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4-5-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4-5@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4-@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:3e-7,output:15e-6},description:"Anthropic's most intelligent model. The first hybrid reasoning model on the market with the highest level of intelligence and capability with toggleable extended thinking. Top-tier results in reasoning, coding, multilingual tasks, long-context handling, honesty, and image processing.",extendedThinking:!0,id:"vertex-claude-sonnet-4-@us-east-5",knowledge:null,lastUpdated:null,limit:{context:2e5,output:64e3},modalities:{input:["text","image"],output:["text"]},name:"vertex/claude-sonnet-4@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-22T17:12:51.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@europe-central-2",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@europe-central2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@europe-north-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@europe-north1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@europe-west-4",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@europe-west4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@europe-west-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@europe-west8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@us-central-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@us-central1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@us-east-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@us-east1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@us-east-5",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@us-south-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@us-south1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:75e-9,output:25e-7},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash@us-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash@us-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:3e-5},description:"Google's first hybrid reasoning model which supports a 1M token context window and has thinking budgets. Most balanced Gemini model, optimized for low latency use cases.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-image-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-image-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-09-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@europe-central-2",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@europe-central2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@europe-north-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@europe-north1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@europe-west-4",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@europe-west4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@europe-west-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@europe-west8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@us-central-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@us-central1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@us-east-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@us-east1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@us-east-5",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@us-south-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@us-south1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:1e-7,inputCacheHit:1e-8,output:4e-7},description:"Google's smallest and most cost effective model, built for at scale usage.",extendedThinking:!0,id:"vertex-gemini-2-5-flash-lite@us-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-flash-lite@us-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-20T18:25:24.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@europe-central-2",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@europe-central2",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@europe-north-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@europe-north1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@europe-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@europe-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@europe-west-4",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@europe-west4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@europe-west-8",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@europe-west8",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@us-central-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@us-central1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@us-east-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@us-east1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@us-east-5",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@us-east5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@us-south-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@us-south1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:125e-8,inputCacheHit:31e-8,output:1e-5},description:"Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",extendedThinking:!0,id:"vertex-gemini-2-5-pro@us-west-1",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-2.5-pro@us-west1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:5e-7,inputCacheHit:5e-8,output:3e-6},description:"Gemini 3 Flash Preview is designed to deliver strong agentic capabilities (near-Pro level) at substantial speed and value. Making it perfect for engaging multi-turn chats, and collaborating back and forth with your coding agent without getting out of flow. Compared to 2.5 Flash it delivers significant improvements across the board.",extendedThinking:!0,id:"vertex-gemini-3-flash-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-3-flash-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-18T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-7,output:12e-6},description:"Gemini 3 Pro Image, or Gemini 3 Pro (with Nano Banana), is designed to tackle the most challenging image generation by incorporating state-of-the-art reasoning capabilities. It's the best model for complex and multi-turn image generation and editing, having improved accuracy and enhanced image quality.",extendedThinking:!0,id:"vertex-gemini-3-pro-image-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-3-pro-image-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-7,output:12e-6},description:"Gemini 3 Pro is designed to tackle the most challenging agentic problems with strong coding and state-of-the-art reasoning capabilities. It is the best model for complex multimodal understanding. Compared to Gemini 2.5 Pro, it improves significantly on complex instruction following and delivers outcomes with better output efficiency.",extendedThinking:!0,id:"vertex-gemini-3-pro-preview",knowledge:null,lastUpdated:null,limit:{context:1048576,output:65535},modalities:{input:["text","image"],output:["text"]},name:"vertex/gemini-3-pro-preview",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-05-07T01:41:53.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-6,inputCacheHit:2e-6,output:1e-5},description:`x AI's Our previous generation chat model.
|
|
69
|
+
`,extendedThinking:!1,id:"xai-grok-2-1212",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-2-1212",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:5e-6,inputCacheHit:5e-6,output:25e-6},description:"Excels at enterprise use cases like data extraction, coding, and text summarization. Possesses deep domain knowledge in finance, healthcare, law, and science.",extendedThinking:!1,id:"xai-grok-3",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-3",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:5e-7},description:"A lightweight model that thinks before responding. Fast, smart, and great for logic-based tasks that do not require deep domain knowledge. The raw thinking traces are accessible.",extendedThinking:!0,id:"xai-grok-3-mini",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-3-mini",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:5e-7},description:"A lightweight model that thinks before responding. Fast, smart, and great for logic-based tasks that do not require deep domain knowledge. The raw thinking traces are accessible.",extendedThinking:!0,id:"xai-grok-3-mini:high",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-3-mini:high",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-7,inputCacheHit:3e-7,output:5e-7},description:"A lightweight model that thinks before responding. Fast, smart, and great for logic-based tasks that do not require deep domain knowledge. The raw thinking traces are accessible.",extendedThinking:!0,id:"xai-grok-3-mini:low",knowledge:null,lastUpdated:null,limit:{context:131072,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-3-mini:low",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:3e-6,inputCacheHit:75e-8,output:15e-6},description:"xAI's latest and greatest flagship model, offering unparalleled performance in natural language, math and reasoning - the perfect jack of all trades.",extendedThinking:!1,id:"xai-grok-4",knowledge:null,lastUpdated:null,limit:{context:256e3,output:null},modalities:{input:["text","image"],output:["text"]},name:"xai/grok-4",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:5e-8,output:5e-7},description:"A frontier multimodal model optimized specifically for high-performance agentic tool calling.",extendedThinking:!1,id:"xai-grok-4-1-fast-non-reasoning",knowledge:null,lastUpdated:null,limit:{context:2e6,output:null},modalities:{input:["text","image"],output:["text"]},name:"xai/grok-4-1-fast-non-reasoning",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-12-01T17:14:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:5e-8,output:5e-7},description:"A frontier multimodal model optimized specifically for high-performance agentic tool calling.",extendedThinking:!1,id:"xai-grok-4-1-fast-reasoning",knowledge:null,lastUpdated:null,limit:{context:2e6,output:null},modalities:{input:["text","image"],output:["text"]},name:"xai/grok-4-1-fast-reasoning",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-12-01T17:14:44.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:5e-8,output:5e-7},description:"xAI's latest advancement in cost-efficient reasoning models",extendedThinking:!1,id:"xai-grok-4-fast",knowledge:null,lastUpdated:null,limit:{context:2e6,output:3e4},modalities:{input:["text"],output:["text"]},name:"xai/grok-4-fast",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:5e-8,output:5e-7},description:"xAI's latest advancement in cost-efficient reasoning models",extendedThinking:!1,id:"xai-grok-4-fast-non-reasoning",knowledge:null,lastUpdated:null,limit:{context:2e6,output:null},modalities:{input:["text"],output:["text"]},name:"xai/grok-4-fast-non-reasoning",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:2e-7,inputCacheHit:2e-8,output:15e-7},description:"A speedy and economical reasoning model that excels at agentic coding",extendedThinking:!1,id:"xai-grok-code-fast-1",knowledge:null,lastUpdated:null,limit:{context:256e3,output:1e4},modalities:{input:["text","image"],output:["text"]},name:"xai/grok-code-fast-1",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!1,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!0,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:11e-8,output:22e-7},description:"GLM-4.5 and GLM-4.5-Air are Z AI's latest flagship models, purpose-built as foundational models for agent-oriented applications. Both leverage a Mixture-of-Experts (MoE) architecture. GLM-4.5 has a total parameter count of 355B with 32B active parameters per forward pass, while GLM-4.5-Air adopts a more streamlined design with 106B total parameters and 12B active parameters.",extendedThinking:!0,id:"zai-glm-4-5",knowledge:null,lastUpdated:null,limit:{context:131072,output:98304},modalities:{input:["text"],output:["text"]},name:"zai/GLM-4.5",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:11e-8,output:22e-7},description:`GLM-4.6 is Z AI’s latest flagship model, designed to push agentic and coding performance further. It expands the context window from 128K to 200K tokens, improves reasoning and tool-use capabilities, and delivers stronger results in coding benchmarks and real-world development workflows. GLM-4.6 demonstrates refined writing quality, more capable agent behavior, and higher token efficiency (≈15% fewer tokens vs. GLM-4.5).
|
|
70
|
+
|
|
71
|
+
Evaluations show clear gains over GLM-4.5 across reasoning, agents, and coding, reaching near parity with Claude Sonnet 4 in practical tasks while outperforming other open-source baselines. GLM-4.6 is available through the Z.ai API platform, OpenRouter, coding agents (Claude Code, Roo Code, Cline, Kilo Code), and soon as downloadable weights on HuggingFace and ModelScope.`,extendedThinking:!0,id:"zai-glm-4-6",knowledge:null,lastUpdated:null,limit:{context:2e5,output:128e3},modalities:{input:["text"],output:["text"]},name:"zai/GLM-4.6",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-10-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"},{attachment:!1,cost:{input:6e-7,inputCacheHit:11e-8,output:22e-7},description:`GLM-4.7 is Z AI’s latest flagship model, designed to push agentic and coding performance further. It expands the context window from 128K to 200K tokens, improves reasoning and tool-use capabilities, and delivers stronger results in coding benchmarks and real-world development workflows. GLM-4.6 demonstrates refined writing quality, more capable agent behavior, and higher token efficiency (≈15% fewer tokens vs. GLM-4.5).
|
|
72
|
+
|
|
73
|
+
Evaluations show clear gains over GLM-4.5 across reasoning, agents, and coding, reaching near parity with Claude Sonnet 4 in practical tasks while outperforming other open-source baselines. GLM-4.6 is available through the Z.ai API platform, OpenRouter, coding agents (Claude Code, Roo Code, Cline, Kilo Code), and soon as downloadable weights on HuggingFace and ModelScope.`,extendedThinking:!0,id:"zai-glm-4-7",knowledge:null,lastUpdated:null,limit:{context:2e5,output:128e3},modalities:{input:["text"],output:["text"]},name:"zai/GLM-4.7",openWeights:!1,ownedBy:"system",provider:"Requesty",providerDoc:"https://docs.requesty.ai/",providerEnv:["REQUESTY_API_KEY"],providerId:"requesty",providerModelsDevId:"requesty",providerNpm:"@ai-sdk/requesty",reasoning:!0,releaseDate:"2025-12-16T10:16:28.000Z",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"requesty"}];const i={metadata:e,models:t};export{i as default,e as metadata,t as models};
|