@anolilab/ai-model-registry 4.0.2 → 4.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (159) hide show
  1. package/CHANGELOG.md +21 -0
  2. package/README.md +123 -0
  3. package/dist/icons-sprite.d.ts +1 -1
  4. package/dist/icons-sprite.js +3 -6
  5. package/dist/index.js +1 -1
  6. package/dist/packem_shared/{aihubmix-CjuksD_A.js → aihubmix-CKC2Vphr.js} +29 -25
  7. package/dist/packem_shared/alibaba-BLF61-pB.js +1 -0
  8. package/dist/packem_shared/{amazon-bedrock-BEpdRaRK.js → amazon-bedrock-CdFZeHv4.js} +1 -1
  9. package/dist/packem_shared/anthropic-Cj3HbHOC.js +5 -0
  10. package/dist/packem_shared/{azure-open-ai-C6QP5dSw.js → azure-open-ai-DlsrRCEy.js} +1 -1
  11. package/dist/packem_shared/cerebras-BotOp5tD.js +1 -0
  12. package/dist/packem_shared/chutes-BN981Qp1.js +2 -0
  13. package/dist/packem_shared/cloudflare-DOp6zAqg.js +3 -0
  14. package/dist/packem_shared/{deep-infra-BbtbGruj.js → deep-infra-B_BEr6MG.js} +1 -1
  15. package/dist/packem_shared/{deep-seek-CWqPVvJA.js → deep-seek-BxgLsWKo.js} +1 -1
  16. package/dist/packem_shared/fireworks-ai-SnVqnrNl.js +1 -0
  17. package/dist/packem_shared/git-hub-copilot-B9NwoDx6.js +1 -0
  18. package/dist/packem_shared/{google-CF3J5A9V.js → google-LG4qP1Ni.js} +1 -1
  19. package/dist/packem_shared/{google-partner-DV4aTl-d.js → google-partner-BEfKq6Z4.js} +1 -1
  20. package/dist/packem_shared/google-vertex-DtU07xbf.js +1 -0
  21. package/dist/packem_shared/{groq-BiW3tuc8.js → groq-ClFSzTLI.js} +1 -1
  22. package/dist/packem_shared/hugging-face-zI_cB94C.js +1 -0
  23. package/dist/packem_shared/inception-D2tOK5hm.js +1 -0
  24. package/dist/packem_shared/{inference-DeVnQ2rP.js → inference-BhdnsOKc.js} +1 -1
  25. package/dist/packem_shared/{meta-DgZfkCHT.js → meta-iORQYhl7.js} +1 -1
  26. package/dist/packem_shared/{mistral-Bqk6rtoD.js → mistral-D1nJG57r.js} +1 -1
  27. package/dist/packem_shared/model-scope-Dp4WKn9q.js +3 -0
  28. package/dist/packem_shared/ollama-cloud-DCEPhpO7.js +1 -0
  29. package/dist/packem_shared/{open-ai-byQBxiD0.js → open-ai-DRCCDDUM.js} +1 -1
  30. package/dist/packem_shared/open-router-DbfuPYJD.js +1 -0
  31. package/dist/packem_shared/requesty-CPVpB5vf.js +72 -0
  32. package/dist/packem_shared/{together-ai-DGMgKPuZ.js → together-ai-owzQDpv4.js} +1 -1
  33. package/dist/packem_shared/{upstage-CdH3YrFR.js → upstage-Bs8ieFx_.js} +1 -1
  34. package/dist/packem_shared/v0-BoOzNNt8.js +1 -0
  35. package/dist/packem_shared/{venice-C3HQT6s8.js → venice-ByXSis7D.js} +1 -1
  36. package/dist/packem_shared/vercel-cgwTxewS.js +1 -0
  37. package/dist/packem_shared/weights-_-biases-CRDc_nTA.js +1 -0
  38. package/dist/packem_shared/xai-DO2i-deZ.js +3 -0
  39. package/dist/providers/aihubmix.js +1 -1
  40. package/dist/providers/alibaba.js +1 -1
  41. package/dist/providers/amazon-bedrock.js +1 -1
  42. package/dist/providers/anthropic.js +1 -1
  43. package/dist/providers/azure-open-ai.js +1 -1
  44. package/dist/providers/cerebras.js +1 -1
  45. package/dist/providers/chutes.js +1 -1
  46. package/dist/providers/cloudflare.js +1 -1
  47. package/dist/providers/deep-infra.js +1 -1
  48. package/dist/providers/deep-seek.js +1 -1
  49. package/dist/providers/fireworks-ai.js +1 -1
  50. package/dist/providers/git-hub-copilot.js +1 -1
  51. package/dist/providers/google-partner.js +1 -1
  52. package/dist/providers/google-vertex.js +1 -1
  53. package/dist/providers/google.js +1 -1
  54. package/dist/providers/groq.js +1 -1
  55. package/dist/providers/hugging-face.js +1 -1
  56. package/dist/providers/inception.js +1 -1
  57. package/dist/providers/inference.js +1 -1
  58. package/dist/providers/meta.js +1 -1
  59. package/dist/providers/mistral.js +1 -1
  60. package/dist/providers/model-scope.js +1 -1
  61. package/dist/providers/ollama-cloud.js +1 -1
  62. package/dist/providers/open-ai.js +1 -1
  63. package/dist/providers/open-router.js +1 -1
  64. package/dist/providers/requesty.js +1 -1
  65. package/dist/providers/together-ai.js +1 -1
  66. package/dist/providers/upstage.js +1 -1
  67. package/dist/providers/v0.js +1 -1
  68. package/dist/providers/venice.js +1 -1
  69. package/dist/providers/vercel.js +1 -1
  70. package/dist/providers/weights-_-biases.js +1 -1
  71. package/dist/providers/xai.js +1 -1
  72. package/dist/schema.d.ts +2 -1
  73. package/dist/schema.js +1 -1
  74. package/dist/types/ai-hub-mix.d.ts +1 -1
  75. package/dist/types/anthropic.d.ts +1 -1
  76. package/dist/types/chutes.d.ts +1 -1
  77. package/dist/types/cloudflare.d.ts +1 -1
  78. package/dist/types/deep-infra.d.ts +1 -1
  79. package/dist/types/fireworks-ai.d.ts +1 -1
  80. package/dist/types/git-hub-copilot.d.ts +1 -1
  81. package/dist/types/google-vertex.d.ts +1 -1
  82. package/dist/types/hugging-face.d.ts +1 -1
  83. package/dist/types/inception.d.ts +1 -1
  84. package/dist/types/mistral.d.ts +1 -1
  85. package/dist/types/model-scope.d.ts +1 -1
  86. package/dist/types/ollama-cloud.d.ts +2 -2
  87. package/dist/types/open-router.d.ts +1 -1
  88. package/dist/types/providers.d.ts +2 -2
  89. package/dist/types/providers.js +1 -1
  90. package/dist/types/requesty.d.ts +1 -1
  91. package/dist/types/upstage.d.ts +1 -1
  92. package/dist/types/v0.d.ts +1 -1
  93. package/dist/types/vercel.d.ts +1 -1
  94. package/dist/types/weights-_-biases.d.ts +1 -1
  95. package/package.json +2 -2
  96. package/public/ai-hub-mix.json +2691 -327
  97. package/public/alibaba.json +23 -23
  98. package/public/amazon-bedrock.json +4 -4
  99. package/public/anthropic.json +114 -5
  100. package/public/azure-open-ai.json +2 -2
  101. package/public/cerebras.json +11 -11
  102. package/public/chutes.json +349 -9
  103. package/public/cloudflare.json +246 -9
  104. package/public/deep-infra.json +111 -6
  105. package/public/deep-seek.json +2 -2
  106. package/public/fireworks-ai.json +169 -5615
  107. package/public/git-hub-copilot.json +178 -3
  108. package/public/google-partner.json +2 -2
  109. package/public/google-vertex.json +38 -3
  110. package/public/google.json +2 -2
  111. package/public/groq.json +2 -2
  112. package/public/hugging-face.json +285 -5
  113. package/public/inception.json +73 -3
  114. package/public/inference.json +2 -2
  115. package/public/meta.json +2 -2
  116. package/public/mistral.json +38 -3
  117. package/public/model-scope.json +17735 -6066
  118. package/public/ollama-cloud.json +209 -961
  119. package/public/open-ai.json +2 -2
  120. package/public/open-router.json +2074 -744
  121. package/public/providers.json +4 -40
  122. package/public/requesty.json +5404 -2892
  123. package/public/together-ai.json +2 -2
  124. package/public/upstage.json +89 -54
  125. package/public/v0.json +73 -3
  126. package/public/venice.json +2 -2
  127. package/public/vercel.json +1587 -257
  128. package/public/weights-&-biases.json +349 -119
  129. package/public/xai.json +3 -2
  130. package/dist/packem_shared/alibaba-KCpNTbPH.js +0 -1
  131. package/dist/packem_shared/anthropic-D5m7WwK5.js +0 -2
  132. package/dist/packem_shared/cerebras-BjaINteN.js +0 -1
  133. package/dist/packem_shared/chutes-DJskRgMs.js +0 -3
  134. package/dist/packem_shared/cloudflare-C83X5Lbf.js +0 -3
  135. package/dist/packem_shared/fireworks-ai-D3bvzN8A.js +0 -1
  136. package/dist/packem_shared/git-hub-copilot-B0xJ4jvl.js +0 -1
  137. package/dist/packem_shared/git-hub-models-Cyf4N1hP.js +0 -1
  138. package/dist/packem_shared/google-vertex-CbWZUmxZ.js +0 -1
  139. package/dist/packem_shared/hugging-face-CWivgPFf.js +0 -1
  140. package/dist/packem_shared/inception-C2RuJCqh.js +0 -1
  141. package/dist/packem_shared/model-scope-B97EzMOm.js +0 -1
  142. package/dist/packem_shared/morph-CBEkhIbg.js +0 -1
  143. package/dist/packem_shared/ollama-cloud-AMbVVbEe.js +0 -1
  144. package/dist/packem_shared/open-router-Bgk_Eg85.js +0 -1
  145. package/dist/packem_shared/requesty-jpAC9INf.js +0 -73
  146. package/dist/packem_shared/v0-BjlxXJua.js +0 -1
  147. package/dist/packem_shared/vercel-CtkHacHk.js +0 -1
  148. package/dist/packem_shared/weights-_-biases-BJw-U5-q.js +0 -1
  149. package/dist/packem_shared/xai-ojVHf5mV.js +0 -1
  150. package/dist/providers/git-hub-models.d.ts +0 -49
  151. package/dist/providers/git-hub-models.js +0 -1
  152. package/dist/providers/morph.d.ts +0 -49
  153. package/dist/providers/morph.js +0 -1
  154. package/dist/types/git-hub-models.d.ts +0 -6
  155. package/dist/types/git-hub-models.js +0 -0
  156. package/dist/types/morph.d.ts +0 -6
  157. package/dist/types/morph.js +0 -0
  158. package/public/git-hub-models.json +0 -257
  159. package/public/morph.json +0 -47
@@ -0,0 +1 @@
1
+ var n=Object.defineProperty;var a=(e,t)=>n(e,"name",{value:t,configurable:!0});var d=[{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"ollama/glm-5",knowledge:null,lastUpdated:"2026-02-11",limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"glm-5",openWeights:!0,provider:"Ollama Cloud",providerDoc:"https://docs.ollama.com/cloud",providerEnv:["OLLAMA_API_KEY"],providerId:"ollama-cloud",providerModelsDevId:"ollama",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2026-02-11",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"ollama/kimi-k2.5",knowledge:null,lastUpdated:"2026-01-26",limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"kimi-k2.5",openWeights:!0,provider:"Ollama Cloud",providerDoc:"https://docs.ollama.com/cloud",providerEnv:["OLLAMA_API_KEY"],providerId:"ollama-cloud",providerModelsDevId:"ollama",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2026-01-26",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"ollama/minimax-m2.5",knowledge:null,lastUpdated:"2026-02-12",limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"minimax-m2.5",openWeights:!0,provider:"Ollama Cloud",providerDoc:"https://docs.ollama.com/cloud",providerEnv:["OLLAMA_API_KEY"],providerId:"ollama-cloud",providerModelsDevId:"ollama",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2026-02-12",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"ollama/qwen3.5:397b",knowledge:null,lastUpdated:"2026-02-16",limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"qwen3.5:397b",openWeights:!0,provider:"Ollama Cloud",providerDoc:"https://docs.ollama.com/cloud",providerEnv:["OLLAMA_API_KEY"],providerId:"ollama-cloud",providerModelsDevId:"ollama",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2026-02-16",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1},{attachment:!1,cost:{input:null,inputCacheHit:null,output:null},extendedThinking:!1,id:"ollama/qwen3-coder-next",knowledge:null,lastUpdated:"2025-02-04",limit:{context:null,output:null},modalities:{input:["text"],output:["text"]},name:"qwen3-coder-next",openWeights:!0,provider:"Ollama Cloud",providerDoc:"https://docs.ollama.com/cloud",providerEnv:["OLLAMA_API_KEY"],providerId:"ollama-cloud",providerModelsDevId:"ollama",providerNpm:"@ai-sdk/openai-compatible",reasoning:!1,releaseDate:"2025-02-04",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1}];const r={models:d};var u=Object.defineProperty,l=a((e,t)=>u(e,"name",{value:t,configurable:!0}),"t");const{models:o}=r,p=l(()=>[...o],"getModels"),s=l(e=>{if(e?.trim())return o.find(t=>t.id===e)},"getModelById"),m=l(e=>o.filter(t=>!(e.vision!==void 0&&t.vision!==e.vision||e.reasoning!==void 0&&t.reasoning!==e.reasoning||e.tool_call!==void 0&&t.toolCall!==e.tool_call||e.streaming_supported!==void 0&&t.streamingSupported!==e.streaming_supported||e.preview!==void 0&&t.preview!==e.preview||e.modalities?.input&&!e.modalities.input.every(i=>t.modalities.input.includes(i))||e.modalities?.output&&!e.modalities.output.every(i=>t.modalities.output.includes(i))||e.context_min!==void 0&&(!t.limit.context||t.limit.context<e.context_min)||e.context_max!==void 0&&(!t.limit.context||t.limit.context>e.context_max))),"searchModels"),c=l(()=>o.length,"getModelCount"),g=Object.defineProperty({__proto__:null,getModelById:s,getModelCount:c,getModels:p,searchModels:m},Symbol.toStringTag,{value:"Module"});export{g as V,c as a,p as b,s as g,m as s};
@@ -1,3 +1,3 @@
1
1
  var r=Object.defineProperty;var a=(e,t)=>r(e,"name",{value:t,configurable:!0});var p=[{attachment:!0,cost:{input:1.5,inputCacheHit:.375,output:6},extendedThinking:!0,id:"codex-mini-latest",knowledge:"2024-04",lastUpdated:"2025-05-16",limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"Codex Mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2025-05-16",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!1,icon:"openai",description:`Only supports v1/responses API calls.https://docs.aihubmix.com/en/api/Responses-API
2
2
  codex-mini-latest is a fine-tuned version of o4-mini specifically for use in Codex CLI. For direct use in the API, we recommend starting with gpt-4.1.`,preview:!1},{attachment:!1,cost:{input:.5,inputCacheHit:1.25,output:1.5},extendedThinking:!1,id:"gpt-3.5-turbo",knowledge:"2021-09-01",lastUpdated:"2023-11-06",limit:{context:16385,output:4096},modalities:{input:["text"],output:["text"]},name:"GPT-3.5-turbo",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2023-03-01",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"openai",description:"Since the GPT-3.5-turbo model has been officially deprecated, all requests targeting this model will be automatically routed to GPT-40-mini. We recommend using GPT-40-mini directly as a replacement.",preview:!1},{attachment:!0,cost:{input:10,inputCacheHit:null,output:30},extendedThinking:!1,id:"gpt-4",knowledge:"2023-11",lastUpdated:"2024-04-09",limit:{context:8192,output:8192},modalities:{input:["text"],output:["text"]},name:"GPT-4",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2023-11-06",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"openai",preview:!1},{attachment:!0,cost:{input:2,inputCacheHit:.5,output:8},extendedThinking:!1,id:"gpt-4.1",knowledge:"2024-05",lastUpdated:"2025-04-14",limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"GPT-4.1",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2025-04-14",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",description:"The latest flagship multimodal model supports million-token context, with encoding capability (SWE-bench 54.6%) and instruction-following (Scale AI 38.3%) performance significantly surpassing GPT-4o, while reducing costs by 26%, making it suitable for complex tasks. Its automatic caching mechanism offers a 75% cost reduction on cache hits.",preview:!1},{attachment:!0,cost:{input:.4,inputCacheHit:.1,output:1.6},extendedThinking:!1,id:"gpt-4.1-mini",knowledge:"2024-04",lastUpdated:"2025-04-14",limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"GPT-4.1 mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2025-04-14",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",description:"Lightweight, high-performance model with million-token context and near-flagship-level encoding and image understanding capabilities, while reducing costs by 83%. It is suitable for rapid development and small to medium-sized applications. The automatic caching mechanism provides a 75% cost reduction on cache hits.",preview:!1},{attachment:!0,cost:{input:.1,inputCacheHit:.03,output:.4},extendedThinking:!1,id:"gpt-4.1-nano",knowledge:"2024-04",lastUpdated:"2025-04-14",limit:{context:1047576,output:32768},modalities:{input:["text","image"],output:["text"]},name:"GPT-4.1 nano",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2025-04-14",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",description:"Ultra-lightweight model with million-token context, optimized for speed and low latency, costing only $0.10 per million input tokens. It is suitable for edge computing and real-time interaction. The automatic caching mechanism offers a 75% cost reduction on cache hits.",preview:!1},{attachment:!1,cost:{input:60,inputCacheHit:null,output:120},extendedThinking:!1,id:"gpt-4-32k",knowledge:"2023-11",lastUpdated:"2023-03-14",limit:{context:32768,output:32768},modalities:{input:["text"],output:["text"]},name:"GPT-4 32K",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2023-03-14",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!1,icon:"openai",description:"The smartest version of GPT-4; OpenAI no longer offers it officially. All the 32k versions on this site are provided by Microsoft, deployed on Azure OpenAI by the official Microsoft service.",preview:!1},{attachment:!0,cost:{input:2.5,inputCacheHit:1.25,output:10},extendedThinking:!1,id:"gpt-4o",knowledge:"2023-09",lastUpdated:"2024-05-13",limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"GPT-4o",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2024-05-13",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",description:"GPT-4o (“o” stands for “omni”) is a new-generation multimodal model designed for more natural human–computer interaction. It can accept any combination of text, audio, image, and video as input, and generate multimodal outputs including text, audio, and images. With audio response latency as low as 232 milliseconds on average around 320 milliseconds, it approaches real human conversational speed. The model delivers strong performance in English text and code, significantly improved multilingual understanding, and outstanding capabilities in visual and audio perception, while offering faster API performance and substantially reduced cost for real-time and complex multimodal applications.",preview:!1},{attachment:!0,cost:{input:.15,inputCacheHit:.08,output:.6},extendedThinking:!1,id:"gpt-4o-mini",knowledge:"2023-09",lastUpdated:"2024-07-18",limit:{context:128e3,output:16384},modalities:{input:["text","image"],output:["text"]},name:"GPT-4o mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2024-07-18",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",description:"The lightweight version of GPT-4o, which is affordable and fast, suitable for handling simple tasks; our site supports the official automatic caching for this model, and charges for cache hits will be automatically halved.",preview:!1},{attachment:!0,cost:{input:10,inputCacheHit:null,output:30},extendedThinking:!1,id:"gpt-4-turbo",knowledge:"2023-11",lastUpdated:"2024-04-09",limit:{context:128e3,output:4096},modalities:{input:["text","image"],output:["text"]},name:"GPT-4 Turbo",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!1,releaseDate:"2023-11-06",streamingSupported:!0,temperature:!0,toolCall:!0,vision:!0,icon:"openai",preview:!1},{attachment:!1,cost:{input:15,inputCacheHit:7.5,output:60},extendedThinking:!0,id:"o1",knowledge:"2023-09",lastUpdated:"2024-12-05",limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"o1",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2024-12-05",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!0,icon:"openai",description:"OpenAI's most powerful O-series model supports official cache hits that halve the input cost.",preview:!1},{attachment:!1,cost:{input:1.1,inputCacheHit:.55,output:4.4},extendedThinking:!0,id:"o1-mini",knowledge:"2023-09",lastUpdated:"2024-09-12",limit:{context:128e3,output:65536},modalities:{input:["text"],output:["text"]},name:"o1-mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2024-09-12",streamingSupported:!0,temperature:!1,toolCall:!1,vision:!1,icon:"openai",description:"o1-mini is faster and 80% cheaper, and is competitive with o1-preview on coding tasks. AiHubMix uses both OpenAI and Microsoft Azure OpenAI channels simultaneously.",preview:!1},{attachment:!1,cost:{input:15,inputCacheHit:7.5,output:60},extendedThinking:!0,id:"o1-preview",knowledge:"2023-09",lastUpdated:"2024-09-12",limit:{context:128e3,output:32768},modalities:{input:["text"],output:["text"]},name:"o1-preview",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2024-09-12",streamingSupported:!0,temperature:!0,toolCall:!1,vision:!1,icon:"openai",description:"The latest and most powerful inference model from OpenAI; AiHubMix uses both OpenAI and Microsoft Azure OpenAI channels simultaneously to achieve high-concurrency load balancing.",preview:!0},{attachment:!0,cost:{input:2,inputCacheHit:.5,output:8},extendedThinking:!0,id:"o3",knowledge:"2024-05",lastUpdated:"2025-04-16",limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"o3",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2025-04-16",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!0,icon:"openai",description:"OpenAI o3 is a powerful model across multiple domains, setting a new standard for coding, math, science, and visual reasoning tasks.",preview:!1},{attachment:!1,cost:{input:1.1,inputCacheHit:.55,output:4.4},extendedThinking:!0,id:"o3-mini",knowledge:"2024-05",lastUpdated:"2025-01-29",limit:{context:2e5,output:1e5},modalities:{input:["text"],output:["text"]},name:"o3-mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2024-12-20",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!1,icon:"openai",description:"OpenAI's latest fast inference model excels at STEAM tasks and offers exceptional cost-effectiveness. Official support for cache hits reduces input prices by half.",preview:!1},{attachment:!0,cost:{input:20,inputCacheHit:null,output:80},extendedThinking:!0,id:"o3-pro",knowledge:"2024-05",lastUpdated:"2025-06-10",limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"o3-pro",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2025-06-10",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!0,icon:"openai",description:`o3-pro
3
- This model only supports Requests API interface requests.The model's thinking time is relatively long, so the response will be slow.`,preview:!1},{attachment:!0,cost:{input:1.1,inputCacheHit:.28,output:4.4},extendedThinking:!0,id:"o4-mini",knowledge:"2024-05",lastUpdated:"2025-04-16",limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"o4-mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2025-04-16",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!0,icon:"openai",description:"o4-mini is a remarkably smart model for its speed and cost-efficiency. This allows it to support significantly higher usage limits than o3, making it a strong high-volume, high-throughput option for everyone with questions that benefit from reasoning.",preview:!1}];const s={models:p};var d=Object.defineProperty,i=a((e,t)=>d(e,"name",{value:t,configurable:!0}),"t");const{models:o}=s,l=i(()=>[...o],"getModels"),u=i(e=>{if(e?.trim())return o.find(t=>t.id===e)},"getModelById"),m=i(e=>o.filter(t=>!(e.vision!==void 0&&t.vision!==e.vision||e.reasoning!==void 0&&t.reasoning!==e.reasoning||e.tool_call!==void 0&&t.toolCall!==e.tool_call||e.streaming_supported!==void 0&&t.streamingSupported!==e.streaming_supported||e.preview!==void 0&&t.preview!==e.preview||e.modalities?.input&&!e.modalities.input.every(n=>t.modalities.input.includes(n))||e.modalities?.output&&!e.modalities.output.every(n=>t.modalities.output.includes(n))||e.context_min!==void 0&&(!t.limit.context||t.limit.context<e.context_min)||e.context_max!==void 0&&(!t.limit.context||t.limit.context>e.context_max))),"searchModels"),c=i(()=>o.length,"getModelCount"),g=Object.defineProperty({__proto__:null,getModelById:u,getModelCount:c,getModels:l,searchModels:m},Symbol.toStringTag,{value:"Module"});export{u as a,c as b,l as g,m as s,g as z};
3
+ This model only supports Requests API interface requests.The model's thinking time is relatively long, so the response will be slow.`,preview:!1},{attachment:!0,cost:{input:1.1,inputCacheHit:.28,output:4.4},extendedThinking:!0,id:"o4-mini",knowledge:"2024-05",lastUpdated:"2025-04-16",limit:{context:2e5,output:1e5},modalities:{input:["text","image"],output:["text"]},name:"o4-mini",openWeights:!1,provider:"OpenAI",providerDoc:"https://platform.openai.com/docs/models",providerEnv:["OPENAI_API_KEY"],providerId:"open-ai",providerModelsDevId:"openai",providerNpm:"@ai-sdk/openai",reasoning:!0,releaseDate:"2025-04-16",streamingSupported:!0,temperature:!1,toolCall:!0,vision:!0,icon:"openai",description:"o4-mini is a remarkably smart model for its speed and cost-efficiency. This allows it to support significantly higher usage limits than o3, making it a strong high-volume, high-throughput option for everyone with questions that benefit from reasoning.",preview:!1}];const s={models:p};var d=Object.defineProperty,i=a((e,t)=>d(e,"name",{value:t,configurable:!0}),"t");const{models:o}=s,l=i(()=>[...o],"getModels"),u=i(e=>{if(e?.trim())return o.find(t=>t.id===e)},"getModelById"),m=i(e=>o.filter(t=>!(e.vision!==void 0&&t.vision!==e.vision||e.reasoning!==void 0&&t.reasoning!==e.reasoning||e.tool_call!==void 0&&t.toolCall!==e.tool_call||e.streaming_supported!==void 0&&t.streamingSupported!==e.streaming_supported||e.preview!==void 0&&t.preview!==e.preview||e.modalities?.input&&!e.modalities.input.every(n=>t.modalities.input.includes(n))||e.modalities?.output&&!e.modalities.output.every(n=>t.modalities.output.includes(n))||e.context_min!==void 0&&(!t.limit.context||t.limit.context<e.context_min)||e.context_max!==void 0&&(!t.limit.context||t.limit.context>e.context_max))),"searchModels"),c=i(()=>o.length,"getModelCount"),g=Object.defineProperty({__proto__:null,getModelById:u,getModelCount:c,getModels:l,searchModels:m},Symbol.toStringTag,{value:"Module"});export{c as a,l as b,u as g,g as q,m as s};