@anolilab/ai-model-registry 4.0.2 → 4.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (159) hide show
  1. package/CHANGELOG.md +21 -0
  2. package/README.md +123 -0
  3. package/dist/icons-sprite.d.ts +1 -1
  4. package/dist/icons-sprite.js +3 -6
  5. package/dist/index.js +1 -1
  6. package/dist/packem_shared/{aihubmix-CjuksD_A.js → aihubmix-CKC2Vphr.js} +29 -25
  7. package/dist/packem_shared/alibaba-BLF61-pB.js +1 -0
  8. package/dist/packem_shared/{amazon-bedrock-BEpdRaRK.js → amazon-bedrock-CdFZeHv4.js} +1 -1
  9. package/dist/packem_shared/anthropic-Cj3HbHOC.js +5 -0
  10. package/dist/packem_shared/{azure-open-ai-C6QP5dSw.js → azure-open-ai-DlsrRCEy.js} +1 -1
  11. package/dist/packem_shared/cerebras-BotOp5tD.js +1 -0
  12. package/dist/packem_shared/chutes-BN981Qp1.js +2 -0
  13. package/dist/packem_shared/cloudflare-DOp6zAqg.js +3 -0
  14. package/dist/packem_shared/{deep-infra-BbtbGruj.js → deep-infra-B_BEr6MG.js} +1 -1
  15. package/dist/packem_shared/{deep-seek-CWqPVvJA.js → deep-seek-BxgLsWKo.js} +1 -1
  16. package/dist/packem_shared/fireworks-ai-SnVqnrNl.js +1 -0
  17. package/dist/packem_shared/git-hub-copilot-B9NwoDx6.js +1 -0
  18. package/dist/packem_shared/{google-CF3J5A9V.js → google-LG4qP1Ni.js} +1 -1
  19. package/dist/packem_shared/{google-partner-DV4aTl-d.js → google-partner-BEfKq6Z4.js} +1 -1
  20. package/dist/packem_shared/google-vertex-DtU07xbf.js +1 -0
  21. package/dist/packem_shared/{groq-BiW3tuc8.js → groq-ClFSzTLI.js} +1 -1
  22. package/dist/packem_shared/hugging-face-zI_cB94C.js +1 -0
  23. package/dist/packem_shared/inception-D2tOK5hm.js +1 -0
  24. package/dist/packem_shared/{inference-DeVnQ2rP.js → inference-BhdnsOKc.js} +1 -1
  25. package/dist/packem_shared/{meta-DgZfkCHT.js → meta-iORQYhl7.js} +1 -1
  26. package/dist/packem_shared/{mistral-Bqk6rtoD.js → mistral-D1nJG57r.js} +1 -1
  27. package/dist/packem_shared/model-scope-Dp4WKn9q.js +3 -0
  28. package/dist/packem_shared/ollama-cloud-DCEPhpO7.js +1 -0
  29. package/dist/packem_shared/{open-ai-byQBxiD0.js → open-ai-DRCCDDUM.js} +1 -1
  30. package/dist/packem_shared/open-router-DbfuPYJD.js +1 -0
  31. package/dist/packem_shared/requesty-CPVpB5vf.js +72 -0
  32. package/dist/packem_shared/{together-ai-DGMgKPuZ.js → together-ai-owzQDpv4.js} +1 -1
  33. package/dist/packem_shared/{upstage-CdH3YrFR.js → upstage-Bs8ieFx_.js} +1 -1
  34. package/dist/packem_shared/v0-BoOzNNt8.js +1 -0
  35. package/dist/packem_shared/{venice-C3HQT6s8.js → venice-ByXSis7D.js} +1 -1
  36. package/dist/packem_shared/vercel-cgwTxewS.js +1 -0
  37. package/dist/packem_shared/weights-_-biases-CRDc_nTA.js +1 -0
  38. package/dist/packem_shared/xai-DO2i-deZ.js +3 -0
  39. package/dist/providers/aihubmix.js +1 -1
  40. package/dist/providers/alibaba.js +1 -1
  41. package/dist/providers/amazon-bedrock.js +1 -1
  42. package/dist/providers/anthropic.js +1 -1
  43. package/dist/providers/azure-open-ai.js +1 -1
  44. package/dist/providers/cerebras.js +1 -1
  45. package/dist/providers/chutes.js +1 -1
  46. package/dist/providers/cloudflare.js +1 -1
  47. package/dist/providers/deep-infra.js +1 -1
  48. package/dist/providers/deep-seek.js +1 -1
  49. package/dist/providers/fireworks-ai.js +1 -1
  50. package/dist/providers/git-hub-copilot.js +1 -1
  51. package/dist/providers/google-partner.js +1 -1
  52. package/dist/providers/google-vertex.js +1 -1
  53. package/dist/providers/google.js +1 -1
  54. package/dist/providers/groq.js +1 -1
  55. package/dist/providers/hugging-face.js +1 -1
  56. package/dist/providers/inception.js +1 -1
  57. package/dist/providers/inference.js +1 -1
  58. package/dist/providers/meta.js +1 -1
  59. package/dist/providers/mistral.js +1 -1
  60. package/dist/providers/model-scope.js +1 -1
  61. package/dist/providers/ollama-cloud.js +1 -1
  62. package/dist/providers/open-ai.js +1 -1
  63. package/dist/providers/open-router.js +1 -1
  64. package/dist/providers/requesty.js +1 -1
  65. package/dist/providers/together-ai.js +1 -1
  66. package/dist/providers/upstage.js +1 -1
  67. package/dist/providers/v0.js +1 -1
  68. package/dist/providers/venice.js +1 -1
  69. package/dist/providers/vercel.js +1 -1
  70. package/dist/providers/weights-_-biases.js +1 -1
  71. package/dist/providers/xai.js +1 -1
  72. package/dist/schema.d.ts +2 -1
  73. package/dist/schema.js +1 -1
  74. package/dist/types/ai-hub-mix.d.ts +1 -1
  75. package/dist/types/anthropic.d.ts +1 -1
  76. package/dist/types/chutes.d.ts +1 -1
  77. package/dist/types/cloudflare.d.ts +1 -1
  78. package/dist/types/deep-infra.d.ts +1 -1
  79. package/dist/types/fireworks-ai.d.ts +1 -1
  80. package/dist/types/git-hub-copilot.d.ts +1 -1
  81. package/dist/types/google-vertex.d.ts +1 -1
  82. package/dist/types/hugging-face.d.ts +1 -1
  83. package/dist/types/inception.d.ts +1 -1
  84. package/dist/types/mistral.d.ts +1 -1
  85. package/dist/types/model-scope.d.ts +1 -1
  86. package/dist/types/ollama-cloud.d.ts +2 -2
  87. package/dist/types/open-router.d.ts +1 -1
  88. package/dist/types/providers.d.ts +2 -2
  89. package/dist/types/providers.js +1 -1
  90. package/dist/types/requesty.d.ts +1 -1
  91. package/dist/types/upstage.d.ts +1 -1
  92. package/dist/types/v0.d.ts +1 -1
  93. package/dist/types/vercel.d.ts +1 -1
  94. package/dist/types/weights-_-biases.d.ts +1 -1
  95. package/package.json +2 -2
  96. package/public/ai-hub-mix.json +2691 -327
  97. package/public/alibaba.json +23 -23
  98. package/public/amazon-bedrock.json +4 -4
  99. package/public/anthropic.json +114 -5
  100. package/public/azure-open-ai.json +2 -2
  101. package/public/cerebras.json +11 -11
  102. package/public/chutes.json +349 -9
  103. package/public/cloudflare.json +246 -9
  104. package/public/deep-infra.json +111 -6
  105. package/public/deep-seek.json +2 -2
  106. package/public/fireworks-ai.json +169 -5615
  107. package/public/git-hub-copilot.json +178 -3
  108. package/public/google-partner.json +2 -2
  109. package/public/google-vertex.json +38 -3
  110. package/public/google.json +2 -2
  111. package/public/groq.json +2 -2
  112. package/public/hugging-face.json +285 -5
  113. package/public/inception.json +73 -3
  114. package/public/inference.json +2 -2
  115. package/public/meta.json +2 -2
  116. package/public/mistral.json +38 -3
  117. package/public/model-scope.json +17735 -6066
  118. package/public/ollama-cloud.json +209 -961
  119. package/public/open-ai.json +2 -2
  120. package/public/open-router.json +2074 -744
  121. package/public/providers.json +4 -40
  122. package/public/requesty.json +5404 -2892
  123. package/public/together-ai.json +2 -2
  124. package/public/upstage.json +89 -54
  125. package/public/v0.json +73 -3
  126. package/public/venice.json +2 -2
  127. package/public/vercel.json +1587 -257
  128. package/public/weights-&-biases.json +349 -119
  129. package/public/xai.json +3 -2
  130. package/dist/packem_shared/alibaba-KCpNTbPH.js +0 -1
  131. package/dist/packem_shared/anthropic-D5m7WwK5.js +0 -2
  132. package/dist/packem_shared/cerebras-BjaINteN.js +0 -1
  133. package/dist/packem_shared/chutes-DJskRgMs.js +0 -3
  134. package/dist/packem_shared/cloudflare-C83X5Lbf.js +0 -3
  135. package/dist/packem_shared/fireworks-ai-D3bvzN8A.js +0 -1
  136. package/dist/packem_shared/git-hub-copilot-B0xJ4jvl.js +0 -1
  137. package/dist/packem_shared/git-hub-models-Cyf4N1hP.js +0 -1
  138. package/dist/packem_shared/google-vertex-CbWZUmxZ.js +0 -1
  139. package/dist/packem_shared/hugging-face-CWivgPFf.js +0 -1
  140. package/dist/packem_shared/inception-C2RuJCqh.js +0 -1
  141. package/dist/packem_shared/model-scope-B97EzMOm.js +0 -1
  142. package/dist/packem_shared/morph-CBEkhIbg.js +0 -1
  143. package/dist/packem_shared/ollama-cloud-AMbVVbEe.js +0 -1
  144. package/dist/packem_shared/open-router-Bgk_Eg85.js +0 -1
  145. package/dist/packem_shared/requesty-jpAC9INf.js +0 -73
  146. package/dist/packem_shared/v0-BjlxXJua.js +0 -1
  147. package/dist/packem_shared/vercel-CtkHacHk.js +0 -1
  148. package/dist/packem_shared/weights-_-biases-BJw-U5-q.js +0 -1
  149. package/dist/packem_shared/xai-ojVHf5mV.js +0 -1
  150. package/dist/providers/git-hub-models.d.ts +0 -49
  151. package/dist/providers/git-hub-models.js +0 -1
  152. package/dist/providers/morph.d.ts +0 -49
  153. package/dist/providers/morph.js +0 -1
  154. package/dist/types/git-hub-models.d.ts +0 -6
  155. package/dist/types/git-hub-models.js +0 -0
  156. package/dist/types/morph.d.ts +0 -6
  157. package/dist/types/morph.js +0 -0
  158. package/public/git-hub-models.json +0 -257
  159. package/public/morph.json +0 -47
@@ -1,11 +1,11 @@
1
1
  {
2
2
  "metadata": {
3
3
  "description": "AI Models API - Models from Cloudflare",
4
- "lastUpdated": "2026-01-13T21:44:10.322Z",
4
+ "lastUpdated": "2026-03-02T07:56:54.908Z",
5
5
  "pricingUnit": "1K",
6
6
  "provider": "Cloudflare",
7
- "totalModels": 76,
8
- "version": "4.0.2"
7
+ "totalModels": 82,
8
+ "version": "4.1.0"
9
9
  },
10
10
  "models": [
11
11
  {
@@ -90,6 +90,47 @@
90
90
  "extendedThinking": false,
91
91
  "preview": false
92
92
  },
93
+ {
94
+ "attachment": false,
95
+ "audioGeneration": false,
96
+ "cost": {
97
+ "input": null,
98
+ "inputCacheHit": null,
99
+ "output": null
100
+ },
101
+ "description": "The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks.",
102
+ "id": "qwen3-embedding-0.6b",
103
+ "knowledge": null,
104
+ "lastUpdated": null,
105
+ "limit": {
106
+ "context": 8192000,
107
+ "output": null
108
+ },
109
+ "modalities": {
110
+ "input": ["text"],
111
+ "output": ["embedding"]
112
+ },
113
+ "name": "qwen3-embedding-0.6b",
114
+ "openWeights": false,
115
+ "provider": "Cloudflare",
116
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
117
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
118
+ "providerId": "cloudflare",
119
+ "providerNpm": "@ai-sdk/cloudflare",
120
+ "reasoning": false,
121
+ "releaseDate": null,
122
+ "streamingSupported": true,
123
+ "structuredOutputs": true,
124
+ "supportsStructuredOutput": true,
125
+ "supportsTools": true,
126
+ "temperature": true,
127
+ "toolCall": true,
128
+ "version": null,
129
+ "vision": false,
130
+ "icon": "cloudflare",
131
+ "extendedThinking": false,
132
+ "preview": false
133
+ },
93
134
  {
94
135
  "attachment": false,
95
136
  "audioGeneration": true,
@@ -126,7 +167,7 @@
126
167
  "temperature": true,
127
168
  "toolCall": true,
128
169
  "version": null,
129
- "vision": true,
170
+ "vision": false,
130
171
  "icon": "cloudflare"
131
172
  },
132
173
  {
@@ -165,7 +206,7 @@
165
206
  "temperature": true,
166
207
  "toolCall": true,
167
208
  "version": null,
168
- "vision": true,
209
+ "vision": false,
169
210
  "icon": "cloudflare"
170
211
  },
171
212
  {
@@ -204,7 +245,7 @@
204
245
  "temperature": true,
205
246
  "toolCall": true,
206
247
  "version": null,
207
- "vision": true,
248
+ "vision": false,
208
249
  "icon": "cloudflare"
209
250
  },
210
251
  {
@@ -259,7 +300,7 @@
259
300
  "knowledge": null,
260
301
  "lastUpdated": null,
261
302
  "limit": {
262
- "context": null,
303
+ "context": 153600000,
263
304
  "output": null
264
305
  },
265
306
  "modalities": {
@@ -534,7 +575,7 @@
534
575
  "lastUpdated": null,
535
576
  "limit": {
536
577
  "context": 80000000,
537
- "output": null
578
+ "output": 32768
538
579
  },
539
580
  "modalities": {
540
581
  "input": ["text"],
@@ -949,6 +990,45 @@
949
990
  "vision": false,
950
991
  "icon": "cloudflare"
951
992
  },
993
+ {
994
+ "attachment": false,
995
+ "audioGeneration": true,
996
+ "cost": {
997
+ "input": null,
998
+ "inputCacheHit": null,
999
+ "output": null
1000
+ },
1001
+ "description": "GLM-4.7-Flash is a fast and efficient multilingual text generation model with a 131,072 token context window. Optimized for dialogue, instruction-following, and multi-turn tool calling across 100+ languages.",
1002
+ "id": "glm-4.7-flash",
1003
+ "knowledge": null,
1004
+ "lastUpdated": null,
1005
+ "limit": {
1006
+ "context": 131072000,
1007
+ "output": null
1008
+ },
1009
+ "modalities": {
1010
+ "input": ["text", "image"],
1011
+ "output": ["text"]
1012
+ },
1013
+ "name": "glm-4.7-flash",
1014
+ "openWeights": false,
1015
+ "provider": "Cloudflare",
1016
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1017
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1018
+ "providerId": "cloudflare",
1019
+ "providerNpm": "@ai-sdk/cloudflare",
1020
+ "reasoning": false,
1021
+ "releaseDate": null,
1022
+ "streamingSupported": true,
1023
+ "structuredOutputs": true,
1024
+ "supportsStructuredOutput": true,
1025
+ "supportsTools": true,
1026
+ "temperature": true,
1027
+ "toolCall": true,
1028
+ "version": null,
1029
+ "vision": true,
1030
+ "icon": "cloudflare"
1031
+ },
952
1032
  {
953
1033
  "attachment": false,
954
1034
  "audioGeneration": false,
@@ -988,6 +1068,45 @@
988
1068
  "vision": false,
989
1069
  "icon": "cloudflare"
990
1070
  },
1071
+ {
1072
+ "attachment": false,
1073
+ "audioGeneration": false,
1074
+ "cost": {
1075
+ "input": 0.0003,
1076
+ "inputCacheHit": null,
1077
+ "output": 0.0003
1078
+ },
1079
+ "description": "Hermes 2 Pro on Mistral 7B is the new flagship 7B Hermes! Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
1080
+ "id": "hermes-2-pro-mistral-7b",
1081
+ "knowledge": null,
1082
+ "lastUpdated": null,
1083
+ "limit": {
1084
+ "context": 24000000,
1085
+ "output": null
1086
+ },
1087
+ "modalities": {
1088
+ "input": ["text"],
1089
+ "output": ["text"]
1090
+ },
1091
+ "name": "hermes-2-pro-mistral-7b",
1092
+ "openWeights": false,
1093
+ "provider": "Cloudflare",
1094
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1095
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1096
+ "providerId": "cloudflare",
1097
+ "providerNpm": "@ai-sdk/cloudflare",
1098
+ "reasoning": false,
1099
+ "releaseDate": null,
1100
+ "streamingSupported": true,
1101
+ "structuredOutputs": true,
1102
+ "supportsStructuredOutput": true,
1103
+ "supportsTools": true,
1104
+ "temperature": true,
1105
+ "toolCall": true,
1106
+ "version": null,
1107
+ "vision": false,
1108
+ "icon": "cloudflare"
1109
+ },
991
1110
  {
992
1111
  "attachment": false,
993
1112
  "audioGeneration": false,
@@ -1534,6 +1653,45 @@
1534
1653
  "vision": false,
1535
1654
  "icon": "cloudflare"
1536
1655
  },
1656
+ {
1657
+ "attachment": false,
1658
+ "audioGeneration": false,
1659
+ "cost": {
1660
+ "input": 0.00004,
1661
+ "inputCacheHit": null,
1662
+ "output": 0.00004
1663
+ },
1664
+ "description": "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.",
1665
+ "id": "llama-3-8b-instruct",
1666
+ "knowledge": null,
1667
+ "lastUpdated": null,
1668
+ "limit": {
1669
+ "context": 7968000,
1670
+ "output": null
1671
+ },
1672
+ "modalities": {
1673
+ "input": ["text"],
1674
+ "output": ["text"]
1675
+ },
1676
+ "name": "llama-3-8b-instruct",
1677
+ "openWeights": false,
1678
+ "provider": "Cloudflare",
1679
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1680
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1681
+ "providerId": "cloudflare",
1682
+ "providerNpm": "@ai-sdk/cloudflare",
1683
+ "reasoning": false,
1684
+ "releaseDate": null,
1685
+ "streamingSupported": true,
1686
+ "structuredOutputs": true,
1687
+ "supportsStructuredOutput": true,
1688
+ "supportsTools": true,
1689
+ "temperature": true,
1690
+ "toolCall": true,
1691
+ "version": null,
1692
+ "vision": false,
1693
+ "icon": "cloudflare"
1694
+ },
1537
1695
  {
1538
1696
  "attachment": false,
1539
1697
  "audioGeneration": false,
@@ -1573,6 +1731,46 @@
1573
1731
  "vision": false,
1574
1732
  "icon": "cloudflare"
1575
1733
  },
1734
+ {
1735
+ "attachment": false,
1736
+ "audioGeneration": false,
1737
+ "cost": {
1738
+ "input": 0.0002,
1739
+ "inputCacheHit": null,
1740
+ "output": 0.0002
1741
+ },
1742
+ "description": "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.",
1743
+ "extendedThinking": false,
1744
+ "id": "llama-guard-3-8b",
1745
+ "knowledge": null,
1746
+ "lastUpdated": null,
1747
+ "limit": {
1748
+ "context": 131072000,
1749
+ "output": null
1750
+ },
1751
+ "modalities": {
1752
+ "input": ["text"],
1753
+ "output": ["text"]
1754
+ },
1755
+ "name": "llama-guard-3-8b",
1756
+ "openWeights": false,
1757
+ "provider": "Cloudflare",
1758
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
1759
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
1760
+ "providerId": "cloudflare",
1761
+ "providerNpm": "@ai-sdk/cloudflare",
1762
+ "reasoning": false,
1763
+ "releaseDate": "2025-02-12",
1764
+ "streamingSupported": true,
1765
+ "structuredOutputs": true,
1766
+ "supportsStructuredOutput": true,
1767
+ "supportsTools": true,
1768
+ "temperature": true,
1769
+ "toolCall": true,
1770
+ "version": null,
1771
+ "vision": false,
1772
+ "icon": "cloudflare"
1773
+ },
1576
1774
  {
1577
1775
  "attachment": false,
1578
1776
  "audioGeneration": false,
@@ -2407,7 +2605,7 @@
2407
2605
  "lastUpdated": null,
2408
2606
  "limit": {
2409
2607
  "context": 32768000,
2410
- "output": 32768
2608
+ "output": 8192
2411
2609
  },
2412
2610
  "modalities": {
2413
2611
  "input": ["text"],
@@ -2471,6 +2669,45 @@
2471
2669
  "vision": false,
2472
2670
  "icon": "cloudflare"
2473
2671
  },
2672
+ {
2673
+ "attachment": false,
2674
+ "audioGeneration": false,
2675
+ "cost": {
2676
+ "input": 0.00003,
2677
+ "inputCacheHit": null,
2678
+ "output": 0.00011
2679
+ },
2680
+ "description": "QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.",
2681
+ "id": "qwq-32b",
2682
+ "knowledge": null,
2683
+ "lastUpdated": null,
2684
+ "limit": {
2685
+ "context": 24000000,
2686
+ "output": null
2687
+ },
2688
+ "modalities": {
2689
+ "input": ["text", "image"],
2690
+ "output": ["text"]
2691
+ },
2692
+ "name": "qwq-32b",
2693
+ "openWeights": false,
2694
+ "provider": "Cloudflare",
2695
+ "providerDoc": "https://developers.cloudflare.com/workers-ai/models",
2696
+ "providerEnv": ["CLOUDFLARE_API_TOKEN"],
2697
+ "providerId": "cloudflare",
2698
+ "providerNpm": "@ai-sdk/cloudflare",
2699
+ "reasoning": false,
2700
+ "releaseDate": null,
2701
+ "streamingSupported": true,
2702
+ "structuredOutputs": true,
2703
+ "supportsStructuredOutput": true,
2704
+ "supportsTools": true,
2705
+ "temperature": true,
2706
+ "toolCall": true,
2707
+ "version": null,
2708
+ "vision": true,
2709
+ "icon": "cloudflare"
2710
+ },
2474
2711
  {
2475
2712
  "attachment": false,
2476
2713
  "audioGeneration": false,
@@ -1,11 +1,11 @@
1
1
  {
2
2
  "metadata": {
3
3
  "description": "AI Models API - Models from Deep Infra",
4
- "lastUpdated": "2026-01-13T21:44:10.328Z",
4
+ "lastUpdated": "2026-03-02T07:56:54.912Z",
5
5
  "pricingUnit": "1K",
6
6
  "provider": "Deep Infra",
7
- "totalModels": 26,
8
- "version": "4.0.2"
7
+ "totalModels": 29,
8
+ "version": "4.1.0"
9
9
  },
10
10
  "models": [
11
11
  {
@@ -328,7 +328,7 @@
328
328
  "cost": {
329
329
  "input": 0.00026,
330
330
  "inputCacheHit": null,
331
- "output": 0.00039
331
+ "output": 0.00038
332
332
  },
333
333
  "extendedThinking": false,
334
334
  "id": "deepseek-ai/DeepSeek-V3.2",
@@ -603,6 +603,41 @@
603
603
  "vision": false,
604
604
  "icon": "deepinfra"
605
605
  },
606
+ {
607
+ "attachment": false,
608
+ "cost": {
609
+ "input": 0.00027,
610
+ "inputCacheHit": null,
611
+ "output": 0.00095
612
+ },
613
+ "extendedThinking": false,
614
+ "id": "MiniMaxAI/MiniMax-M2.5",
615
+ "knowledge": null,
616
+ "lastUpdated": null,
617
+ "limit": {
618
+ "context": 196608,
619
+ "output": null
620
+ },
621
+ "modalities": {
622
+ "input": ["text"],
623
+ "output": ["text"]
624
+ },
625
+ "name": "MiniMax-M2.5",
626
+ "openWeights": false,
627
+ "provider": "Deep Infra",
628
+ "providerDoc": "https://deepinfra.com/docs",
629
+ "providerEnv": [],
630
+ "providerId": "deep-infra",
631
+ "providerModelsDevId": "deepinfra",
632
+ "providerNpm": "@deepinfra/sdk",
633
+ "reasoning": false,
634
+ "releaseDate": null,
635
+ "streamingSupported": true,
636
+ "temperature": true,
637
+ "toolCall": false,
638
+ "vision": false,
639
+ "icon": "deepinfra"
640
+ },
606
641
  {
607
642
  "attachment": false,
608
643
  "cost": {
@@ -673,6 +708,41 @@
673
708
  "vision": false,
674
709
  "icon": "deepinfra"
675
710
  },
711
+ {
712
+ "attachment": false,
713
+ "cost": {
714
+ "input": 0.00045,
715
+ "inputCacheHit": null,
716
+ "output": 0.00225
717
+ },
718
+ "extendedThinking": false,
719
+ "id": "moonshotai/Kimi-K2.5",
720
+ "knowledge": null,
721
+ "lastUpdated": null,
722
+ "limit": {
723
+ "context": 262144,
724
+ "output": null
725
+ },
726
+ "modalities": {
727
+ "input": ["text"],
728
+ "output": ["text"]
729
+ },
730
+ "name": "Kimi-K2.5",
731
+ "openWeights": false,
732
+ "provider": "Deep Infra",
733
+ "providerDoc": "https://deepinfra.com/docs",
734
+ "providerEnv": [],
735
+ "providerId": "deep-infra",
736
+ "providerModelsDevId": "deepinfra",
737
+ "providerNpm": "@deepinfra/sdk",
738
+ "reasoning": false,
739
+ "releaseDate": null,
740
+ "streamingSupported": true,
741
+ "temperature": true,
742
+ "toolCall": false,
743
+ "vision": false,
744
+ "icon": "deepinfra"
745
+ },
676
746
  {
677
747
  "attachment": false,
678
748
  "cost": {
@@ -711,9 +781,9 @@
711
781
  {
712
782
  "attachment": false,
713
783
  "cost": {
714
- "input": 0.00006,
784
+ "input": 0.00005,
715
785
  "inputCacheHit": null,
716
- "output": 0.00024
786
+ "output": 0.0002
717
787
  },
718
788
  "extendedThinking": false,
719
789
  "id": "nvidia/Nemotron-3-Nano-30B-A3B",
@@ -917,6 +987,41 @@
917
987
  "toolCall": false,
918
988
  "vision": false,
919
989
  "icon": "deepinfra"
990
+ },
991
+ {
992
+ "attachment": false,
993
+ "cost": {
994
+ "input": 0.00006,
995
+ "inputCacheHit": null,
996
+ "output": 0.0004
997
+ },
998
+ "extendedThinking": false,
999
+ "id": "zai-org/GLM-4.7-Flash",
1000
+ "knowledge": null,
1001
+ "lastUpdated": null,
1002
+ "limit": {
1003
+ "context": 202752,
1004
+ "output": null
1005
+ },
1006
+ "modalities": {
1007
+ "input": ["text"],
1008
+ "output": ["text"]
1009
+ },
1010
+ "name": "GLM-4.7-Flash",
1011
+ "openWeights": false,
1012
+ "provider": "Deep Infra",
1013
+ "providerDoc": "https://deepinfra.com/docs",
1014
+ "providerEnv": [],
1015
+ "providerId": "deep-infra",
1016
+ "providerModelsDevId": "deepinfra",
1017
+ "providerNpm": "@deepinfra/sdk",
1018
+ "reasoning": false,
1019
+ "releaseDate": null,
1020
+ "streamingSupported": true,
1021
+ "temperature": true,
1022
+ "toolCall": false,
1023
+ "vision": false,
1024
+ "icon": "deepinfra"
920
1025
  }
921
1026
  ]
922
1027
  }
@@ -1,11 +1,11 @@
1
1
  {
2
2
  "metadata": {
3
3
  "description": "AI Models API - Models from DeepSeek",
4
- "lastUpdated": "2026-01-13T21:44:10.328Z",
4
+ "lastUpdated": "2026-03-02T07:56:54.912Z",
5
5
  "pricingUnit": "1K",
6
6
  "provider": "DeepSeek",
7
7
  "totalModels": 1,
8
- "version": "4.0.2"
8
+ "version": "4.1.0"
9
9
  },
10
10
  "models": [
11
11
  {