@lobehub/lobehub 2.1.2 → 2.1.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.env.example +4 -6
- package/.env.example.development +0 -3
- package/.github/workflows/release-desktop-stable.yml +1 -1
- package/CHANGELOG.md +59 -0
- package/Dockerfile +6 -4
- package/README.md +2 -3
- package/README.zh-CN.md +2 -3
- package/changelog/v2.json +14 -0
- package/docker-compose/deploy/.env.example +3 -1
- package/docker-compose/deploy/.env.zh-CN.example +4 -1
- package/docker-compose/local/.env.example +0 -1
- package/docker-compose/local/.env.zh-CN.example +0 -1
- package/docker-compose/local/grafana/.env.example +0 -1
- package/docker-compose/local/grafana/.env.zh-CN.example +0 -1
- package/docker-compose/local/logto/docker-compose.yml +0 -1
- package/docker-compose/local/zitadel/.env.example +1 -2
- package/docker-compose/local/zitadel/.env.zh-CN.example +1 -2
- package/docker-compose/production/grafana/.env.example +0 -1
- package/docker-compose/production/grafana/.env.zh-CN.example +0 -1
- package/docker-compose/production/logto/.env.example +0 -2
- package/docker-compose/production/logto/.env.zh-CN.example +0 -2
- package/docker-compose/production/zitadel/.env.example +0 -2
- package/docker-compose/production/zitadel/.env.zh-CN.example +0 -2
- package/docker-compose/setup.sh +16 -2
- package/docs/development/basic/folder-structure.mdx +23 -14
- package/docs/development/basic/folder-structure.zh-CN.mdx +23 -14
- package/docs/development/basic/work-with-server-side-database.mdx +0 -1
- package/docs/development/basic/work-with-server-side-database.zh-CN.mdx +0 -1
- package/docs/development/start.mdx +19 -12
- package/docs/development/start.zh-CN.mdx +19 -12
- package/docs/self-hosting/advanced/s3/cloudflare-r2.mdx +0 -5
- package/docs/self-hosting/advanced/s3/cloudflare-r2.zh-CN.mdx +0 -5
- package/docs/self-hosting/advanced/s3/rustfs.mdx +0 -2
- package/docs/self-hosting/advanced/s3/rustfs.zh-CN.mdx +0 -2
- package/docs/self-hosting/advanced/s3/tencent-cloud.mdx +0 -1
- package/docs/self-hosting/advanced/s3/tencent-cloud.zh-CN.mdx +0 -2
- package/docs/self-hosting/advanced/s3.mdx +0 -9
- package/docs/self-hosting/advanced/s3.zh-CN.mdx +0 -8
- package/docs/self-hosting/auth/providers/password.mdx +112 -0
- package/docs/self-hosting/auth/providers/password.zh-CN.mdx +103 -0
- package/docs/self-hosting/auth.mdx +12 -0
- package/docs/self-hosting/auth.zh-CN.mdx +12 -0
- package/docs/self-hosting/environment-variables/auth.mdx +7 -0
- package/docs/self-hosting/environment-variables/auth.zh-CN.mdx +7 -0
- package/docs/self-hosting/environment-variables/basic.mdx +0 -7
- package/docs/self-hosting/environment-variables/basic.zh-CN.mdx +0 -7
- package/docs/self-hosting/environment-variables/s3.mdx +0 -7
- package/docs/self-hosting/environment-variables/s3.zh-CN.mdx +0 -7
- package/docs/self-hosting/examples/azure-openai.mdx +0 -1
- package/docs/self-hosting/examples/azure-openai.zh-CN.mdx +0 -1
- package/docs/self-hosting/platform/docker-compose.mdx +0 -1
- package/docs/self-hosting/platform/docker-compose.zh-CN.mdx +0 -1
- package/docs/self-hosting/platform/docker.mdx +5 -3
- package/docs/self-hosting/platform/docker.zh-CN.mdx +5 -4
- package/docs/self-hosting/platform/dokploy.mdx +0 -2
- package/docs/self-hosting/platform/dokploy.zh-CN.mdx +0 -2
- package/docs/self-hosting/platform/vercel.mdx +0 -7
- package/docs/self-hosting/platform/vercel.zh-CN.mdx +0 -7
- package/e2e/src/steps/home/sidebarAgent.steps.ts +56 -24
- package/locales/ar/authError.json +1 -0
- package/locales/ar/models.json +25 -22
- package/locales/ar/providers.json +0 -1
- package/locales/ar/setting.json +16 -0
- package/locales/bg-BG/authError.json +1 -0
- package/locales/bg-BG/models.json +18 -21
- package/locales/bg-BG/providers.json +0 -1
- package/locales/bg-BG/setting.json +16 -0
- package/locales/de-DE/authError.json +1 -0
- package/locales/de-DE/models.json +20 -20
- package/locales/de-DE/providers.json +0 -1
- package/locales/de-DE/setting.json +16 -0
- package/locales/en-US/auth.json +1 -0
- package/locales/en-US/models.json +22 -22
- package/locales/en-US/providers.json +0 -1
- package/locales/es-ES/authError.json +1 -0
- package/locales/es-ES/models.json +84 -20
- package/locales/es-ES/providers.json +0 -1
- package/locales/es-ES/setting.json +16 -0
- package/locales/fa-IR/authError.json +1 -0
- package/locales/fa-IR/models.json +43 -20
- package/locales/fa-IR/providers.json +0 -1
- package/locales/fa-IR/setting.json +16 -0
- package/locales/fr-FR/authError.json +1 -0
- package/locales/fr-FR/models.json +19 -21
- package/locales/fr-FR/providers.json +0 -1
- package/locales/fr-FR/setting.json +16 -0
- package/locales/it-IT/authError.json +1 -0
- package/locales/it-IT/models.json +17 -19
- package/locales/it-IT/providers.json +0 -1
- package/locales/it-IT/setting.json +16 -0
- package/locales/ja-JP/authError.json +1 -0
- package/locales/ja-JP/models.json +43 -22
- package/locales/ja-JP/providers.json +0 -1
- package/locales/ja-JP/setting.json +16 -0
- package/locales/ko-KR/authError.json +1 -0
- package/locales/ko-KR/models.json +41 -20
- package/locales/ko-KR/providers.json +0 -1
- package/locales/ko-KR/setting.json +16 -0
- package/locales/nl-NL/authError.json +1 -0
- package/locales/nl-NL/models.json +48 -20
- package/locales/nl-NL/providers.json +0 -1
- package/locales/nl-NL/setting.json +16 -0
- package/locales/pl-PL/authError.json +1 -0
- package/locales/pl-PL/models.json +19 -22
- package/locales/pl-PL/providers.json +0 -1
- package/locales/pl-PL/setting.json +16 -0
- package/locales/pt-BR/authError.json +1 -0
- package/locales/pt-BR/models.json +21 -21
- package/locales/pt-BR/providers.json +0 -1
- package/locales/pt-BR/setting.json +16 -0
- package/locales/ru-RU/authError.json +1 -0
- package/locales/ru-RU/models.json +23 -20
- package/locales/ru-RU/providers.json +0 -1
- package/locales/ru-RU/setting.json +16 -0
- package/locales/tr-TR/authError.json +1 -0
- package/locales/tr-TR/models.json +37 -20
- package/locales/tr-TR/providers.json +0 -1
- package/locales/tr-TR/setting.json +16 -0
- package/locales/vi-VN/authError.json +1 -0
- package/locales/vi-VN/models.json +15 -19
- package/locales/vi-VN/providers.json +0 -1
- package/locales/vi-VN/setting.json +16 -0
- package/locales/zh-CN/auth.json +1 -0
- package/locales/zh-CN/models.json +20 -20
- package/locales/zh-CN/providers.json +0 -1
- package/locales/zh-TW/authError.json +1 -0
- package/locales/zh-TW/models.json +20 -20
- package/locales/zh-TW/providers.json +0 -1
- package/locales/zh-TW/setting.json +16 -0
- package/netlify.toml +0 -1
- package/package.json +1 -1
- package/packages/model-bank/src/aiModels/google.ts +0 -19
- package/packages/model-bank/src/aiModels/moonshot.ts +56 -5
- package/packages/model-bank/src/aiModels/ollamacloud.ts +14 -0
- package/packages/model-bank/src/aiModels/openrouter.ts +0 -14
- package/packages/model-bank/src/aiModels/qwen.ts +105 -4
- package/packages/model-bank/src/aiModels/siliconcloud.ts +39 -0
- package/packages/model-bank/src/aiModels/wenxin.ts +0 -99
- package/packages/model-runtime/src/core/contextBuilders/openai.test.ts +24 -0
- package/packages/model-runtime/src/core/contextBuilders/openai.ts +22 -5
- package/packages/model-runtime/src/core/openaiCompatibleFactory/index.ts +10 -3
- package/packages/model-runtime/src/core/streams/google/google-ai.test.ts +54 -13
- package/packages/model-runtime/src/core/streams/google/index.ts +1 -4
- package/packages/model-runtime/src/providers/moonshot/index.ts +24 -2
- package/packages/model-runtime/src/providers/qwen/index.ts +16 -15
- package/packages/types/src/serverConfig.ts +1 -0
- package/src/app/[variants]/(auth)/signin/SignInEmailStep.tsx +56 -49
- package/src/app/[variants]/(auth)/signin/page.tsx +2 -0
- package/src/app/[variants]/(auth)/signin/useSignIn.ts +2 -0
- package/src/app/[variants]/(auth)/signup/[[...signup]]/page.tsx +7 -0
- package/src/envs/app.ts +0 -2
- package/src/envs/auth.ts +3 -0
- package/src/libs/better-auth/define-config.ts +1 -1
- package/src/libs/next/proxy/define-config.ts +0 -1
- package/src/locales/default/auth.ts +2 -0
- package/src/server/globalConfig/index.ts +1 -0
- package/src/server/routers/lambda/__tests__/integration/aiAgent/execAgent.integration.test.ts +3 -2
- package/src/store/chat/slices/topic/action.ts +1 -1
- package/src/store/electron/actions/settings.ts +7 -7
- package/src/store/electron/actions/sync.ts +11 -11
- package/src/store/global/actions/general.ts +12 -12
- package/src/store/global/initialState.ts +11 -11
- package/src/store/global/selectors/clientDB.ts +1 -1
- package/src/store/global/selectors/systemStatus.ts +1 -1
- package/src/store/image/slices/generationConfig/action.ts +12 -12
- package/src/store/image/utils/size.ts +11 -11
- package/src/store/library/slices/ragEval/actions/dataset.ts +1 -1
- package/src/store/serverConfig/selectors.ts +1 -0
- package/src/store/session/slices/session/initialState.ts +6 -6
- package/src/store/session/slices/session/reducers.ts +1 -1
- package/src/store/session/slices/sessionGroup/initialState.ts +2 -2
- package/src/store/tool/slices/customPlugin/action.ts +2 -2
- package/src/store/tool/slices/oldStore/action.ts +5 -5
- package/src/store/userMemory/slices/preference/action.ts +6 -6
|
@@ -90,6 +90,7 @@
|
|
|
90
90
|
"Phi-3-small-8k-instruct.description": "一個擁有 70 億參數的模型,品質優於 Phi-3-mini,專注於高品質、需推理的資料。",
|
|
91
91
|
"Phi-3.5-mini-instruct.description": "Phi-3-mini 模型的更新版本。",
|
|
92
92
|
"Phi-3.5-vision-instrust.description": "Phi-3-vision 模型的更新版本。",
|
|
93
|
+
"Pro/MiniMaxAI/MiniMax-M2.1.description": "MiniMax-M2.1 是一款開源的大型語言模型,專為代理能力進行優化,擅長程式設計、工具使用、指令遵循與長期規劃。該模型支援多語言軟體開發與複雜的多步驟工作流程執行,在 SWE-bench Verified 測試中獲得 74.0 分,並在多語言場景中超越 Claude Sonnet 4.5。",
|
|
93
94
|
"Pro/Qwen/Qwen2-7B-Instruct.description": "Qwen2-7B-Instruct 是 Qwen2 系列中的一款 70 億參數指令微調大型語言模型。它採用 Transformer 架構,結合 SwiGLU、注意力 QKV 偏置與分組查詢注意力機制,能處理大規模輸入內容。該模型在語言理解、生成、多語言任務、程式碼、數學與推理等方面表現優異,超越多數開源模型,並可與商業模型競爭。在多項基準測試中表現優於 Qwen1.5-7B-Chat。",
|
|
94
95
|
"Pro/Qwen/Qwen2.5-7B-Instruct.description": "Qwen2.5-7B-Instruct 是阿里雲最新大型語言模型系列的一部分。此 70 億參數模型在程式碼與數學方面有顯著提升,支援超過 29 種語言,並加強了指令遵循、結構化資料理解與結構化輸出(特別是 JSON)能力。",
|
|
95
96
|
"Pro/Qwen/Qwen2.5-Coder-7B-Instruct.description": "Qwen2.5-Coder-7B-Instruct 是阿里雲最新專注於程式碼的語言模型。基於 Qwen2.5 架構並訓練於 5.5 兆詞元上,顯著提升了程式碼生成、推理與修復能力,同時保有數學與通用能力,為開發智能程式代理提供堅實基礎。",
|
|
@@ -271,20 +272,20 @@
|
|
|
271
272
|
"chatgpt-4o-latest.description": "ChatGPT-4o 是一款即時更新的動態模型,結合強大的理解與生成能力,適用於客服、教育與技術支援等大規模應用場景。",
|
|
272
273
|
"claude-2.0.description": "Claude 2 提供企業級關鍵改進,包括領先的 20 萬 token 上下文、降低幻覺、系統提示與新測試功能:工具調用。",
|
|
273
274
|
"claude-2.1.description": "Claude 2 提供企業級關鍵改進,包括領先的 20 萬 token 上下文、降低幻覺、系統提示與新測試功能:工具調用。",
|
|
274
|
-
"claude-3-5-haiku-20241022.description": "Claude 3.5 Haiku 是 Anthropic
|
|
275
|
+
"claude-3-5-haiku-20241022.description": "Claude 3.5 Haiku 是 Anthropic 推出的次世代最快模型。相較於 Claude 3 Haiku,其在多項技能上皆有提升,並在多項智能基準測試中超越先前最大模型 Claude 3 Opus。",
|
|
275
276
|
"claude-3-5-haiku-latest.description": "Claude 3.5 Haiku 提供快速回應,適用於輕量任務。",
|
|
276
|
-
"claude-3-7-sonnet-20250219.description": "Claude
|
|
277
|
+
"claude-3-7-sonnet-20250219.description": "Claude 3.7 Sonnet 是 Anthropic 最智慧的模型,也是市場上首款混合推理模型。它能夠即時回應,或進行可視化的逐步推理。Sonnet 尤其擅長程式設計、資料科學、視覺處理與代理任務。",
|
|
277
278
|
"claude-3-7-sonnet-latest.description": "Claude 3.7 Sonnet 是 Anthropic 最新且最強大的模型,適用於高度複雜任務,具備卓越的效能、智慧、流暢度與理解力。",
|
|
278
279
|
"claude-3-haiku-20240307.description": "Claude 3 Haiku 是 Anthropic 推出的最快速且最精簡的模型,設計用於即時回應,具備快速且準確的表現。",
|
|
279
280
|
"claude-3-opus-20240229.description": "Claude 3 Opus 是 Anthropic 最強大的模型,適用於高度複雜任務,具備卓越的效能、智慧、流暢度與理解力。",
|
|
280
281
|
"claude-3-sonnet-20240229.description": "Claude 3 Sonnet 在智慧與速度之間取得平衡,適合企業工作負載,提供高效能與低成本的大規模部署。",
|
|
281
|
-
"claude-haiku-4-5-20251001.description": "Claude Haiku 4.5 是 Anthropic
|
|
282
|
+
"claude-haiku-4-5-20251001.description": "Claude Haiku 4.5 是 Anthropic 目前最快且最智慧的 Haiku 模型,具備閃電般的速度與延伸推理能力。",
|
|
282
283
|
"claude-opus-4-1-20250805-thinking.description": "Claude Opus 4.1 Thinking 是一個進階版本,能夠揭示其推理過程。",
|
|
283
|
-
"claude-opus-4-1-20250805.description": "Claude Opus 4.1 是 Anthropic
|
|
284
|
-
"claude-opus-4-20250514.description": "Claude Opus 4 是 Anthropic
|
|
284
|
+
"claude-opus-4-1-20250805.description": "Claude Opus 4.1 是 Anthropic 最新且最強大的模型,專為處理高度複雜任務而設計,表現出色,具備卓越的智能、流暢度與理解力。",
|
|
285
|
+
"claude-opus-4-20250514.description": "Claude Opus 4 是 Anthropic 最強大的模型,專為處理高度複雜任務而設計,表現卓越,具備優異的智能、流暢度與理解能力。",
|
|
285
286
|
"claude-opus-4-5-20251101.description": "Claude Opus 4.5 是 Anthropic 的旗艦模型,結合卓越智慧與可擴展效能,適合需要最高品質回應與推理的複雜任務。",
|
|
286
287
|
"claude-sonnet-4-20250514-thinking.description": "Claude Sonnet 4 Thinking 可產生即時回應或延伸的逐步思考,並顯示其推理過程。",
|
|
287
|
-
"claude-sonnet-4-20250514.description": "Claude Sonnet 4
|
|
288
|
+
"claude-sonnet-4-20250514.description": "Claude Sonnet 4 能夠即時回應,或進行可視化的逐步思考過程。",
|
|
288
289
|
"claude-sonnet-4-5-20250929.description": "Claude Sonnet 4.5 是 Anthropic 迄今最智慧的模型。",
|
|
289
290
|
"codegeex-4.description": "CodeGeeX-4 是一款強大的 AI 程式輔助工具,支援多語言問答與程式碼補全,能有效提升開發者的生產力。",
|
|
290
291
|
"codegeex4-all-9b.description": "CodeGeeX4-ALL-9B 是一個多語言程式碼生成模型,支援程式碼補全與生成、程式碼解釋器、網頁搜尋、函式呼叫與倉庫層級的程式碼問答,涵蓋多種軟體開發場景。它是參數數量低於 100 億的頂尖程式碼模型之一。",
|
|
@@ -351,11 +352,10 @@
|
|
|
351
352
|
"deepseek-ai/DeepSeek-V3.2-Exp.description": "DeepSeek-V3.2-Exp 是 V3.2 的實驗版本,銜接下一代架構。在 V3.1-Terminus 基礎上加入 DeepSeek Sparse Attention(DSA),提升長上下文訓練與推理效率,並針對工具使用、長文理解與多步推理進行優化,適合探索高效推理與大上下文應用。",
|
|
352
353
|
"deepseek-ai/DeepSeek-V3.description": "DeepSeek-V3 是一款擁有 671B 參數的 MoE 模型,採用 MLA 與 DeepSeekMoE 架構,具備無損負載平衡,訓練與推理效率高。預訓練資料達 14.8T 高品質 token,並經過 SFT 與 RL 微調,表現超越其他開源模型,接近封閉模型領先水準。",
|
|
353
354
|
"deepseek-ai/deepseek-llm-67b-chat.description": "DeepSeek LLM Chat(67B)是一款創新模型,具備深層語言理解與互動能力。",
|
|
354
|
-
"deepseek-ai/deepseek-r1.description": "一款高效能的先進大模型,擅長推理、數學與程式設計。",
|
|
355
355
|
"deepseek-ai/deepseek-v3.1-terminus.description": "DeepSeek V3.1 是新一代推理模型,具備更強的複雜推理與思維鏈能力,適用於深度分析任務。",
|
|
356
356
|
"deepseek-ai/deepseek-v3.1.description": "DeepSeek V3.1 是新一代推理模型,具備更強的複雜推理與思維鏈能力,適用於深度分析任務。",
|
|
357
357
|
"deepseek-ai/deepseek-vl2.description": "DeepSeek-VL2 是一款基於 DeepSeekMoE-27B 的 MoE 視覺語言模型,採用稀疏激活,僅使用 4.5B 活躍參數即可達到強大表現。擅長視覺問答、OCR、文件/表格/圖表理解與視覺對齊。",
|
|
358
|
-
"deepseek-chat.description": "DeepSeek
|
|
358
|
+
"deepseek-chat.description": "一款結合通用與程式能力的新型開源模型。它保留了聊天模型的對話能力與程式模型的強大編碼能力,並具備更佳的偏好對齊。DeepSeek-V2.5 同時提升了寫作與指令遵循能力。",
|
|
359
359
|
"deepseek-coder-33B-instruct.description": "DeepSeek Coder 33B 是一款程式語言模型,訓練於 2T token(87% 程式碼,13% 中英文文本),支援 16K 上下文視窗與中間填充任務,提供專案級程式補全與片段填充功能。",
|
|
360
360
|
"deepseek-coder-v2.description": "DeepSeek Coder V2 是一款開源 MoE 程式模型,在程式任務中表現強勁,媲美 GPT-4 Turbo。",
|
|
361
361
|
"deepseek-coder-v2:236b.description": "DeepSeek Coder V2 是一款開源 MoE 程式模型,在程式任務中表現強勁,媲美 GPT-4 Turbo。",
|
|
@@ -378,7 +378,7 @@
|
|
|
378
378
|
"deepseek-r1-fast-online.description": "DeepSeek R1 快速全量版,支援即時網頁搜尋,結合 671B 規模能力與快速回應。",
|
|
379
379
|
"deepseek-r1-online.description": "DeepSeek R1 全量版擁有 671B 參數與即時網頁搜尋功能,提供更強的理解與生成能力。",
|
|
380
380
|
"deepseek-r1.description": "DeepSeek-R1 在強化學習前使用冷啟動資料,於數學、程式碼與推理任務中表現可媲美 OpenAI-o1。",
|
|
381
|
-
"deepseek-reasoner.description": "DeepSeek V3.2
|
|
381
|
+
"deepseek-reasoner.description": "DeepSeek V3.2 推理模式在給出最終答案前會輸出思考鏈,以提升準確性。",
|
|
382
382
|
"deepseek-v2.description": "DeepSeek V2 是一款高效的 MoE 模型,適用於具成本效益的處理任務。",
|
|
383
383
|
"deepseek-v2:236b.description": "DeepSeek V2 236B 是 DeepSeek 專注於程式碼生成的模型,具備強大能力。",
|
|
384
384
|
"deepseek-v3-0324.description": "DeepSeek-V3-0324 是一款擁有 671B 參數的 MoE 模型,在程式設計、技術能力、語境理解與長文本處理方面表現出色。",
|
|
@@ -472,8 +472,7 @@
|
|
|
472
472
|
"ernie-tiny-8k.description": "ERNIE Tiny 8K 是一款超輕量模型,適合簡單問答、分類與低成本推理。",
|
|
473
473
|
"ernie-x1-turbo-32k.description": "ERNIE X1 Turbo 32K 是一款快速思考模型,具備 32K 上下文,適合複雜推理與多輪對話。",
|
|
474
474
|
"ernie-x1.1-preview.description": "ERNIE X1.1 預覽版是一款思考模型預覽,用於評估與測試。",
|
|
475
|
-
"fal-ai/bytedance/seedream/v4.
|
|
476
|
-
"fal-ai/bytedance/seedream/v4.description": "Seedream 4.0 由字節跳動 Seed 團隊打造,支援文字與圖像輸入,實現高度可控且高品質的圖像生成。",
|
|
475
|
+
"fal-ai/bytedance/seedream/v4.description": "Seedream 4.0 是來自字節跳動 Seed 團隊的圖像生成模型,支援文字與圖像輸入,能夠高品質且高度可控地生成圖像,並可根據文字提示生成圖像。",
|
|
477
476
|
"fal-ai/flux-kontext/dev.description": "FLUX.1 模型專注於圖像編輯,支援文字與圖像輸入。",
|
|
478
477
|
"fal-ai/flux-pro/kontext.description": "FLUX.1 Kontext [pro] 接受文字與參考圖像輸入,實現目標區域編輯與複雜場景轉換。",
|
|
479
478
|
"fal-ai/flux/krea.description": "Flux Krea [dev] 是一款圖像生成模型,偏好更真實自然的美學風格。",
|
|
@@ -481,7 +480,7 @@
|
|
|
481
480
|
"fal-ai/hunyuan-image/v3.description": "一款強大的原生多模態圖像生成模型。",
|
|
482
481
|
"fal-ai/imagen4/preview.description": "來自 Google 的高品質圖像生成模型。",
|
|
483
482
|
"fal-ai/nano-banana.description": "Nano Banana 是 Google 最新、最快且最高效的原生多模態模型,支援透過對話進行圖像生成與編輯。",
|
|
484
|
-
"fal-ai/qwen-image-edit.description": "來自 Qwen
|
|
483
|
+
"fal-ai/qwen-image-edit.description": "來自 Qwen 團隊的專業圖像編輯模型,支援語義與外觀編輯,能精確編輯中英文文字,並實現風格轉換、物體旋轉等高品質編輯效果。",
|
|
485
484
|
"fal-ai/qwen-image.description": "來自 Qwen 團隊的強大圖像生成模型,具備優異的中文文字渲染能力與多樣化的視覺風格。",
|
|
486
485
|
"flux-1-schnell.description": "來自黑森林實驗室的 12B 參數文字轉圖像模型,透過潛在對抗擴散蒸餾技術,在 1 至 4 步內生成高品質圖像。其表現媲美封閉式替代方案,並以 Apache-2.0 授權釋出,供個人、研究與商業用途。",
|
|
487
486
|
"flux-dev.description": "FLUX.1 [dev] 是一款開放權重的蒸餾模型,僅限非商業用途。它保有接近專業水準的圖像品質與指令遵循能力,同時運行更高效,資源使用優於同等大小的標準模型。",
|
|
@@ -514,8 +513,6 @@
|
|
|
514
513
|
"gemini-2.0-flash-lite-001.description": "Gemini 2.0 Flash 的一個變體,針對成本效益與低延遲進行最佳化。",
|
|
515
514
|
"gemini-2.0-flash-lite.description": "Gemini 2.0 Flash 的一個變體,針對成本效益與低延遲進行最佳化。",
|
|
516
515
|
"gemini-2.0-flash.description": "Gemini 2.0 Flash 提供次世代功能,包括極速處理、原生工具使用、多模態生成,以及 100 萬 token 的上下文視窗。",
|
|
517
|
-
"gemini-2.5-flash-image-preview.description": "Nano Banana 是 Google 最新、最快且最有效率的原生多模態模型,支援對話式圖像生成與編輯。",
|
|
518
|
-
"gemini-2.5-flash-image-preview:image.description": "Nano Banana 是 Google 最新、最快且最有效率的原生多模態模型,支援對話式圖像生成與編輯。",
|
|
519
516
|
"gemini-2.5-flash-image.description": "Nano Banana 是 Google 最新、最快且最有效率的原生多模態模型,支援對話式圖像生成與編輯。",
|
|
520
517
|
"gemini-2.5-flash-image:image.description": "Nano Banana 是 Google 最新、最快且最有效率的原生多模態模型,支援對話式圖像生成與編輯。",
|
|
521
518
|
"gemini-2.5-flash-lite-preview-06-17.description": "Gemini 2.5 Flash-Lite Preview 是 Google 體積最小、性價比最高的模型,專為大規模應用而設計。",
|
|
@@ -530,7 +527,7 @@
|
|
|
530
527
|
"gemini-2.5-pro.description": "Gemini 2.5 Pro 是 Google 的旗艦推理模型,支援長上下文,適用於處理複雜任務。",
|
|
531
528
|
"gemini-3-flash-preview.description": "Gemini 3 Flash 是一款以速度為核心的智慧模型,結合尖端智能與卓越的搜尋基礎能力。",
|
|
532
529
|
"gemini-3-pro-image-preview.description": "Gemini 3 Pro Image(Nano Banana Pro)是 Google 的圖像生成模型,支援多模態對話。",
|
|
533
|
-
"gemini-3-pro-image-preview:image.description": "Gemini 3 Pro Image(Nano Banana Pro)是 Google
|
|
530
|
+
"gemini-3-pro-image-preview:image.description": "Gemini 3 Pro Image(Nano Banana Pro)是 Google 的圖像生成模型,同時支援多模態聊天功能。",
|
|
534
531
|
"gemini-3-pro-preview.description": "Gemini 3 Pro 是 Google 最強大的智能代理與情境編碼模型,具備頂尖推理能力、豐富視覺表現與深度互動。",
|
|
535
532
|
"gemini-flash-latest.description": "Gemini Flash 最新版本",
|
|
536
533
|
"gemini-flash-lite-latest.description": "Gemini Flash-Lite 最新版本",
|
|
@@ -543,7 +540,7 @@
|
|
|
543
540
|
"generalv3.5.description": "Spark Max 是功能最完整的版本,支援網頁搜尋與多種內建插件。其核心能力、系統角色與函數調用皆經過全面優化,能在複雜應用場景中展現卓越表現。",
|
|
544
541
|
"generalv3.description": "Spark Pro 是一款針對專業領域優化的高效能大型語言模型,專注於數學、程式設計、醫療與教育,支援網頁搜尋與天氣、日期等內建插件。其在複雜知識問答、語言理解與高階文本創作方面表現出色,是專業應用的理想選擇。",
|
|
545
542
|
"glm-4-0520.description": "GLM-4-0520 是最新版本模型,專為高度複雜與多樣化任務設計,具備卓越效能。",
|
|
546
|
-
"glm-4-
|
|
543
|
+
"glm-4-7.description": "GLM-4.7 是智譜 AI 最新的旗艦模型。GLM-4.7 強化了程式能力、長期任務規劃與工具協作,適用於 Agentic Coding 場景,在多項公開基準測試中於開源模型中表現領先。其通用能力亦有所提升,回應更簡潔自然,寫作更具沉浸感。在複雜代理任務中,工具調用時的指令遵循能力更強,並進一步提升 Artifacts 與 Agentic Coding 前端的美學與長期任務完成效率。• 更強的程式能力:多語言編碼與終端代理表現顯著提升;GLM-4.7 現可在 Claude Code、Kilo Code、TRAE、Cline、Roo Code 等框架中實現「先思考,後行動」機制,處理複雜任務更穩定。• 前端美學提升:GLM-4.7 在前端生成品質上有顯著進展,能生成更具視覺吸引力的網站、簡報與海報。• 更強的工具調用能力:GLM-4.7 在 BrowseComp 網頁任務評估中得分 67,在 τ²-Bench 互動工具調用評估中得分 84.7,超越 Claude Sonnet 4.5,成為開源 SOTA。• 推理能力提升:數學與推理能力大幅增強,在 HLE(「人類最後考試」)基準測試中得分 42.8%,較 GLM-4.6 提升 41%,超越 GPT-5.1。• 通用能力增強:GLM-4.7 對話更簡潔、智慧且具人性;寫作與角色扮演更具文學性與沉浸感。",
|
|
547
544
|
"glm-4-9b-chat.description": "GLM-4-9B-Chat 在語義、數學、推理、程式與知識方面表現優異,並支援網頁瀏覽、程式執行、自訂工具調用與長文本推理,支援包括日語、韓語、德語在內的 26 種語言。",
|
|
548
545
|
"glm-4-air-250414.description": "GLM-4-Air 是一款高性價比選擇,效能接近 GLM-4,速度快且成本低。",
|
|
549
546
|
"glm-4-air.description": "GLM-4-Air 是一款高性價比選擇,效能接近 GLM-4,速度快且成本低。",
|
|
@@ -558,11 +555,12 @@
|
|
|
558
555
|
"glm-4.1v-thinking-flashx.description": "GLM-4.1V-Thinking 是目前已知最強的約 10B 視覺語言模型,涵蓋影片理解、圖像問答、學科解題、OCR、文件與圖表閱讀、GUI 智能體、前端編碼與語義對齊等 SOTA 任務。透過先進強化學習與思考鏈推理,提升準確性與豐富度,超越傳統非思考模型的結果與可解釋性。",
|
|
559
556
|
"glm-4.5-air.description": "GLM-4.5 輕量版,兼顧效能與成本,支援靈活的混合思考模式。",
|
|
560
557
|
"glm-4.5-airx.description": "GLM-4.5-Air 快速版,回應更迅速,適用於大規模高速應用。",
|
|
561
|
-
"glm-4.5-flash.description": "免費的 GLM-4.5 版本,在推理、程式與智能體任務中表現強勁。",
|
|
562
558
|
"glm-4.5-x.description": "GLM-4.5 快速版,生成速度高達每秒 100 個 token,效能出色。",
|
|
563
559
|
"glm-4.5.description": "智譜旗艦模型,支援可切換思考模式,整體表現達開源 SOTA,支援最多 128K 上下文。",
|
|
564
560
|
"glm-4.5v.description": "智譜新一代 MoE 視覺推理模型,總參數 106B,啟用 12B,於圖像、影片、文件理解與 GUI 任務中,在同級開源多模態模型中達到 SOTA 水準。",
|
|
565
561
|
"glm-4.6.description": "智譜最新旗艦模型 GLM-4.6(3550 億參數)在高階編碼、長文本處理、推理與智能體能力上全面超越前代,程式能力與 Claude Sonnet 4 相當,成為中國頂尖的編碼模型。",
|
|
562
|
+
"glm-4.7-flash.description": "GLM-4.7-Flash 作為 30B 級別的 SOTA 模型,提供在效能與效率間取得平衡的新選擇。它強化了程式能力、長期任務規劃與工具協作,適用於 Agentic Coding 場景,在多項當前基準排行榜中於同級開源模型中表現領先。在執行複雜智慧代理任務時,其工具調用的指令遵循能力更強,並進一步提升 Artifacts 與 Agentic Coding 前端的美學與長期任務完成效率。",
|
|
563
|
+
"glm-4.7-flashx.description": "GLM-4.7-Flash 作為 30B 級別的 SOTA 模型,提供在效能與效率間取得平衡的新選擇。它強化了程式能力、長期任務規劃與工具協作,適用於 Agentic Coding 場景,在多項當前基準排行榜中於同級開源模型中表現領先。在執行複雜智慧代理任務時,其工具調用的指令遵循能力更強,並進一步提升 Artifacts 與 Agentic Coding 前端的美學與長期任務完成效率。",
|
|
566
564
|
"glm-4.7.description": "GLM-4.7 是智譜最新旗艦模型,針對智能體編碼場景進行強化,提升了編碼能力、長期任務規劃與工具協作能力。在多個公開基準測試中於開源模型中表現領先。通用能力方面,回應更簡潔自然,寫作更具沉浸感。對於複雜智能體任務,工具調用時的指令遵循能力更強,並進一步提升了 Artifacts 與智能體編碼的前端美學與長期任務完成效率。",
|
|
567
565
|
"glm-4.description": "GLM-4 是 2024 年 1 月推出的舊版旗艦模型,現已由更強的 GLM-4-0520 取代。",
|
|
568
566
|
"glm-4v-flash.description": "GLM-4V-Flash 專注於單張圖像的高效理解,適用於即時或批次圖像處理等快速分析場景。",
|
|
@@ -654,8 +652,8 @@
|
|
|
654
652
|
"gpt-5.1-codex-mini.description": "GPT-5.1 Codex mini:較小且成本更低的 Codex 變體,針對智能編碼任務進行最佳化。",
|
|
655
653
|
"gpt-5.1-codex.description": "GPT-5.1 Codex:GPT-5.1 的變體,針對複雜程式碼與智能體工作流程進行最佳化,適用於 Responses API。",
|
|
656
654
|
"gpt-5.1.description": "GPT-5.1 — 旗艦模型,針對編碼與智能體任務進行最佳化,具備可調整的推理深度與更長的上下文支援。",
|
|
657
|
-
"gpt-5.2-chat-latest.description": "GPT-5.2 Chat 是
|
|
658
|
-
"gpt-5.2-pro.description": "GPT-5.2 Pro
|
|
655
|
+
"gpt-5.2-chat-latest.description": "GPT-5.2 Chat 是 ChatGPT 的最新對話版本,帶來最新的對話體驗改進。",
|
|
656
|
+
"gpt-5.2-pro.description": "GPT-5.2 Pro:更智慧、更精確的 GPT-5.2 變體(僅限 Responses API),適用於高難度問題與多輪長推理。",
|
|
659
657
|
"gpt-5.2.description": "GPT-5.2 是針對編碼與智能體工作流程的旗艦模型,具備更強推理能力與長上下文表現。",
|
|
660
658
|
"gpt-5.description": "跨領域編碼與智能體任務的最佳模型。GPT-5 在準確性、速度、推理、上下文理解、結構化思維與問題解決方面實現飛躍。",
|
|
661
659
|
"gpt-audio.description": "GPT Audio 是一款通用聊天模型,支援音訊輸入/輸出,並整合於 Chat Completions API 中。",
|
|
@@ -966,6 +964,9 @@
|
|
|
966
964
|
"openai/gpt-5.1-codex-mini.description": "GPT-5.1-Codex-Mini 是 GPT-5.1-Codex 的小型快速版本,適合對延遲與成本敏感的程式設計場景。",
|
|
967
965
|
"openai/gpt-5.1-codex.description": "GPT-5.1-Codex 是針對軟體工程與程式工作流程優化的 GPT-5.1 變體,適用於大型重構、複雜除錯與長時間自動程式任務。",
|
|
968
966
|
"openai/gpt-5.1.description": "GPT-5.1 是 GPT-5 系列的最新旗艦版本,在通用推理、指令遵循與對話自然度方面相較 GPT-5 有顯著提升,適用於廣泛任務。",
|
|
967
|
+
"openai/gpt-5.2-chat.description": "GPT-5.2 Chat 是 ChatGPT 的變體,用於體驗最新的對話改進。",
|
|
968
|
+
"openai/gpt-5.2-pro.description": "GPT-5.2 Pro:更智慧、更精確的 GPT-5.2 變體(僅限 Responses API),適用於更困難的問題與更長的多輪推理。",
|
|
969
|
+
"openai/gpt-5.2.description": "GPT-5.2 是一款旗艦模型,專為程式設計與代理工作流程打造,具備更強的推理與長上下文處理能力。",
|
|
969
970
|
"openai/gpt-5.description": "GPT-5 是 OpenAI 的高效能模型,適用於各類生產與研究任務。",
|
|
970
971
|
"openai/gpt-oss-120b.description": "一款具備強大可控推理能力的通用大型語言模型。",
|
|
971
972
|
"openai/gpt-oss-20b.description": "一款緊湊型開源權重語言模型,針對低延遲與資源受限環境(如本地與邊緣部署)進行優化。",
|
|
@@ -1223,7 +1224,6 @@
|
|
|
1223
1224
|
"z-ai/glm-4.5.description": "GLM 4.5 是 Z.AI 的旗艦模型,採用混合推理設計,針對工程與長上下文任務進行優化。",
|
|
1224
1225
|
"z-ai/glm-4.6.description": "GLM 4.6 是 Z.AI 的旗艦模型,擴展上下文長度並增強編碼能力。",
|
|
1225
1226
|
"z-ai/glm-4.7.description": "GLM-4.7 是智譜最新旗艦模型,具備更強通用能力、回應更自然簡潔,並提供更具沉浸感的寫作體驗。",
|
|
1226
|
-
"zai-glm-4.6.description": "在編碼與推理任務中表現優異,支援串流與工具調用,適合代理式編碼與複雜推理。",
|
|
1227
1227
|
"zai-org/GLM-4.5-Air.description": "GLM-4.5-Air 是一款基於專家混合架構的代理應用基礎模型,針對工具使用、網頁瀏覽、軟體工程與前端編碼進行優化,並可與 Claude Code、Roo Code 等程式代理整合。採用混合推理處理複雜與日常任務。",
|
|
1228
1228
|
"zai-org/GLM-4.5.description": "GLM-4.5 是一款基於專家混合架構的代理應用基礎模型,深度優化工具使用、網頁瀏覽、軟體工程與前端編碼,並可與 Claude Code、Roo Code 等程式代理整合。採用混合推理處理複雜與日常任務。",
|
|
1229
1229
|
"zai-org/GLM-4.5V.description": "GLM-4.5V 是智譜 AI 最新 VLM,基於 GLM-4.5-Air 旗艦文本模型(總參數 106B,啟用 12B),採用 MoE 架構,在成本較低的情況下提供強大效能。延續 GLM-4.1V-Thinking 路線,加入 3D-RoPE 提升三維空間推理能力。透過預訓練、SFT 與強化學習優化,支援圖像、影片與長文檔,在 41 項公開多模態基準中名列前茅。提供「思考模式」切換,讓用戶在速度與深度間取得平衡。",
|
|
@@ -29,7 +29,6 @@
|
|
|
29
29
|
"internlm.description": "一個專注於大型模型研究與工具的開源組織,提供高效、易用的平台,讓尖端模型與演算法更易於取得。",
|
|
30
30
|
"jina.description": "Jina AI 成立於 2020 年,是領先的搜尋 AI 公司。其搜尋技術堆疊包含向量模型、重排序器與小型語言模型,打造可靠且高品質的生成式與多模態搜尋應用。",
|
|
31
31
|
"lmstudio.description": "LM Studio 是一款桌面應用程式,可在本機開發與實驗大型語言模型。",
|
|
32
|
-
"lobehub.description": "LobeHub Cloud 使用官方 API 存取 AI 模型,並以與模型代幣相關的點數(Credits)來計算使用量。",
|
|
33
32
|
"minimax.description": "MiniMax 成立於 2021 年,致力於打造通用 AI,擁有多模態基礎模型,包括兆級參數的 MoE 文本模型、語音模型與視覺模型,並推出如海螺 AI 等應用。",
|
|
34
33
|
"mistral.description": "Mistral 提供先進的通用、專業與研究模型,支援複雜推理、多語言任務與程式碼生成,並支援函式呼叫以實現自訂整合。",
|
|
35
34
|
"modelscope.description": "ModelScope 是阿里雲的模型即服務平台,提供多樣化的 AI 模型與推理服務。",
|
|
@@ -34,11 +34,20 @@
|
|
|
34
34
|
"agentCronJobs.empty.description": "建立您的第一個排程任務以自動化代理程式",
|
|
35
35
|
"agentCronJobs.empty.title": "尚無排程任務",
|
|
36
36
|
"agentCronJobs.enable": "啟用",
|
|
37
|
+
"agentCronJobs.form.at": "在",
|
|
37
38
|
"agentCronJobs.form.content.placeholder": "輸入給代理程式的提示或指令",
|
|
39
|
+
"agentCronJobs.form.every": "每",
|
|
40
|
+
"agentCronJobs.form.frequency": "頻率",
|
|
41
|
+
"agentCronJobs.form.hours": "小時",
|
|
42
|
+
"agentCronJobs.form.maxExecutions": "執行次數上限",
|
|
38
43
|
"agentCronJobs.form.maxExecutions.placeholder": "留空表示無限制",
|
|
39
44
|
"agentCronJobs.form.name.placeholder": "輸入任務名稱",
|
|
45
|
+
"agentCronJobs.form.time": "時間",
|
|
40
46
|
"agentCronJobs.form.timeRange.end": "結束時間",
|
|
41
47
|
"agentCronJobs.form.timeRange.start": "開始時間",
|
|
48
|
+
"agentCronJobs.form.times": "次",
|
|
49
|
+
"agentCronJobs.form.timezone": "時區",
|
|
50
|
+
"agentCronJobs.form.unlimited": "持續執行",
|
|
42
51
|
"agentCronJobs.form.validation.contentRequired": "任務內容為必填",
|
|
43
52
|
"agentCronJobs.form.validation.invalidTimeRange": "開始時間必須早於結束時間",
|
|
44
53
|
"agentCronJobs.form.validation.nameRequired": "任務名稱為必填",
|
|
@@ -83,6 +92,13 @@
|
|
|
83
92
|
"agentCronJobs.weekday.tuesday": "星期二",
|
|
84
93
|
"agentCronJobs.weekday.wednesday": "星期三",
|
|
85
94
|
"agentCronJobs.weekdays": "平日",
|
|
95
|
+
"agentCronJobs.weekdays.fri": "週五",
|
|
96
|
+
"agentCronJobs.weekdays.mon": "週一",
|
|
97
|
+
"agentCronJobs.weekdays.sat": "週六",
|
|
98
|
+
"agentCronJobs.weekdays.sun": "週日",
|
|
99
|
+
"agentCronJobs.weekdays.thu": "週四",
|
|
100
|
+
"agentCronJobs.weekdays.tue": "週二",
|
|
101
|
+
"agentCronJobs.weekdays.wed": "週三",
|
|
86
102
|
"agentInfoDescription.basic.avatar": "頭像",
|
|
87
103
|
"agentInfoDescription.basic.description": "描述",
|
|
88
104
|
"agentInfoDescription.basic.name": "名稱",
|
package/netlify.toml
CHANGED
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@lobehub/lobehub",
|
|
3
|
-
"version": "2.1.
|
|
3
|
+
"version": "2.1.4",
|
|
4
4
|
"description": "LobeHub - an open-source,comprehensive AI Agent framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"framework",
|
|
@@ -651,25 +651,6 @@ const googleChatModels: AIChatModelCard[] = [
|
|
|
651
651
|
releasedAt: '2025-02-05',
|
|
652
652
|
type: 'chat',
|
|
653
653
|
},
|
|
654
|
-
{
|
|
655
|
-
abilities: {
|
|
656
|
-
imageOutput: true,
|
|
657
|
-
vision: true,
|
|
658
|
-
},
|
|
659
|
-
contextWindowTokens: 1_048_576 + 8192,
|
|
660
|
-
description: 'A Gemini 2.0 Flash variant optimized for cost efficiency and low latency.',
|
|
661
|
-
displayName: 'Gemini 2.0 Flash Exp',
|
|
662
|
-
id: 'gemini-2.0-flash-exp',
|
|
663
|
-
maxOutput: 8192,
|
|
664
|
-
pricing: {
|
|
665
|
-
units: [
|
|
666
|
-
{ name: 'textInput', rate: 0, strategy: 'fixed', unit: 'millionTokens' },
|
|
667
|
-
{ name: 'textOutput', rate: 0, strategy: 'fixed', unit: 'millionTokens' },
|
|
668
|
-
],
|
|
669
|
-
},
|
|
670
|
-
releasedAt: '2025-02-05',
|
|
671
|
-
type: 'chat',
|
|
672
|
-
},
|
|
673
654
|
{
|
|
674
655
|
abilities: {
|
|
675
656
|
vision: true,
|
|
@@ -2,6 +2,34 @@ import { AIChatModelCard } from '../types/aiModel';
|
|
|
2
2
|
|
|
3
3
|
// https://platform.moonshot.cn/docs/pricing/chat
|
|
4
4
|
const moonshotChatModels: AIChatModelCard[] = [
|
|
5
|
+
{
|
|
6
|
+
abilities: {
|
|
7
|
+
functionCall: true,
|
|
8
|
+
reasoning: true,
|
|
9
|
+
structuredOutput: true,
|
|
10
|
+
vision: true,
|
|
11
|
+
},
|
|
12
|
+
contextWindowTokens: 262_144,
|
|
13
|
+
description:
|
|
14
|
+
'Kimi K2.5 is Kimi\'s most versatile model to date, featuring a native multimodal architecture that supports both vision and text inputs, "thinking" and "non-thinking" modes, and both conversational and agent tasks.',
|
|
15
|
+
displayName: 'Kimi K2.5',
|
|
16
|
+
enabled: true,
|
|
17
|
+
id: 'kimi-k2.5',
|
|
18
|
+
maxOutput: 32_768,
|
|
19
|
+
pricing: {
|
|
20
|
+
currency: 'CNY',
|
|
21
|
+
units: [
|
|
22
|
+
{ name: 'textInput_cacheRead', rate: 0.7, strategy: 'fixed', unit: 'millionTokens' },
|
|
23
|
+
{ name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
|
|
24
|
+
{ name: 'textOutput', rate: 21, strategy: 'fixed', unit: 'millionTokens' },
|
|
25
|
+
],
|
|
26
|
+
},
|
|
27
|
+
releasedAt: '2026-01-27',
|
|
28
|
+
settings: {
|
|
29
|
+
extendParams: ['enableReasoning'],
|
|
30
|
+
},
|
|
31
|
+
type: 'chat',
|
|
32
|
+
},
|
|
5
33
|
{
|
|
6
34
|
abilities: {
|
|
7
35
|
functionCall: true,
|
|
@@ -12,7 +40,6 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
|
12
40
|
description:
|
|
13
41
|
'K2 long-thinking model with 256k context, supporting multi-step tool use and reasoning for complex problems.',
|
|
14
42
|
displayName: 'Kimi K2 Thinking',
|
|
15
|
-
enabled: true,
|
|
16
43
|
id: 'kimi-k2-thinking',
|
|
17
44
|
maxOutput: 65_536,
|
|
18
45
|
pricing: {
|
|
@@ -58,7 +85,6 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
|
58
85
|
description:
|
|
59
86
|
'kimi-k2-0905-preview offers a 256k context window, stronger agentic coding, better front-end code quality, and improved context understanding.',
|
|
60
87
|
displayName: 'Kimi K2 0905',
|
|
61
|
-
enabled: true,
|
|
62
88
|
id: 'kimi-k2-0905-preview',
|
|
63
89
|
pricing: {
|
|
64
90
|
currency: 'CNY',
|
|
@@ -126,8 +152,32 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
|
126
152
|
currency: 'CNY',
|
|
127
153
|
units: [
|
|
128
154
|
{ name: 'textInput_cacheRead', rate: 1, strategy: 'fixed', unit: 'millionTokens' },
|
|
129
|
-
{
|
|
130
|
-
|
|
155
|
+
{
|
|
156
|
+
lookup: {
|
|
157
|
+
prices: {
|
|
158
|
+
'[0, 0.008_192]': 2,
|
|
159
|
+
'[0.008_193, 0.032_768]': 5,
|
|
160
|
+
'[0.032_769, 0.131_072]': 10,
|
|
161
|
+
},
|
|
162
|
+
pricingParams: ['textInput'],
|
|
163
|
+
},
|
|
164
|
+
name: 'textInput',
|
|
165
|
+
strategy: 'lookup',
|
|
166
|
+
unit: 'millionTokens',
|
|
167
|
+
},
|
|
168
|
+
{
|
|
169
|
+
lookup: {
|
|
170
|
+
prices: {
|
|
171
|
+
'[0, 0.008_192]': 10,
|
|
172
|
+
'[0.008_193, 0.032_768]': 20,
|
|
173
|
+
'[0.032_769, 0.131_072]': 30,
|
|
174
|
+
},
|
|
175
|
+
pricingParams: ['textInput'],
|
|
176
|
+
},
|
|
177
|
+
name: 'textOutput',
|
|
178
|
+
strategy: 'lookup',
|
|
179
|
+
unit: 'millionTokens',
|
|
180
|
+
},
|
|
131
181
|
],
|
|
132
182
|
},
|
|
133
183
|
releasedAt: '2025-02-17',
|
|
@@ -138,7 +188,8 @@ const moonshotChatModels: AIChatModelCard[] = [
|
|
|
138
188
|
functionCall: true,
|
|
139
189
|
},
|
|
140
190
|
contextWindowTokens: 131_072,
|
|
141
|
-
description:
|
|
191
|
+
description:
|
|
192
|
+
'Moonshot V1 Auto selects the appropriate model based on current context token usage.',
|
|
142
193
|
displayName: 'Moonshot V1 Auto',
|
|
143
194
|
id: 'moonshot-v1-auto',
|
|
144
195
|
pricing: {
|
|
@@ -1,6 +1,20 @@
|
|
|
1
1
|
import { AIChatModelCard } from '../types/aiModel';
|
|
2
2
|
|
|
3
3
|
const ollamaCloudModels: AIChatModelCard[] = [
|
|
4
|
+
{
|
|
5
|
+
abilities: {
|
|
6
|
+
functionCall: true,
|
|
7
|
+
reasoning: true,
|
|
8
|
+
vision: true,
|
|
9
|
+
},
|
|
10
|
+
contextWindowTokens: 262_144,
|
|
11
|
+
description:
|
|
12
|
+
'Kimi K2.5 is an open-source, native multimodal agentic model that seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as conversational and agentic paradigms.',
|
|
13
|
+
displayName: 'Kimi K2.5',
|
|
14
|
+
enabled: true,
|
|
15
|
+
id: 'kimi-k2.5',
|
|
16
|
+
type: 'chat',
|
|
17
|
+
},
|
|
4
18
|
{
|
|
5
19
|
abilities: {
|
|
6
20
|
functionCall: true,
|
|
@@ -1079,20 +1079,6 @@ const openrouterChatModels: AIChatModelCard[] = [
|
|
|
1079
1079
|
id: 'google/gemma-2-9b-it:free',
|
|
1080
1080
|
type: 'chat',
|
|
1081
1081
|
},
|
|
1082
|
-
{
|
|
1083
|
-
abilities: {
|
|
1084
|
-
functionCall: true,
|
|
1085
|
-
vision: true,
|
|
1086
|
-
},
|
|
1087
|
-
contextWindowTokens: 1_048_576 + 8192,
|
|
1088
|
-
description:
|
|
1089
|
-
'Gemini 2.0 Flash Experimental is Google’s latest experimental multimodal AI model with quality improvements over prior versions, especially in world knowledge, code, and long context.',
|
|
1090
|
-
displayName: 'Gemini 2.0 Flash Experimental (Free)',
|
|
1091
|
-
id: 'google/gemini-2.0-flash-exp:free',
|
|
1092
|
-
maxOutput: 8192,
|
|
1093
|
-
releasedAt: '2024-12-11',
|
|
1094
|
-
type: 'chat',
|
|
1095
|
-
},
|
|
1096
1082
|
];
|
|
1097
1083
|
|
|
1098
1084
|
export const allModels = [...openrouterChatModels];
|
|
@@ -3,6 +3,55 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
|
|
|
3
3
|
// https://help.aliyun.com/zh/model-studio/models?spm=a2c4g.11186623
|
|
4
4
|
|
|
5
5
|
const qwenChatModels: AIChatModelCard[] = [
|
|
6
|
+
{
|
|
7
|
+
abilities: {
|
|
8
|
+
functionCall: true,
|
|
9
|
+
reasoning: true,
|
|
10
|
+
vision: true,
|
|
11
|
+
},
|
|
12
|
+
contextWindowTokens: 262_144,
|
|
13
|
+
description:
|
|
14
|
+
'Kimi K2.5 is the most capable Kimi model, delivering open-source SOTA in agent tasks, coding, and vision understanding. It supports multimodal inputs and both thinking and non-thinking modes.',
|
|
15
|
+
displayName: 'Kimi K2.5',
|
|
16
|
+
id: 'kimi-k2.5',
|
|
17
|
+
maxOutput: 32_768,
|
|
18
|
+
organization: 'Qwen',
|
|
19
|
+
pricing: {
|
|
20
|
+
currency: 'CNY',
|
|
21
|
+
units: [
|
|
22
|
+
{ name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
|
|
23
|
+
{ name: 'textOutput', rate: 21, strategy: 'fixed', unit: 'millionTokens' },
|
|
24
|
+
],
|
|
25
|
+
},
|
|
26
|
+
settings: {
|
|
27
|
+
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
28
|
+
},
|
|
29
|
+
type: 'chat',
|
|
30
|
+
},
|
|
31
|
+
{
|
|
32
|
+
abilities: {
|
|
33
|
+
functionCall: true,
|
|
34
|
+
reasoning: true,
|
|
35
|
+
search: true,
|
|
36
|
+
},
|
|
37
|
+
contextWindowTokens: 204_800,
|
|
38
|
+
description:
|
|
39
|
+
'MiniMax-M2.1 is a flagship open-source large model from MiniMax, focusing on solving complex real-world tasks. Its core strengths are multi-language programming capabilities and the ability to solve complex tasks as an Agent.',
|
|
40
|
+
displayName: 'MiniMax-M2.1',
|
|
41
|
+
id: 'MiniMax-M2.1',
|
|
42
|
+
maxOutput: 32_768,
|
|
43
|
+
pricing: {
|
|
44
|
+
currency: 'CNY',
|
|
45
|
+
units: [
|
|
46
|
+
{ name: 'textInput', rate: 2.1, strategy: 'fixed', unit: 'millionTokens' },
|
|
47
|
+
{ name: 'textOutput', rate: 8.4, strategy: 'fixed', unit: 'millionTokens' },
|
|
48
|
+
],
|
|
49
|
+
},
|
|
50
|
+
settings: {
|
|
51
|
+
searchImpl: 'params',
|
|
52
|
+
},
|
|
53
|
+
type: 'chat',
|
|
54
|
+
},
|
|
6
55
|
{
|
|
7
56
|
abilities: {
|
|
8
57
|
reasoning: true,
|
|
@@ -61,7 +110,7 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
|
61
110
|
vision: true,
|
|
62
111
|
},
|
|
63
112
|
config: {
|
|
64
|
-
deploymentName: 'qwen3-vl-flash-
|
|
113
|
+
deploymentName: 'qwen3-vl-flash-2026-01-22',
|
|
65
114
|
},
|
|
66
115
|
contextWindowTokens: 262_144,
|
|
67
116
|
description:
|
|
@@ -101,7 +150,6 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
|
101
150
|
},
|
|
102
151
|
],
|
|
103
152
|
},
|
|
104
|
-
releasedAt: '2025-10-15',
|
|
105
153
|
settings: {
|
|
106
154
|
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
107
155
|
},
|
|
@@ -1019,7 +1067,8 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
|
1019
1067
|
deploymentName: 'qwen-plus-2025-12-01',
|
|
1020
1068
|
},
|
|
1021
1069
|
contextWindowTokens: 1_000_000,
|
|
1022
|
-
description:
|
|
1070
|
+
description:
|
|
1071
|
+
'Enhanced ultra-large Qwen model supporting Chinese, English, and other languages.',
|
|
1023
1072
|
displayName: 'Qwen Plus',
|
|
1024
1073
|
enabled: true,
|
|
1025
1074
|
id: 'qwen-plus',
|
|
@@ -1078,6 +1127,57 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
|
1078
1127
|
},
|
|
1079
1128
|
type: 'chat',
|
|
1080
1129
|
},
|
|
1130
|
+
{
|
|
1131
|
+
abilities: {
|
|
1132
|
+
functionCall: true,
|
|
1133
|
+
search: true,
|
|
1134
|
+
reasoning: true,
|
|
1135
|
+
},
|
|
1136
|
+
contextWindowTokens: 262_144,
|
|
1137
|
+
description:
|
|
1138
|
+
'Qwen3 Max models deliver large gains over the 2.5 series in general ability, Chinese/English understanding, complex instruction following, subjective open tasks, multilingual ability, and tool use, with fewer hallucinations. The latest qwen3-max improves agentic programming and tool use over qwen3-max-preview. This release reaches field SOTA and targets more complex agent needs.',
|
|
1139
|
+
displayName: 'Qwen3 Max Thinking',
|
|
1140
|
+
id: 'qwen3-max-2026-01-23',
|
|
1141
|
+
maxOutput: 65_536,
|
|
1142
|
+
organization: 'Qwen',
|
|
1143
|
+
pricing: {
|
|
1144
|
+
currency: 'CNY',
|
|
1145
|
+
units: [
|
|
1146
|
+
{
|
|
1147
|
+
lookup: {
|
|
1148
|
+
prices: {
|
|
1149
|
+
'[0, 0.032]': 2.5,
|
|
1150
|
+
'[0.032, 0.128]': 4,
|
|
1151
|
+
'[0.128, 0.252]': 7,
|
|
1152
|
+
},
|
|
1153
|
+
pricingParams: ['textInputRange'],
|
|
1154
|
+
},
|
|
1155
|
+
name: 'textInput',
|
|
1156
|
+
strategy: 'lookup',
|
|
1157
|
+
unit: 'millionTokens',
|
|
1158
|
+
},
|
|
1159
|
+
{
|
|
1160
|
+
lookup: {
|
|
1161
|
+
prices: {
|
|
1162
|
+
'[0, 0.032]': 10,
|
|
1163
|
+
'[0.032, 0.128]': 16,
|
|
1164
|
+
'[0.128, 0.252]': 28,
|
|
1165
|
+
},
|
|
1166
|
+
pricingParams: ['textInputRange'],
|
|
1167
|
+
},
|
|
1168
|
+
name: 'textOutput',
|
|
1169
|
+
strategy: 'lookup',
|
|
1170
|
+
unit: 'millionTokens',
|
|
1171
|
+
},
|
|
1172
|
+
],
|
|
1173
|
+
},
|
|
1174
|
+
releasedAt: '2026-01-23',
|
|
1175
|
+
settings: {
|
|
1176
|
+
extendParams: ['enableReasoning', 'reasoningBudgetToken'],
|
|
1177
|
+
searchImpl: 'params',
|
|
1178
|
+
},
|
|
1179
|
+
type: 'chat',
|
|
1180
|
+
},
|
|
1081
1181
|
{
|
|
1082
1182
|
abilities: {
|
|
1083
1183
|
functionCall: true,
|
|
@@ -1477,7 +1577,8 @@ const qwenChatModels: AIChatModelCard[] = [
|
|
|
1477
1577
|
vision: true,
|
|
1478
1578
|
},
|
|
1479
1579
|
contextWindowTokens: 131_072,
|
|
1480
|
-
description:
|
|
1580
|
+
description:
|
|
1581
|
+
'Qwen3 VL 8B non-thinking mode (Instruct) for standard multimodal generation and recognition.',
|
|
1481
1582
|
displayName: 'Qwen3 VL 8B Instruct',
|
|
1482
1583
|
id: 'qwen3-vl-8b-instruct',
|
|
1483
1584
|
maxOutput: 32_768,
|
|
@@ -2,6 +2,45 @@ import { AIChatModelCard, AIImageModelCard } from '../types/aiModel';
|
|
|
2
2
|
|
|
3
3
|
// https://siliconflow.cn/zh-cn/models
|
|
4
4
|
const siliconcloudChatModels: AIChatModelCard[] = [
|
|
5
|
+
{
|
|
6
|
+
abilities: {
|
|
7
|
+
functionCall: true,
|
|
8
|
+
reasoning: true,
|
|
9
|
+
vision: true,
|
|
10
|
+
},
|
|
11
|
+
contextWindowTokens: 262_144,
|
|
12
|
+
description:
|
|
13
|
+
'Kimi K2.5 is an open-source native multimodal agent model, built on Kimi-K2-Base, trained on approximately 1.5 trillion mixed vision and text tokens. The model adopts an MoE architecture with 1T total parameters and 32B active parameters, supporting a 256K context window, seamlessly integrating vision and language understanding capabilities.',
|
|
14
|
+
displayName: 'Kimi-K2.5 (Pro)',
|
|
15
|
+
id: 'Pro/moonshotai/Kimi-K2.5',
|
|
16
|
+
pricing: {
|
|
17
|
+
currency: 'CNY',
|
|
18
|
+
units: [
|
|
19
|
+
{ name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
|
|
20
|
+
{ name: 'textOutput', rate: 21, strategy: 'fixed', unit: 'millionTokens' },
|
|
21
|
+
],
|
|
22
|
+
},
|
|
23
|
+
releasedAt: '2026-01-27',
|
|
24
|
+
type: 'chat',
|
|
25
|
+
},
|
|
26
|
+
{
|
|
27
|
+
abilities: {
|
|
28
|
+
vision: true,
|
|
29
|
+
},
|
|
30
|
+
description:
|
|
31
|
+
'PaddleOCR-VL-1.5 is an upgraded version of the PaddleOCR-VL series, achieving 94.5% accuracy on the OmniDocBench v1.5 document parsing benchmark, surpassing leading general large models and specialized document parsing models. It innovatively supports irregular bounding box localization for document elements, handling scanned, tilted, and screen-captured images effectively.',
|
|
32
|
+
displayName: 'PaddleOCR-VL 1.5',
|
|
33
|
+
id: 'PaddlePaddle/PaddleOCR-VL-1.5',
|
|
34
|
+
pricing: {
|
|
35
|
+
currency: 'CNY',
|
|
36
|
+
units: [
|
|
37
|
+
{ name: 'textInput', rate: 0, strategy: 'fixed', unit: 'millionTokens' },
|
|
38
|
+
{ name: 'textOutput', rate: 0, strategy: 'fixed', unit: 'millionTokens' },
|
|
39
|
+
],
|
|
40
|
+
},
|
|
41
|
+
releasedAt: '2026-01-29',
|
|
42
|
+
type: 'chat',
|
|
43
|
+
},
|
|
5
44
|
{
|
|
6
45
|
abilities: {
|
|
7
46
|
functionCall: true,
|