@lobehub/lobehub 2.0.0-next.51 → 2.0.0-next.53

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (107) hide show
  1. package/CHANGELOG.md +58 -0
  2. package/README.md +8 -8
  3. package/README.zh-CN.md +8 -8
  4. package/apps/desktop/package.json +1 -1
  5. package/apps/desktop/src/main/controllers/LocalFileCtr.ts +25 -5
  6. package/apps/desktop/src/main/controllers/__tests__/LocalFileCtr.test.ts +4 -1
  7. package/apps/desktop/src/main/modules/fileSearch/__tests__/macOS.integration.test.ts +357 -0
  8. package/apps/desktop/src/main/modules/fileSearch/impl/macOS.ts +30 -22
  9. package/changelog/v1.json +21 -0
  10. package/locales/ar/models.json +119 -126
  11. package/locales/ar/plugin.json +1 -1
  12. package/locales/bg-BG/models.json +104 -132
  13. package/locales/bg-BG/plugin.json +1 -1
  14. package/locales/de-DE/models.json +119 -126
  15. package/locales/de-DE/plugin.json +1 -1
  16. package/locales/en-US/models.json +167 -126
  17. package/locales/en-US/plugin.json +1 -1
  18. package/locales/es-ES/models.json +119 -126
  19. package/locales/es-ES/plugin.json +1 -1
  20. package/locales/fa-IR/models.json +119 -126
  21. package/locales/fa-IR/plugin.json +1 -1
  22. package/locales/fr-FR/models.json +119 -126
  23. package/locales/fr-FR/plugin.json +1 -1
  24. package/locales/it-IT/models.json +119 -126
  25. package/locales/it-IT/plugin.json +1 -1
  26. package/locales/ja-JP/models.json +119 -126
  27. package/locales/ja-JP/plugin.json +1 -1
  28. package/locales/ko-KR/models.json +119 -126
  29. package/locales/ko-KR/plugin.json +1 -1
  30. package/locales/nl-NL/models.json +119 -126
  31. package/locales/nl-NL/plugin.json +1 -1
  32. package/locales/pl-PL/models.json +119 -126
  33. package/locales/pl-PL/plugin.json +1 -1
  34. package/locales/pt-BR/models.json +119 -126
  35. package/locales/pt-BR/plugin.json +1 -1
  36. package/locales/ru-RU/models.json +119 -126
  37. package/locales/ru-RU/plugin.json +1 -1
  38. package/locales/tr-TR/models.json +119 -126
  39. package/locales/tr-TR/plugin.json +1 -1
  40. package/locales/vi-VN/models.json +119 -126
  41. package/locales/vi-VN/plugin.json +1 -1
  42. package/locales/zh-CN/models.json +173 -80
  43. package/locales/zh-CN/plugin.json +1 -1
  44. package/locales/zh-TW/models.json +119 -126
  45. package/locales/zh-TW/plugin.json +1 -1
  46. package/package.json +2 -2
  47. package/packages/const/src/models.ts +2 -0
  48. package/packages/electron-client-ipc/src/types/localSystem.ts +26 -2
  49. package/packages/electron-server-ipc/src/ipcClient.ts +31 -31
  50. package/packages/electron-server-ipc/src/ipcServer.ts +15 -15
  51. package/packages/model-bank/src/aiModels/aihubmix.ts +106 -2
  52. package/packages/model-bank/src/aiModels/openai.ts +107 -3
  53. package/packages/model-bank/src/aiModels/qwen.ts +76 -7
  54. package/packages/model-bank/src/types/aiModel.ts +1 -0
  55. package/packages/model-runtime/src/core/contextBuilders/openai.test.ts +58 -0
  56. package/packages/model-runtime/src/core/contextBuilders/openai.ts +24 -10
  57. package/packages/model-runtime/src/core/openaiCompatibleFactory/index.ts +3 -2
  58. package/packages/model-runtime/src/providers/openai/index.test.ts +44 -0
  59. package/packages/types/src/agent/chatConfig.ts +9 -0
  60. package/packages/types/src/tool/builtin.ts +6 -4
  61. package/src/app/[variants]/(main)/chat/components/WorkspaceLayout.tsx +32 -23
  62. package/src/features/ChatInput/ActionBar/Model/ControlsForm.tsx +12 -0
  63. package/src/features/ChatInput/ActionBar/Model/GPT51ReasoningEffortSlider.tsx +58 -0
  64. package/src/features/ChatItem/components/MessageContent.tsx +3 -1
  65. package/src/features/Conversation/Messages/Assistant/Tool/Render/LoadingPlaceholder/index.tsx +3 -3
  66. package/src/features/Conversation/Messages/Group/Tool/Render/Intervention/ApprovalActions.tsx +34 -13
  67. package/src/features/Conversation/Messages/Group/Tool/Render/Intervention/index.tsx +2 -2
  68. package/src/features/Conversation/Messages/Group/Tool/Render/LoadingPlaceholder/index.tsx +3 -3
  69. package/src/features/Conversation/Messages/User/index.tsx +11 -1
  70. package/src/features/PluginsUI/Render/BuiltinType/index.test.tsx +10 -4
  71. package/src/features/PluginsUI/Render/BuiltinType/index.tsx +2 -2
  72. package/src/libs/mcp/__tests__/__snapshots__/index.test.ts.snap +0 -6
  73. package/src/locales/default/chat.ts +2 -0
  74. package/src/locales/default/plugin.ts +1 -1
  75. package/src/services/chat/chat.test.ts +1 -0
  76. package/src/services/chat/index.ts +7 -0
  77. package/src/store/aiInfra/slices/aiProvider/__tests__/selectors.test.ts +62 -0
  78. package/src/store/aiInfra/slices/aiProvider/selectors.ts +1 -1
  79. package/src/store/chat/slices/aiChat/actions/conversationControl.ts +42 -0
  80. package/src/tools/code-interpreter/Render/index.tsx +1 -1
  81. package/src/tools/interventions.ts +28 -4
  82. package/src/tools/local-system/Intervention/RunCommand/index.tsx +4 -5
  83. package/src/tools/local-system/Placeholder/ListFiles.tsx +3 -5
  84. package/src/tools/local-system/Placeholder/SearchFiles.tsx +2 -5
  85. package/src/tools/local-system/Render/ListFiles/index.tsx +16 -21
  86. package/src/tools/local-system/Render/ReadLocalFile/ReadFileView.tsx +2 -1
  87. package/src/tools/local-system/Render/RenameLocalFile/index.tsx +15 -20
  88. package/src/tools/local-system/Render/RunCommand/index.tsx +67 -70
  89. package/src/tools/local-system/Render/SearchFiles/SearchQuery/index.tsx +0 -1
  90. package/src/tools/local-system/Render/SearchFiles/index.tsx +15 -20
  91. package/src/tools/local-system/Render/WriteFile/index.tsx +2 -8
  92. package/src/tools/local-system/index.ts +5 -4
  93. package/src/tools/local-system/systemRole.ts +1 -1
  94. package/src/tools/placeholders.ts +39 -8
  95. package/src/tools/renders.ts +56 -9
  96. package/src/tools/web-browsing/Placeholder/{PageContent.tsx → CrawlMultiPages.tsx} +4 -1
  97. package/src/tools/web-browsing/Placeholder/CrawlSinglePage.tsx +12 -0
  98. package/src/tools/web-browsing/Placeholder/Search.tsx +4 -4
  99. package/src/tools/web-browsing/Render/CrawlMultiPages.tsx +15 -0
  100. package/src/tools/web-browsing/Render/CrawlSinglePage.tsx +15 -0
  101. package/src/tools/web-browsing/Render/Search/index.tsx +39 -44
  102. package/packages/database/migrations/0044_add_tool_intervention.sql +0 -1
  103. package/src/tools/local-system/Intervention/index.tsx +0 -17
  104. package/src/tools/local-system/Placeholder/index.tsx +0 -25
  105. package/src/tools/local-system/Render/index.tsx +0 -42
  106. package/src/tools/web-browsing/Placeholder/index.tsx +0 -40
  107. package/src/tools/web-browsing/Render/index.tsx +0 -57
@@ -1049,6 +1049,9 @@
1049
1049
  "deepseek-r1-0528": {
1050
1050
  "description": "The full-capacity 685B model released on May 28, 2025. DeepSeek-R1 extensively employs reinforcement learning during post-training, significantly enhancing reasoning capabilities with minimal labeled data. It demonstrates strong performance in mathematics, coding, and natural language reasoning tasks."
1051
1051
  },
1052
+ "deepseek-r1-250528": {
1053
+ "description": "DeepSeek R1 250528, the full-performance DeepSeek-R1 inference model, ideal for complex mathematical and logical tasks."
1054
+ },
1052
1055
  "deepseek-r1-70b-fast-online": {
1053
1056
  "description": "DeepSeek R1 70B fast version, supporting real-time online search, providing faster response times while maintaining model performance."
1054
1057
  },
@@ -1059,31 +1062,34 @@
1059
1062
  "description": "deepseek-r1-distill-llama is a model distilled from DeepSeek-R1 based on Llama."
1060
1063
  },
1061
1064
  "deepseek-r1-distill-llama-70b": {
1062
- "description": "DeepSeek R1—the larger and smarter model in the DeepSeek suite—has been distilled into the Llama 70B architecture. Based on benchmark tests and human evaluations, this model is smarter than the original Llama 70B, especially excelling in tasks requiring mathematical and factual accuracy."
1065
+ "description": "DeepSeek R1 Distill Llama 70B, a distilled model combining general R1 inference capabilities with the Llama ecosystem."
1063
1066
  },
1064
1067
  "deepseek-r1-distill-llama-8b": {
1065
- "description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
1068
+ "description": "DeepSeek-R1-Distill-Llama-8B is a distilled large language model based on Llama-3.1-8B, utilizing outputs from DeepSeek R1."
1066
1069
  },
1067
- "deepseek-r1-distill-qianfan-llama-70b": {
1068
- "description": "First released on February 14, 2025, distilled by the Qianfan model development team using Llama3_70B as the base model (Built with Meta Llama), with Qianfan's corpus also added to the distilled data."
1070
+ "deepseek-r1-distill-qianfan-70b": {
1071
+ "description": "DeepSeek R1 Distill Qianfan 70B, a cost-effective R1 distilled model based on Qianfan-70B."
1069
1072
  },
1070
- "deepseek-r1-distill-qianfan-llama-8b": {
1071
- "description": "First released on February 14, 2025, distilled by the Qianfan model development team using Llama3_8B as the base model (Built with Meta Llama), with Qianfan's corpus also added to the distilled data."
1073
+ "deepseek-r1-distill-qianfan-8b": {
1074
+ "description": "DeepSeek R1 Distill Qianfan 8B, an R1 distilled model based on Qianfan-8B, suitable for small to medium-scale applications."
1075
+ },
1076
+ "deepseek-r1-distill-qianfan-llama-70b": {
1077
+ "description": "DeepSeek R1 Distill Qianfan Llama 70B, an R1 distilled model based on Llama-70B."
1072
1078
  },
1073
1079
  "deepseek-r1-distill-qwen": {
1074
1080
  "description": "deepseek-r1-distill-qwen is a model distilled from DeepSeek-R1 based on Qwen."
1075
1081
  },
1076
1082
  "deepseek-r1-distill-qwen-1.5b": {
1077
- "description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
1083
+ "description": "DeepSeek R1 Distill Qwen 1.5B, an ultra-lightweight R1 distilled model designed for extremely low-resource environments."
1078
1084
  },
1079
1085
  "deepseek-r1-distill-qwen-14b": {
1080
- "description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
1086
+ "description": "DeepSeek R1 Distill Qwen 14B, a mid-sized R1 distilled model suitable for multi-scenario deployment."
1081
1087
  },
1082
1088
  "deepseek-r1-distill-qwen-32b": {
1083
- "description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
1089
+ "description": "DeepSeek R1 Distill Qwen 32B, an R1 distilled model based on Qwen-32B, balancing performance and cost."
1084
1090
  },
1085
1091
  "deepseek-r1-distill-qwen-7b": {
1086
- "description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
1092
+ "description": "DeepSeek R1 Distill Qwen 7B, a lightweight R1 distilled model ideal for edge computing and enterprise private deployments."
1087
1093
  },
1088
1094
  "deepseek-r1-fast-online": {
1089
1095
  "description": "DeepSeek R1 full fast version, supporting real-time online search, combining the powerful capabilities of 671B parameters with faster response times."
@@ -1112,12 +1118,24 @@
1112
1118
  "deepseek-v3.1-terminus": {
1113
1119
  "description": "DeepSeek-V3.1-Terminus is an optimized large language model developed by DeepSeek, specifically tailored for terminal devices."
1114
1120
  },
1121
+ "deepseek-v3.1-think-250821": {
1122
+ "description": "DeepSeek V3.1 Think 250821, a deep reasoning model aligned with the Terminus version, suitable for high-performance inference scenarios."
1123
+ },
1115
1124
  "deepseek-v3.1:671b": {
1116
1125
  "description": "DeepSeek V3.1: The next-generation reasoning model that enhances complex reasoning and chain-of-thought capabilities, suitable for tasks requiring in-depth analysis."
1117
1126
  },
1118
1127
  "deepseek-v3.2-exp": {
1119
1128
  "description": "deepseek-v3.2-exp introduces a sparse attention mechanism designed to enhance training and inference efficiency when processing long texts, priced lower than deepseek-v3.1."
1120
1129
  },
1130
+ "deepseek-v3.2-think": {
1131
+ "description": "DeepSeek V3.2 Think, a full-performance deep reasoning model with enhanced long-chain reasoning capabilities."
1132
+ },
1133
+ "deepseek-vl2": {
1134
+ "description": "DeepSeek VL2, a multimodal model supporting image-text understanding and fine-grained visual question answering."
1135
+ },
1136
+ "deepseek-vl2-small": {
1137
+ "description": "DeepSeek VL2 Small, a lightweight multimodal version designed for resource-constrained and high-concurrency scenarios."
1138
+ },
1121
1139
  "deepseek/deepseek-chat-v3-0324": {
1122
1140
  "description": "DeepSeek V3 is a 685B parameter expert mixture model, the latest iteration in the DeepSeek team's flagship chat model series.\n\nIt inherits from the [DeepSeek V3](/deepseek/deepseek-chat-v3) model and performs excellently across various tasks."
1123
1141
  },
@@ -1253,83 +1271,89 @@
1253
1271
  "emohaa": {
1254
1272
  "description": "Emohaa is a psychological model with professional counseling capabilities, helping users understand emotional issues."
1255
1273
  },
1256
- "ernie-3.5-128k": {
1257
- "description": "Baidu's flagship large-scale language model, covering a vast amount of Chinese and English corpus, possesses strong general capabilities to meet the requirements of most dialogue Q&A, creative generation, and plugin application scenarios; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information."
1258
- },
1259
- "ernie-3.5-8k": {
1260
- "description": "Baidu's flagship large-scale language model, covering a vast amount of Chinese and English corpus, possesses strong general capabilities to meet the requirements of most dialogue Q&A, creative generation, and plugin application scenarios; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information."
1261
- },
1262
- "ernie-3.5-8k-preview": {
1263
- "description": "Baidu's flagship large-scale language model, covering a vast amount of Chinese and English corpus, possesses strong general capabilities to meet the requirements of most dialogue Q&A, creative generation, and plugin application scenarios; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information."
1264
- },
1265
- "ernie-4.0-8k-latest": {
1266
- "description": "Baidu's flagship ultra-large-scale language model, which has achieved a comprehensive upgrade in model capabilities compared to ERNIE 3.5, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information."
1267
- },
1268
- "ernie-4.0-8k-preview": {
1269
- "description": "Baidu's flagship ultra-large-scale language model, which has achieved a comprehensive upgrade in model capabilities compared to ERNIE 3.5, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information."
1270
- },
1271
- "ernie-4.0-turbo-128k": {
1272
- "description": "Baidu's flagship ultra-large-scale language model, demonstrating outstanding overall performance, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information. It performs better than ERNIE 4.0 in terms of performance."
1273
- },
1274
- "ernie-4.0-turbo-8k-latest": {
1275
- "description": "Baidu's flagship ultra-large-scale language model, demonstrating outstanding overall performance, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information. It performs better than ERNIE 4.0 in terms of performance."
1276
- },
1277
- "ernie-4.0-turbo-8k-preview": {
1278
- "description": "Baidu's flagship ultra-large-scale language model, demonstrating outstanding overall performance, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information. It performs better than ERNIE 4.0 in terms of performance."
1274
+ "ernie-4.5-0.3b": {
1275
+ "description": "ERNIE 4.5 0.3B, an open-source lightweight model suitable for local and customized deployments."
1279
1276
  },
1280
1277
  "ernie-4.5-21b-a3b": {
1281
- "description": "ERNIE 4.5 21B A3B is a Mixture of Experts model from Baidu's Wenxin series, offering strong reasoning and multilingual capabilities."
1278
+ "description": "ERNIE 4.5 21B A3B, an open-source large-parameter model with stronger performance in understanding and generation tasks."
1282
1279
  },
1283
1280
  "ernie-4.5-300b-a47b": {
1284
1281
  "description": "ERNIE 4.5 300B A47B is a large-scale Mixture of Experts model from Baidu's Wenxin series, delivering exceptional reasoning performance."
1285
1282
  },
1286
1283
  "ernie-4.5-8k-preview": {
1287
- "description": "ERNIE 4.5 is Baidu's self-developed next-generation native multimodal foundational model, achieving collaborative optimization through joint modeling of multiple modalities, with excellent multimodal understanding capabilities; it features enhanced language abilities, with significant improvements in understanding, generation, logic, and memory, as well as reduced hallucinations and improved logical reasoning and coding capabilities."
1284
+ "description": "ERNIE 4.5 8K Preview, an 8K context preview model for experiencing and testing ERNIE 4.5 capabilities."
1288
1285
  },
1289
1286
  "ernie-4.5-turbo-128k": {
1290
- "description": "Wenxin 4.5 Turbo shows significant enhancements in reducing hallucinations, logical reasoning, and coding capabilities. Compared to Wenxin 4.5, it is faster and more cost-effective. The model's capabilities have been comprehensively improved to better meet the needs of multi-turn long history dialogue processing and long document understanding Q&A tasks."
1287
+ "description": "ERNIE 4.5 Turbo 128K, a high-performance general-purpose model supporting search augmentation and tool invocation, suitable for Q&A, coding, agents, and more."
1288
+ },
1289
+ "ernie-4.5-turbo-128k-preview": {
1290
+ "description": "ERNIE 4.5 Turbo 128K Preview, offering the same capabilities as the official version, ideal for integration testing and staging."
1291
1291
  },
1292
1292
  "ernie-4.5-turbo-32k": {
1293
- "description": "Wenxin 4.5 Turbo has also shown significant enhancements in reducing hallucinations, logical reasoning, and coding capabilities. Compared to Wenxin 4.5, it is faster and more cost-effective. There are notable improvements in text creation and knowledge Q&A capabilities. The output length and sentence delay have increased compared to ERNIE 4.5."
1293
+ "description": "ERNIE 4.5 Turbo 32K, a medium-to-long context version suitable for Q&A, knowledge retrieval, and multi-turn conversations."
1294
+ },
1295
+ "ernie-4.5-turbo-latest": {
1296
+ "description": "ERNIE 4.5 Turbo Latest, optimized for overall performance, ideal as a general-purpose production model."
1297
+ },
1298
+ "ernie-4.5-turbo-vl": {
1299
+ "description": "ERNIE 4.5 Turbo VL, a mature multimodal model for image-text understanding and recognition in production environments."
1294
1300
  },
1295
1301
  "ernie-4.5-turbo-vl-32k": {
1296
- "description": "A brand new version of the Wenxin large model, with significant improvements in image understanding, creation, translation, and coding capabilities, now supports a context length of 32K for the first time, with a significant reduction in first token delay."
1302
+ "description": "ERNIE 4.5 Turbo VL 32K, a medium-to-long text multimodal version for joint understanding of long documents and images."
1303
+ },
1304
+ "ernie-4.5-turbo-vl-32k-preview": {
1305
+ "description": "ERNIE 4.5 Turbo VL 32K Preview, a 32K multimodal preview version for evaluating long-context visual capabilities."
1306
+ },
1307
+ "ernie-4.5-turbo-vl-latest": {
1308
+ "description": "ERNIE 4.5 Turbo VL Latest, the latest multimodal version offering improved image-text understanding and reasoning."
1309
+ },
1310
+ "ernie-4.5-turbo-vl-preview": {
1311
+ "description": "ERNIE 4.5 Turbo VL Preview, a multimodal preview model supporting image-text understanding and generation, ideal for visual Q&A and content comprehension."
1312
+ },
1313
+ "ernie-4.5-vl-28b-a3b": {
1314
+ "description": "ERNIE 4.5 VL 28B A3B, an open-source multimodal model supporting image-text understanding and reasoning tasks."
1315
+ },
1316
+ "ernie-5.0-thinking-preview": {
1317
+ "description": "ERNIE 5.0 Thinking Preview, a native all-modality flagship model supporting unified modeling of text, image, audio, and video, with comprehensive capability upgrades for complex Q&A, creative tasks, and agent scenarios."
1297
1318
  },
1298
1319
  "ernie-char-8k": {
1299
- "description": "Baidu's vertical scene large language model, suitable for applications such as game NPCs, customer service dialogues, and role-playing conversations, with a more distinct and consistent character style, stronger instruction-following capabilities, and superior inference performance."
1320
+ "description": "ERNIE Character 8K, a persona dialogue model ideal for IP character building and long-term companion conversations."
1300
1321
  },
1301
1322
  "ernie-char-fiction-8k": {
1302
- "description": "Baidu's vertical scene large language model, suitable for applications such as game NPCs, customer service dialogues, and role-playing conversations, with a more distinct and consistent character style, stronger instruction-following capabilities, and superior inference performance."
1323
+ "description": "ERNIE Character Fiction 8K, a persona model for novel and story creation, suitable for generating long-form narratives."
1324
+ },
1325
+ "ernie-char-fiction-8k-preview": {
1326
+ "description": "ERNIE Character Fiction 8K Preview, a preview model for character and story creation, designed for feature testing and experience."
1303
1327
  },
1304
1328
  "ernie-irag-edit": {
1305
- "description": "Baidu's self-developed ERNIE iRAG Edit image editing model supports operations such as erase (object removal), repaint (object redrawing), and variation (variant generation) based on images."
1329
+ "description": "ERNIE iRAG Edit, an image editing model supporting image erasure, redrawing, and variant generation."
1306
1330
  },
1307
1331
  "ernie-lite-8k": {
1308
- "description": "ERNIE Lite is Baidu's lightweight large language model, balancing excellent model performance with inference efficiency, suitable for low-power AI acceleration card inference."
1332
+ "description": "ERNIE Lite 8K, a lightweight general-purpose model suitable for cost-sensitive daily Q&A and content generation."
1309
1333
  },
1310
1334
  "ernie-lite-pro-128k": {
1311
- "description": "Baidu's lightweight large language model, balancing excellent model performance with inference efficiency, offering better performance than ERNIE Lite, suitable for low-power AI acceleration card inference."
1335
+ "description": "ERNIE Lite Pro 128K, a lightweight high-performance model ideal for latency- and cost-sensitive business scenarios."
1312
1336
  },
1313
1337
  "ernie-novel-8k": {
1314
- "description": "Baidu's general-purpose large language model, which has a significant advantage in novel continuation capabilities and can also be used in short plays, movies, and other scenarios."
1338
+ "description": "ERNIE Novel 8K, a model for long-form novel and IP story creation, skilled in multi-character and multi-threaded storytelling."
1315
1339
  },
1316
1340
  "ernie-speed-128k": {
1317
- "description": "Baidu's latest self-developed high-performance large language model released in 2024, with excellent general capabilities, suitable as a base model for fine-tuning to better address specific scenario issues while also demonstrating excellent inference performance."
1341
+ "description": "ERNIE Speed 128K, a large model with no input/output cost, suitable for long-text understanding and large-scale trials."
1342
+ },
1343
+ "ernie-speed-8k": {
1344
+ "description": "ERNIE Speed 8K, a free and fast model ideal for daily conversations and lightweight text tasks."
1318
1345
  },
1319
1346
  "ernie-speed-pro-128k": {
1320
- "description": "Baidu's latest self-developed high-performance large language model released in 2024, with excellent general capabilities, offering better performance than ERNIE Speed, suitable as a base model for fine-tuning to better address specific scenario issues while also demonstrating excellent inference performance."
1347
+ "description": "ERNIE Speed Pro 128K, a high-concurrency, cost-effective model suitable for large-scale online services and enterprise applications."
1321
1348
  },
1322
1349
  "ernie-tiny-8k": {
1323
- "description": "ERNIE Tiny is Baidu's ultra-high-performance large language model, with the lowest deployment and fine-tuning costs among the Wenxin series models."
1324
- },
1325
- "ernie-x1-32k": {
1326
- "description": "Possesses stronger abilities in understanding, planning, reflection, and evolution. As a more comprehensive deep thinking model, Wenxin X1 combines accuracy, creativity, and eloquence, excelling in areas such as Chinese knowledge Q&A, literary creation, document writing, daily conversation, logical reasoning, complex calculations, and tool invocation."
1327
- },
1328
- "ernie-x1-32k-preview": {
1329
- "description": "The ERNIE X1 model possesses stronger understanding, planning, reflection, and evolution capabilities. As a more comprehensive deep thinking model, ERNIE X1 excels in accuracy, creativity, and eloquence, particularly in Chinese knowledge Q&A, literary creation, document writing, daily conversation, logical reasoning, complex calculations, and tool invocation."
1350
+ "description": "ERNIE Tiny 8K, an ultra-lightweight model for simple Q&A, classification, and other low-cost inference scenarios."
1330
1351
  },
1331
1352
  "ernie-x1-turbo-32k": {
1332
- "description": "The model performs better in terms of effectiveness and performance compared to ERNIE-X1-32K."
1353
+ "description": "ERNIE X1 Turbo 32K, a high-speed reasoning model with 32K long context, ideal for complex reasoning and multi-turn dialogue."
1354
+ },
1355
+ "ernie-x1.1-preview": {
1356
+ "description": "ERNIE X1.1 Preview, a preview version of the ERNIE X1.1 reasoning model, suitable for capability validation and testing."
1333
1357
  },
1334
1358
  "fal-ai/bytedance/seedream/v4": {
1335
1359
  "description": "Seedream 4.0 image generation model developed by ByteDance Seed team supports text and image inputs, providing a highly controllable and high-quality image generation experience. Generates images based on text prompts."
@@ -1389,7 +1413,7 @@
1389
1413
  "description": "FLUX.1 [schnell], currently the most advanced open-source few-step model, surpasses competitors and even powerful non-distilled models like Midjourney v6.0 and DALL·E 3 (HD). Finely tuned to retain the full output diversity from pretraining, FLUX.1 [schnell] significantly enhances visual quality, instruction compliance, size/aspect ratio variation, font handling, and output diversity compared to state-of-the-art models on the market, offering users a richer and more diverse creative image generation experience."
1390
1414
  },
1391
1415
  "flux.1-schnell": {
1392
- "description": "A 12-billion-parameter rectified flow transformer capable of generating images based on text descriptions."
1416
+ "description": "FLUX.1-schnell, a high-performance image generation model for fast creation of multi-style images."
1393
1417
  },
1394
1418
  "gemini-1.0-pro-001": {
1395
1419
  "description": "Gemini 1.0 Pro 001 (Tuning) offers stable and tunable performance, making it an ideal choice for complex task solutions."
@@ -1538,6 +1562,9 @@
1538
1562
  "glm-4-0520": {
1539
1563
  "description": "GLM-4-0520 is the latest model version designed for highly complex and diverse tasks, demonstrating outstanding performance."
1540
1564
  },
1565
+ "glm-4-32b-0414": {
1566
+ "description": "GLM-4 32B 0414, a general-purpose large model from the GLM series, supporting multi-task text generation and understanding."
1567
+ },
1541
1568
  "glm-4-9b-chat": {
1542
1569
  "description": "GLM-4-9B-Chat demonstrates high performance across semantics, mathematics, reasoning, coding, and knowledge. It also supports web browsing, code execution, custom tool invocation, and long-text reasoning. Supports 26 languages including Japanese, Korean, and German."
1543
1570
  },
@@ -2036,14 +2063,26 @@
2036
2063
  "internlm3-latest": {
2037
2064
  "description": "Our latest model series boasts exceptional inference performance, leading the pack among open-source models of similar scale. It defaults to our most recently released InternLM3 series models."
2038
2065
  },
2066
+ "internvl2.5-38b-mpo": {
2067
+ "description": "InternVL2.5 38B MPO, a multimodal pre-trained model supporting complex image-text reasoning tasks."
2068
+ },
2039
2069
  "internvl2.5-latest": {
2040
2070
  "description": "The InternVL2.5 version we continue to maintain, offering excellent and stable performance. It defaults to our latest released InternVL2.5 series model, currently pointing to internvl2.5-78b."
2041
2071
  },
2072
+ "internvl3-14b": {
2073
+ "description": "InternVL3 14B, a mid-sized multimodal model balancing performance and cost."
2074
+ },
2075
+ "internvl3-1b": {
2076
+ "description": "InternVL3 1B, a lightweight multimodal model suitable for deployment in resource-constrained environments."
2077
+ },
2078
+ "internvl3-38b": {
2079
+ "description": "InternVL3 38B, a large-scale open-source multimodal model for high-precision image-text understanding tasks."
2080
+ },
2042
2081
  "internvl3-latest": {
2043
2082
  "description": "Our latest released multimodal large model, featuring enhanced image-text understanding capabilities and long-sequence image comprehension, performs on par with top proprietary models. It defaults to our latest released InternVL series model, currently pointing to internvl3-78b."
2044
2083
  },
2045
2084
  "irag-1.0": {
2046
- "description": "Baidu's self-developed iRAG (image-based Retrieval-Augmented Generation) technology combines Baidu Search's hundreds of millions of image resources with powerful foundational model capabilities to generate ultra-realistic images. The overall effect far surpasses native text-to-image systems, eliminating the AI-generated feel while maintaining low cost. iRAG features hallucination-free, ultra-realistic, and instant retrieval characteristics."
2085
+ "description": "ERNIE iRAG, an image retrieval-augmented generation model supporting image search, image-text retrieval, and content generation."
2047
2086
  },
2048
2087
  "jamba-large": {
2049
2088
  "description": "Our most powerful and advanced model, designed for handling complex enterprise-level tasks with exceptional performance."
@@ -2064,7 +2103,7 @@
2064
2103
  "description": "The kimi-k2-0905-preview model has a context length of 256k, featuring stronger Agentic Coding capabilities, more outstanding aesthetics and practicality of frontend code, and better context understanding."
2065
2104
  },
2066
2105
  "kimi-k2-instruct": {
2067
- "description": "Kimi K2 Instruct is a large language model developed by Moonshot AI, featuring ultra-long context processing capabilities."
2106
+ "description": "Kimi K2 Instruct, the official Kimi inference model supporting long context, code, Q&A, and more."
2068
2107
  },
2069
2108
  "kimi-k2-turbo-preview": {
2070
2109
  "description": "Kimi-K2 is a Mixture-of-Experts (MoE) foundation model with exceptional coding and agent capabilities, featuring 1T total parameters and 32B activated parameters. In benchmark evaluations across core categories — general knowledge reasoning, programming, mathematics, and agent tasks — the K2 model outperforms other leading open-source models."
@@ -2735,6 +2774,54 @@
2735
2774
  "pro-deepseek-v3": {
2736
2775
  "description": "Enterprise-exclusive service model with concurrent service support."
2737
2776
  },
2777
+ "qianfan-70b": {
2778
+ "description": "Qianfan 70B, a large-parameter Chinese model suitable for high-quality content generation and complex reasoning tasks."
2779
+ },
2780
+ "qianfan-8b": {
2781
+ "description": "Qianfan 8B, a medium-sized general-purpose model balancing cost and performance for text generation and Q&A."
2782
+ },
2783
+ "qianfan-agent-intent-32k": {
2784
+ "description": "Qianfan Agent Intent 32K, a model for intent recognition and agent orchestration, supporting long-context scenarios."
2785
+ },
2786
+ "qianfan-agent-lite-8k": {
2787
+ "description": "Qianfan Agent Lite 8K, a lightweight agent model for low-cost multi-turn dialogue and business orchestration."
2788
+ },
2789
+ "qianfan-agent-speed-32k": {
2790
+ "description": "Qianfan Agent Speed 32K, a high-throughput agent model for large-scale, multi-task agent applications."
2791
+ },
2792
+ "qianfan-agent-speed-8k": {
2793
+ "description": "Qianfan Agent Speed 8K, a high-concurrency agent model for short-to-medium conversations and fast response."
2794
+ },
2795
+ "qianfan-check-vl": {
2796
+ "description": "Qianfan Check VL, a multimodal content moderation and detection model supporting image-text compliance and recognition tasks."
2797
+ },
2798
+ "qianfan-composition": {
2799
+ "description": "Qianfan Composition, a multimodal creative model supporting integrated image-text understanding and generation."
2800
+ },
2801
+ "qianfan-engcard-vl": {
2802
+ "description": "Qianfan EngCard VL, a multimodal recognition model focused on English-language scenarios."
2803
+ },
2804
+ "qianfan-lightning-128b-a19b": {
2805
+ "description": "Qianfan Lightning 128B A19B, a high-performance Chinese general-purpose model for complex Q&A and large-scale reasoning tasks."
2806
+ },
2807
+ "qianfan-llama-vl-8b": {
2808
+ "description": "Qianfan Llama VL 8B, a multimodal model based on Llama for general image-text understanding tasks."
2809
+ },
2810
+ "qianfan-multipicocr": {
2811
+ "description": "Qianfan MultiPicOCR, a multi-image OCR model supporting text detection and recognition across multiple images."
2812
+ },
2813
+ "qianfan-qi-vl": {
2814
+ "description": "Qianfan QI VL, a multimodal Q&A model supporting accurate retrieval and question answering in complex image-text scenarios."
2815
+ },
2816
+ "qianfan-singlepicocr": {
2817
+ "description": "Qianfan SinglePicOCR, a single-image OCR model supporting high-precision character recognition."
2818
+ },
2819
+ "qianfan-vl-70b": {
2820
+ "description": "Qianfan VL 70B, a large-parameter vision-language model for complex image-text understanding scenarios."
2821
+ },
2822
+ "qianfan-vl-8b": {
2823
+ "description": "Qianfan VL 8B, a lightweight vision-language model suitable for daily image-text Q&A and analysis."
2824
+ },
2738
2825
  "qvq-72b-preview": {
2739
2826
  "description": "The QVQ model is an experimental research model developed by the Qwen team, focusing on enhancing visual reasoning capabilities, particularly in the field of mathematical reasoning."
2740
2827
  },
@@ -2885,9 +2972,7 @@
2885
2972
  "qwen2.5-72b-instruct": {
2886
2973
  "description": "The 72B model of Tongyi Qianwen 2.5 is open-sourced."
2887
2974
  },
2888
- "qwen2.5-7b-instruct": {
2889
- "description": "The 7B model of Tongyi Qianwen 2.5 is open-sourced."
2890
- },
2975
+ "qwen2.5-7b-instruct": {},
2891
2976
  "qwen2.5-coder-1.5b-instruct": {
2892
2977
  "description": "Open-source version of the Qwen coding model."
2893
2978
  },
@@ -2918,15 +3003,11 @@
2918
3003
  "qwen2.5-omni-7b": {
2919
3004
  "description": "The Qwen-Omni series models support input of various modalities, including video, audio, images, and text, and output audio and text."
2920
3005
  },
2921
- "qwen2.5-vl-32b-instruct": {
2922
- "description": "The Qwen2.5-VL model series enhances the model's intelligence level, practicality, and applicability, delivering superior performance in scenarios such as natural conversations, content creation, professional knowledge services, and code development. The 32B version employs reinforcement learning techniques to optimize the model, offering more human-preferred output styles, enhanced reasoning capabilities for complex mathematical problems, and fine-grained image understanding and reasoning compared to other models in the Qwen2.5-VL series."
2923
- },
3006
+ "qwen2.5-vl-32b-instruct": {},
2924
3007
  "qwen2.5-vl-72b-instruct": {
2925
3008
  "description": "This version enhances instruction following, mathematics, problem-solving, and coding capabilities, improving the ability to recognize various formats and accurately locate visual elements. It supports understanding long video files (up to 10 minutes) and pinpointing events in seconds, comprehending the sequence and speed of time, and based on parsing and locating capabilities, it supports controlling OS or Mobile agents. It has strong key information extraction and JSON output capabilities, and this version is the most powerful in the series at 72B."
2926
3009
  },
2927
- "qwen2.5-vl-7b-instruct": {
2928
- "description": "This version enhances instruction following, mathematics, problem-solving, and coding capabilities, improving the ability to recognize various formats and accurately locate visual elements. It supports understanding long video files (up to 10 minutes) and pinpointing events in seconds, comprehending the sequence and speed of time, and based on parsing and locating capabilities, it supports controlling OS or Mobile agents. It has strong key information extraction and JSON output capabilities, and this version is the most powerful in the series at 72B."
2929
- },
3010
+ "qwen2.5-vl-7b-instruct": {},
2930
3011
  "qwen2.5-vl-instruct": {
2931
3012
  "description": "Qwen2.5-VL is the latest version of the visual language model in the Qwen model family."
2932
3013
  },
@@ -2951,48 +3032,22 @@
2951
3032
  "qwen3": {
2952
3033
  "description": "Qwen3 is Alibaba's next-generation large-scale language model, designed to support diverse application needs with outstanding performance."
2953
3034
  },
2954
- "qwen3-0.6b": {
2955
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2956
- },
2957
- "qwen3-1.7b": {
2958
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2959
- },
2960
- "qwen3-14b": {
2961
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2962
- },
2963
- "qwen3-235b-a22b": {
2964
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2965
- },
2966
- "qwen3-235b-a22b-instruct-2507": {
2967
- "description": "An open-source non-thinking mode model based on Qwen3, with slight improvements in subjective creativity and model safety compared to the previous version (Tongyi Qianwen 3-235B-A22B)."
2968
- },
2969
- "qwen3-235b-a22b-thinking-2507": {
2970
- "description": "An open-source thinking mode model based on Qwen3, with significant improvements in logical ability, general capabilities, knowledge enhancement, and creativity compared to the previous version (Tongyi Qianwen 3-235B-A22B), suitable for high-difficulty and strong reasoning scenarios."
2971
- },
2972
- "qwen3-30b-a3b": {
2973
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2974
- },
2975
- "qwen3-30b-a3b-instruct-2507": {
2976
- "description": "Compared to the previous version (Qwen3-30B-A3B), this version shows substantial improvements in overall general capabilities in both Chinese and multilingual contexts. It features specialized optimizations for subjective and open-ended tasks, aligning significantly better with user preferences and providing more helpful responses."
2977
- },
2978
- "qwen3-30b-a3b-thinking-2507": {
2979
- "description": "An open-source thinking mode model based on Qwen3, this version shows significant enhancements over the previous release (Tongyi Qianwen 3-30B-A3B) in logical ability, general capability, knowledge augmentation, and creative capacity. It is suitable for challenging scenarios requiring strong reasoning."
2980
- },
2981
- "qwen3-32b": {
2982
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2983
- },
2984
- "qwen3-4b": {
2985
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2986
- },
2987
- "qwen3-8b": {
2988
- "description": "Qwen3 is a next-generation model with significantly enhanced capabilities, achieving industry-leading levels in reasoning, general tasks, agent functionality, and multilingual support, while also supporting mode switching."
2989
- },
3035
+ "qwen3-0.6b": {},
3036
+ "qwen3-1.7b": {},
3037
+ "qwen3-14b": {},
3038
+ "qwen3-235b-a22b": {},
3039
+ "qwen3-235b-a22b-instruct-2507": {},
3040
+ "qwen3-235b-a22b-thinking-2507": {},
3041
+ "qwen3-30b-a3b": {},
3042
+ "qwen3-30b-a3b-instruct-2507": {},
3043
+ "qwen3-30b-a3b-thinking-2507": {},
3044
+ "qwen3-32b": {},
3045
+ "qwen3-4b": {},
3046
+ "qwen3-8b": {},
2990
3047
  "qwen3-coder-30b-a3b-instruct": {
2991
3048
  "description": "The open-source version of the Tongyi Qianwen code model. The latest qwen3-coder-30b-a3b-instruct is a code generation model based on Qwen3, featuring powerful Coding Agent capabilities. It excels at tool usage and environment interaction, enabling autonomous programming with outstanding coding and general abilities."
2992
3049
  },
2993
- "qwen3-coder-480b-a35b-instruct": {
2994
- "description": "Open-source version of Tongyi Qianwen's code model. The latest qwen3-coder-480b-a35b-instruct is a code generation model based on Qwen3, featuring powerful Coding Agent capabilities, proficient in tool invocation and environment interaction, enabling autonomous programming with excellent coding and general capabilities."
2995
- },
3050
+ "qwen3-coder-480b-a35b-instruct": {},
2996
3051
  "qwen3-coder-flash": {
2997
3052
  "description": "Tongyi Qianwen code model. The latest Qwen3-Coder series models are code generation models based on Qwen3, equipped with powerful Coding Agent capabilities, proficient in tool invocation and environment interaction, enabling autonomous programming with excellent coding skills alongside general capabilities."
2998
3053
  },
@@ -3008,30 +3063,16 @@
3008
3063
  "qwen3-next-80b-a3b-instruct": {
3009
3064
  "description": "A new generation of non-thinking mode open-source model based on Qwen3. Compared to the previous version (Tongyi Qianwen 3-235B-A22B-Instruct-2507), it offers better Chinese text comprehension, enhanced logical reasoning abilities, and improved performance in text generation tasks."
3010
3065
  },
3011
- "qwen3-next-80b-a3b-thinking": {
3012
- "description": "A new generation of thinking mode open-source model based on Qwen3. Compared to the previous version (Tongyi Qianwen 3-235B-A22B-Thinking-2507), it features improved instruction-following capabilities and more concise model-generated summaries."
3013
- },
3066
+ "qwen3-next-80b-a3b-thinking": {},
3014
3067
  "qwen3-omni-flash": {
3015
3068
  "description": "The Qwen-Omni model accepts multimodal input including text, images, audio, and video, and generates responses in text or speech. It offers a variety of human-like voice tones, supports multilingual and dialectal speech output, and is applicable to scenarios such as text creation, visual recognition, and voice assistants."
3016
3069
  },
3017
- "qwen3-vl-235b-a22b-instruct": {
3018
- "description": "Qwen3 VL 235B A22B Instruct Mode (non-thinking), designed for instruction-following scenarios without deep reasoning, while maintaining strong visual understanding capabilities."
3019
- },
3020
- "qwen3-vl-235b-a22b-thinking": {
3021
- "description": "Qwen3 VL 235B A22B Thinking Mode (open-source version), tailored for complex reasoning and long video understanding tasks, offering top-tier visual and textual reasoning performance."
3022
- },
3023
- "qwen3-vl-30b-a3b-instruct": {
3024
- "description": "Qwen3 VL 30B Instruct Mode (non-thinking), designed for general instruction-following scenarios, with strong multimodal understanding and generation capabilities."
3025
- },
3026
- "qwen3-vl-30b-a3b-thinking": {
3027
- "description": "The open-source Qwen-VL model provides visual understanding and text generation capabilities. It supports agent interaction, visual encoding, spatial awareness, long video comprehension, and deep reasoning, with enhanced text recognition and multilingual support in complex scenarios."
3028
- },
3029
- "qwen3-vl-8b-instruct": {
3030
- "description": "Qwen3 VL 8B Instruct Mode (non-thinking), suitable for standard multimodal generation and recognition tasks."
3031
- },
3032
- "qwen3-vl-8b-thinking": {
3033
- "description": "Qwen3 VL 8B Thinking Mode, designed for lightweight multimodal reasoning and interaction scenarios, while retaining long-context understanding capabilities."
3034
- },
3070
+ "qwen3-vl-235b-a22b-instruct": {},
3071
+ "qwen3-vl-235b-a22b-thinking": {},
3072
+ "qwen3-vl-30b-a3b-instruct": {},
3073
+ "qwen3-vl-30b-a3b-thinking": {},
3074
+ "qwen3-vl-8b-instruct": {},
3075
+ "qwen3-vl-8b-thinking": {},
3035
3076
  "qwen3-vl-flash": {
3036
3077
  "description": "Qwen3 VL Flash: a lightweight, high-speed inference version ideal for latency-sensitive or high-volume request scenarios."
3037
3078
  },
@@ -259,7 +259,7 @@
259
259
  "searchLocalFiles": "Search Files",
260
260
  "writeLocalFile": "Write File"
261
261
  },
262
- "title": "Local Files"
262
+ "title": "Local System"
263
263
  },
264
264
  "mcpInstall": {
265
265
  "CHECKING_INSTALLATION": "Checking installation environment...",