@lobehub/chat 1.128.1 → 1.128.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,32 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.128.2](https://github.com/lobehub/lobe-chat/compare/v1.128.1...v1.128.2)
6
+
7
+ <sup>Released on **2025-09-13**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Update i18n, Update model configs.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Update i18n, closes [#9237](https://github.com/lobehub/lobe-chat/issues/9237) ([642dc3b](https://github.com/lobehub/lobe-chat/commit/642dc3b))
21
+ - **misc**: Update model configs, closes [#9170](https://github.com/lobehub/lobe-chat/issues/9170) ([f89b730](https://github.com/lobehub/lobe-chat/commit/f89b730))
22
+
23
+ </details>
24
+
25
+ <div align="right">
26
+
27
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
28
+
29
+ </div>
30
+
5
31
  ### [Version 1.128.1](https://github.com/lobehub/lobe-chat/compare/v1.128.0...v1.128.1)
6
32
 
7
33
  <sup>Released on **2025-09-12**</sup>
package/changelog/v1.json CHANGED
@@ -1,4 +1,13 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "improvements": [
5
+ "Update i18n, Update model configs."
6
+ ]
7
+ },
8
+ "date": "2025-09-13",
9
+ "version": "1.128.2"
10
+ },
2
11
  {
3
12
  "children": {
4
13
  "improvements": [
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "نسخة المعاينة لنموذج Qwen 3 Max من سلسلة Tongyi Qianwen، مع تحسينات كبيرة في القدرات العامة مقارنة بسلسلة 2.5، بما في ذلك فهم النصوص العامة باللغتين الصينية والإنجليزية، الالتزام بالتعليمات المعقدة، المهام المفتوحة الذاتية، القدرات متعددة اللغات، واستدعاء الأدوات؛ مع تقليل الأوهام المعرفية للنموذج."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "نموذج مفتوح المصدر من الجيل الجديد لوضع عدم التفكير مبني على Qwen3، يتميز بفهم أفضل للنصوص الصينية مقارنة بالإصدار السابق (Tongyi Qianwen 3-235B-A22B-Instruct-2507)، مع تعزيز في قدرات الاستدلال المنطقي وأداء أفضل في مهام توليد النصوص."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "نموذج مفتوح المصدر من الجيل الجديد لوضع التفكير مبني على Qwen3، يتميز بتحسين في الالتزام بالتعليمات مقارنة بالإصدار السابق (Tongyi Qianwen 3-235B-A22B-Thinking-2507)، مع ردود ملخصة وأكثر إيجازًا من النموذج."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ هو نموذج بحث تجريبي يركز على تحسين قدرات الاستدلال للذكاء الاصطناعي."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Preview версия на Max модела от серията Qwen 3, с голямо подобрение спрямо серия 2.5 в общите универсални способности, разбиране на текст на китайски и английски, следване на сложни инструкции, субективни отворени задачи, многоезични способности и използване на инструменти; моделът има по-малко халюцинации на знания."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Базирано на Qwen3, ново поколение отворен модел без мисловен режим, който предлага по-добро разбиране на китайски текстове, подобрени логически умения и по-добри резултати при задачи за генериране на текст в сравнение с предишната версия (Tongyi Qianwen 3-235B-A22B-Instruct-2507)."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Базирано на Qwen3, ново поколение отворен модел с мисловен режим, който подобрява спазването на инструкции и предоставя по-кратки и точни обобщения в сравнение с предишната версия (Tongyi Qianwen 3-235B-A22B-Thinking-2507)."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ е експериментален изследователски модел, който се фокусира върху подобряване на AI разсъдъчните способности."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Die Preview-Version des Max-Modells der Tongyi Qianwen 3-Serie zeigt im Vergleich zur 2.5-Serie eine deutliche Steigerung der allgemeinen Fähigkeiten, einschließlich verbesserter chinesisch- und englischsprachiger Textverständnisfähigkeiten, komplexer Befolgung von Anweisungen, subjektiver offener Aufgaben, Mehrsprachigkeit und Werkzeugaufruf-Fähigkeiten; das Modell zeigt weniger Wissenshalluzinationen."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Ein neues Open-Source-Modell der nächsten Generation im Nicht-Denk-Modus basierend auf Qwen3. Im Vergleich zur vorherigen Version (Tongyi Qianwen 3-235B-A22B-Instruct-2507) bietet es eine verbesserte chinesische Textverständnisfähigkeit, verstärkte logische Schlussfolgerungen und bessere Leistung bei textgenerierenden Aufgaben."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Ein neues Open-Source-Modell der nächsten Generation im Denkmodus basierend auf Qwen3. Im Vergleich zur vorherigen Version (Tongyi Qianwen 3-235B-A22B-Thinking-2507) wurde die Befehlsbefolgung verbessert und die Modellantworten sind prägnanter zusammengefasst."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ ist ein experimentelles Forschungsmodell, das sich auf die Verbesserung der KI-Inferenzfähigkeiten konzentriert."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "The Qwen 3 series Max model preview version shows a significant overall improvement compared to the 2.5 series in general capabilities, including Chinese and English text understanding, complex instruction adherence, subjective open tasks, multilingual capabilities, and tool invocation. The model also exhibits fewer knowledge hallucinations."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "A new generation of non-thinking mode open-source model based on Qwen3. Compared to the previous version (Tongyi Qianwen 3-235B-A22B-Instruct-2507), it offers better Chinese text comprehension, enhanced logical reasoning abilities, and improved performance in text generation tasks."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "A new generation of thinking mode open-source model based on Qwen3. Compared to the previous version (Tongyi Qianwen 3-235B-A22B-Thinking-2507), it features improved instruction-following capabilities and more concise model-generated summaries."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ is an experimental research model focused on improving AI reasoning capabilities."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Versión preliminar del modelo Max de la serie Qwen 3, que presenta una mejora significativa en la capacidad general en comparación con la serie 2.5, incluyendo comprensión de texto en chino e inglés, cumplimiento de instrucciones complejas, tareas abiertas subjetivas, capacidades multilingües y llamadas a herramientas; además, reduce notablemente las alucinaciones de conocimiento del modelo."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Modelo de código abierto de nueva generación basado en Qwen3 en modo no reflexivo, que ofrece una mejor comprensión del texto en chino, mayor capacidad de razonamiento lógico y un mejor desempeño en tareas de generación de texto en comparación con la versión anterior (Tongyi Qianwen 3-235B-A22B-Instruct-2507)."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Modelo de código abierto de nueva generación basado en Qwen3 en modo reflexivo, que mejora la capacidad de seguir instrucciones y ofrece respuestas más concisas en comparación con la versión anterior (Tongyi Qianwen 3-235B-A22B-Thinking-2507)."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ es un modelo de investigación experimental que se centra en mejorar la capacidad de razonamiento de la IA."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "نسخه پیش‌نمایش مدل Max از سری Qwen 3، نسبت به سری 2.5 بهبود قابل توجهی در توانایی‌های عمومی، درک متن‌های چندزبانه چینی و انگلیسی، پیروی از دستورات پیچیده، انجام وظایف باز و ذهنی، پشتیبانی چندزبانه و فراخوانی ابزارها دارد؛ همچنین خطاهای دانش مدل کاهش یافته است."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "مدل متن‌باز نسل جدید بدون حالت تفکر مبتنی بر Qwen3، که نسبت به نسخه قبلی (Tongyi Qianwen 3-235B-A22B-Instruct-2507) در درک متن‌های چینی بهتر است، توانایی استدلال منطقی بهبود یافته و عملکرد بهتری در وظایف تولید متن دارد."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "مدل متن‌باز نسل جدید با حالت تفکر مبتنی بر Qwen3، که نسبت به نسخه قبلی (Tongyi Qianwen 3-235B-A22B-Thinking-2507) در پیروی از دستورات پیشرفت داشته و پاسخ‌های مدل خلاصه‌تر شده‌اند."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ یک مدل تحقیقاتی تجربی است که بر بهبود توانایی استدلال AI تمرکز دارد."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Version Preview du modèle Max de la série Qwen 3, avec une amélioration significative des capacités générales par rapport à la série 2.5, notamment en compréhension de texte général bilingue (chinois et anglais), respect des instructions complexes, tâches ouvertes subjectives, capacités multilingues et appels d'outils ; le modèle présente également moins d'hallucinations de connaissances."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Modèle open source de nouvelle génération en mode non réflexif basé sur Qwen3, offrant une meilleure compréhension du texte en chinois, des capacités de raisonnement logique renforcées et de meilleures performances dans les tâches de génération de texte par rapport à la version précédente (Tongyi Qianwen 3-235B-A22B-Instruct-2507)."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Modèle open source de nouvelle génération en mode réflexif basé sur Qwen3, avec une meilleure conformité aux instructions et des réponses plus concises dans les résumés par rapport à la version précédente (Tongyi Qianwen 3-235B-A22B-Thinking-2507)."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ est un modèle de recherche expérimental, axé sur l'amélioration des capacités de raisonnement de l'IA."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Versione preview del modello Max della serie Qwen 3, che presenta un miglioramento significativo rispetto alla serie 2.5 nelle capacità generali, comprensione del testo in cinese e inglese, rispetto di istruzioni complesse, compiti soggettivi aperti, capacità multilingue e chiamata di strumenti; il modello presenta meno allucinazioni di conoscenza."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Modello open source di nuova generazione basato su Qwen3 in modalità non riflessiva, con una migliore comprensione del testo in cinese rispetto alla versione precedente (Tongyi Qianwen 3-235B-A22B-Instruct-2507), capacità di ragionamento logico potenziate e prestazioni superiori nelle attività di generazione di testo."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Modello open source di nuova generazione basato su Qwen3 in modalità riflessiva, con migliorata capacità di seguire le istruzioni rispetto alla versione precedente (Tongyi Qianwen 3-235B-A22B-Thinking-2507) e risposte di sintesi più concise."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ è un modello di ricerca sperimentale, focalizzato sul miglioramento delle capacità di ragionamento dell'IA."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "通義千問3シリーズMaxモデルのプレビュー版で、2.5シリーズに比べて全体的な汎用能力が大幅に向上し、中日両言語の汎用テキスト理解能力、複雑な指示遵守能力、主観的なオープンタスク能力、多言語能力、ツール呼び出し能力が著しく強化されました。モデルの知識幻覚も減少しています。"
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Qwen3に基づく次世代の非思考モードのオープンソースモデルで、前バージョン(通義千問3-235B-A22B-Instruct-2507)と比べて中国語テキストの理解能力が向上し、論理推論能力が強化され、テキスト生成タスクのパフォーマンスがより優れています。"
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Qwen3に基づく次世代の思考モードのオープンソースモデルで、前バージョン(通義千問3-235B-A22B-Thinking-2507)と比べて指示遵守能力が向上し、モデルの要約応答がより簡潔になっています。"
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQはAIの推論能力を向上させることに特化した実験的研究モデルです。"
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "통의천문3 시리즈 Max 모델 프리뷰 버전으로, 2.5 시리즈에 비해 전반적인 범용 능력이 크게 향상되었으며, 중영문 범용 텍스트 이해 능력, 복잡한 지시 준수 능력, 주관적 개방형 작업 능력, 다국어 능력, 도구 호출 능력이 모두 크게 강화되었습니다. 모델의 지식 환각도 감소하였습니다."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Qwen3 기반의 차세대 비사고 모드 오픈 소스 모델로, 이전 버전(통의천문3-235B-A22B-Instruct-2507)과 비교하여 중국어 텍스트 이해 능력이 향상되었고, 논리 추론 능력이 강화되었으며, 텍스트 생성 작업에서 더 우수한 성능을 보입니다."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Qwen3 기반의 차세대 사고 모드 오픈 소스 모델로, 이전 버전(통의천문3-235B-A22B-Thinking-2507)과 비교하여 명령 준수 능력이 향상되었고, 모델의 요약 응답이 더욱 간결해졌습니다."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ는 AI 추론 능력을 향상시키는 데 중점을 둔 실험 연구 모델입니다."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Previewversie van het Qwen 3 Max-model uit de Tongyi Qianwen 3-serie, met aanzienlijke verbeteringen ten opzichte van de 2.5-serie in algemene capaciteiten, tweetalig tekstbegrip (Chinees en Engels), complexe instructienaleving, subjectieve open taken, meertalige vaardigheden en toolaanroepen; het model vertoont minder kennisillusies."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Een nieuwe generatie open-source model zonder denkmodus gebaseerd op Qwen3, met verbeterde Chinese tekstbegrip, versterkte logische redeneervaardigheden en betere prestaties bij tekstgeneratietaken vergeleken met de vorige versie (Tongyi Qianwen 3-235B-A22B-Instruct-2507)."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Een nieuwe generatie open-source model met denkmodus gebaseerd op Qwen3, met verbeterde instructienaleving en meer beknopte samenvattende antwoorden vergeleken met de vorige versie (Tongyi Qianwen 3-235B-A22B-Thinking-2507)."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ is een experimenteel onderzoeksmodel dat zich richt op het verbeteren van de AI-redeneringscapaciteiten."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Preview modelu Max z serii Qwen 3, który w porównaniu do serii 2.5 znacząco poprawia ogólne zdolności, w tym rozumienie tekstu w języku chińskim i angielskim, przestrzeganie złożonych instrukcji, zdolności do zadań otwartych i subiektywnych, wielojęzyczność oraz wywoływanie narzędzi; model generuje mniej halucynacji wiedzy."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Nowa generacja otwartego modelu bez trybu myślenia oparta na Qwen3, która w porównaniu z poprzednią wersją (Tongyi Qianwen 3-235B-A22B-Instruct-2507) cechuje się lepszym rozumieniem tekstu w języku chińskim, wzmocnionymi zdolnościami wnioskowania logicznego oraz lepszą wydajnością w zadaniach generowania tekstu."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Nowa generacja otwartego modelu z trybem myślenia oparta na Qwen3, która w porównaniu z poprzednią wersją (Tongyi Qianwen 3-235B-A22B-Thinking-2507) wykazuje poprawę w przestrzeganiu instrukcji oraz bardziej zwięzłe podsumowania w odpowiedziach modelu."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ to eksperymentalny model badawczy, skoncentrowany na zwiększeniu zdolności wnioskowania AI."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Versão Preview do modelo Max da série Qwen 3, com melhorias significativas em relação à série 2.5 em capacidades gerais, compreensão de texto em chinês e inglês, conformidade com instruções complexas, tarefas subjetivas abertas, multilinguismo e chamadas de ferramentas; o modelo apresenta menos alucinações de conhecimento."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Modelo open source de nova geração no modo não reflexivo baseado no Qwen3, que apresenta melhor compreensão de texto em chinês, capacidades aprimoradas de raciocínio lógico e desempenho superior em tarefas de geração de texto em comparação com a versão anterior (Tongyi Qianwen 3-235B-A22B-Instruct-2507)."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Modelo open source de nova geração no modo reflexivo baseado no Qwen3, que oferece melhor conformidade com instruções e respostas mais concisas em resumos, em comparação com a versão anterior (Tongyi Qianwen 3-235B-A22B-Thinking-2507)."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ é um modelo de pesquisa experimental, focado em melhorar a capacidade de raciocínio da IA."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Предварительная версия модели серии Qwen 3 Max, которая значительно превосходит серию 2.5 по универсальным возможностям, включая понимание текста на китайском и английском языках, выполнение сложных инструкций, способности к субъективным открытым задачам, многоязычность и вызов инструментов; модель демонстрирует меньше искажений знаний."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Новая генерация открытой модели без режима мышления на базе Qwen3, которая по сравнению с предыдущей версией (通义千问3-235B-A22B-Instruct-2507) обладает улучшенными способностями понимания китайского текста, усиленными логическими рассуждениями и лучшими результатами в задачах генерации текста."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Новая генерация открытой модели с режимом мышления на базе Qwen3, которая по сравнению с предыдущей версией (通义千问3-235B-A22B-Thinking-2507) демонстрирует улучшенное следование инструкциям и более лаконичные ответы модели."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ — это экспериментальная исследовательская модель, сосредоточенная на повышении возможностей вывода ИИ."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Tongyi Qianwen 3 serisi Max modelinin önizleme sürümüdür. 2.5 serisine kıyasla genel yeteneklerde büyük gelişmeler göstermiştir; Çince ve İngilizce genel metin anlama, karmaşık talimat uyumu, öznel açık görevler, çok dilli yetenekler ve araç çağrısı yetenekleri belirgin şekilde artmıştır; model bilgi yanılgıları daha azdır."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Qwen3 tabanlı yeni nesil düşünmeden çalışan açık kaynak modeli, önceki sürüme (Tongyi Qianwen 3-235B-A22B-Instruct-2507) kıyasla Çince metin anlama yeteneği daha iyi, mantıksal çıkarım yeteneği geliştirilmiş ve metin üretimi görevlerinde daha başarılıdır."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Qwen3 tabanlı yeni nesil düşünme modlu açık kaynak modeli, önceki sürüme (Tongyi Qianwen 3-235B-A22B-Thinking-2507) kıyasla komutlara uyum yeteneği artırılmış ve modelin özetleyici yanıtları daha özlü hale getirilmiştir."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ, AI akıl yürütme yeteneklerini artırmaya odaklanan deneysel bir araştırma modelidir."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "Phiên bản xem trước mô hình Max của dòng Qwen 3, so với dòng 2.5 có sự cải thiện lớn về năng lực tổng quát, khả năng hiểu văn bản song ngữ Trung-Anh, tuân thủ chỉ thị phức tạp, thực hiện nhiệm vụ mở chủ quan, đa ngôn ngữ và gọi công cụ đều được tăng cường rõ rệt; mô hình giảm thiểu ảo giác kiến thức."
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "Mô hình mã nguồn mở thế hệ mới không có chế độ suy nghĩ dựa trên Qwen3, so với phiên bản trước (Thông Nghĩa Thiên Vấn 3-235B-A22B-Instruct-2507) có khả năng hiểu văn bản tiếng Trung tốt hơn, năng lực suy luận logic được cải thiện, và hiệu suất trong các nhiệm vụ tạo văn bản cũng tốt hơn."
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "Mô hình mã nguồn mở thế hệ mới có chế độ suy nghĩ dựa trên Qwen3, so với phiên bản trước (Thông Nghĩa Thiên Vấn 3-235B-A22B-Thinking-2507) có khả năng tuân thủ chỉ dẫn được nâng cao, và các phản hồi tóm tắt của mô hình trở nên ngắn gọn hơn."
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ là một mô hình nghiên cứu thử nghiệm, tập trung vào việc nâng cao khả năng suy luận của AI."
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "通义千问3系列Max模型Preview版本,相较2.5系列整体通用能力有大幅度提升,中英文通用文本理解能力、复杂指令遵循能力、主观开放任务能力、多语言能力、工具调用能力均显著增强;模型知识幻觉更少。"
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "基于 Qwen3 的新一代非思考模式开源模型,相较上一版本(通义千问3-235B-A22B-Instruct-2507)中文文本理解能力更佳、逻辑推理能力有增强、文本生成类任务表现更好。"
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "基于 Qwen3 的新一代思考模式开源模型,相较上一版本(通义千问3-235B-A22B-Thinking-2507)指令遵循能力有提升、模型总结回复更加精简。"
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ 是 Qwen 系列的推理模型。与传统的指令调优模型相比,QwQ 具备思考和推理的能力,能够在下游任务中,尤其是困难问题上,显著提升性能。QwQ-32B 是中型推理模型,能够在与最先进的推理模型(如 DeepSeek-R1、o1-mini)竞争时取得可观的表现。"
2620
2626
  },
@@ -2615,6 +2615,12 @@
2615
2615
  "qwen3-max-preview": {
2616
2616
  "description": "通義千問3系列 Max 模型 Preview 版本,相較 2.5 系列整體通用能力有大幅度提升,中英文通用文本理解能力、複雜指令遵循能力、主觀開放任務能力、多語言能力、工具調用能力均顯著增強;模型知識幻覺更少。"
2617
2617
  },
2618
+ "qwen3-next-80b-a3b-instruct": {
2619
+ "description": "基於 Qwen3 的新一代非思考模式開源模型,相較上一版本(通義千問3-235B-A22B-Instruct-2507)中文文本理解能力更佳、邏輯推理能力有增強、文本生成類任務表現更好。"
2620
+ },
2621
+ "qwen3-next-80b-a3b-thinking": {
2622
+ "description": "基於 Qwen3 的新一代思考模式開源模型,相較上一版本(通義千問3-235B-A22B-Thinking-2507)指令遵循能力有提升、模型總結回覆更加精簡。"
2623
+ },
2618
2624
  "qwq": {
2619
2625
  "description": "QwQ 是一個實驗研究模型,專注於提高 AI 推理能力。"
2620
2626
  },
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.128.1",
3
+ "version": "1.128.2",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -17,8 +17,8 @@ const deepseekChatModels: AIChatModelCard[] = [
17
17
  currency: 'CNY',
18
18
  units: [
19
19
  { name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
20
- { name: 'textInput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
21
- { name: 'textOutput', rate: 8, strategy: 'fixed', unit: 'millionTokens' },
20
+ { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
21
+ { name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
22
22
  ],
23
23
  },
24
24
  releasedAt: '2025-08-21',
@@ -39,9 +39,9 @@ const deepseekChatModels: AIChatModelCard[] = [
39
39
  pricing: {
40
40
  currency: 'CNY',
41
41
  units: [
42
- { name: 'textInput_cacheRead', rate: 1, strategy: 'fixed', unit: 'millionTokens' },
42
+ { name: 'textInput_cacheRead', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
43
43
  { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
44
- { name: 'textOutput', rate: 16, strategy: 'fixed', unit: 'millionTokens' },
44
+ { name: 'textOutput', rate: 12, strategy: 'fixed', unit: 'millionTokens' },
45
45
  ],
46
46
  },
47
47
  releasedAt: '2025-08-21',
@@ -18,7 +18,15 @@ const googleChatModels: AIChatModelCard[] = [
18
18
  maxOutput: 65_536,
19
19
  pricing: {
20
20
  units: [
21
- { name: 'textInput_cacheRead', rate: 0.31, strategy: 'fixed', unit: 'millionTokens' },
21
+ {
22
+ name: 'textInput_cacheRead',
23
+ strategy: 'tiered',
24
+ tiers: [
25
+ { rate: 0.31, upTo: 200_000 },
26
+ { rate: 0.625, upTo: 'infinity' },
27
+ ],
28
+ unit: 'millionTokens',
29
+ },
22
30
  {
23
31
  name: 'textInput',
24
32
  strategy: 'tiered',
@@ -57,12 +65,20 @@ const googleChatModels: AIChatModelCard[] = [
57
65
  contextWindowTokens: 1_048_576 + 65_536,
58
66
  description:
59
67
  'Gemini 2.5 Pro Preview 是 Google 最先进的思维模型,能够对代码、数学和STEM领域的复杂问题进行推理,以及使用长上下文分析大型数据集、代码库和文档。',
60
- displayName: 'Gemini 2.5 Pro Preview 06-05 (Paid)',
68
+ displayName: 'Gemini 2.5 Pro Preview 06-05',
61
69
  id: 'gemini-2.5-pro-preview-06-05',
62
70
  maxOutput: 65_536,
63
71
  pricing: {
64
72
  units: [
65
- { name: 'textInput_cacheRead', rate: 0.31, strategy: 'fixed', unit: 'millionTokens' },
73
+ {
74
+ name: 'textInput_cacheRead',
75
+ strategy: 'tiered',
76
+ tiers: [
77
+ { rate: 0.31, upTo: 200_000 },
78
+ { rate: 0.625, upTo: 'infinity' },
79
+ ],
80
+ unit: 'millionTokens',
81
+ },
66
82
  {
67
83
  name: 'textInput',
68
84
  strategy: 'tiered',
@@ -101,12 +117,20 @@ const googleChatModels: AIChatModelCard[] = [
101
117
  contextWindowTokens: 1_048_576 + 65_536,
102
118
  description:
103
119
  'Gemini 2.5 Pro Preview 是 Google 最先进的思维模型,能够对代码、数学和STEM领域的复杂问题进行推理,以及使用长上下文分析大型数据集、代码库和文档。',
104
- displayName: 'Gemini 2.5 Pro Preview 05-06 (Paid)',
120
+ displayName: 'Gemini 2.5 Pro Preview 05-06',
105
121
  id: 'gemini-2.5-pro-preview-05-06',
106
122
  maxOutput: 65_536,
107
123
  pricing: {
108
124
  units: [
109
- { name: 'textInput_cacheRead', rate: 0.31, strategy: 'fixed', unit: 'millionTokens' },
125
+ {
126
+ name: 'textInput_cacheRead',
127
+ strategy: 'tiered',
128
+ tiers: [
129
+ { rate: 0.31, upTo: 200_000 },
130
+ { rate: 0.625, upTo: 'infinity' },
131
+ ],
132
+ unit: 'millionTokens',
133
+ },
110
134
  {
111
135
  name: 'textInput',
112
136
  strategy: 'tiered',
@@ -479,7 +503,7 @@ const googleChatModels: AIChatModelCard[] = [
479
503
  contextWindowTokens: 2_008_192,
480
504
  description:
481
505
  'Gemini 1.5 Pro 002 是最新的生产就绪模型,提供更高质量的输出,特别在数学、长上下文和视觉任务方面有显著提升。',
482
- displayName: 'Gemini 1.5 Pro 002 (Paid)',
506
+ displayName: 'Gemini 1.5 Pro 002',
483
507
  id: 'gemini-1.5-pro-002', // Deprecated on 2025-09-24
484
508
  maxOutput: 8192,
485
509
  pricing: {
@@ -73,24 +73,6 @@ const groqChatModels: AIChatModelCard[] = [
73
73
  displayName: 'Kimi K2 0905',
74
74
  enabled: true,
75
75
  id: 'moonshotai/kimi-k2-instruct-0905',
76
- pricing: {
77
- units: [
78
- { name: 'textInput', rate: 1, strategy: 'fixed', unit: 'millionTokens' },
79
- { name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
80
- ],
81
- },
82
- releasedAt: '2025-09-05',
83
- type: 'chat',
84
- },
85
- {
86
- abilities: {
87
- functionCall: true,
88
- },
89
- contextWindowTokens: 131_072,
90
- description:
91
- 'kimi-k2 是一款具备超强代码和 Agent 能力的 MoE 架构基础模型,总参数 1T,激活参数 32B。在通用知识推理、编程、数学、Agent 等主要类别的基准性能测试中,K2 模型的性能超过其他主流开源模型。',
92
- displayName: 'Kimi K2 0711',
93
- id: 'moonshotai/kimi-k2-instruct',
94
76
  maxOutput: 16_384,
95
77
  pricing: {
96
78
  units: [
@@ -98,7 +80,7 @@ const groqChatModels: AIChatModelCard[] = [
98
80
  { name: 'textOutput', rate: 3, strategy: 'fixed', unit: 'millionTokens' },
99
81
  ],
100
82
  },
101
- releasedAt: '2025-07-11',
83
+ releasedAt: '2025-09-05',
102
84
  type: 'chat',
103
85
  },
104
86
  {
@@ -4,14 +4,20 @@ const modelscopeChatModels: AIChatModelCard[] = [
4
4
  {
5
5
  abilities: {
6
6
  functionCall: true,
7
+ reasoning: true,
7
8
  },
8
- contextWindowTokens: 262_144,
9
- description:
10
- 'kimi-k2-0905-preview 模型上下文长度为 256k,具备更强的 Agentic Coding 能力、更突出的前端代码的美观度和实用性、以及更好的上下文理解能力。',
11
- displayName: 'Kimi K2 0905',
12
- enabled: true,
13
- id: 'moonshotai/Kimi-K2-Instruct-0905',
14
- releasedAt: '2025-09-05',
9
+ contextWindowTokens: 131_072,
10
+ displayName: 'Qwen3 Next 80B A3B Thinking',
11
+ id: 'Qwen/Qwen3-Next-80B-A3B-Thinking',
12
+ type: 'chat',
13
+ },
14
+ {
15
+ abilities: {
16
+ functionCall: true,
17
+ },
18
+ contextWindowTokens: 131_072,
19
+ displayName: 'Qwen3 Next 80B A3B Instruct',
20
+ id: 'Qwen/Qwen3-Next-80B-A3B-Instruct',
15
21
  type: 'chat',
16
22
  },
17
23
  {
@@ -20,8 +26,8 @@ const modelscopeChatModels: AIChatModelCard[] = [
20
26
  reasoning: true,
21
27
  },
22
28
  contextWindowTokens: 131_072,
23
- description: 'DeepSeek-V3.1 模型为混合推理架构模型,同时支持思考模式与非思考模式。',
24
- displayName: 'DeepSeek-V3.1',
29
+ description: 'DeepSeek V3.1 模型为混合推理架构模型,同时支持思考模式与非思考模式。',
30
+ displayName: 'DeepSeek V3.1',
25
31
  enabled: true,
26
32
  id: 'deepseek-ai/DeepSeek-V3.1',
27
33
  type: 'chat',
@@ -34,7 +40,7 @@ const modelscopeChatModels: AIChatModelCard[] = [
34
40
  contextWindowTokens: 131_072,
35
41
  description:
36
42
  'DeepSeek R1 通过利用增加的计算资源和在后训练过程中引入算法优化机制,显著提高了其推理和推断能力的深度。该模型在各种基准评估中表现出色,包括数学、编程和一般逻辑方面。其整体性能现已接近领先模型,如 O3 和 Gemini 2.5 Pro。',
37
- displayName: 'DeepSeek-R1-0528',
43
+ displayName: 'DeepSeek R1 0528',
38
44
  id: 'deepseek-ai/DeepSeek-R1-0528',
39
45
  type: 'chat',
40
46
  },
@@ -44,8 +50,8 @@ const modelscopeChatModels: AIChatModelCard[] = [
44
50
  reasoning: true,
45
51
  },
46
52
  contextWindowTokens: 131_072,
47
- description: 'DeepSeek-R1是DeepSeek最新的推理模型,专注于复杂推理任务。',
48
- displayName: 'DeepSeek-R1',
53
+ description: 'DeepSeek R1是DeepSeek最新的推理模型,专注于复杂推理任务。',
54
+ displayName: 'DeepSeek R1',
49
55
  id: 'deepseek-ai/DeepSeek-R1',
50
56
  type: 'chat',
51
57
  },
@@ -54,8 +60,8 @@ const modelscopeChatModels: AIChatModelCard[] = [
54
60
  functionCall: true,
55
61
  },
56
62
  contextWindowTokens: 131_072,
57
- description: 'DeepSeek-V3是DeepSeek第三代模型的最新版本,具有强大的推理和对话能力。',
58
- displayName: 'DeepSeek-V3',
63
+ description: 'DeepSeek V3是DeepSeek第三代模型的最新版本,具有强大的推理和对话能力。',
64
+ displayName: 'DeepSeek V3',
59
65
  id: 'deepseek-ai/DeepSeek-V3',
60
66
  type: 'chat',
61
67
  },
@@ -64,8 +70,8 @@ const modelscopeChatModels: AIChatModelCard[] = [
64
70
  functionCall: true,
65
71
  },
66
72
  contextWindowTokens: 131_072,
67
- description: 'Qwen3-235B-A22B是通义千问3代超大规模模型,提供顶级的AI能力。',
68
- displayName: 'Qwen3-235B-A22B',
73
+ description: 'Qwen3 235B A22B是通义千问3代超大规模模型,提供顶级的AI能力。',
74
+ displayName: 'Qwen3 235B A22B',
69
75
  id: 'Qwen/Qwen3-235B-A22B',
70
76
  type: 'chat',
71
77
  },
@@ -74,8 +80,8 @@ const modelscopeChatModels: AIChatModelCard[] = [
74
80
  functionCall: true,
75
81
  },
76
82
  contextWindowTokens: 131_072,
77
- description: 'Qwen3-32B是通义千问3代模型,具有强大的推理和对话能力。',
78
- displayName: 'Qwen3-32B',
83
+ description: 'Qwen3 32B是通义千问3代模型,具有强大的推理和对话能力。',
84
+ displayName: 'Qwen3 32B',
79
85
  id: 'Qwen/Qwen3-32B',
80
86
  type: 'chat',
81
87
  },
@@ -2,6 +2,52 @@ import { AIChatModelCard } from '../types/aiModel';
2
2
 
3
3
  // https://novita.ai/pricing
4
4
  const novitaChatModels: AIChatModelCard[] = [
5
+ {
6
+ abilities: {
7
+ functionCall: true,
8
+ reasoning: true,
9
+ },
10
+ contextWindowTokens: 65_536,
11
+ displayName: 'Qwen3 Next 80B A3B Thinking',
12
+ id: 'qwen/qwen3-next-80b-a3b-thinking',
13
+ maxOutput: 65_536,
14
+ pricing: {
15
+ units: [
16
+ { name: 'textInput', rate: 0.15, strategy: 'fixed', unit: 'millionTokens' },
17
+ { name: 'textOutput', rate: 1.5, strategy: 'fixed', unit: 'millionTokens' },
18
+ ],
19
+ },
20
+ type: 'chat',
21
+ },
22
+ {
23
+ abilities: {
24
+ functionCall: true,
25
+ },
26
+ contextWindowTokens: 65_536,
27
+ displayName: 'Qwen3 Next 80B A3B Instruct',
28
+ id: 'qwen/qwen3-next-80b-a3b-instruct',
29
+ maxOutput: 65_536,
30
+ pricing: {
31
+ units: [
32
+ { name: 'textInput', rate: 0.15, strategy: 'fixed', unit: 'millionTokens' },
33
+ { name: 'textOutput', rate: 1.5, strategy: 'fixed', unit: 'millionTokens' },
34
+ ],
35
+ },
36
+ type: 'chat',
37
+ },
38
+ {
39
+ contextWindowTokens: 4096,
40
+ displayName: 'Qwen MT Plus',
41
+ id: 'qwen/qwen-mt-plus',
42
+ maxOutput: 2048,
43
+ pricing: {
44
+ units: [
45
+ { name: 'textInput', rate: 0.25, strategy: 'fixed', unit: 'millionTokens' },
46
+ { name: 'textOutput', rate: 0.75, strategy: 'fixed', unit: 'millionTokens' },
47
+ ],
48
+ },
49
+ type: 'chat',
50
+ },
5
51
  {
6
52
  abilities: {
7
53
  functionCall: true,
@@ -41,13 +87,13 @@ const novitaChatModels: AIChatModelCard[] = [
41
87
  abilities: {
42
88
  functionCall: true,
43
89
  },
44
- contextWindowTokens: 262_144,
90
+ contextWindowTokens: 65_536,
45
91
  displayName: 'Qwen3 Coder 480B A35B Instruct',
46
92
  id: 'qwen/qwen3-coder-480b-a35b-instruct',
47
93
  pricing: {
48
94
  units: [
49
- { name: 'textInput', rate: 0.35, strategy: 'fixed', unit: 'millionTokens' },
50
- { name: 'textOutput', rate: 1.5, strategy: 'fixed', unit: 'millionTokens' },
95
+ { name: 'textInput', rate: 0.29, strategy: 'fixed', unit: 'millionTokens' },
96
+ { name: 'textOutput', rate: 1.2, strategy: 'fixed', unit: 'millionTokens' },
51
97
  ],
52
98
  },
53
99
  type: 'chat',
@@ -83,6 +129,24 @@ const novitaChatModels: AIChatModelCard[] = [
83
129
  },
84
130
  type: 'chat',
85
131
  },
132
+ {
133
+ abilities: {
134
+ functionCall: true,
135
+ reasoning: true,
136
+ vision: true,
137
+ },
138
+ contextWindowTokens: 65_536,
139
+ displayName: 'GLM-4.5V',
140
+ id: 'zai-org/glm-4.5v',
141
+ maxOutput: 16_384,
142
+ pricing: {
143
+ units: [
144
+ { name: 'textInput', rate: 0.6, strategy: 'fixed', unit: 'millionTokens' },
145
+ { name: 'textOutput', rate: 1.8, strategy: 'fixed', unit: 'millionTokens' },
146
+ ],
147
+ },
148
+ type: 'chat',
149
+ },
86
150
  {
87
151
  abilities: {
88
152
  functionCall: true,
@@ -91,6 +155,7 @@ const novitaChatModels: AIChatModelCard[] = [
91
155
  contextWindowTokens: 131_072,
92
156
  displayName: 'GLM-4.5',
93
157
  id: 'zai-org/glm-4.5',
158
+ maxOutput: 98_304,
94
159
  pricing: {
95
160
  units: [
96
161
  { name: 'textInput', rate: 0.6, strategy: 'fixed', unit: 'millionTokens' },
@@ -103,9 +168,10 @@ const novitaChatModels: AIChatModelCard[] = [
103
168
  abilities: {
104
169
  functionCall: true,
105
170
  },
106
- contextWindowTokens: 262_144,
171
+ contextWindowTokens: 131_072,
107
172
  displayName: 'Qwen3 235B A22B Instruct 2507',
108
173
  id: 'qwen/qwen3-235b-a22b-instruct-2507',
174
+ maxOutput: 16_384,
109
175
  pricing: {
110
176
  units: [
111
177
  { name: 'textInput', rate: 0.15, strategy: 'fixed', unit: 'millionTokens' },
@@ -717,6 +783,7 @@ const novitaChatModels: AIChatModelCard[] = [
717
783
  contextWindowTokens: 32_000,
718
784
  displayName: 'Qwen2.5 72B Instruct',
719
785
  id: 'qwen/qwen-2.5-72b-instruct',
786
+ maxOutput: 8192,
720
787
  pricing: {
721
788
  units: [
722
789
  { name: 'textInput', rate: 0.38, strategy: 'fixed', unit: 'millionTokens' },
@@ -770,7 +837,6 @@ const novitaChatModels: AIChatModelCard[] = [
770
837
  },
771
838
  contextWindowTokens: 32_768,
772
839
  displayName: 'Qwen2.5 VL 72B Instruct',
773
- enabled: true,
774
840
  id: 'qwen/qwen2.5-vl-72b-instruct',
775
841
  pricing: {
776
842
  units: [
@@ -48,6 +48,92 @@ const qwenChatModels: AIChatModelCard[] = [
48
48
  },
49
49
  type: 'chat',
50
50
  },
51
+ {
52
+ abilities: {
53
+ reasoning: true,
54
+ },
55
+ contextWindowTokens: 131_072,
56
+ description: 'GLM-4.5系列模型是智谱AI专为智能体设计的混合推理模型,提供思考与非思考两种模式。',
57
+ displayName: 'GLM-4.5',
58
+ id: 'glm-4.5',
59
+ maxOutput: 16_384,
60
+ pricing: {
61
+ currency: 'CNY',
62
+ units: [
63
+ {
64
+ lookup: {
65
+ prices: {
66
+ '[0, 32_000]': 3,
67
+ '[32_000, infinity]': 4,
68
+ },
69
+ pricingParams: ['textInputRange'],
70
+ },
71
+ name: 'textInput',
72
+ strategy: 'lookup',
73
+ unit: 'millionTokens',
74
+ },
75
+ {
76
+ lookup: {
77
+ prices: {
78
+ '[0, 32_000]': 14,
79
+ '[32_000, infinity]': 16,
80
+ },
81
+ pricingParams: ['textInputRange'],
82
+ },
83
+ name: 'textOutput',
84
+ strategy: 'lookup',
85
+ unit: 'millionTokens',
86
+ },
87
+ ],
88
+ },
89
+ settings: {
90
+ extendParams: ['enableReasoning', 'reasoningBudgetToken'],
91
+ },
92
+ type: 'chat',
93
+ },
94
+ {
95
+ abilities: {
96
+ reasoning: true,
97
+ },
98
+ contextWindowTokens: 131_072,
99
+ description: 'GLM-4.5系列模型是智谱AI专为智能体设计的混合推理模型,提供思考与非思考两种模式。',
100
+ displayName: 'GLM-4.5-Air',
101
+ id: 'glm-4.5-air',
102
+ maxOutput: 16_384,
103
+ pricing: {
104
+ currency: 'CNY',
105
+ units: [
106
+ {
107
+ lookup: {
108
+ prices: {
109
+ '[0, 32_000]': 0.8,
110
+ '[32_000, infinity]': 1.2,
111
+ },
112
+ pricingParams: ['textInputRange'],
113
+ },
114
+ name: 'textInput',
115
+ strategy: 'lookup',
116
+ unit: 'millionTokens',
117
+ },
118
+ {
119
+ lookup: {
120
+ prices: {
121
+ '[0, 32_000]': 6,
122
+ '[32_000, infinity]': 8,
123
+ },
124
+ pricingParams: ['textInputRange'],
125
+ },
126
+ name: 'textOutput',
127
+ strategy: 'lookup',
128
+ unit: 'millionTokens',
129
+ },
130
+ ],
131
+ },
132
+ settings: {
133
+ extendParams: ['enableReasoning', 'reasoningBudgetToken'],
134
+ },
135
+ type: 'chat',
136
+ },
51
137
  {
52
138
  abilities: {
53
139
  functionCall: true,
@@ -583,7 +669,7 @@ const qwenChatModels: AIChatModelCard[] = [
583
669
  search: true,
584
670
  },
585
671
  config: {
586
- deploymentName: 'qwen-plus-2025-07-28',
672
+ deploymentName: 'qwen-plus-2025-09-11',
587
673
  },
588
674
  contextWindowTokens: 1_000_000,
589
675
  description: '通义千问超大规模语言模型增强版,支持中文、英文等不同语言输入。',
@@ -627,7 +713,6 @@ const qwenChatModels: AIChatModelCard[] = [
627
713
  '[0, 128_000]_[false]': 2,
628
714
  '[0, 128_000]_[true]': 8,
629
715
  '[128_000, 256_000]_[false]': 20,
630
-
631
716
  '[128_000, 256_000]_[true]': 24,
632
717
  '[256_000, infinity]_[false]': 48,
633
718
  '[256_000, infinity]_[true]': 64,
@@ -47,6 +47,25 @@ const siliconcloudChatModels: AIChatModelCard[] = [
47
47
  },
48
48
  type: 'chat',
49
49
  },
50
+ {
51
+ abilities: {
52
+ functionCall: true,
53
+ },
54
+ contextWindowTokens: 131_072,
55
+ description:
56
+ 'Ling-mini-2.0 是一款基于 MoE 架构的小尺寸高性能大语言模型。它拥有 16B 总参数,但每个 token 仅激活 1.4B(non-embedding 789M),从而实现了极高的生成速度。得益于高效的 MoE 设计与大规模高质量训练数据,尽管激活参数仅为 1.4B,Ling-mini-2.0 依然在下游任务中展现出可媲美 10B 以下 dense LLM 及更大规模 MoE 模型的顶尖性能',
57
+ displayName: 'Ling Mini 2.0',
58
+ id: 'inclusionAI/Ling-mini-2.0',
59
+ pricing: {
60
+ currency: 'CNY',
61
+ units: [
62
+ { name: 'textInput', rate: 0.5, strategy: 'fixed', unit: 'millionTokens' },
63
+ { name: 'textOutput', rate: 2, strategy: 'fixed', unit: 'millionTokens' },
64
+ ],
65
+ },
66
+ releasedAt: '2025-09-09',
67
+ type: 'chat',
68
+ },
50
69
  {
51
70
  abilities: {
52
71
  functionCall: true,
@@ -223,10 +242,51 @@ const siliconcloudChatModels: AIChatModelCard[] = [
223
242
  type: 'chat',
224
243
  },
225
244
  {
245
+ abilities: {
246
+ functionCall: true,
247
+ },
248
+ contextWindowTokens: 262_144,
249
+ description:
250
+ 'Kimi K2-Instruct-0905 是 Kimi K2 最新、最强大的版本。它是一款顶尖的混合专家(MoE)语言模型,拥有 1 万亿的总参数和 320 亿的激活参数。该模型的主要特性包括:增强的智能体编码智能,在公开基准测试和真实世界的编码智能体任务中表现出显著的性能提升;改进的前端编码体验,在前端编程的美观性和实用性方面均有进步。',
251
+ displayName: 'Kimi K2 0905',
252
+ id: 'moonshotai/Kimi-K2-Instruct-0905',
253
+ pricing: {
254
+ currency: 'CNY',
255
+ units: [
256
+ { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
257
+ { name: 'textOutput', rate: 16, strategy: 'fixed', unit: 'millionTokens' },
258
+ ],
259
+ },
260
+ releasedAt: '2025-09-05',
261
+ type: 'chat',
262
+ },
263
+ {
264
+ abilities: {
265
+ functionCall: true,
266
+ },
267
+ contextWindowTokens: 262_144,
268
+ description:
269
+ 'Kimi K2-Instruct-0905 是 Kimi K2 最新、最强大的版本。它是一款顶尖的混合专家(MoE)语言模型,拥有 1 万亿的总参数和 320 亿的激活参数。该模型的主要特性包括:增强的智能体编码智能,在公开基准测试和真实世界的编码智能体任务中表现出显著的性能提升;改进的前端编码体验,在前端编程的美观性和实用性方面均有进步。',
270
+ displayName: 'Kimi K2 0905 (Pro)',
271
+ id: 'Pro/moonshotai/Kimi-K2-Instruct-0905',
272
+ pricing: {
273
+ currency: 'CNY',
274
+ units: [
275
+ { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
276
+ { name: 'textOutput', rate: 16, strategy: 'fixed', unit: 'millionTokens' },
277
+ ],
278
+ },
279
+ releasedAt: '2025-09-05',
280
+ type: 'chat',
281
+ },
282
+ {
283
+ abilities: {
284
+ functionCall: true,
285
+ },
226
286
  contextWindowTokens: 131_072,
227
287
  description:
228
288
  'Kimi K2 是一款具备超强代码和 Agent 能力的 MoE 架构基础模型,总参数 1T,激活参数 32B。在通用知识推理、编程、数学、Agent 等主要类别的基准性能测试中,K2 模型的性能超过其他主流开源模型。',
229
- displayName: 'Kimi K2 Instruct',
289
+ displayName: 'Kimi K2 0711',
230
290
  id: 'moonshotai/Kimi-K2-Instruct',
231
291
  pricing: {
232
292
  currency: 'CNY',
@@ -239,10 +299,13 @@ const siliconcloudChatModels: AIChatModelCard[] = [
239
299
  type: 'chat',
240
300
  },
241
301
  {
302
+ abilities: {
303
+ functionCall: true,
304
+ },
242
305
  contextWindowTokens: 131_072,
243
306
  description:
244
307
  'Kimi K2 是一款具备超强代码和 Agent 能力的 MoE 架构基础模型,总参数 1T,激活参数 32B。在通用知识推理、编程、数学、Agent 等主要类别的基准性能测试中,K2 模型的性能超过其他主流开源模型。',
245
- displayName: 'Kimi K2 Instruct (Pro)',
308
+ displayName: 'Kimi K2 0711 (Pro)',
246
309
  id: 'Pro/moonshotai/Kimi-K2-Instruct',
247
310
  pricing: {
248
311
  currency: 'CNY',
@@ -35,17 +35,18 @@ const doubaoChatModels: AIChatModelCard[] = [
35
35
  functionCall: true,
36
36
  },
37
37
  config: {
38
- deploymentName: 'kimi-k2-250711',
38
+ deploymentName: 'kimi-k2-250905',
39
39
  },
40
- contextWindowTokens: 256_000,
40
+ contextWindowTokens: 131_072,
41
41
  description:
42
42
  'Kimi-K2 是一款Moonshot AI推出的具备超强代码和 Agent 能力的 MoE 架构基础模型,总参数 1T,激活参数 32B。在通用知识推理、编程、数学、Agent 等主要类别的基准性能测试中,K2 模型的性能超过其他主流开源模型。',
43
43
  displayName: 'Kimi K2',
44
44
  id: 'kimi-k2',
45
- maxOutput: 16_384,
45
+ maxOutput: 32_768,
46
46
  pricing: {
47
47
  currency: 'CNY',
48
48
  units: [
49
+ { name: 'textInput_cacheRead', rate: 0.8, strategy: 'fixed', unit: 'millionTokens' },
49
50
  { name: 'textInput', rate: 4, strategy: 'fixed', unit: 'millionTokens' },
50
51
  { name: 'textOutput', rate: 16, strategy: 'fixed', unit: 'millionTokens' },
51
52
  ],
@@ -53,7 +53,7 @@ export const MODEL_LIST_CONFIGS = {
53
53
  'qwen2.5',
54
54
  'qwen3',
55
55
  ],
56
- reasoningKeywords: ['qvq', 'qwq', 'qwen3', '!-instruct-', '!-coder-'],
56
+ reasoningKeywords: ['qvq', 'qwq', 'qwen3', '!-instruct-', '!-coder-', '!-max-'],
57
57
  visionKeywords: ['qvq', 'vl'],
58
58
  },
59
59
  v0: {
@@ -303,9 +303,9 @@ const processModelCard = (
303
303
  )
304
304
  ? 'image'
305
305
  : isKeywordListMatch(
306
- model.id.toLowerCase(),
307
- EMBEDDING_MODEL_KEYWORDS.map((k) => k.toLowerCase()),
308
- )
306
+ model.id.toLowerCase(),
307
+ EMBEDDING_MODEL_KEYWORDS.map((k) => k.toLowerCase()),
308
+ )
309
309
  ? 'embedding'
310
310
  : 'chat');
311
311