@lobehub/chat 1.58.0 → 1.59.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +25 -0
- package/changelog/v1.json +9 -0
- package/locales/ar/providers.json +3 -0
- package/locales/bg-BG/providers.json +3 -0
- package/locales/de-DE/providers.json +3 -0
- package/locales/en-US/providers.json +3 -0
- package/locales/es-ES/providers.json +3 -0
- package/locales/fa-IR/modelProvider.json +5 -11
- package/locales/fa-IR/providers.json +3 -0
- package/locales/fr-FR/providers.json +3 -0
- package/locales/it-IT/providers.json +3 -0
- package/locales/ja-JP/providers.json +3 -0
- package/locales/ko-KR/providers.json +3 -0
- package/locales/nl-NL/providers.json +3 -0
- package/locales/pl-PL/providers.json +3 -0
- package/locales/pt-BR/providers.json +3 -0
- package/locales/ru-RU/providers.json +3 -0
- package/locales/tr-TR/modelProvider.json +5 -11
- package/locales/tr-TR/providers.json +3 -0
- package/locales/vi-VN/providers.json +3 -0
- package/locales/zh-CN/models.json +6 -6
- package/locales/zh-CN/providers.json +3 -0
- package/locales/zh-TW/providers.json +3 -0
- package/next.config.ts +5 -0
- package/package.json +1 -1
- package/src/app/[variants]/(main)/settings/provider/(list)/ProviderGrid/index.tsx +6 -3
- package/src/config/aiModels/index.ts +3 -0
- package/src/config/aiModels/volcengine.ts +83 -0
- package/src/config/llm.ts +6 -0
- package/src/config/modelProviders/index.ts +3 -0
- package/src/config/modelProviders/volcengine.ts +23 -0
- package/src/libs/agent-runtime/AgentRuntime.ts +4 -2
- package/src/libs/agent-runtime/index.ts +1 -1
- package/src/libs/agent-runtime/types/type.ts +4 -0
- package/src/libs/agent-runtime/{doubao → volcengine}/index.ts +3 -3
- package/src/services/chat.ts +7 -4
- package/src/types/aiProvider.ts +4 -0
- package/src/types/user/settings/keyVaults.ts +1 -0
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,31 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
## [Version 1.59.0](https://github.com/lobehub/lobe-chat/compare/v1.58.0...v1.59.0)
|
6
|
+
|
7
|
+
<sup>Released on **2025-02-16**</sup>
|
8
|
+
|
9
|
+
#### ✨ Features
|
10
|
+
|
11
|
+
- **misc**: Add volcengine as a new provider.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### What's improved
|
19
|
+
|
20
|
+
- **misc**: Add volcengine as a new provider, closes [#6221](https://github.com/lobehub/lobe-chat/issues/6221) ([09bf8f0](https://github.com/lobehub/lobe-chat/commit/09bf8f0))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
5
30
|
## [Version 1.58.0](https://github.com/lobehub/lobe-chat/compare/v1.57.1...v1.58.0)
|
6
31
|
|
7
32
|
<sup>Released on **2025-02-16**</sup>
|
package/changelog/v1.json
CHANGED
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM هو مكتبة سريعة وسهلة الاستخدام لاستدلال LLM والخدمات."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "منصة تطوير خدمات النماذج الكبيرة التي أطلقتها بايت دانس، تقدم خدمات استدعاء نماذج غنية بالوظائف وآمنة وتنافسية من حيث الأسعار، كما توفر بيانات النماذج، والتعديل الدقيق، والاستدلال، والتقييم، وغيرها من الوظائف الشاملة، لضمان تطوير تطبيقات الذكاء الاصطناعي الخاصة بك بشكل كامل."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "منصة تطوير وخدمات النماذج الكبيرة والتطبيقات الأصلية للذكاء الاصطناعي على مستوى المؤسسات، تقدم مجموعة شاملة وسهلة الاستخدام من أدوات تطوير النماذج الذكية التوليدية وأدوات تطوير التطبيقات على مدار العملية بأكملها."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM е бърза и лесна за използване библиотека за LLM инференция и услуги."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Платформа за разработка на услуги с големи модели, пусната от ByteDance, предлагаща богати на функции, безопасни и конкурентни по цена услуги за извикване на модели. Освен това предоставя край до край функции като данни за модели, фина настройка, инференция и оценка, за да осигури всестранна подкрепа за разработката на вашите AI приложения."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Платформа за разработка и услуги на корпоративно ниво, предлагаща цялостно решение за разработка на генеративни модели на изкуствен интелект и приложения, с най-пълния и лесен за използване инструментариум за целия процес на разработка на модели и приложения."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM ist eine schnelle und benutzerfreundliche Bibliothek für LLM-Inferenz und -Dienste."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Die von ByteDance eingeführte Entwicklungsplattform für große Modellservices bietet funktionsreiche, sichere und preislich wettbewerbsfähige Modellaufrufdienste. Sie bietet zudem End-to-End-Funktionen wie Moduldaten, Feinabstimmung, Inferenz und Bewertung, um die Entwicklung Ihrer KI-Anwendungen umfassend zu unterstützen."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Eine unternehmensweite, umfassende Plattform für die Entwicklung und den Service von großen Modellen und KI-nativen Anwendungen, die die vollständigsten und benutzerfreundlichsten Werkzeuge für die Entwicklung generativer KI-Modelle und den gesamten Anwendungsentwicklungsprozess bietet."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM is a fast and easy-to-use library for LLM inference and serving."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "A development platform for large model services launched by ByteDance, offering feature-rich, secure, and competitively priced model invocation services. It also provides end-to-end functionalities such as model data, fine-tuning, inference, and evaluation, ensuring comprehensive support for the development and implementation of your AI applications."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "An enterprise-level one-stop platform for large model and AI-native application development and services, providing the most comprehensive and user-friendly toolchain for the entire process of generative artificial intelligence model development and application development."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM es una biblioteca rápida y fácil de usar para la inferencia y el servicio de LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Plataforma de desarrollo de servicios de modelos grandes lanzada por ByteDance, que ofrece servicios de invocación de modelos ricos en funciones, seguros y competitivos en precio, al tiempo que proporciona datos de modelos, ajuste fino, inferencia, evaluación y otras funciones de extremo a extremo, garantizando de manera integral el desarrollo y la implementación de sus aplicaciones de IA."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Plataforma de desarrollo y servicios de modelos grandes y aplicaciones nativas de IA de nivel empresarial, que ofrece la cadena de herramientas más completa y fácil de usar para el desarrollo de modelos de inteligencia artificial generativa y el desarrollo de aplicaciones en todo el proceso."
|
124
127
|
},
|
@@ -110,17 +110,6 @@
|
|
110
110
|
"required": "لطفاً نام ارائهدهنده را وارد کنید",
|
111
111
|
"title": "نام ارائهدهنده"
|
112
112
|
},
|
113
|
-
"providerModels": {
|
114
|
-
"item": {
|
115
|
-
"modelConfig": {
|
116
|
-
"deployName": {
|
117
|
-
"extra": "این فیلد به عنوان شناسه مدل هنگام ارسال درخواست استفاده میشود",
|
118
|
-
"placeholder": "لطفاً نام یا شناسه واقعی استقرار مدل را وارد کنید",
|
119
|
-
"title": "نام استقرار مدل"
|
120
|
-
}
|
121
|
-
}
|
122
|
-
}
|
123
|
-
},
|
124
113
|
"proxyUrl": {
|
125
114
|
"required": "لطفاً آدرس پروکسی را وارد کنید",
|
126
115
|
"title": "آدرس پروکسی"
|
@@ -240,6 +229,11 @@
|
|
240
229
|
"placeholder": "لطفاً نام استقرار مدل در Azure را وارد کنید",
|
241
230
|
"title": "نام استقرار مدل"
|
242
231
|
},
|
232
|
+
"deployName": {
|
233
|
+
"extra": "این فیلد به عنوان شناسه مدل هنگام ارسال درخواست استفاده میشود",
|
234
|
+
"placeholder": "لطفاً نام یا شناسه واقعی مدل را وارد کنید",
|
235
|
+
"title": "نام مدل برای استقرار"
|
236
|
+
},
|
243
237
|
"displayName": {
|
244
238
|
"placeholder": "لطفاً نام نمایشی مدل را وارد کنید، مانند ChatGPT، GPT-4 و غیره",
|
245
239
|
"title": "نام نمایشی مدل"
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM یک کتابخانه سریع و آسان برای استفاده است که برای استنتاج و خدمات LLM طراحی شده است."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "پلتفرم توسعه خدمات مدلهای بزرگ که توسط بایتدANCE راهاندازی شده است، خدمات فراوان، ایمن و با قیمت رقابتی برای فراخوانی مدلها را ارائه میدهد. همچنین امکاناتی از جمله دادههای مدل، تنظیم دقیق، استنتاج و ارزیابی را به صورت end-to-end فراهم میکند و به طور جامع از توسعه و پیادهسازی برنامههای هوش مصنوعی شما حمایت میکند."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "پلتفرم جامع توسعه و خدمات مدلهای بزرگ و برنامههای بومی هوش مصنوعی در سطح سازمانی، ارائهدهنده کاملترین و کاربرپسندترین زنجیره ابزارهای توسعه مدلهای هوش مصنوعی مولد و توسعه برنامهها"
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM est une bibliothèque rapide et facile à utiliser pour l'inférence et les services LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "La plateforme de développement des services de grands modèles lancée par ByteDance, offrant des services d'appel de modèles riches en fonctionnalités, sécurisés et compétitifs en termes de prix. Elle propose également des fonctionnalités de bout en bout telles que les données de modèle, le réglage fin, l'inférence et l'évaluation, garantissant ainsi le succès de votre développement d'applications AI."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Plateforme de développement et de services d'applications AI natives et de modèles de grande envergure, tout-en-un pour les entreprises, offrant la chaîne d'outils la plus complète et facile à utiliser pour le développement de modèles d'intelligence artificielle générative et le développement d'applications."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM è una libreria veloce e facile da usare per l'inferenza e i servizi LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "La piattaforma di sviluppo dei servizi di modelli di grandi dimensioni lanciata da ByteDance, offre servizi di invocazione di modelli ricchi di funzionalità, sicuri e competitivi in termini di prezzo, fornendo anche dati sui modelli, messa a punto, inferenza, valutazione e altre funzionalità end-to-end, garantendo in modo completo lo sviluppo e l'implementazione delle vostre applicazioni AI."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Piattaforma di sviluppo e servizi per modelli di grandi dimensioni e applicazioni AI native, a livello aziendale, che offre la catena di strumenti completa e facile da usare per lo sviluppo di modelli di intelligenza artificiale generativa e per l'intero processo di sviluppo delle applicazioni."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLMは、LLM推論とサービスのための迅速で使いやすいライブラリです。"
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "バイトダンスが提供する大規模モデルサービスの開発プラットフォームで、機能が豊富で安全性が高く、価格競争力のあるモデル呼び出しサービスを提供します。また、モデルデータ、ファインチューニング、推論、評価などのエンドツーエンド機能を提供し、AIアプリケーションの開発を全面的にサポートします。"
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "企業向けのワンストップ大規模モデルとAIネイティブアプリケーションの開発およびサービスプラットフォームで、最も包括的で使いやすい生成的人工知能モデルの開発とアプリケーション開発の全プロセスツールチェーンを提供します。"
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM은 LLM 추론 및 서비스를 위한 빠르고 사용하기 쉬운 라이브러리입니다."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "바이트댄스가 출시한 대형 모델 서비스 개발 플랫폼으로, 기능이 풍부하고 안전하며 가격 경쟁력이 있는 모델 호출 서비스를 제공합니다. 또한 모델 데이터, 세밀 조정, 추론, 평가 등 엔드 투 엔드 기능을 제공하여 귀하의 AI 애플리케이션 개발을 전방위적으로 지원합니다."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "기업용 원스톱 대형 모델 및 AI 네이티브 애플리케이션 개발 및 서비스 플랫폼으로, 가장 포괄적이고 사용하기 쉬운 생성적 인공지능 모델 개발 및 애플리케이션 개발 전체 프로세스 도구 체인을 제공합니다."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM is een snelle en gebruiksvriendelijke bibliotheek voor LLM-inferentie en -diensten."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Het ontwikkelingsplatform voor de grote modellenservice van ByteDance, dat een breed scala aan functies biedt, veilig is en concurrerende prijzen heeft voor modelaanroepdiensten. Het biedt ook end-to-end functionaliteiten zoals modelgegevens, fine-tuning, inferentie en evaluatie, om de ontwikkeling van uw AI-toepassingen volledig te ondersteunen."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Een enterprise-grade, alles-in-één platform voor de ontwikkeling en service van grote modellen en AI-native applicaties, dat de meest uitgebreide en gebruiksvriendelijke toolchain biedt voor de ontwikkeling van generatieve kunstmatige intelligentiemodellen en applicaties."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM to szybka i łatwa w użyciu biblioteka do wnioskowania i usług LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Platforma deweloperska usług dużych modeli wprowadzona przez ByteDance, oferująca bogate w funkcje, bezpieczne i konkurencyjne cenowo usługi wywoływania modeli. Oferuje również dane modelowe, dostosowywanie, wnioskowanie, ocenę i inne funkcje end-to-end, zapewniając kompleksowe wsparcie dla rozwoju aplikacji AI."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Platforma do rozwoju i usług aplikacji AI oraz dużych modeli w skali przedsiębiorstwa, oferująca najbardziej kompleksowy i łatwy w użyciu zestaw narzędzi do rozwoju modeli sztucznej inteligencji generatywnej oraz całego procesu tworzenia aplikacji."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM é uma biblioteca rápida e fácil de usar para inferência e serviços de LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "A plataforma de desenvolvimento de serviços de grandes modelos lançada pela ByteDance, que oferece serviços de chamada de modelos ricos em funcionalidades, seguros e com preços competitivos, além de fornecer dados de modelos, ajuste fino, inferência, avaliação e outras funcionalidades de ponta a ponta, garantindo de forma abrangente a implementação do seu desenvolvimento de aplicações de IA."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Plataforma de desenvolvimento e serviços de aplicativos nativos de IA e modelos de grande escala, voltada para empresas, que oferece a mais completa e fácil ferramenta de cadeia de ferramentas para o desenvolvimento de modelos de inteligência artificial generativa e aplicativos."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM — это быстрая и простая в использовании библиотека для вывода и обслуживания LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Платформа разработки сервисов больших моделей, запущенная ByteDance, предлагает функционально богатые, безопасные и конкурентоспособные по цене услуги вызова моделей, а также предоставляет полные функции от данных моделей, тонкой настройки, вывода до оценки, обеспечивая всестороннюю поддержку разработки ваших AI приложений."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Корпоративная платформа для разработки и обслуживания крупных моделей и нативных приложений ИИ, предлагающая самый полный и удобный инструментарий для разработки генеративных моделей искусственного интеллекта и полного процесса разработки приложений."
|
124
127
|
},
|
@@ -110,17 +110,6 @@
|
|
110
110
|
"required": "Lütfen hizmet sağlayıcı adını girin",
|
111
111
|
"title": "Hizmet Sağlayıcı Adı"
|
112
112
|
},
|
113
|
-
"providerModels": {
|
114
|
-
"item": {
|
115
|
-
"modelConfig": {
|
116
|
-
"deployName": {
|
117
|
-
"extra": "İstek gönderildiğinde bu alan model ID'si olarak kullanılacaktır",
|
118
|
-
"placeholder": "Lütfen modelin gerçek dağıtım adını veya kimliğini girin",
|
119
|
-
"title": "Model Dağıtım Adı"
|
120
|
-
}
|
121
|
-
}
|
122
|
-
}
|
123
|
-
},
|
124
113
|
"proxyUrl": {
|
125
114
|
"required": "Lütfen proxy adresini girin",
|
126
115
|
"title": "Proxy Adresi"
|
@@ -240,6 +229,11 @@
|
|
240
229
|
"placeholder": "Lütfen Azure'daki model dağıtım adını girin",
|
241
230
|
"title": "Model Dağıtım Adı"
|
242
231
|
},
|
232
|
+
"deployName": {
|
233
|
+
"extra": "Bu alan, isteği gönderirken model kimliği olarak kullanılacaktır",
|
234
|
+
"placeholder": "Modelin gerçek dağıtım adını veya kimliğini girin",
|
235
|
+
"title": "Model Dağıtım Adı"
|
236
|
+
},
|
243
237
|
"displayName": {
|
244
238
|
"placeholder": "Lütfen modelin gösterim adını girin, örneğin ChatGPT, GPT-4 vb.",
|
245
239
|
"title": "Model Gösterim Adı"
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM, LLM çıkarımı ve hizmetleri için hızlı ve kullanımı kolay bir kütüphanedir."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "ByteDance tarafından sunulan büyük model hizmetleri geliştirme platformu, zengin özellikler, güvenlik ve rekabetçi fiyatlarla model çağırma hizmetleri sunar. Ayrıca model verileri, ince ayar, çıkarım, değerlendirme gibi uçtan uca işlevler sağlar ve AI uygulama geliştirme sürecinizi her yönüyle güvence altına alır."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Kurumsal düzeyde tek duraklı büyük model ve AI yerel uygulama geliştirme ve hizmet platformu, en kapsamlı ve kullanımı kolay üretken yapay zeka modeli geliştirme, uygulama geliştirme için tam süreç araç zinciri sunar."
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM là một thư viện nhanh chóng và dễ sử dụng cho suy diễn và dịch vụ LLM."
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "Nền tảng phát triển dịch vụ mô hình lớn do ByteDance phát triển, cung cấp dịch vụ gọi mô hình phong phú, an toàn và có giá cả cạnh tranh, đồng thời cung cấp dữ liệu mô hình, tinh chỉnh, suy diễn, đánh giá và các chức năng đầu cuối khác, đảm bảo toàn diện cho việc phát triển ứng dụng AI của bạn."
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "Nền tảng phát triển và dịch vụ ứng dụng AI gốc với mô hình lớn một cửa dành cho doanh nghiệp, cung cấp chuỗi công cụ toàn diện và dễ sử dụng cho phát triển mô hình trí tuệ nhân tạo sinh sinh và phát triển ứng dụng."
|
124
127
|
},
|
@@ -60,22 +60,22 @@
|
|
60
60
|
"description": "基于 Qwen2.5-Math-7B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。"
|
61
61
|
},
|
62
62
|
"Doubao-lite-128k": {
|
63
|
-
"description": "
|
63
|
+
"description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 128k 上下文窗口的推理和精调。"
|
64
64
|
},
|
65
65
|
"Doubao-lite-32k": {
|
66
|
-
"description": "
|
66
|
+
"description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 32k 上下文窗口的推理和精调。"
|
67
67
|
},
|
68
68
|
"Doubao-lite-4k": {
|
69
|
-
"description": "
|
69
|
+
"description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。"
|
70
70
|
},
|
71
71
|
"Doubao-pro-128k": {
|
72
|
-
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持128k上下文窗口的推理和精调。"
|
72
|
+
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 128k 上下文窗口的推理和精调。"
|
73
73
|
},
|
74
74
|
"Doubao-pro-32k": {
|
75
|
-
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持32k上下文窗口的推理和精调。"
|
75
|
+
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 32k 上下文窗口的推理和精调。"
|
76
76
|
},
|
77
77
|
"Doubao-pro-4k": {
|
78
|
-
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持4k上下文窗口的推理和精调。"
|
78
|
+
"description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 4k 上下文窗口的推理和精调。"
|
79
79
|
},
|
80
80
|
"ERNIE-3.5-128K": {
|
81
81
|
"description": "百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。"
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM 是一个快速且易于使用的库,用于 LLM 推理和服务。"
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "字节跳动推出的大模型服务的开发平台,提供功能丰富、安全以及具备价格竞争力的模型调用服务,同时提供模型数据、精调、推理、评测等端到端功能,全方位保障您的 AI 应用开发落地。"
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "企业级一站式大模型与AI原生应用开发及服务平台,提供最全面易用的生成式人工智能模型开发、应用开发全流程工具链"
|
124
127
|
},
|
@@ -119,6 +119,9 @@
|
|
119
119
|
"vllm": {
|
120
120
|
"description": "vLLM 是一個快速且易於使用的庫,用於 LLM 推理和服務。"
|
121
121
|
},
|
122
|
+
"volcengine": {
|
123
|
+
"description": "字節跳動推出的大模型服務的開發平台,提供功能豐富、安全以及具備價格競爭力的模型調用服務,同時提供模型數據、精調、推理、評測等端到端功能,全方位保障您的 AI 應用開發落地。"
|
124
|
+
},
|
122
125
|
"wenxin": {
|
123
126
|
"description": "企業級一站式大模型與AI原生應用開發及服務平台,提供最全面易用的生成式人工智慧模型開發、應用開發全流程工具鏈"
|
124
127
|
},
|
package/next.config.ts
CHANGED
@@ -170,6 +170,11 @@ const nextConfig: NextConfig = {
|
|
170
170
|
permanent: true,
|
171
171
|
source: '/welcome',
|
172
172
|
},
|
173
|
+
{
|
174
|
+
destination: '/settings/provider/volcengine',
|
175
|
+
permanent: true,
|
176
|
+
source: '/settings/provider/doubao',
|
177
|
+
},
|
173
178
|
// we need back /repos url in the further
|
174
179
|
{
|
175
180
|
destination: '/files',
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.
|
3
|
+
"version": "1.59.0",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -79,9 +79,12 @@ const List = memo(() => {
|
|
79
79
|
</Grid>
|
80
80
|
</Flexbox>
|
81
81
|
<Flexbox gap={12}>
|
82
|
-
<
|
83
|
-
{
|
84
|
-
|
82
|
+
<Flexbox align={'center'} gap={4} horizontal>
|
83
|
+
<Typography.Text style={{ fontSize: 16, fontWeight: 'bold' }}>
|
84
|
+
{t('list.title.disabled')}
|
85
|
+
</Typography.Text>
|
86
|
+
<Center className={styles.count}>{disabledList.length}</Center>
|
87
|
+
</Flexbox>
|
85
88
|
<Grid className={styles.grid}>
|
86
89
|
{disabledList.map((item) => (
|
87
90
|
<Card {...item} key={item.id} />
|
@@ -40,6 +40,7 @@ import { default as tencentcloud } from './tencentcloud';
|
|
40
40
|
import { default as togetherai } from './togetherai';
|
41
41
|
import { default as upstage } from './upstage';
|
42
42
|
import { default as vllm } from './vllm';
|
43
|
+
import { default as volcengine } from './volcengine';
|
43
44
|
import { default as wenxin } from './wenxin';
|
44
45
|
import { default as xai } from './xai';
|
45
46
|
import { default as zeroone } from './zeroone';
|
@@ -105,6 +106,7 @@ export const LOBE_DEFAULT_MODEL_LIST = buildDefaultModelList({
|
|
105
106
|
togetherai,
|
106
107
|
upstage,
|
107
108
|
vllm,
|
109
|
+
volcengine,
|
108
110
|
wenxin,
|
109
111
|
xai,
|
110
112
|
zeroone,
|
@@ -151,6 +153,7 @@ export { default as tencentcloud } from './tencentcloud';
|
|
151
153
|
export { default as togetherai } from './togetherai';
|
152
154
|
export { default as upstage } from './upstage';
|
153
155
|
export { default as vllm } from './vllm';
|
156
|
+
export { default as volcengine } from './volcengine';
|
154
157
|
export { default as wenxin } from './wenxin';
|
155
158
|
export { default as xai } from './xai';
|
156
159
|
export { default as zeroone } from './zeroone';
|
@@ -0,0 +1,83 @@
|
|
1
|
+
import { AIChatModelCard } from '@/types/aiModel';
|
2
|
+
|
3
|
+
const doubaoChatModels: AIChatModelCard[] = [
|
4
|
+
{
|
5
|
+
abilities: {
|
6
|
+
reasoning: true,
|
7
|
+
},
|
8
|
+
contextWindowTokens: 65_536,
|
9
|
+
description:
|
10
|
+
'拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。',
|
11
|
+
displayName: 'DeepSeek R1',
|
12
|
+
enabled: true,
|
13
|
+
id: 'deepseek-r1',
|
14
|
+
type: 'chat',
|
15
|
+
},
|
16
|
+
{
|
17
|
+
abilities: {
|
18
|
+
functionCall: true,
|
19
|
+
},
|
20
|
+
contextWindowTokens: 65_536,
|
21
|
+
description:
|
22
|
+
'DeepSeek-V3 是一款由深度求索公司自研的MoE模型。DeepSeek-V3 多项评测成绩超越了 Qwen2.5-72B 和 Llama-3.1-405B 等其他开源模型,并在性能上和世界顶尖的闭源模型 GPT-4o 以及 Claude-3.5-Sonnet 不分伯仲。',
|
23
|
+
displayName: 'DeepSeek V3',
|
24
|
+
enabled: true,
|
25
|
+
id: 'deepseek-v3',
|
26
|
+
type: 'chat',
|
27
|
+
},
|
28
|
+
{
|
29
|
+
contextWindowTokens: 4096,
|
30
|
+
description:
|
31
|
+
'拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。',
|
32
|
+
displayName: 'Doubao Lite 4k',
|
33
|
+
id: 'Doubao-lite-4k',
|
34
|
+
type: 'chat',
|
35
|
+
},
|
36
|
+
{
|
37
|
+
contextWindowTokens: 32_768,
|
38
|
+
description:
|
39
|
+
'拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 32k 上下文窗口的推理和精调。',
|
40
|
+
displayName: 'Doubao Lite 32k',
|
41
|
+
id: 'Doubao-lite-32k',
|
42
|
+
type: 'chat',
|
43
|
+
},
|
44
|
+
{
|
45
|
+
contextWindowTokens: 128_000,
|
46
|
+
description:
|
47
|
+
'拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 128k 上下文窗口的推理和精调。',
|
48
|
+
displayName: 'Doubao Lite 128k',
|
49
|
+
id: 'Doubao-lite-128k',
|
50
|
+
type: 'chat',
|
51
|
+
},
|
52
|
+
{
|
53
|
+
contextWindowTokens: 4096,
|
54
|
+
description:
|
55
|
+
'效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 4k 上下文窗口的推理和精调。',
|
56
|
+
displayName: 'Doubao Pro 4k',
|
57
|
+
id: 'Doubao-pro-4k',
|
58
|
+
type: 'chat',
|
59
|
+
},
|
60
|
+
{
|
61
|
+
config: {
|
62
|
+
deploymentName: 'Doubao-pro-test',
|
63
|
+
},
|
64
|
+
contextWindowTokens: 32_768,
|
65
|
+
description:
|
66
|
+
'效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 32k 上下文窗口的推理和精调。',
|
67
|
+
displayName: 'Doubao Pro 32k',
|
68
|
+
id: 'Doubao-pro-32k',
|
69
|
+
type: 'chat',
|
70
|
+
},
|
71
|
+
{
|
72
|
+
contextWindowTokens: 128_000,
|
73
|
+
description:
|
74
|
+
'效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 128k 上下文窗口的推理和精调。',
|
75
|
+
displayName: 'Doubao Pro 128k',
|
76
|
+
id: 'Doubao-pro-128k',
|
77
|
+
type: 'chat',
|
78
|
+
},
|
79
|
+
];
|
80
|
+
|
81
|
+
export const allModels = [...doubaoChatModels];
|
82
|
+
|
83
|
+
export default allModels;
|
package/src/config/llm.ts
CHANGED
@@ -132,6 +132,9 @@ export const getLLMConfig = () => {
|
|
132
132
|
ENABLED_DOUBAO: z.boolean(),
|
133
133
|
DOUBAO_API_KEY: z.string().optional(),
|
134
134
|
|
135
|
+
ENABLED_VOLCENGINE: z.boolean(),
|
136
|
+
VOLCENGINE_API_KEY: z.string().optional(),
|
137
|
+
|
135
138
|
ENABLED_TENCENT_CLOUD: z.boolean(),
|
136
139
|
TENCENT_CLOUD_API_KEY: z.string().optional(),
|
137
140
|
|
@@ -158,6 +161,9 @@ export const getLLMConfig = () => {
|
|
158
161
|
ENABLED_GOOGLE: !!process.env.GOOGLE_API_KEY,
|
159
162
|
GOOGLE_API_KEY: process.env.GOOGLE_API_KEY,
|
160
163
|
|
164
|
+
ENABLED_VOLCENGINE: !!process.env.VOLCENGINE_API_KEY,
|
165
|
+
VOLCENGINE_API_KEY: process.env.VOLCENGINE_API_KEY,
|
166
|
+
|
161
167
|
ENABLED_PERPLEXITY: !!process.env.PERPLEXITY_API_KEY,
|
162
168
|
PERPLEXITY_API_KEY: process.env.PERPLEXITY_API_KEY,
|
163
169
|
|
@@ -40,6 +40,7 @@ import TencentcloudProvider from './tencentcloud';
|
|
40
40
|
import TogetherAIProvider from './togetherai';
|
41
41
|
import UpstageProvider from './upstage';
|
42
42
|
import VLLMProvider from './vllm';
|
43
|
+
import VolcengineProvider from './volcengine';
|
43
44
|
import WenxinProvider from './wenxin';
|
44
45
|
import XAIProvider from './xai';
|
45
46
|
import ZeroOneProvider from './zeroone';
|
@@ -127,6 +128,7 @@ export const DEFAULT_MODEL_PROVIDER_LIST = [
|
|
127
128
|
StepfunProvider,
|
128
129
|
MoonshotProvider,
|
129
130
|
BaichuanProvider,
|
131
|
+
VolcengineProvider,
|
130
132
|
MinimaxProvider,
|
131
133
|
LMStudioProvider,
|
132
134
|
InternLMProvider,
|
@@ -186,6 +188,7 @@ export { default as TencentCloudProviderCard } from './tencentcloud';
|
|
186
188
|
export { default as TogetherAIProviderCard } from './togetherai';
|
187
189
|
export { default as UpstageProviderCard } from './upstage';
|
188
190
|
export { default as VLLMProviderCard } from './vllm';
|
191
|
+
export { default as VolcengineProviderCard } from './volcengine';
|
189
192
|
export { default as WenxinProviderCard } from './wenxin';
|
190
193
|
export { default as XAIProviderCard } from './xai';
|
191
194
|
export { default as ZeroOneProviderCard } from './zeroone';
|
@@ -0,0 +1,23 @@
|
|
1
|
+
import { ModelProviderCard } from '@/types/llm';
|
2
|
+
|
3
|
+
// ref https://www.volcengine.com/docs/82379/1330310
|
4
|
+
const Doubao: ModelProviderCard = {
|
5
|
+
chatModels: [],
|
6
|
+
description:
|
7
|
+
'字节跳动推出的大模型服务的开发平台,提供功能丰富、安全以及具备价格竞争力的模型调用服务,同时提供模型数据、精调、推理、评测等端到端功能,全方位保障您的 AI 应用开发落地。',
|
8
|
+
id: 'volcengine',
|
9
|
+
modelsUrl: 'https://www.volcengine.com/docs/82379/1330310',
|
10
|
+
name: '火山引擎',
|
11
|
+
settings: {
|
12
|
+
disableBrowserRequest: true, // CORS error
|
13
|
+
sdkType: 'openai',
|
14
|
+
showDeployName: true,
|
15
|
+
smoothing: {
|
16
|
+
speed: 2,
|
17
|
+
text: true,
|
18
|
+
},
|
19
|
+
},
|
20
|
+
url: 'https://www.volcengine.com/product/ark',
|
21
|
+
};
|
22
|
+
|
23
|
+
export default Doubao;
|
@@ -12,7 +12,6 @@ import { LobeBaichuanAI } from './baichuan';
|
|
12
12
|
import { LobeBedrockAI, LobeBedrockAIParams } from './bedrock';
|
13
13
|
import { LobeCloudflareAI, LobeCloudflareParams } from './cloudflare';
|
14
14
|
import { LobeDeepSeekAI } from './deepseek';
|
15
|
-
import { LobeDoubaoAI } from './doubao';
|
16
15
|
import { LobeFireworksAI } from './fireworksai';
|
17
16
|
import { LobeGiteeAI } from './giteeai';
|
18
17
|
import { LobeGithubAI } from './github';
|
@@ -52,6 +51,7 @@ import {
|
|
52
51
|
} from './types';
|
53
52
|
import { LobeUpstageAI } from './upstage';
|
54
53
|
import { LobeVLLMAI } from './vllm';
|
54
|
+
import { LobeVolcengineAI } from './volcengine';
|
55
55
|
import { LobeWenxinAI } from './wenxin';
|
56
56
|
import { LobeXAI } from './xai';
|
57
57
|
import { LobeZeroOneAI } from './zeroone';
|
@@ -178,6 +178,7 @@ class AgentRuntime {
|
|
178
178
|
togetherai: Partial<ClientOptions>;
|
179
179
|
upstage: Partial<ClientOptions>;
|
180
180
|
vllm: Partial<ClientOptions>;
|
181
|
+
volcengine: Partial<ClientOptions>;
|
181
182
|
wenxin: Partial<ClientOptions>;
|
182
183
|
xai: Partial<ClientOptions>;
|
183
184
|
zeroone: Partial<ClientOptions>;
|
@@ -403,8 +404,9 @@ class AgentRuntime {
|
|
403
404
|
break;
|
404
405
|
}
|
405
406
|
|
407
|
+
case ModelProvider.Volcengine:
|
406
408
|
case ModelProvider.Doubao: {
|
407
|
-
runtimeModel = new
|
409
|
+
runtimeModel = new LobeVolcengineAI(params.volcengine || params.doubao);
|
408
410
|
break;
|
409
411
|
}
|
410
412
|
|
@@ -5,7 +5,6 @@ export { LobeAzureOpenAI } from './azureOpenai';
|
|
5
5
|
export * from './BaseAI';
|
6
6
|
export { LobeBedrockAI } from './bedrock';
|
7
7
|
export { LobeDeepSeekAI } from './deepseek';
|
8
|
-
export { LobeDoubaoAI } from './doubao';
|
9
8
|
export * from './error';
|
10
9
|
export { LobeGoogleAI } from './google';
|
11
10
|
export { LobeGroq } from './groq';
|
@@ -20,5 +19,6 @@ export { LobeQwenAI } from './qwen';
|
|
20
19
|
export { LobeTogetherAI } from './togetherai';
|
21
20
|
export * from './types';
|
22
21
|
export { AgentRuntimeError } from './utils/createError';
|
22
|
+
export { LobeVolcengineAI } from './volcengine';
|
23
23
|
export { LobeZeroOneAI } from './zeroone';
|
24
24
|
export { LobeZhipuAI } from './zhipu';
|
@@ -31,6 +31,9 @@ export enum ModelProvider {
|
|
31
31
|
Bedrock = 'bedrock',
|
32
32
|
Cloudflare = 'cloudflare',
|
33
33
|
DeepSeek = 'deepseek',
|
34
|
+
/**
|
35
|
+
* @deprecated
|
36
|
+
*/
|
34
37
|
Doubao = 'doubao',
|
35
38
|
FireworksAI = 'fireworksai',
|
36
39
|
GiteeAI = 'giteeai',
|
@@ -62,6 +65,7 @@ export enum ModelProvider {
|
|
62
65
|
TogetherAI = 'togetherai',
|
63
66
|
Upstage = 'upstage',
|
64
67
|
VLLM = 'vllm',
|
68
|
+
Volcengine = 'volcengine',
|
65
69
|
Wenxin = 'wenxin',
|
66
70
|
XAI = 'xai',
|
67
71
|
ZeroOne = 'zeroone',
|
@@ -1,10 +1,10 @@
|
|
1
1
|
import { ModelProvider } from '../types';
|
2
2
|
import { LobeOpenAICompatibleFactory } from '../utils/openaiCompatibleFactory';
|
3
3
|
|
4
|
-
export const
|
4
|
+
export const LobeVolcengineAI = LobeOpenAICompatibleFactory({
|
5
5
|
baseURL: 'https://ark.cn-beijing.volces.com/api/v3',
|
6
6
|
debug: {
|
7
7
|
chatCompletion: () => process.env.DEBUG_DOUBAO_CHAT_COMPLETION === '1',
|
8
8
|
},
|
9
|
-
provider: ModelProvider.
|
10
|
-
});
|
9
|
+
provider: ModelProvider.Volcengine,
|
10
|
+
});
|
package/src/services/chat.ts
CHANGED
@@ -51,9 +51,9 @@ const isCanUseFC = (model: string, provider: string) => {
|
|
51
51
|
};
|
52
52
|
|
53
53
|
/**
|
54
|
-
* TODO: we need to update this function to auto find deploymentName
|
54
|
+
* TODO: we need to update this function to auto find deploymentName with provider setting config
|
55
55
|
*/
|
56
|
-
const findDeploymentName = (model: string) => {
|
56
|
+
const findDeploymentName = (model: string, provider: string) => {
|
57
57
|
let deploymentId = model;
|
58
58
|
|
59
59
|
// TODO: remove isDeprecatedEdition condition in V2.0
|
@@ -66,7 +66,9 @@ const findDeploymentName = (model: string) => {
|
|
66
66
|
if (deploymentName) deploymentId = deploymentName;
|
67
67
|
} else {
|
68
68
|
// find the model by id
|
69
|
-
const modelItem = useAiInfraStore
|
69
|
+
const modelItem = useAiInfraStore
|
70
|
+
.getState()
|
71
|
+
.enabledAiModels?.find((i) => i.id === model && i.providerId === provider);
|
70
72
|
|
71
73
|
if (modelItem && modelItem.config?.deploymentName) {
|
72
74
|
deploymentId = modelItem.config?.deploymentName;
|
@@ -227,12 +229,13 @@ class ChatService {
|
|
227
229
|
// if the provider is Azure, get the deployment name as the request model
|
228
230
|
const providersWithDeploymentName = [
|
229
231
|
ModelProvider.Azure,
|
232
|
+
ModelProvider.Volcengine,
|
230
233
|
ModelProvider.Doubao,
|
231
234
|
ModelProvider.AzureAI,
|
232
235
|
] as string[];
|
233
236
|
|
234
237
|
if (providersWithDeploymentName.includes(provider)) {
|
235
|
-
model = findDeploymentName(model);
|
238
|
+
model = findDeploymentName(model, provider);
|
236
239
|
}
|
237
240
|
|
238
241
|
const payload = merge(
|
package/src/types/aiProvider.ts
CHANGED
@@ -19,11 +19,15 @@ export const AiProviderSDKEnum = {
|
|
19
19
|
AzureAI: 'azureai',
|
20
20
|
Bedrock: 'bedrock',
|
21
21
|
Cloudflare: 'cloudflare',
|
22
|
+
/**
|
23
|
+
* @deprecated
|
24
|
+
*/
|
22
25
|
Doubao: 'doubao',
|
23
26
|
Google: 'google',
|
24
27
|
Huggingface: 'huggingface',
|
25
28
|
Ollama: 'ollama',
|
26
29
|
Openai: 'openai',
|
30
|
+
Volcengine: 'volcengine',
|
27
31
|
} as const;
|
28
32
|
|
29
33
|
export type AiProviderSDKType = (typeof AiProviderSDKEnum)[keyof typeof AiProviderSDKEnum];
|
@@ -68,6 +68,7 @@ export interface UserKeyVaults {
|
|
68
68
|
togetherai?: OpenAICompatibleKeyVault;
|
69
69
|
upstage?: OpenAICompatibleKeyVault;
|
70
70
|
vllm?: OpenAICompatibleKeyVault;
|
71
|
+
volcengine?: OpenAICompatibleKeyVault;
|
71
72
|
wenxin?: OpenAICompatibleKeyVault;
|
72
73
|
xai?: OpenAICompatibleKeyVault;
|
73
74
|
zeroone?: OpenAICompatibleKeyVault;
|