@lobehub/chat 1.58.0 → 1.60.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (44) hide show
  1. package/CHANGELOG.md +50 -0
  2. package/Dockerfile +2 -0
  3. package/Dockerfile.database +2 -0
  4. package/changelog/v1.json +18 -0
  5. package/locales/ar/providers.json +3 -0
  6. package/locales/bg-BG/providers.json +3 -0
  7. package/locales/de-DE/providers.json +3 -0
  8. package/locales/en-US/providers.json +3 -0
  9. package/locales/es-ES/providers.json +3 -0
  10. package/locales/fa-IR/modelProvider.json +5 -11
  11. package/locales/fa-IR/providers.json +3 -0
  12. package/locales/fr-FR/providers.json +3 -0
  13. package/locales/it-IT/providers.json +3 -0
  14. package/locales/ja-JP/providers.json +3 -0
  15. package/locales/ko-KR/providers.json +3 -0
  16. package/locales/nl-NL/providers.json +3 -0
  17. package/locales/pl-PL/providers.json +3 -0
  18. package/locales/pt-BR/providers.json +3 -0
  19. package/locales/ru-RU/providers.json +3 -0
  20. package/locales/tr-TR/modelProvider.json +5 -11
  21. package/locales/tr-TR/providers.json +3 -0
  22. package/locales/vi-VN/providers.json +3 -0
  23. package/locales/zh-CN/models.json +6 -6
  24. package/locales/zh-CN/providers.json +3 -0
  25. package/locales/zh-TW/providers.json +3 -0
  26. package/next.config.ts +5 -0
  27. package/package.json +1 -1
  28. package/src/app/[variants]/(main)/settings/llm/ProviderList/providers.tsx +2 -0
  29. package/src/app/[variants]/(main)/settings/provider/(list)/ProviderGrid/index.tsx +6 -3
  30. package/src/config/aiModels/index.ts +6 -0
  31. package/src/config/aiModels/sambanova.ts +195 -0
  32. package/src/config/aiModels/volcengine.ts +83 -0
  33. package/src/config/llm.ts +12 -0
  34. package/src/config/modelProviders/index.ts +7 -0
  35. package/src/config/modelProviders/sambanova.ts +18 -0
  36. package/src/config/modelProviders/volcengine.ts +23 -0
  37. package/src/libs/agent-runtime/AgentRuntime.ts +11 -2
  38. package/src/libs/agent-runtime/index.ts +1 -1
  39. package/src/libs/agent-runtime/sambanova/index.ts +10 -0
  40. package/src/libs/agent-runtime/types/type.ts +5 -0
  41. package/src/libs/agent-runtime/{doubao → volcengine}/index.ts +3 -3
  42. package/src/services/chat.ts +7 -4
  43. package/src/types/aiProvider.ts +4 -0
  44. package/src/types/user/settings/keyVaults.ts +2 -0
package/CHANGELOG.md CHANGED
@@ -2,6 +2,56 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ## [Version 1.60.0](https://github.com/lobehub/lobe-chat/compare/v1.59.0...v1.60.0)
6
+
7
+ <sup>Released on **2025-02-17**</sup>
8
+
9
+ #### ✨ Features
10
+
11
+ - **misc**: Add SambaNova provider support.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### What's improved
19
+
20
+ - **misc**: Add SambaNova provider support, closes [#6218](https://github.com/lobehub/lobe-chat/issues/6218) ([a46eadf](https://github.com/lobehub/lobe-chat/commit/a46eadf))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
30
+ ## [Version 1.59.0](https://github.com/lobehub/lobe-chat/compare/v1.58.0...v1.59.0)
31
+
32
+ <sup>Released on **2025-02-16**</sup>
33
+
34
+ #### ✨ Features
35
+
36
+ - **misc**: Add volcengine as a new provider.
37
+
38
+ <br/>
39
+
40
+ <details>
41
+ <summary><kbd>Improvements and Fixes</kbd></summary>
42
+
43
+ #### What's improved
44
+
45
+ - **misc**: Add volcengine as a new provider, closes [#6221](https://github.com/lobehub/lobe-chat/issues/6221) ([09bf8f0](https://github.com/lobehub/lobe-chat/commit/09bf8f0))
46
+
47
+ </details>
48
+
49
+ <div align="right">
50
+
51
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
52
+
53
+ </div>
54
+
5
55
  ## [Version 1.58.0](https://github.com/lobehub/lobe-chat/compare/v1.57.1...v1.58.0)
6
56
 
7
57
  <sup>Released on **2025-02-16**</sup>
package/Dockerfile CHANGED
@@ -205,6 +205,8 @@ ENV \
205
205
  PERPLEXITY_API_KEY="" PERPLEXITY_MODEL_LIST="" PERPLEXITY_PROXY_URL="" \
206
206
  # Qwen
207
207
  QWEN_API_KEY="" QWEN_MODEL_LIST="" QWEN_PROXY_URL="" \
208
+ # SambaNova
209
+ SAMBANOVA_API_KEY="" SAMBANOVA_MODEL_LIST="" \
208
210
  # SenseNova
209
211
  SENSENOVA_API_KEY="" SENSENOVA_MODEL_LIST="" \
210
212
  # SiliconCloud
@@ -242,6 +242,8 @@ ENV \
242
242
  PERPLEXITY_API_KEY="" PERPLEXITY_MODEL_LIST="" PERPLEXITY_PROXY_URL="" \
243
243
  # Qwen
244
244
  QWEN_API_KEY="" QWEN_MODEL_LIST="" QWEN_PROXY_URL="" \
245
+ # SambaNova
246
+ SAMBANOVA_API_KEY="" SAMBANOVA_MODEL_LIST="" \
245
247
  # SenseNova
246
248
  SENSENOVA_API_KEY="" SENSENOVA_MODEL_LIST="" \
247
249
  # SiliconCloud
package/changelog/v1.json CHANGED
@@ -1,4 +1,22 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "features": [
5
+ "Add SambaNova provider support."
6
+ ]
7
+ },
8
+ "date": "2025-02-17",
9
+ "version": "1.60.0"
10
+ },
11
+ {
12
+ "children": {
13
+ "features": [
14
+ "Add volcengine as a new provider."
15
+ ]
16
+ },
17
+ "date": "2025-02-16",
18
+ "version": "1.59.0"
19
+ },
2
20
  {
3
21
  "children": {
4
22
  "features": [
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM هو مكتبة سريعة وسهلة الاستخدام لاستدلال LLM والخدمات."
121
121
  },
122
+ "volcengine": {
123
+ "description": "منصة تطوير خدمات النماذج الكبيرة التي أطلقتها بايت دانس، تقدم خدمات استدعاء نماذج غنية بالوظائف وآمنة وتنافسية من حيث الأسعار، كما توفر بيانات النماذج، والتعديل الدقيق، والاستدلال، والتقييم، وغيرها من الوظائف الشاملة، لضمان تطوير تطبيقات الذكاء الاصطناعي الخاصة بك بشكل كامل."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "منصة تطوير وخدمات النماذج الكبيرة والتطبيقات الأصلية للذكاء الاصطناعي على مستوى المؤسسات، تقدم مجموعة شاملة وسهلة الاستخدام من أدوات تطوير النماذج الذكية التوليدية وأدوات تطوير التطبيقات على مدار العملية بأكملها."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM е бърза и лесна за използване библиотека за LLM инференция и услуги."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Платформа за разработка на услуги с големи модели, пусната от ByteDance, предлагаща богати на функции, безопасни и конкурентни по цена услуги за извикване на модели. Освен това предоставя край до край функции като данни за модели, фина настройка, инференция и оценка, за да осигури всестранна подкрепа за разработката на вашите AI приложения."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Платформа за разработка и услуги на корпоративно ниво, предлагаща цялостно решение за разработка на генеративни модели на изкуствен интелект и приложения, с най-пълния и лесен за използване инструментариум за целия процес на разработка на модели и приложения."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM ist eine schnelle und benutzerfreundliche Bibliothek für LLM-Inferenz und -Dienste."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Die von ByteDance eingeführte Entwicklungsplattform für große Modellservices bietet funktionsreiche, sichere und preislich wettbewerbsfähige Modellaufrufdienste. Sie bietet zudem End-to-End-Funktionen wie Moduldaten, Feinabstimmung, Inferenz und Bewertung, um die Entwicklung Ihrer KI-Anwendungen umfassend zu unterstützen."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Eine unternehmensweite, umfassende Plattform für die Entwicklung und den Service von großen Modellen und KI-nativen Anwendungen, die die vollständigsten und benutzerfreundlichsten Werkzeuge für die Entwicklung generativer KI-Modelle und den gesamten Anwendungsentwicklungsprozess bietet."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM is a fast and easy-to-use library for LLM inference and serving."
121
121
  },
122
+ "volcengine": {
123
+ "description": "A development platform for large model services launched by ByteDance, offering feature-rich, secure, and competitively priced model invocation services. It also provides end-to-end functionalities such as model data, fine-tuning, inference, and evaluation, ensuring comprehensive support for the development and implementation of your AI applications."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "An enterprise-level one-stop platform for large model and AI-native application development and services, providing the most comprehensive and user-friendly toolchain for the entire process of generative artificial intelligence model development and application development."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM es una biblioteca rápida y fácil de usar para la inferencia y el servicio de LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Plataforma de desarrollo de servicios de modelos grandes lanzada por ByteDance, que ofrece servicios de invocación de modelos ricos en funciones, seguros y competitivos en precio, al tiempo que proporciona datos de modelos, ajuste fino, inferencia, evaluación y otras funciones de extremo a extremo, garantizando de manera integral el desarrollo y la implementación de sus aplicaciones de IA."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Plataforma de desarrollo y servicios de modelos grandes y aplicaciones nativas de IA de nivel empresarial, que ofrece la cadena de herramientas más completa y fácil de usar para el desarrollo de modelos de inteligencia artificial generativa y el desarrollo de aplicaciones en todo el proceso."
124
127
  },
@@ -110,17 +110,6 @@
110
110
  "required": "لطفاً نام ارائه‌دهنده را وارد کنید",
111
111
  "title": "نام ارائه‌دهنده"
112
112
  },
113
- "providerModels": {
114
- "item": {
115
- "modelConfig": {
116
- "deployName": {
117
- "extra": "این فیلد به عنوان شناسه مدل هنگام ارسال درخواست استفاده می‌شود",
118
- "placeholder": "لطفاً نام یا شناسه واقعی استقرار مدل را وارد کنید",
119
- "title": "نام استقرار مدل"
120
- }
121
- }
122
- }
123
- },
124
113
  "proxyUrl": {
125
114
  "required": "لطفاً آدرس پروکسی را وارد کنید",
126
115
  "title": "آدرس پروکسی"
@@ -240,6 +229,11 @@
240
229
  "placeholder": "لطفاً نام استقرار مدل در Azure را وارد کنید",
241
230
  "title": "نام استقرار مدل"
242
231
  },
232
+ "deployName": {
233
+ "extra": "این فیلد به عنوان شناسه مدل هنگام ارسال درخواست استفاده می‌شود",
234
+ "placeholder": "لطفاً نام یا شناسه واقعی مدل را وارد کنید",
235
+ "title": "نام مدل برای استقرار"
236
+ },
243
237
  "displayName": {
244
238
  "placeholder": "لطفاً نام نمایشی مدل را وارد کنید، مانند ChatGPT، GPT-4 و غیره",
245
239
  "title": "نام نمایشی مدل"
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM یک کتابخانه سریع و آسان برای استفاده است که برای استنتاج و خدمات LLM طراحی شده است."
121
121
  },
122
+ "volcengine": {
123
+ "description": "پلتفرم توسعه خدمات مدل‌های بزرگ که توسط بایت‌دANCE راه‌اندازی شده است، خدمات فراوان، ایمن و با قیمت رقابتی برای فراخوانی مدل‌ها را ارائه می‌دهد. همچنین امکاناتی از جمله داده‌های مدل، تنظیم دقیق، استنتاج و ارزیابی را به صورت end-to-end فراهم می‌کند و به طور جامع از توسعه و پیاده‌سازی برنامه‌های هوش مصنوعی شما حمایت می‌کند."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "پلتفرم جامع توسعه و خدمات مدل‌های بزرگ و برنامه‌های بومی هوش مصنوعی در سطح سازمانی، ارائه‌دهنده کامل‌ترین و کاربرپسندترین زنجیره ابزارهای توسعه مدل‌های هوش مصنوعی مولد و توسعه برنامه‌ها"
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM est une bibliothèque rapide et facile à utiliser pour l'inférence et les services LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "La plateforme de développement des services de grands modèles lancée par ByteDance, offrant des services d'appel de modèles riches en fonctionnalités, sécurisés et compétitifs en termes de prix. Elle propose également des fonctionnalités de bout en bout telles que les données de modèle, le réglage fin, l'inférence et l'évaluation, garantissant ainsi le succès de votre développement d'applications AI."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Plateforme de développement et de services d'applications AI natives et de modèles de grande envergure, tout-en-un pour les entreprises, offrant la chaîne d'outils la plus complète et facile à utiliser pour le développement de modèles d'intelligence artificielle générative et le développement d'applications."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM è una libreria veloce e facile da usare per l'inferenza e i servizi LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "La piattaforma di sviluppo dei servizi di modelli di grandi dimensioni lanciata da ByteDance, offre servizi di invocazione di modelli ricchi di funzionalità, sicuri e competitivi in termini di prezzo, fornendo anche dati sui modelli, messa a punto, inferenza, valutazione e altre funzionalità end-to-end, garantendo in modo completo lo sviluppo e l'implementazione delle vostre applicazioni AI."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Piattaforma di sviluppo e servizi per modelli di grandi dimensioni e applicazioni AI native, a livello aziendale, che offre la catena di strumenti completa e facile da usare per lo sviluppo di modelli di intelligenza artificiale generativa e per l'intero processo di sviluppo delle applicazioni."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLMは、LLM推論とサービスのための迅速で使いやすいライブラリです。"
121
121
  },
122
+ "volcengine": {
123
+ "description": "バイトダンスが提供する大規模モデルサービスの開発プラットフォームで、機能が豊富で安全性が高く、価格競争力のあるモデル呼び出しサービスを提供します。また、モデルデータ、ファインチューニング、推論、評価などのエンドツーエンド機能を提供し、AIアプリケーションの開発を全面的にサポートします。"
124
+ },
122
125
  "wenxin": {
123
126
  "description": "企業向けのワンストップ大規模モデルとAIネイティブアプリケーションの開発およびサービスプラットフォームで、最も包括的で使いやすい生成的人工知能モデルの開発とアプリケーション開発の全プロセスツールチェーンを提供します。"
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM은 LLM 추론 및 서비스를 위한 빠르고 사용하기 쉬운 라이브러리입니다."
121
121
  },
122
+ "volcengine": {
123
+ "description": "바이트댄스가 출시한 대형 모델 서비스 개발 플랫폼으로, 기능이 풍부하고 안전하며 가격 경쟁력이 있는 모델 호출 서비스를 제공합니다. 또한 모델 데이터, 세밀 조정, 추론, 평가 등 엔드 투 엔드 기능을 제공하여 귀하의 AI 애플리케이션 개발을 전방위적으로 지원합니다."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "기업용 원스톱 대형 모델 및 AI 네이티브 애플리케이션 개발 및 서비스 플랫폼으로, 가장 포괄적이고 사용하기 쉬운 생성적 인공지능 모델 개발 및 애플리케이션 개발 전체 프로세스 도구 체인을 제공합니다."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM is een snelle en gebruiksvriendelijke bibliotheek voor LLM-inferentie en -diensten."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Het ontwikkelingsplatform voor de grote modellenservice van ByteDance, dat een breed scala aan functies biedt, veilig is en concurrerende prijzen heeft voor modelaanroepdiensten. Het biedt ook end-to-end functionaliteiten zoals modelgegevens, fine-tuning, inferentie en evaluatie, om de ontwikkeling van uw AI-toepassingen volledig te ondersteunen."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Een enterprise-grade, alles-in-één platform voor de ontwikkeling en service van grote modellen en AI-native applicaties, dat de meest uitgebreide en gebruiksvriendelijke toolchain biedt voor de ontwikkeling van generatieve kunstmatige intelligentiemodellen en applicaties."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM to szybka i łatwa w użyciu biblioteka do wnioskowania i usług LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Platforma deweloperska usług dużych modeli wprowadzona przez ByteDance, oferująca bogate w funkcje, bezpieczne i konkurencyjne cenowo usługi wywoływania modeli. Oferuje również dane modelowe, dostosowywanie, wnioskowanie, ocenę i inne funkcje end-to-end, zapewniając kompleksowe wsparcie dla rozwoju aplikacji AI."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Platforma do rozwoju i usług aplikacji AI oraz dużych modeli w skali przedsiębiorstwa, oferująca najbardziej kompleksowy i łatwy w użyciu zestaw narzędzi do rozwoju modeli sztucznej inteligencji generatywnej oraz całego procesu tworzenia aplikacji."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM é uma biblioteca rápida e fácil de usar para inferência e serviços de LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "A plataforma de desenvolvimento de serviços de grandes modelos lançada pela ByteDance, que oferece serviços de chamada de modelos ricos em funcionalidades, seguros e com preços competitivos, além de fornecer dados de modelos, ajuste fino, inferência, avaliação e outras funcionalidades de ponta a ponta, garantindo de forma abrangente a implementação do seu desenvolvimento de aplicações de IA."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Plataforma de desenvolvimento e serviços de aplicativos nativos de IA e modelos de grande escala, voltada para empresas, que oferece a mais completa e fácil ferramenta de cadeia de ferramentas para o desenvolvimento de modelos de inteligência artificial generativa e aplicativos."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM — это быстрая и простая в использовании библиотека для вывода и обслуживания LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Платформа разработки сервисов больших моделей, запущенная ByteDance, предлагает функционально богатые, безопасные и конкурентоспособные по цене услуги вызова моделей, а также предоставляет полные функции от данных моделей, тонкой настройки, вывода до оценки, обеспечивая всестороннюю поддержку разработки ваших AI приложений."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Корпоративная платформа для разработки и обслуживания крупных моделей и нативных приложений ИИ, предлагающая самый полный и удобный инструментарий для разработки генеративных моделей искусственного интеллекта и полного процесса разработки приложений."
124
127
  },
@@ -110,17 +110,6 @@
110
110
  "required": "Lütfen hizmet sağlayıcı adını girin",
111
111
  "title": "Hizmet Sağlayıcı Adı"
112
112
  },
113
- "providerModels": {
114
- "item": {
115
- "modelConfig": {
116
- "deployName": {
117
- "extra": "İstek gönderildiğinde bu alan model ID'si olarak kullanılacaktır",
118
- "placeholder": "Lütfen modelin gerçek dağıtım adını veya kimliğini girin",
119
- "title": "Model Dağıtım Adı"
120
- }
121
- }
122
- }
123
- },
124
113
  "proxyUrl": {
125
114
  "required": "Lütfen proxy adresini girin",
126
115
  "title": "Proxy Adresi"
@@ -240,6 +229,11 @@
240
229
  "placeholder": "Lütfen Azure'daki model dağıtım adını girin",
241
230
  "title": "Model Dağıtım Adı"
242
231
  },
232
+ "deployName": {
233
+ "extra": "Bu alan, isteği gönderirken model kimliği olarak kullanılacaktır",
234
+ "placeholder": "Modelin gerçek dağıtım adını veya kimliğini girin",
235
+ "title": "Model Dağıtım Adı"
236
+ },
243
237
  "displayName": {
244
238
  "placeholder": "Lütfen modelin gösterim adını girin, örneğin ChatGPT, GPT-4 vb.",
245
239
  "title": "Model Gösterim Adı"
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM, LLM çıkarımı ve hizmetleri için hızlı ve kullanımı kolay bir kütüphanedir."
121
121
  },
122
+ "volcengine": {
123
+ "description": "ByteDance tarafından sunulan büyük model hizmetleri geliştirme platformu, zengin özellikler, güvenlik ve rekabetçi fiyatlarla model çağırma hizmetleri sunar. Ayrıca model verileri, ince ayar, çıkarım, değerlendirme gibi uçtan uca işlevler sağlar ve AI uygulama geliştirme sürecinizi her yönüyle güvence altına alır."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Kurumsal düzeyde tek duraklı büyük model ve AI yerel uygulama geliştirme ve hizmet platformu, en kapsamlı ve kullanımı kolay üretken yapay zeka modeli geliştirme, uygulama geliştirme için tam süreç araç zinciri sunar."
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM là một thư viện nhanh chóng và dễ sử dụng cho suy diễn và dịch vụ LLM."
121
121
  },
122
+ "volcengine": {
123
+ "description": "Nền tảng phát triển dịch vụ mô hình lớn do ByteDance phát triển, cung cấp dịch vụ gọi mô hình phong phú, an toàn và có giá cả cạnh tranh, đồng thời cung cấp dữ liệu mô hình, tinh chỉnh, suy diễn, đánh giá và các chức năng đầu cuối khác, đảm bảo toàn diện cho việc phát triển ứng dụng AI của bạn."
124
+ },
122
125
  "wenxin": {
123
126
  "description": "Nền tảng phát triển và dịch vụ ứng dụng AI gốc với mô hình lớn một cửa dành cho doanh nghiệp, cung cấp chuỗi công cụ toàn diện và dễ sử dụng cho phát triển mô hình trí tuệ nhân tạo sinh sinh và phát triển ứng dụng."
124
127
  },
@@ -60,22 +60,22 @@
60
60
  "description": "基于 Qwen2.5-Math-7B 的 DeepSeek-R1 蒸馏模型,通过强化学习与冷启动数据优化推理性能,开源模型刷新多任务标杆。"
61
61
  },
62
62
  "Doubao-lite-128k": {
63
- "description": "Doubao-lite 拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持128k上下文窗口的推理和精调。"
63
+ "description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 128k 上下文窗口的推理和精调。"
64
64
  },
65
65
  "Doubao-lite-32k": {
66
- "description": "Doubao-lite拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持32k上下文窗口的推理和精调。"
66
+ "description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 32k 上下文窗口的推理和精调。"
67
67
  },
68
68
  "Doubao-lite-4k": {
69
- "description": "Doubao-lite拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持4k上下文窗口的推理和精调。"
69
+ "description": "拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。"
70
70
  },
71
71
  "Doubao-pro-128k": {
72
- "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持128k上下文窗口的推理和精调。"
72
+ "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 128k 上下文窗口的推理和精调。"
73
73
  },
74
74
  "Doubao-pro-32k": {
75
- "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持32k上下文窗口的推理和精调。"
75
+ "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 32k 上下文窗口的推理和精调。"
76
76
  },
77
77
  "Doubao-pro-4k": {
78
- "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持4k上下文窗口的推理和精调。"
78
+ "description": "效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 4k 上下文窗口的推理和精调。"
79
79
  },
80
80
  "ERNIE-3.5-128K": {
81
81
  "description": "百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。"
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM 是一个快速且易于使用的库,用于 LLM 推理和服务。"
121
121
  },
122
+ "volcengine": {
123
+ "description": "字节跳动推出的大模型服务的开发平台,提供功能丰富、安全以及具备价格竞争力的模型调用服务,同时提供模型数据、精调、推理、评测等端到端功能,全方位保障您的 AI 应用开发落地。"
124
+ },
122
125
  "wenxin": {
123
126
  "description": "企业级一站式大模型与AI原生应用开发及服务平台,提供最全面易用的生成式人工智能模型开发、应用开发全流程工具链"
124
127
  },
@@ -119,6 +119,9 @@
119
119
  "vllm": {
120
120
  "description": "vLLM 是一個快速且易於使用的庫,用於 LLM 推理和服務。"
121
121
  },
122
+ "volcengine": {
123
+ "description": "字節跳動推出的大模型服務的開發平台,提供功能豐富、安全以及具備價格競爭力的模型調用服務,同時提供模型數據、精調、推理、評測等端到端功能,全方位保障您的 AI 應用開發落地。"
124
+ },
122
125
  "wenxin": {
123
126
  "description": "企業級一站式大模型與AI原生應用開發及服務平台,提供最全面易用的生成式人工智慧模型開發、應用開發全流程工具鏈"
124
127
  },
package/next.config.ts CHANGED
@@ -170,6 +170,11 @@ const nextConfig: NextConfig = {
170
170
  permanent: true,
171
171
  source: '/welcome',
172
172
  },
173
+ {
174
+ destination: '/settings/provider/volcengine',
175
+ permanent: true,
176
+ source: '/settings/provider/doubao',
177
+ },
173
178
  // we need back /repos url in the further
174
179
  {
175
180
  destination: '/files',
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.58.0",
3
+ "version": "1.60.0",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -22,6 +22,7 @@ import {
22
22
  OpenRouterProviderCard,
23
23
  PerplexityProviderCard,
24
24
  QwenProviderCard,
25
+ SambaNovaProviderCard,
25
26
  SenseNovaProviderCard,
26
27
  SiliconCloudProviderCard,
27
28
  SparkProviderCard,
@@ -79,6 +80,7 @@ export const useProviderList = (): ProviderItem[] => {
79
80
  UpstageProviderCard,
80
81
  XAIProviderCard,
81
82
  JinaProviderCard,
83
+ SambaNovaProviderCard,
82
84
  QwenProviderCard,
83
85
  WenxinProviderCard,
84
86
  HunyuanProviderCard,
@@ -79,9 +79,12 @@ const List = memo(() => {
79
79
  </Grid>
80
80
  </Flexbox>
81
81
  <Flexbox gap={12}>
82
- <Typography.Text style={{ fontSize: 16, fontWeight: 'bold' }}>
83
- {t('list.title.disabled')}
84
- </Typography.Text>
82
+ <Flexbox align={'center'} gap={4} horizontal>
83
+ <Typography.Text style={{ fontSize: 16, fontWeight: 'bold' }}>
84
+ {t('list.title.disabled')}
85
+ </Typography.Text>
86
+ <Center className={styles.count}>{disabledList.length}</Center>
87
+ </Flexbox>
85
88
  <Grid className={styles.grid}>
86
89
  {disabledList.map((item) => (
87
90
  <Card {...item} key={item.id} />
@@ -31,6 +31,7 @@ import { default as openai } from './openai';
31
31
  import { default as openrouter } from './openrouter';
32
32
  import { default as perplexity } from './perplexity';
33
33
  import { default as qwen } from './qwen';
34
+ import { default as sambanova } from './sambanova';
34
35
  import { default as sensenova } from './sensenova';
35
36
  import { default as siliconcloud } from './siliconcloud';
36
37
  import { default as spark } from './spark';
@@ -40,6 +41,7 @@ import { default as tencentcloud } from './tencentcloud';
40
41
  import { default as togetherai } from './togetherai';
41
42
  import { default as upstage } from './upstage';
42
43
  import { default as vllm } from './vllm';
44
+ import { default as volcengine } from './volcengine';
43
45
  import { default as wenxin } from './wenxin';
44
46
  import { default as xai } from './xai';
45
47
  import { default as zeroone } from './zeroone';
@@ -96,6 +98,7 @@ export const LOBE_DEFAULT_MODEL_LIST = buildDefaultModelList({
96
98
  openrouter,
97
99
  perplexity,
98
100
  qwen,
101
+ sambanova,
99
102
  sensenova,
100
103
  siliconcloud,
101
104
  spark,
@@ -105,6 +108,7 @@ export const LOBE_DEFAULT_MODEL_LIST = buildDefaultModelList({
105
108
  togetherai,
106
109
  upstage,
107
110
  vllm,
111
+ volcengine,
108
112
  wenxin,
109
113
  xai,
110
114
  zeroone,
@@ -142,6 +146,7 @@ export { default as openai } from './openai';
142
146
  export { default as openrouter } from './openrouter';
143
147
  export { default as perplexity } from './perplexity';
144
148
  export { default as qwen } from './qwen';
149
+ export { default as sambanova } from './sambanova';
145
150
  export { default as sensenova } from './sensenova';
146
151
  export { default as siliconcloud } from './siliconcloud';
147
152
  export { default as spark } from './spark';
@@ -151,6 +156,7 @@ export { default as tencentcloud } from './tencentcloud';
151
156
  export { default as togetherai } from './togetherai';
152
157
  export { default as upstage } from './upstage';
153
158
  export { default as vllm } from './vllm';
159
+ export { default as volcengine } from './volcengine';
154
160
  export { default as wenxin } from './wenxin';
155
161
  export { default as xai } from './xai';
156
162
  export { default as zeroone } from './zeroone';
@@ -0,0 +1,195 @@
1
+ import { AIChatModelCard } from '@/types/aiModel';
2
+
3
+ const sambanovaChatModels: AIChatModelCard[] = [
4
+ {
5
+ abilities: {
6
+ functionCall: true,
7
+ },
8
+ contextWindowTokens: 16_000,
9
+ description: 'Llama 3.3 是 Llama 系列最先进的多语言开源大型语言模型,以极低成本体验媲美 405B 模型的性能。基于 Transformer 结构,并通过监督微调(SFT)和人类反馈强化学习(RLHF)提升有用性和安全性。其指令调优版本专为多语言对话优化,在多项行业基准上表现优于众多开源和封闭聊天模型。知识截止日期为 2023 年 12 月',
10
+ displayName: 'Meta Llama 3.3 70B Instruct',
11
+ enabled: true,
12
+ id: 'Meta-Llama-3.3-70B-Instruct',
13
+ pricing: {
14
+ input: 0.6,
15
+ output: 1.2
16
+ },
17
+ type: 'chat'
18
+ },
19
+ {
20
+ contextWindowTokens: 16_000,
21
+ description: '先进的最尖端小型语言模型,具备语言理解、卓越的推理能力和文本生成能力。',
22
+ displayName: 'Meta Llama 3.2 1B Instruct',
23
+ id: 'Meta-Llama-3.2-1B-Instruct',
24
+ pricing: {
25
+ input: 0.04,
26
+ output: 0.08
27
+ },
28
+ type: 'chat'
29
+ },
30
+ {
31
+ contextWindowTokens: 8000,
32
+ description: '先进的最尖端小型语言模型,具备语言理解、卓越的推理能力和文本生成能力。',
33
+ displayName: 'Meta Llama 3.2 3B Instruct',
34
+ id: 'Meta-Llama-3.2-3B-Instruct',
35
+ pricing: {
36
+ input: 0.08,
37
+ output: 0.16
38
+ },
39
+ type: 'chat'
40
+ },
41
+ {
42
+ abilities: {
43
+ vision: true,
44
+ },
45
+ contextWindowTokens: 4000,
46
+ description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。',
47
+ displayName: 'Meta Llama 3.2 11B Vision Instruct',
48
+ enabled: true,
49
+ id: 'Llama-3.2-11B-Vision-Instruct',
50
+ pricing: {
51
+ input: 0.15,
52
+ output: 0.3
53
+ },
54
+ type: 'chat'
55
+ },
56
+ {
57
+ abilities: {
58
+ vision: true,
59
+ },
60
+ contextWindowTokens: 4000,
61
+ description: '适用于视觉理解代理应用的高级图像推理能力。',
62
+ displayName: 'Meta Llama 3.2 90B Vision Instruct',
63
+ enabled: true,
64
+ id: 'Llama-3.2-90B-Vision-Instruct ',
65
+ pricing: {
66
+ input: 0.8,
67
+ output: 1.6
68
+ },
69
+ type: 'chat'
70
+ },
71
+ {
72
+ abilities: {
73
+ functionCall: true,
74
+ },
75
+ contextWindowTokens: 16_000,
76
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
77
+ displayName: 'Meta Llama 3.1 8B Instruct',
78
+ id: 'Meta-Llama-3.1-8B-Instruct',
79
+ pricing: {
80
+ input: 0.1,
81
+ output: 0.2
82
+ },
83
+ type: 'chat'
84
+ },
85
+ {
86
+ abilities: {
87
+ functionCall: true,
88
+ },
89
+ contextWindowTokens: 128_000,
90
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
91
+ displayName: 'Meta Llama 3.1 70B Instruct',
92
+ id: 'Meta-Llama-3.1-70B-Instruct',
93
+ pricing: {
94
+ input: 0.6,
95
+ output: 1.2
96
+ },
97
+ type: 'chat'
98
+ },
99
+ {
100
+ abilities: {
101
+ functionCall: true,
102
+ },
103
+ contextWindowTokens: 16_000,
104
+ description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。',
105
+ displayName: 'Meta Llama 3.1 405B Instruct',
106
+ id: 'Meta-Llama-3.1-405B-Instruct',
107
+ pricing: {
108
+ input: 5,
109
+ output: 10
110
+ },
111
+ type: 'chat'
112
+ },
113
+ {
114
+ contextWindowTokens: 16_000,
115
+ displayName: 'Llama 3.1 Tulu 3 405B',
116
+ id: 'Llama-3.1-Tulu-3-405B',
117
+ pricing: {
118
+ input: 0.7,
119
+ output: 1.4
120
+ },
121
+ type: 'chat'
122
+ },
123
+ {
124
+ abilities: {
125
+ reasoning: true,
126
+ },
127
+ contextWindowTokens: 4000,
128
+ description: '最先进的高效 LLM,擅长推理、数学和编程。',
129
+ displayName: 'DeepSeek R1',
130
+ id: 'DeepSeek-R1',
131
+ pricing: {
132
+ input: 5,
133
+ output: 7
134
+ },
135
+ type: 'chat'
136
+ },
137
+ {
138
+ abilities: {
139
+ reasoning: true,
140
+ },
141
+ contextWindowTokens: 32_000,
142
+ description: 'DeepSeek R1——DeepSeek 套件中更大更智能的模型——被蒸馏到 Llama 70B 架构中。基于基准测试和人工评估,该模型比原始 Llama 70B 更智能,尤其在需要数学和事实精确性的任务上表现出色。',
143
+ displayName: 'DeepSeek R1 Distill Llama 70B',
144
+ enabled: true,
145
+ id: 'DeepSeek-R1-Distill-Llama-70B',
146
+ pricing: {
147
+ input: 0.7,
148
+ output: 1.4
149
+ },
150
+ type: 'chat'
151
+ },
152
+ {
153
+ abilities: {
154
+ reasoning: true,
155
+ },
156
+ contextWindowTokens: 16_000,
157
+ description: 'Qwen QwQ 是由 Qwen 团队开发的实验研究模型,专注于提升AI推理能力。',
158
+ displayName: 'QwQ 32B Preview',
159
+ enabled: true,
160
+ id: 'QwQ-32B-Preview',
161
+ pricing: {
162
+ input: 1.5,
163
+ output: 3
164
+ },
165
+ type: 'chat'
166
+ },
167
+ {
168
+ contextWindowTokens: 16_000,
169
+ description: '面向中文和英文的 LLM,针对语言、编程、数学、推理等领域。',
170
+ displayName: 'Qwen2.5 72B Instruct',
171
+ enabled: true,
172
+ id: 'Qwen2.5-72B-Instruct',
173
+ pricing: {
174
+ input: 2,
175
+ output: 4
176
+ },
177
+ type: 'chat'
178
+ },
179
+ {
180
+ contextWindowTokens: 16_000,
181
+ description: '高级 LLM,支持代码生成、推理和修复,涵盖主流编程语言。',
182
+ displayName: 'Qwen2.5 Coder 32B Instruct',
183
+ enabled: true,
184
+ id: 'Qwen2.5-Coder-32B-Instruct',
185
+ pricing: {
186
+ input: 1.5,
187
+ output: 3
188
+ },
189
+ type: 'chat'
190
+ },
191
+ ]
192
+
193
+ export const allModels = [...sambanovaChatModels];
194
+
195
+ export default allModels;
@@ -0,0 +1,83 @@
1
+ import { AIChatModelCard } from '@/types/aiModel';
2
+
3
+ const doubaoChatModels: AIChatModelCard[] = [
4
+ {
5
+ abilities: {
6
+ reasoning: true,
7
+ },
8
+ contextWindowTokens: 65_536,
9
+ description:
10
+ '拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。',
11
+ displayName: 'DeepSeek R1',
12
+ enabled: true,
13
+ id: 'deepseek-r1',
14
+ type: 'chat',
15
+ },
16
+ {
17
+ abilities: {
18
+ functionCall: true,
19
+ },
20
+ contextWindowTokens: 65_536,
21
+ description:
22
+ 'DeepSeek-V3 是一款由深度求索公司自研的MoE模型。DeepSeek-V3 多项评测成绩超越了 Qwen2.5-72B 和 Llama-3.1-405B 等其他开源模型,并在性能上和世界顶尖的闭源模型 GPT-4o 以及 Claude-3.5-Sonnet 不分伯仲。',
23
+ displayName: 'DeepSeek V3',
24
+ enabled: true,
25
+ id: 'deepseek-v3',
26
+ type: 'chat',
27
+ },
28
+ {
29
+ contextWindowTokens: 4096,
30
+ description:
31
+ '拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 4k 上下文窗口的推理和精调。',
32
+ displayName: 'Doubao Lite 4k',
33
+ id: 'Doubao-lite-4k',
34
+ type: 'chat',
35
+ },
36
+ {
37
+ contextWindowTokens: 32_768,
38
+ description:
39
+ '拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 32k 上下文窗口的推理和精调。',
40
+ displayName: 'Doubao Lite 32k',
41
+ id: 'Doubao-lite-32k',
42
+ type: 'chat',
43
+ },
44
+ {
45
+ contextWindowTokens: 128_000,
46
+ description:
47
+ '拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持 128k 上下文窗口的推理和精调。',
48
+ displayName: 'Doubao Lite 128k',
49
+ id: 'Doubao-lite-128k',
50
+ type: 'chat',
51
+ },
52
+ {
53
+ contextWindowTokens: 4096,
54
+ description:
55
+ '效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 4k 上下文窗口的推理和精调。',
56
+ displayName: 'Doubao Pro 4k',
57
+ id: 'Doubao-pro-4k',
58
+ type: 'chat',
59
+ },
60
+ {
61
+ config: {
62
+ deploymentName: 'Doubao-pro-test',
63
+ },
64
+ contextWindowTokens: 32_768,
65
+ description:
66
+ '效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 32k 上下文窗口的推理和精调。',
67
+ displayName: 'Doubao Pro 32k',
68
+ id: 'Doubao-pro-32k',
69
+ type: 'chat',
70
+ },
71
+ {
72
+ contextWindowTokens: 128_000,
73
+ description:
74
+ '效果最好的主力模型,适合处理复杂任务,在参考问答、总结摘要、创作、文本分类、角色扮演等场景都有很好的效果。支持 128k 上下文窗口的推理和精调。',
75
+ displayName: 'Doubao Pro 128k',
76
+ id: 'Doubao-pro-128k',
77
+ type: 'chat',
78
+ },
79
+ ];
80
+
81
+ export const allModels = [...doubaoChatModels];
82
+
83
+ export default allModels;
package/src/config/llm.ts CHANGED
@@ -132,11 +132,17 @@ export const getLLMConfig = () => {
132
132
  ENABLED_DOUBAO: z.boolean(),
133
133
  DOUBAO_API_KEY: z.string().optional(),
134
134
 
135
+ ENABLED_VOLCENGINE: z.boolean(),
136
+ VOLCENGINE_API_KEY: z.string().optional(),
137
+
135
138
  ENABLED_TENCENT_CLOUD: z.boolean(),
136
139
  TENCENT_CLOUD_API_KEY: z.string().optional(),
137
140
 
138
141
  ENABLED_JINA: z.boolean(),
139
142
  JINA_API_KEY: z.string().optional(),
143
+
144
+ ENABLED_SAMBANOVA: z.boolean(),
145
+ SAMBANOVA_API_KEY: z.string().optional(),
140
146
  },
141
147
  runtimeEnv: {
142
148
  API_KEY_SELECT_MODE: process.env.API_KEY_SELECT_MODE,
@@ -158,6 +164,9 @@ export const getLLMConfig = () => {
158
164
  ENABLED_GOOGLE: !!process.env.GOOGLE_API_KEY,
159
165
  GOOGLE_API_KEY: process.env.GOOGLE_API_KEY,
160
166
 
167
+ ENABLED_VOLCENGINE: !!process.env.VOLCENGINE_API_KEY,
168
+ VOLCENGINE_API_KEY: process.env.VOLCENGINE_API_KEY,
169
+
161
170
  ENABLED_PERPLEXITY: !!process.env.PERPLEXITY_API_KEY,
162
171
  PERPLEXITY_API_KEY: process.env.PERPLEXITY_API_KEY,
163
172
 
@@ -272,6 +281,9 @@ export const getLLMConfig = () => {
272
281
 
273
282
  ENABLED_JINA: !!process.env.JINA_API_KEY,
274
283
  JINA_API_KEY: process.env.JINA_API_KEY,
284
+
285
+ ENABLED_SAMBANOVA: !!process.env.SAMBANOVA_API_KEY,
286
+ SAMBANOVA_API_KEY: process.env.SAMBANOVA_API_KEY,
275
287
  },
276
288
  });
277
289
  };
@@ -31,6 +31,7 @@ import OpenAIProvider from './openai';
31
31
  import OpenRouterProvider from './openrouter';
32
32
  import PerplexityProvider from './perplexity';
33
33
  import QwenProvider from './qwen';
34
+ import SambaNovaProvider from './sambanova';
34
35
  import SenseNovaProvider from './sensenova';
35
36
  import SiliconCloudProvider from './siliconcloud';
36
37
  import SparkProvider from './spark';
@@ -40,6 +41,7 @@ import TencentcloudProvider from './tencentcloud';
40
41
  import TogetherAIProvider from './togetherai';
41
42
  import UpstageProvider from './upstage';
42
43
  import VLLMProvider from './vllm';
44
+ import VolcengineProvider from './volcengine';
43
45
  import WenxinProvider from './wenxin';
44
46
  import XAIProvider from './xai';
45
47
  import ZeroOneProvider from './zeroone';
@@ -70,6 +72,7 @@ export const LOBE_DEFAULT_MODEL_LIST: ChatModelCard[] = [
70
72
  HuggingFaceProvider.chatModels,
71
73
  XAIProvider.chatModels,
72
74
  JinaProvider.chatModels,
75
+ SambaNovaProvider.chatModels,
73
76
  ZeroOneProvider.chatModels,
74
77
  StepfunProvider.chatModels,
75
78
  NovitaProvider.chatModels,
@@ -115,6 +118,7 @@ export const DEFAULT_MODEL_PROVIDER_LIST = [
115
118
  UpstageProvider,
116
119
  XAIProvider,
117
120
  JinaProvider,
121
+ SambaNovaProvider,
118
122
  QwenProvider,
119
123
  WenxinProvider,
120
124
  TencentcloudProvider,
@@ -127,6 +131,7 @@ export const DEFAULT_MODEL_PROVIDER_LIST = [
127
131
  StepfunProvider,
128
132
  MoonshotProvider,
129
133
  BaichuanProvider,
134
+ VolcengineProvider,
130
135
  MinimaxProvider,
131
136
  LMStudioProvider,
132
137
  InternLMProvider,
@@ -177,6 +182,7 @@ export { default as OpenAIProviderCard } from './openai';
177
182
  export { default as OpenRouterProviderCard } from './openrouter';
178
183
  export { default as PerplexityProviderCard } from './perplexity';
179
184
  export { default as QwenProviderCard } from './qwen';
185
+ export { default as SambaNovaProviderCard } from './sambanova';
180
186
  export { default as SenseNovaProviderCard } from './sensenova';
181
187
  export { default as SiliconCloudProviderCard } from './siliconcloud';
182
188
  export { default as SparkProviderCard } from './spark';
@@ -186,6 +192,7 @@ export { default as TencentCloudProviderCard } from './tencentcloud';
186
192
  export { default as TogetherAIProviderCard } from './togetherai';
187
193
  export { default as UpstageProviderCard } from './upstage';
188
194
  export { default as VLLMProviderCard } from './vllm';
195
+ export { default as VolcengineProviderCard } from './volcengine';
189
196
  export { default as WenxinProviderCard } from './wenxin';
190
197
  export { default as XAIProviderCard } from './xai';
191
198
  export { default as ZeroOneProviderCard } from './zeroone';
@@ -0,0 +1,18 @@
1
+ import { ModelProviderCard } from '@/types/llm';
2
+
3
+ const SambaNova: ModelProviderCard = {
4
+ chatModels: [],
5
+ checkModel: 'Meta-Llama-3.2-1B-Instruct',
6
+ description: 'SambaNova Cloud 可让开发者轻松使用最佳的开源模型,并享受最快的推理速度。',
7
+ disableBrowserRequest: true,
8
+ id: 'sambanova',
9
+ modelsUrl: 'https://cloud.sambanova.ai/plans/pricing',
10
+ name: 'SambaNova',
11
+ settings: {
12
+ disableBrowserRequest: true,
13
+ sdkType: 'openai',
14
+ },
15
+ url: 'https://cloud.sambanova.ai',
16
+ };
17
+
18
+ export default SambaNova;
@@ -0,0 +1,23 @@
1
+ import { ModelProviderCard } from '@/types/llm';
2
+
3
+ // ref https://www.volcengine.com/docs/82379/1330310
4
+ const Doubao: ModelProviderCard = {
5
+ chatModels: [],
6
+ description:
7
+ '字节跳动推出的大模型服务的开发平台,提供功能丰富、安全以及具备价格竞争力的模型调用服务,同时提供模型数据、精调、推理、评测等端到端功能,全方位保障您的 AI 应用开发落地。',
8
+ id: 'volcengine',
9
+ modelsUrl: 'https://www.volcengine.com/docs/82379/1330310',
10
+ name: '火山引擎',
11
+ settings: {
12
+ disableBrowserRequest: true, // CORS error
13
+ sdkType: 'openai',
14
+ showDeployName: true,
15
+ smoothing: {
16
+ speed: 2,
17
+ text: true,
18
+ },
19
+ },
20
+ url: 'https://www.volcengine.com/product/ark',
21
+ };
22
+
23
+ export default Doubao;
@@ -12,7 +12,6 @@ import { LobeBaichuanAI } from './baichuan';
12
12
  import { LobeBedrockAI, LobeBedrockAIParams } from './bedrock';
13
13
  import { LobeCloudflareAI, LobeCloudflareParams } from './cloudflare';
14
14
  import { LobeDeepSeekAI } from './deepseek';
15
- import { LobeDoubaoAI } from './doubao';
16
15
  import { LobeFireworksAI } from './fireworksai';
17
16
  import { LobeGiteeAI } from './giteeai';
18
17
  import { LobeGithubAI } from './github';
@@ -34,6 +33,7 @@ import { LobeOpenAI } from './openai';
34
33
  import { LobeOpenRouterAI } from './openrouter';
35
34
  import { LobePerplexityAI } from './perplexity';
36
35
  import { LobeQwenAI } from './qwen';
36
+ import { LobeSambaNovaAI } from './sambanova';
37
37
  import { LobeSenseNovaAI } from './sensenova';
38
38
  import { LobeSiliconCloudAI } from './siliconcloud';
39
39
  import { LobeSparkAI } from './spark';
@@ -52,6 +52,7 @@ import {
52
52
  } from './types';
53
53
  import { LobeUpstageAI } from './upstage';
54
54
  import { LobeVLLMAI } from './vllm';
55
+ import { LobeVolcengineAI } from './volcengine';
55
56
  import { LobeWenxinAI } from './wenxin';
56
57
  import { LobeXAI } from './xai';
57
58
  import { LobeZeroOneAI } from './zeroone';
@@ -169,6 +170,7 @@ class AgentRuntime {
169
170
  openrouter: Partial<ClientOptions>;
170
171
  perplexity: Partial<ClientOptions>;
171
172
  qwen: Partial<ClientOptions>;
173
+ sambanova: Partial<ClientOptions>;
172
174
  sensenova: Partial<ClientOptions>;
173
175
  siliconcloud: Partial<ClientOptions>;
174
176
  spark: Partial<ClientOptions>;
@@ -178,6 +180,7 @@ class AgentRuntime {
178
180
  togetherai: Partial<ClientOptions>;
179
181
  upstage: Partial<ClientOptions>;
180
182
  vllm: Partial<ClientOptions>;
183
+ volcengine: Partial<ClientOptions>;
181
184
  wenxin: Partial<ClientOptions>;
182
185
  xai: Partial<ClientOptions>;
183
186
  zeroone: Partial<ClientOptions>;
@@ -383,6 +386,11 @@ class AgentRuntime {
383
386
  break;
384
387
  }
385
388
 
389
+ case ModelProvider.SambaNova: {
390
+ runtimeModel = new LobeSambaNovaAI(params.sambanova);
391
+ break;
392
+ }
393
+
386
394
  case ModelProvider.Cloudflare: {
387
395
  runtimeModel = new LobeCloudflareAI(params.cloudflare);
388
396
  break;
@@ -403,8 +411,9 @@ class AgentRuntime {
403
411
  break;
404
412
  }
405
413
 
414
+ case ModelProvider.Volcengine:
406
415
  case ModelProvider.Doubao: {
407
- runtimeModel = new LobeDoubaoAI(params.doubao);
416
+ runtimeModel = new LobeVolcengineAI(params.volcengine || params.doubao);
408
417
  break;
409
418
  }
410
419
 
@@ -5,7 +5,6 @@ export { LobeAzureOpenAI } from './azureOpenai';
5
5
  export * from './BaseAI';
6
6
  export { LobeBedrockAI } from './bedrock';
7
7
  export { LobeDeepSeekAI } from './deepseek';
8
- export { LobeDoubaoAI } from './doubao';
9
8
  export * from './error';
10
9
  export { LobeGoogleAI } from './google';
11
10
  export { LobeGroq } from './groq';
@@ -20,5 +19,6 @@ export { LobeQwenAI } from './qwen';
20
19
  export { LobeTogetherAI } from './togetherai';
21
20
  export * from './types';
22
21
  export { AgentRuntimeError } from './utils/createError';
22
+ export { LobeVolcengineAI } from './volcengine';
23
23
  export { LobeZeroOneAI } from './zeroone';
24
24
  export { LobeZhipuAI } from './zhipu';
@@ -0,0 +1,10 @@
1
+ import { ModelProvider } from '../types';
2
+ import { LobeOpenAICompatibleFactory } from '../utils/openaiCompatibleFactory';
3
+
4
+ export const LobeSambaNovaAI = LobeOpenAICompatibleFactory({
5
+ baseURL: 'https://api.sambanova.ai/v1',
6
+ debug: {
7
+ chatCompletion: () => process.env.DEBUG_SAMBANOVA_CHAT_COMPLETION === '1',
8
+ },
9
+ provider: ModelProvider.SambaNova,
10
+ });
@@ -31,6 +31,9 @@ export enum ModelProvider {
31
31
  Bedrock = 'bedrock',
32
32
  Cloudflare = 'cloudflare',
33
33
  DeepSeek = 'deepseek',
34
+ /**
35
+ * @deprecated
36
+ */
34
37
  Doubao = 'doubao',
35
38
  FireworksAI = 'fireworksai',
36
39
  GiteeAI = 'giteeai',
@@ -53,6 +56,7 @@ export enum ModelProvider {
53
56
  OpenRouter = 'openrouter',
54
57
  Perplexity = 'perplexity',
55
58
  Qwen = 'qwen',
59
+ SambaNova = 'sambanova',
56
60
  SenseNova = 'sensenova',
57
61
  SiliconCloud = 'siliconcloud',
58
62
  Spark = 'spark',
@@ -62,6 +66,7 @@ export enum ModelProvider {
62
66
  TogetherAI = 'togetherai',
63
67
  Upstage = 'upstage',
64
68
  VLLM = 'vllm',
69
+ Volcengine = 'volcengine',
65
70
  Wenxin = 'wenxin',
66
71
  XAI = 'xai',
67
72
  ZeroOne = 'zeroone',
@@ -1,10 +1,10 @@
1
1
  import { ModelProvider } from '../types';
2
2
  import { LobeOpenAICompatibleFactory } from '../utils/openaiCompatibleFactory';
3
3
 
4
- export const LobeDoubaoAI = LobeOpenAICompatibleFactory({
4
+ export const LobeVolcengineAI = LobeOpenAICompatibleFactory({
5
5
  baseURL: 'https://ark.cn-beijing.volces.com/api/v3',
6
6
  debug: {
7
7
  chatCompletion: () => process.env.DEBUG_DOUBAO_CHAT_COMPLETION === '1',
8
8
  },
9
- provider: ModelProvider.Doubao,
10
- });
9
+ provider: ModelProvider.Volcengine,
10
+ });
@@ -51,9 +51,9 @@ const isCanUseFC = (model: string, provider: string) => {
51
51
  };
52
52
 
53
53
  /**
54
- * TODO: we need to update this function to auto find deploymentName
54
+ * TODO: we need to update this function to auto find deploymentName with provider setting config
55
55
  */
56
- const findDeploymentName = (model: string) => {
56
+ const findDeploymentName = (model: string, provider: string) => {
57
57
  let deploymentId = model;
58
58
 
59
59
  // TODO: remove isDeprecatedEdition condition in V2.0
@@ -66,7 +66,9 @@ const findDeploymentName = (model: string) => {
66
66
  if (deploymentName) deploymentId = deploymentName;
67
67
  } else {
68
68
  // find the model by id
69
- const modelItem = useAiInfraStore.getState().enabledAiModels?.find((i) => i.id === model);
69
+ const modelItem = useAiInfraStore
70
+ .getState()
71
+ .enabledAiModels?.find((i) => i.id === model && i.providerId === provider);
70
72
 
71
73
  if (modelItem && modelItem.config?.deploymentName) {
72
74
  deploymentId = modelItem.config?.deploymentName;
@@ -227,12 +229,13 @@ class ChatService {
227
229
  // if the provider is Azure, get the deployment name as the request model
228
230
  const providersWithDeploymentName = [
229
231
  ModelProvider.Azure,
232
+ ModelProvider.Volcengine,
230
233
  ModelProvider.Doubao,
231
234
  ModelProvider.AzureAI,
232
235
  ] as string[];
233
236
 
234
237
  if (providersWithDeploymentName.includes(provider)) {
235
- model = findDeploymentName(model);
238
+ model = findDeploymentName(model, provider);
236
239
  }
237
240
 
238
241
  const payload = merge(
@@ -19,11 +19,15 @@ export const AiProviderSDKEnum = {
19
19
  AzureAI: 'azureai',
20
20
  Bedrock: 'bedrock',
21
21
  Cloudflare: 'cloudflare',
22
+ /**
23
+ * @deprecated
24
+ */
22
25
  Doubao: 'doubao',
23
26
  Google: 'google',
24
27
  Huggingface: 'huggingface',
25
28
  Ollama: 'ollama',
26
29
  Openai: 'openai',
30
+ Volcengine: 'volcengine',
27
31
  } as const;
28
32
 
29
33
  export type AiProviderSDKType = (typeof AiProviderSDKEnum)[keyof typeof AiProviderSDKEnum];
@@ -59,6 +59,7 @@ export interface UserKeyVaults {
59
59
  password?: string;
60
60
  perplexity?: OpenAICompatibleKeyVault;
61
61
  qwen?: OpenAICompatibleKeyVault;
62
+ sambanova?: OpenAICompatibleKeyVault;
62
63
  sensenova?: OpenAICompatibleKeyVault;
63
64
  siliconcloud?: OpenAICompatibleKeyVault;
64
65
  spark?: OpenAICompatibleKeyVault;
@@ -68,6 +69,7 @@ export interface UserKeyVaults {
68
69
  togetherai?: OpenAICompatibleKeyVault;
69
70
  upstage?: OpenAICompatibleKeyVault;
70
71
  vllm?: OpenAICompatibleKeyVault;
72
+ volcengine?: OpenAICompatibleKeyVault;
71
73
  wenxin?: OpenAICompatibleKeyVault;
72
74
  xai?: OpenAICompatibleKeyVault;
73
75
  zeroone?: OpenAICompatibleKeyVault;