@lobehub/chat 1.53.12 → 1.55.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +58 -0
- package/Dockerfile +4 -0
- package/Dockerfile.database +4 -0
- package/README.ja-JP.md +1 -1
- package/README.md +1 -1
- package/README.zh-CN.md +1 -1
- package/README.zh-TW.md +9 -14
- package/changelog/v1.json +21 -0
- package/docs/changelog/2024-11-25-november-providers.mdx +1 -1
- package/docs/changelog/2024-11-25-november-providers.zh-CN.mdx +1 -1
- package/docs/usage/features/multi-ai-providers.mdx +1 -1
- package/docs/usage/features/multi-ai-providers.zh-CN.mdx +1 -1
- package/locales/ar/modelProvider.json +0 -1
- package/locales/ar/setting.json +12 -9
- package/locales/bg-BG/modelProvider.json +0 -1
- package/locales/bg-BG/setting.json +12 -9
- package/locales/de-DE/modelProvider.json +0 -1
- package/locales/de-DE/setting.json +13 -10
- package/locales/en-US/modelProvider.json +0 -1
- package/locales/en-US/setting.json +12 -9
- package/locales/es-ES/modelProvider.json +0 -1
- package/locales/es-ES/setting.json +12 -9
- package/locales/fa-IR/modelProvider.json +0 -1
- package/locales/fa-IR/setting.json +12 -9
- package/locales/fr-FR/modelProvider.json +0 -1
- package/locales/fr-FR/setting.json +12 -9
- package/locales/it-IT/modelProvider.json +0 -1
- package/locales/it-IT/setting.json +13 -10
- package/locales/ja-JP/modelProvider.json +0 -1
- package/locales/ja-JP/setting.json +12 -9
- package/locales/ko-KR/modelProvider.json +0 -1
- package/locales/ko-KR/setting.json +12 -9
- package/locales/nl-NL/modelProvider.json +0 -1
- package/locales/nl-NL/setting.json +12 -9
- package/locales/pl-PL/modelProvider.json +0 -1
- package/locales/pl-PL/setting.json +12 -9
- package/locales/pt-BR/modelProvider.json +0 -1
- package/locales/pt-BR/setting.json +13 -10
- package/locales/ru-RU/modelProvider.json +0 -1
- package/locales/ru-RU/setting.json +12 -9
- package/locales/tr-TR/modelProvider.json +0 -1
- package/locales/tr-TR/setting.json +12 -9
- package/locales/vi-VN/modelProvider.json +0 -1
- package/locales/vi-VN/setting.json +12 -9
- package/locales/zh-CN/modelProvider.json +0 -1
- package/locales/zh-CN/setting.json +13 -10
- package/locales/zh-TW/modelProvider.json +0 -1
- package/locales/zh-TW/setting.json +12 -9
- package/package.json +1 -1
- package/src/app/[variants]/(main)/chat/(workspace)/@conversation/features/ChatInput/Desktop/index.tsx +1 -1
- package/src/app/[variants]/(main)/settings/llm/ProviderList/providers.tsx +4 -0
- package/src/components/InfoTooltip/index.tsx +25 -0
- package/src/components/Loading/UpdateLoading/index.tsx +19 -0
- package/src/config/aiModels/index.ts +6 -0
- package/src/config/aiModels/nvidia.ts +155 -0
- package/src/config/aiModels/vllm.ts +94 -0
- package/src/config/llm.ts +12 -0
- package/src/config/modelProviders/index.ts +8 -0
- package/src/config/modelProviders/nvidia.ts +21 -0
- package/src/config/modelProviders/vllm.ts +20 -0
- package/src/const/url.ts +1 -1
- package/src/features/ChatInput/ActionBar/Params/ParamsControls.tsx +95 -0
- package/src/features/ChatInput/ActionBar/Params/index.tsx +47 -0
- package/src/features/ChatInput/ActionBar/config.ts +3 -2
- package/src/features/ChatInput/Mobile/index.tsx +1 -1
- package/src/features/ModelParamsControl/FrequencyPenalty.tsx +37 -0
- package/src/features/ModelParamsControl/PresencePenalty.tsx +35 -0
- package/src/features/ModelParamsControl/Temperature.tsx +71 -0
- package/src/features/ModelParamsControl/TopP.tsx +39 -0
- package/src/features/ModelParamsControl/index.ts +4 -0
- package/src/libs/agent-runtime/AgentRuntime.ts +14 -0
- package/src/libs/agent-runtime/nvidia/index.ts +44 -0
- package/src/libs/agent-runtime/types/type.ts +2 -0
- package/src/libs/agent-runtime/vllm/index.ts +44 -0
- package/src/locales/default/setting.ts +12 -9
- package/src/types/user/settings/keyVaults.ts +2 -0
- package/src/features/ChatInput/ActionBar/Temperature.tsx +0 -49
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,64 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
## [Version 1.55.0](https://github.com/lobehub/lobe-chat/compare/v1.54.0...v1.55.0)
|
6
|
+
|
7
|
+
<sup>Released on **2025-02-14**</sup>
|
8
|
+
|
9
|
+
#### ✨ Features
|
10
|
+
|
11
|
+
- **misc**: Add vLLM provider support.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### What's improved
|
19
|
+
|
20
|
+
- **misc**: Add vLLM provider support, closes [#6154](https://github.com/lobehub/lobe-chat/issues/6154) ([1708e32](https://github.com/lobehub/lobe-chat/commit/1708e32))
|
21
|
+
|
22
|
+
</details>
|
23
|
+
|
24
|
+
<div align="right">
|
25
|
+
|
26
|
+
[](#readme-top)
|
27
|
+
|
28
|
+
</div>
|
29
|
+
|
30
|
+
## [Version 1.54.0](https://github.com/lobehub/lobe-chat/compare/v1.53.12...v1.54.0)
|
31
|
+
|
32
|
+
<sup>Released on **2025-02-14**</sup>
|
33
|
+
|
34
|
+
#### ✨ Features
|
35
|
+
|
36
|
+
- **misc**: Add Nvidia NIM provider support.
|
37
|
+
|
38
|
+
#### 💄 Styles
|
39
|
+
|
40
|
+
- **misc**: Improve advanced params settings.
|
41
|
+
|
42
|
+
<br/>
|
43
|
+
|
44
|
+
<details>
|
45
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
46
|
+
|
47
|
+
#### What's improved
|
48
|
+
|
49
|
+
- **misc**: Add Nvidia NIM provider support, closes [#6142](https://github.com/lobehub/lobe-chat/issues/6142) ([ab796a7](https://github.com/lobehub/lobe-chat/commit/ab796a7))
|
50
|
+
|
51
|
+
#### Styles
|
52
|
+
|
53
|
+
- **misc**: Improve advanced params settings, closes [#6149](https://github.com/lobehub/lobe-chat/issues/6149) ([bf6699c](https://github.com/lobehub/lobe-chat/commit/bf6699c))
|
54
|
+
|
55
|
+
</details>
|
56
|
+
|
57
|
+
<div align="right">
|
58
|
+
|
59
|
+
[](#readme-top)
|
60
|
+
|
61
|
+
</div>
|
62
|
+
|
5
63
|
### [Version 1.53.12](https://github.com/lobehub/lobe-chat/compare/v1.53.11...v1.53.12)
|
6
64
|
|
7
65
|
<sup>Released on **2025-02-14**</sup>
|
package/Dockerfile
CHANGED
@@ -191,6 +191,8 @@ ENV \
|
|
191
191
|
MOONSHOT_API_KEY="" MOONSHOT_MODEL_LIST="" MOONSHOT_PROXY_URL="" \
|
192
192
|
# Novita
|
193
193
|
NOVITA_API_KEY="" NOVITA_MODEL_LIST="" \
|
194
|
+
# Nvidia NIM
|
195
|
+
NVIDIA_API_KEY="" NVIDIA_MODEL_LIST="" NVIDIA_PROXY_URL="" \
|
194
196
|
# Ollama
|
195
197
|
ENABLED_OLLAMA="" OLLAMA_MODEL_LIST="" OLLAMA_PROXY_URL="" \
|
196
198
|
# OpenAI
|
@@ -215,6 +217,8 @@ ENV \
|
|
215
217
|
TOGETHERAI_API_KEY="" TOGETHERAI_MODEL_LIST="" \
|
216
218
|
# Upstage
|
217
219
|
UPSTAGE_API_KEY="" UPSTAGE_MODEL_LIST="" \
|
220
|
+
# vLLM
|
221
|
+
VLLM_API_KEY="" VLLM_MODEL_LIST="" VLLM_PROXY_URL="" \
|
218
222
|
# Wenxin
|
219
223
|
WENXIN_API_KEY="" WENXIN_MODEL_LIST="" \
|
220
224
|
# xAI
|
package/Dockerfile.database
CHANGED
@@ -228,6 +228,8 @@ ENV \
|
|
228
228
|
MOONSHOT_API_KEY="" MOONSHOT_MODEL_LIST="" MOONSHOT_PROXY_URL="" \
|
229
229
|
# Novita
|
230
230
|
NOVITA_API_KEY="" NOVITA_MODEL_LIST="" \
|
231
|
+
# Nvidia NIM
|
232
|
+
NVIDIA_API_KEY="" NVIDIA_MODEL_LIST="" NVIDIA_PROXY_URL="" \
|
231
233
|
# Ollama
|
232
234
|
ENABLED_OLLAMA="" OLLAMA_MODEL_LIST="" OLLAMA_PROXY_URL="" \
|
233
235
|
# OpenAI
|
@@ -252,6 +254,8 @@ ENV \
|
|
252
254
|
TOGETHERAI_API_KEY="" TOGETHERAI_MODEL_LIST="" \
|
253
255
|
# Upstage
|
254
256
|
UPSTAGE_API_KEY="" UPSTAGE_MODEL_LIST="" \
|
257
|
+
# vLLM
|
258
|
+
VLLM_API_KEY="" VLLM_MODEL_LIST="" VLLM_PROXY_URL="" \
|
255
259
|
# Wenxin
|
256
260
|
WENXIN_API_KEY="" WENXIN_MODEL_LIST="" \
|
257
261
|
# xAI
|
package/README.ja-JP.md
CHANGED
@@ -170,7 +170,7 @@ LobeChat の継続的な開発において、AI 会話サービスを提供す
|
|
170
170
|
|
171
171
|
<!-- PROVIDER LIST -->
|
172
172
|
|
173
|
-
同時に、私たちはさらに多くのモデルサービスプロバイダーをサポートする計画を立てており、サービスプロバイダーのライブラリをさらに充実させる予定です。 LobeChat があなたのお気に入りのサービスプロバイダーをサポートすることを希望する場合は、[💬 コミュニティディスカッション](https://github.com/lobehub/lobe-chat/discussions/
|
173
|
+
同時に、私たちはさらに多くのモデルサービスプロバイダーをサポートする計画を立てており、サービスプロバイダーのライブラリをさらに充実させる予定です。 LobeChat があなたのお気に入りのサービスプロバイダーをサポートすることを希望する場合は、[💬 コミュニティディスカッション](https://github.com/lobehub/lobe-chat/discussions/6157)に参加してください。
|
174
174
|
|
175
175
|
<div align="right">
|
176
176
|
|
package/README.md
CHANGED
@@ -189,7 +189,7 @@ We have implemented support for the following model service providers:
|
|
189
189
|
|
190
190
|
<!-- PROVIDER LIST -->
|
191
191
|
|
192
|
-
At the same time, we are also planning to support more model service providers. If you would like LobeChat to support your favorite service provider, feel free to join our [💬 community discussion](https://github.com/lobehub/lobe-chat/discussions/
|
192
|
+
At the same time, we are also planning to support more model service providers. If you would like LobeChat to support your favorite service provider, feel free to join our [💬 community discussion](https://github.com/lobehub/lobe-chat/discussions/6157).
|
193
193
|
|
194
194
|
<div align="right">
|
195
195
|
|
package/README.zh-CN.md
CHANGED
@@ -189,7 +189,7 @@ LobeChat 支持文件上传与知识库功能,你可以上传文件、图片
|
|
189
189
|
|
190
190
|
<!-- PROVIDER LIST -->
|
191
191
|
|
192
|
-
同时,我们也在计划支持更多的模型服务商,以进一步丰富我们的服务商库。如果你希望让 LobeChat 支持你喜爱的服务商,欢迎加入我们的 [💬 社区讨论](https://github.com/lobehub/lobe-chat/discussions/
|
192
|
+
同时,我们也在计划支持更多的模型服务商,以进一步丰富我们的服务商库。如果你希望让 LobeChat 支持你喜爱的服务商,欢迎加入我们的 [💬 社区讨论](https://github.com/lobehub/lobe-chat/discussions/6157)。
|
193
193
|
|
194
194
|
<div align="right">
|
195
195
|
|
package/README.zh-TW.md
CHANGED
@@ -4,8 +4,8 @@
|
|
4
4
|
|
5
5
|
<h1>Lobe Chat</h1>
|
6
6
|
|
7
|
-
現代風格的開源 ChatGPT/LLMs
|
8
|
-
支持語音合成、多模態、可擴展的([function call][docs-functionc-call]
|
7
|
+
現代風格的開源 ChatGPT/LLMs 使用介面 / 框架\
|
8
|
+
支持語音合成、多模態、可擴展的([function call][docs-functionc-call])插件系統\
|
9
9
|
一鍵**免費**部署私人的 OpenAI ChatGPT/Claude/Gemini/Groq/Ollama 聊天應用
|
10
10
|
|
11
11
|
[English](./README.md) · **繁體中文** · [簡體中文](./README.zh-CN.md) · [日本語](./README.ja-JP.md) · [官網][official-site] · [更新日誌][changelog] · [文檔][docs] · [部落格][blog] · [反饋問題][github-issues-link]
|
@@ -50,14 +50,13 @@
|
|
50
50
|
|
51
51
|
#### TOC
|
52
52
|
|
53
|
-
- [👋🏻 開始使用
|
53
|
+
- [👋🏻 開始使用 & 交流](#-開始使用--交流)
|
54
54
|
- [✨ 特性一覽](#-特性一覽)
|
55
55
|
- [`1` 文件上傳 / 知識庫](#1-文件上傳--知識庫)
|
56
56
|
- [`2` 多模型服務商支持](#2-多模型服務商支持)
|
57
|
-
- [已支持的模型服務商](#已支持的模型服務商)
|
58
57
|
- [`3` 支持本地大型語言模型 (LLM)](#3-支持本地大型語言模型-llm)
|
59
58
|
- [`4` 模型視覺識別 (Model Visual)](#4-模型視覺識別-model-visual)
|
60
|
-
- [`5` TTS
|
59
|
+
- [`5` TTS & STT 語音轉化](#5-tts--stt-語音轉化)
|
61
60
|
- [`6` Text to Image 文生圖](#6-text-to-image-文生圖)
|
62
61
|
- [`7` 插件系統 (Tools Calling)](#7-插件系統-tools-calling)
|
63
62
|
- [`8` 助手市集 (GPTs)](#8-助手市集-gpts)
|
@@ -70,13 +69,9 @@
|
|
70
69
|
- [⚡️ 性能測試](#️-性能測試)
|
71
70
|
- [🛳 開箱即用](#-開箱即用)
|
72
71
|
- [`A` 使用 Vercel、Zeabur 、Sealos 或 Alibaba Cloud 部署](#a-使用-vercelzeabur-sealos-或-alibaba-cloud-部署)
|
73
|
-
- [Fork 之後](#fork-之後)
|
74
|
-
- [保持更新](#保持更新)
|
75
72
|
- [`B` 使用 Docker 部署](#b-使用-docker-部署)
|
76
73
|
- [環境變量](#環境變量)
|
77
74
|
- [獲取 OpenAI API Key](#獲取-openai-api-key)
|
78
|
-
- [`A` 通過 OpenAI 官方渠道](#a-通過-openai-官方渠道)
|
79
|
-
- [`B` 通過 OpenAI 第三方代理商](#b-通過-openai-第三方代理商)
|
80
75
|
- [📦 生態系統](#-生態系統)
|
81
76
|
- [🧩 插件體系](#-插件體系)
|
82
77
|
- [⌨️ 本地開發](#️-本地開發)
|
@@ -93,12 +88,12 @@
|
|
93
88
|
## 👋🏻 開始使用 & 交流
|
94
89
|
|
95
90
|
我們是一群充滿熱情的設計工程師,希望為 AIGC 提供現代化的設計組件和工具,並以開源的方式分享。
|
96
|
-
透過採用Bootstrapping 的方式,我們的目標是為開發人員和使用者提供一個更加開放、透明和使用者友好的產品生態系統。
|
91
|
+
透過採用 Bootstrapping 的方式,我們的目標是為開發人員和使用者提供一個更加開放、透明和使用者友好的產品生態系統。
|
97
92
|
|
98
93
|
LobeHub 旨在成為普通用戶與專業開發者測試 AI 助手的場所。LobeChat 目前正在積極開發中,有任何需求或者問題,歡迎提交 [issues][issues-link]
|
99
94
|
|
100
|
-
| [![][vercel-shield-badge]][vercel-link] | 無需安裝或註冊!訪問我們的網站立刻體驗
|
101
|
-
| :---------------------------------------- |
|
95
|
+
| [![][vercel-shield-badge]][vercel-link] | 無需安裝或註冊!訪問我們的網站立刻體驗 |
|
96
|
+
| :---------------------------------------- | :------------------------------------------------------ |
|
102
97
|
| [![][discord-shield-badge]][discord-link] | 加入我們的 Discord 和開發者交流,和其他用戶們分享心得! |
|
103
98
|
|
104
99
|
> \[!IMPORTANT]
|
@@ -194,7 +189,7 @@ LobeChat 支持文件上傳與知識庫功能,你可以上傳文件、圖片
|
|
194
189
|
|
195
190
|
<!-- PROVIDER LIST -->
|
196
191
|
|
197
|
-
同時,我們也在計劃支持更多的模型服務商,以進一步豐富我們的服務商庫。如果你希望讓 LobeChat 支持你喜愛的服務商,歡迎加入我們的 [💬 社區討論](https://github.com/lobehub/lobe-chat/discussions/
|
192
|
+
同時,我們也在計劃支持更多的模型服務商,以進一步豐富我們的服務商庫。如果你希望讓 LobeChat 支持你喜愛的服務商,歡迎加入我們的 [💬 社區討論](https://github.com/lobehub/lobe-chat/discussions/6157)。
|
198
193
|
|
199
194
|
<div align="right">
|
200
195
|
|
@@ -477,7 +472,7 @@ LobeChat 提供了 Vercel 的 自托管版本 和 [Docker 鏡像][docker-release
|
|
477
472
|
|
478
473
|
<div align="center">
|
479
474
|
|
480
|
-
| 使用 Vercel 部署 | 使用 Zeabur 部署 | 使用 Sealos 部署 |
|
475
|
+
| 使用 Vercel 部署 | 使用 Zeabur 部署 | 使用 Sealos 部署 | 使用 Alibaba Cloud 部署 |
|
481
476
|
| :-------------------------------------: | :---------------------------------------------------------: | :---------------------------------------------------------: | :-----------------------------------------------------------------------: |
|
482
477
|
| [![][deploy-button-image]][deploy-link] | [![][deploy-on-zeabur-button-image]][deploy-on-zeabur-link] | [![][deploy-on-sealos-button-image]][deploy-on-sealos-link] | [![][deploy-on-alibaba-cloud-button-image]][deploy-on-alibaba-cloud-link] |
|
483
478
|
|
package/changelog/v1.json
CHANGED
@@ -1,4 +1,25 @@
|
|
1
1
|
[
|
2
|
+
{
|
3
|
+
"children": {
|
4
|
+
"features": [
|
5
|
+
"Add vLLM provider support."
|
6
|
+
]
|
7
|
+
},
|
8
|
+
"date": "2025-02-14",
|
9
|
+
"version": "1.55.0"
|
10
|
+
},
|
11
|
+
{
|
12
|
+
"children": {
|
13
|
+
"features": [
|
14
|
+
"Add Nvidia NIM provider support."
|
15
|
+
],
|
16
|
+
"improvements": [
|
17
|
+
"Improve advanced params settings."
|
18
|
+
]
|
19
|
+
},
|
20
|
+
"date": "2025-02-14",
|
21
|
+
"version": "1.54.0"
|
22
|
+
},
|
2
23
|
{
|
3
24
|
"children": {
|
4
25
|
"improvements": [
|
@@ -23,4 +23,4 @@ We're excited to announce that LobeChat has expanded its AI model support with t
|
|
23
23
|
|
24
24
|
## Need More Model Providers?
|
25
25
|
|
26
|
-
Feel free to submit your requests at [More Model Provider Support](https://github.com/lobehub/lobe-chat/discussions/
|
26
|
+
Feel free to submit your requests at [More Model Provider Support](https://github.com/lobehub/lobe-chat/discussions/6157).
|
@@ -43,7 +43,7 @@ We have implemented support for the following model service providers:
|
|
43
43
|
- **DeepSeek**: Integrated with the DeepSeek series models, an innovative AI startup from China, The product has been designed to provide a model that balances performance with price. [Learn more](https://www.deepseek.com/)
|
44
44
|
- **Qwen**: Integrated with the Qwen series models, including the latest **qwen-turbo**, **qwen-plus** and **qwen-max**. [Learn more](https://help.aliyun.com/zh/dashscope/developer-reference/model-introduction)
|
45
45
|
|
46
|
-
At the same time, we are also planning to support more model service providers, such as Replicate and Perplexity, to further enrich our service provider library. If you would like LobeChat to support your favorite service provider, feel free to join our [community discussion](https://github.com/lobehub/lobe-chat/discussions/
|
46
|
+
At the same time, we are also planning to support more model service providers, such as Replicate and Perplexity, to further enrich our service provider library. If you would like LobeChat to support your favorite service provider, feel free to join our [community discussion](https://github.com/lobehub/lobe-chat/discussions/6157).
|
47
47
|
|
48
48
|
## Local Model Support
|
49
49
|
|
@@ -43,7 +43,7 @@ tags:
|
|
43
43
|
- **DeepSeek**: 接入了 DeepSeek 的 AI 模型,包括最新的 **DeepSeek-V2**,提供兼顾性能与价格的模型。[了解更多](https://www.deepseek.com/)
|
44
44
|
- **Qwen (通义千问)**: 接入了 Qwen 的 AI 模型,包括最新的 **qwen-turbo**,**qwen-plus** 和 **qwen-max** 等模型。[了解更多](https://help.aliyun.com/zh/dashscope/developer-reference/model-introduction)
|
45
45
|
|
46
|
-
同时,我们也在计划支持更多的模型服务商,如 Replicate 和 Perplexity 等,以进一步丰富我们的服务商库。如果你希望让 LobeChat 支持你喜爱的服务商,欢迎加入我们的[社区讨论](https://github.com/lobehub/lobe-chat/discussions/
|
46
|
+
同时,我们也在计划支持更多的模型服务商,如 Replicate 和 Perplexity 等,以进一步丰富我们的服务商库。如果你希望让 LobeChat 支持你喜爱的服务商,欢迎加入我们的[社区讨论](https://github.com/lobehub/lobe-chat/discussions/6157)。
|
47
47
|
|
48
48
|
## 本地模型支持
|
49
49
|
|
package/locales/ar/setting.json
CHANGED
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "تفعيل ضبط قوة الاستدلال"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "كلما زادت القيمة،
|
208
|
-
"title": "
|
207
|
+
"desc": "كلما زادت القيمة، كانت المفردات أكثر تنوعًا؛ وكلما انخفضت القيمة، كانت المفردات أكثر بساطة ووضوحًا",
|
208
|
+
"title": "تنوع المفردات"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "عدد الرموز الأقصى المستخدمة في التفاعل الواحد",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "{{provider}} نموذج",
|
216
216
|
"title": "النموذج"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "إعدادات متقدمة"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "كلما زادت القيمة،
|
220
|
-
"title": "
|
222
|
+
"desc": "كلما زادت القيمة، زادت الميل إلى استخدام تعبيرات مختلفة، مما يتجنب تكرار المفاهيم؛ وكلما انخفضت القيمة، زادت الميل إلى استخدام المفاهيم أو السرد المتكرر، مما يجعل التعبير أكثر اتساقًا",
|
223
|
+
"title": "تنوع التعبير"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "كلما زادت القيمة، زادت قدرة الاستدلال، ولكن قد يؤدي ذلك إلى زيادة وقت الاستجابة واستهلاك التوكنات",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "قوة الاستدلال"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "كلما زادت القيمة،
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "كلما زادت القيمة، كانت الإجابات أكثر إبداعًا وخيالًا؛ وكلما انخفضت القيمة، كانت الإجابات أكثر دقة",
|
236
|
+
"title": "مستوى الإبداع",
|
237
|
+
"warning": "إذا كانت قيمة مستوى الإبداع مرتفعة جدًا، قد تحتوي المخرجات على تشويش"
|
235
238
|
},
|
236
239
|
"title": "إعدادات النموذج",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "عدد الاحتمالات التي يتم أخذها في الاعتبار، كلما زادت القيمة، زادت احتمالية قبول إجابات متعددة؛ وكلما انخفضت القيمة، زادت الميل لاختيار الإجابة الأكثر احتمالًا. لا يُنصح بتغييرها مع مستوى الإبداع",
|
242
|
+
"title": "مستوى الانفتاح الفكري"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "Активиране на настройките за интензивност на разсъжденията"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "Колкото
|
208
|
-
"title": "
|
207
|
+
"desc": "Колкото по-голяма е стойността, толкова по-богат и разнообразен е речникът; колкото по-ниска е стойността, толкова по-прост и обикновен е речникът.",
|
208
|
+
"title": "Богатство на речника"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "Максималният брой токени, използвани за всяко взаимодействие",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "{{provider}} модел",
|
216
216
|
"title": "Модел"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "Разширени параметри"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "Колкото
|
220
|
-
"title": "
|
222
|
+
"desc": "Колкото по-голяма е стойността, толкова по-склонен е към различни изрази, избягвайки повторение на концепции; колкото по-ниска е стойността, толкова по-склонен е да използва повторение на концепции или разкази, изразявайки по-голяма последователност.",
|
223
|
+
"title": "Разнообразие на изразите"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "Колкото по-висока е стойността, толкова по-силна е способността за разсъждение, но може да увеличи времето за отговор и консумацията на токени",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "Интензивност на разсъжденията"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "Колкото
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "Колкото по-голямо е числото, толкова по-креативни и въображаеми са отговорите; колкото по-малко е числото, толкова по-строги са отговорите",
|
236
|
+
"title": "Креативна активност",
|
237
|
+
"warning": "Ако стойността на креативната активност е твърде голяма, изходът може да съдържа грешки"
|
235
238
|
},
|
236
239
|
"title": "Настройки на модела",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "Колко възможности да се вземат предвид, по-голямата стойност приема повече възможни отговори; по-малката стойност предпочита най-вероятния отговор. Не се препоръчва да се променя заедно с креативната активност",
|
242
|
+
"title": "Отвореност на мисленето"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "Aktivieren Sie die Anpassung der Schlussfolgerungsintensität"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "Je höher der Wert, desto
|
208
|
-
"title": "
|
207
|
+
"desc": "Je höher der Wert, desto vielfältiger und abwechslungsreicher die Wortwahl; je niedriger der Wert, desto einfacher und schlichter die Wortwahl",
|
208
|
+
"title": "Wortvielfalt"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "Maximale Anzahl von Tokens, die pro Interaktion verwendet werden",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "{{provider}} Modell",
|
216
216
|
"title": "Modell"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "Erweiterte Parameter"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "Je höher der Wert, desto
|
220
|
-
"title": "
|
222
|
+
"desc": "Je höher der Wert, desto eher werden unterschiedliche Ausdrucksweisen bevorzugt, um Wiederholungen zu vermeiden; je niedriger der Wert, desto eher werden wiederholte Konzepte oder Erzählungen verwendet, was zu einer konsistenteren Ausdrucksweise führt",
|
223
|
+
"title": "Ausdrucksvielfalt"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "Je höher der Wert, desto stärker die Schlussfolgerungsfähigkeit, aber dies kann die Antwortzeit und den Tokenverbrauch erhöhen.",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "Schlussfolgerungsintensität"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "Je höher der Wert, desto
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "Je höher der Wert, desto kreativer und einfallsreicher die Antworten; je niedriger der Wert, desto strenger die Antworten",
|
236
|
+
"title": "Kreativitätsgrad",
|
237
|
+
"warning": "Ein zu hoher Kreativitätsgrad kann zu unverständlichen Ausgaben führen"
|
235
238
|
},
|
236
|
-
"title": "
|
239
|
+
"title": "Modell Einstellungen",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "Wie viele Möglichkeiten in Betracht gezogen werden, je höher der Wert, desto mehr mögliche Antworten werden akzeptiert; je niedriger der Wert, desto eher wird die wahrscheinlichste Antwort gewählt. Es wird nicht empfohlen, dies zusammen mit dem Kreativitätsgrad zu ändern",
|
242
|
+
"title": "Offenheit des Denkens"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "Enable Reasoning Effort Adjustment"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "The higher the value, the more
|
208
|
-
"title": "
|
207
|
+
"desc": "The higher the value, the more diverse and rich the vocabulary; the lower the value, the simpler and more straightforward the language.",
|
208
|
+
"title": "Vocabulary Richness"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "The maximum number of tokens used for each interaction",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "{{provider}} model",
|
216
216
|
"title": "Model"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "Advanced Parameters"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "The higher the value, the more
|
220
|
-
"title": "
|
222
|
+
"desc": "The higher the value, the more inclined to use different expressions and avoid concept repetition; the lower the value, the more inclined to use repeated concepts or narratives, resulting in more consistent expression.",
|
223
|
+
"title": "Expression Divergence"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "The higher the value, the stronger the reasoning ability, but it may increase response time and token consumption.",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "Reasoning Effort"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "The higher the value, the more
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "The higher the value, the more creative and imaginative the responses; the lower the value, the more rigorous the responses.",
|
236
|
+
"title": "Creativity Level",
|
237
|
+
"warning": "If the creativity level is set too high, the output may become garbled."
|
235
238
|
},
|
236
239
|
"title": "Model Settings",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "How many possibilities to consider; a higher value accepts more potential answers, while a lower value tends to choose the most likely answer. It is not recommended to change this alongside the creativity level.",
|
242
|
+
"title": "Openness to Ideas"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "Activar ajuste de intensidad de razonamiento"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "Cuanto mayor sea el valor, más
|
208
|
-
"title": "
|
207
|
+
"desc": "Cuanto mayor sea el valor, más rica y variada será la elección de palabras; cuanto menor sea el valor, más simples y directas serán las palabras.",
|
208
|
+
"title": "Riqueza del vocabulario"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "Número máximo de tokens utilizados en una interacción",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "{{provider}} modelo",
|
216
216
|
"title": "Modelo"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "Parámetros avanzados"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "Cuanto mayor sea el valor, más
|
220
|
-
"title": "
|
222
|
+
"desc": "Cuanto mayor sea el valor, más se inclinará hacia diferentes formas de expresión, evitando la repetición de conceptos; cuanto menor sea el valor, más se inclinará hacia el uso de conceptos o narrativas repetidas, expresando mayor consistencia.",
|
223
|
+
"title": "Diversidad de expresión"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "Cuanto mayor sea el valor, más fuerte será la capacidad de razonamiento, pero puede aumentar el tiempo de respuesta y el consumo de tokens.",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "Intensidad de razonamiento"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "Cuanto mayor sea el valor, más
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "Cuanto mayor sea el valor, más creativas e imaginativas serán las respuestas; cuanto menor sea el valor, más rigurosas serán las respuestas",
|
236
|
+
"title": "Nivel de creatividad",
|
237
|
+
"warning": "Un valor de creatividad demasiado alto puede generar salidas confusas"
|
235
238
|
},
|
236
239
|
"title": "Configuración del modelo",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "Cuántas posibilidades se consideran, cuanto mayor sea el valor, más respuestas posibles se aceptan; cuanto menor sea el valor, se tiende a elegir la respuesta más probable. No se recomienda cambiarlo junto con el nivel de creatividad",
|
242
|
+
"title": "Apertura mental"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|
@@ -204,8 +204,8 @@
|
|
204
204
|
"title": "فعالسازی تنظیم شدت استدلال"
|
205
205
|
},
|
206
206
|
"frequencyPenalty": {
|
207
|
-
"desc": "
|
208
|
-
"title": "
|
207
|
+
"desc": "هر چه مقدار بزرگتر باشد، واژگان متنوعتر و غنیتری استفاده میشود؛ هر چه مقدار کوچکتر باشد، واژگان سادهتر و عادیتر خواهند بود.",
|
208
|
+
"title": "تنوع واژگان"
|
209
209
|
},
|
210
210
|
"maxTokens": {
|
211
211
|
"desc": "حداکثر تعداد توکنهای استفادهشده در هر تعامل",
|
@@ -215,9 +215,12 @@
|
|
215
215
|
"desc": "مدل {{provider}}",
|
216
216
|
"title": "مدل"
|
217
217
|
},
|
218
|
+
"params": {
|
219
|
+
"title": "پارامترهای پیشرفته"
|
220
|
+
},
|
218
221
|
"presencePenalty": {
|
219
|
-
"desc": "
|
220
|
-
"title": "
|
222
|
+
"desc": "هر چه مقدار بزرگتر باشد، تمایل به استفاده از عبارات مختلف بیشتر میشود و از تکرار مفاهیم جلوگیری میکند؛ هر چه مقدار کوچکتر باشد، تمایل به استفاده از مفاهیم یا روایتهای تکراری بیشتر میشود و بیان یکدستتری خواهد داشت.",
|
223
|
+
"title": "گستردگی بیان"
|
221
224
|
},
|
222
225
|
"reasoningEffort": {
|
223
226
|
"desc": "هرچه مقدار بیشتر باشد، توانایی استدلال قویتر است، اما ممکن است زمان پاسخ و مصرف توکن را افزایش دهد",
|
@@ -229,14 +232,14 @@
|
|
229
232
|
"title": "شدت استدلال"
|
230
233
|
},
|
231
234
|
"temperature": {
|
232
|
-
"desc": "
|
233
|
-
"title": "
|
234
|
-
"
|
235
|
+
"desc": "هر چه عدد بزرگتر باشد، پاسخها خلاقانهتر و تخیلیتر خواهند بود؛ هر چه عدد کوچکتر باشد، پاسخها دقیقتر خواهند بود",
|
236
|
+
"title": "فعالیت خلاقانه",
|
237
|
+
"warning": "اگر عدد فعالیت خلاقانه بیش از حد بزرگ باشد، خروجی ممکن است دچار اختلال شود"
|
235
238
|
},
|
236
239
|
"title": "تنظیمات مدل",
|
237
240
|
"topP": {
|
238
|
-
"desc": "
|
239
|
-
"title": "
|
241
|
+
"desc": "چند احتمال را در نظر میگیرد، هر چه عدد بزرگتر باشد، پاسخهای بیشتری را میپذیرد؛ هر چه عدد کوچکتر باشد، تمایل به انتخاب پاسخهای محتملتر دارد. تغییر همزمان با فعالیت خلاقانه توصیه نمیشود",
|
242
|
+
"title": "باز بودن ذهن"
|
240
243
|
}
|
241
244
|
},
|
242
245
|
"settingPlugin": {
|