@lobehub/chat 1.77.10 → 1.77.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.github/scripts/pr-comment.js +111 -55
- package/CHANGELOG.md +51 -0
- package/changelog/v1.json +18 -0
- package/docs/self-hosting/environment-variables/basic.mdx +39 -0
- package/docs/self-hosting/environment-variables/basic.zh-CN.mdx +40 -1
- package/locales/ar/hotkey.json +4 -0
- package/locales/ar/models.json +6 -0
- package/locales/bg-BG/hotkey.json +4 -0
- package/locales/bg-BG/models.json +6 -0
- package/locales/de-DE/hotkey.json +4 -0
- package/locales/de-DE/models.json +6 -0
- package/locales/en-US/hotkey.json +4 -0
- package/locales/en-US/models.json +6 -0
- package/locales/es-ES/hotkey.json +4 -0
- package/locales/es-ES/models.json +6 -0
- package/locales/fa-IR/hotkey.json +4 -0
- package/locales/fa-IR/models.json +6 -0
- package/locales/fr-FR/hotkey.json +4 -0
- package/locales/fr-FR/models.json +6 -0
- package/locales/it-IT/hotkey.json +4 -0
- package/locales/it-IT/models.json +6 -0
- package/locales/ja-JP/hotkey.json +4 -0
- package/locales/ja-JP/models.json +6 -0
- package/locales/ko-KR/hotkey.json +4 -0
- package/locales/ko-KR/models.json +6 -0
- package/locales/nl-NL/hotkey.json +4 -0
- package/locales/nl-NL/models.json +6 -0
- package/locales/pl-PL/hotkey.json +4 -0
- package/locales/pl-PL/models.json +6 -0
- package/locales/pt-BR/hotkey.json +4 -0
- package/locales/pt-BR/models.json +6 -0
- package/locales/ru-RU/hotkey.json +4 -0
- package/locales/ru-RU/models.json +6 -0
- package/locales/tr-TR/hotkey.json +4 -0
- package/locales/tr-TR/models.json +6 -0
- package/locales/vi-VN/hotkey.json +4 -0
- package/locales/vi-VN/models.json +6 -0
- package/locales/zh-CN/hotkey.json +4 -0
- package/locales/zh-CN/models.json +7 -1
- package/locales/zh-TW/hotkey.json +4 -0
- package/locales/zh-TW/models.json +6 -0
- package/package.json +1 -1
- package/packages/electron-client-ipc/src/dispatch.ts +29 -0
- package/packages/electron-client-ipc/src/events/index.ts +9 -2
- package/packages/electron-client-ipc/src/events/menu.ts +5 -0
- package/packages/electron-client-ipc/src/events/search.ts +4 -0
- package/packages/electron-client-ipc/src/events/system.ts +3 -0
- package/packages/electron-client-ipc/src/events/windows.ts +8 -0
- package/packages/electron-client-ipc/src/index.ts +1 -0
- package/packages/web-crawler/src/crawler.ts +3 -1
- package/scripts/migrateServerDB/index.ts +3 -2
- package/src/app/(backend)/webapi/assistant/store/route.ts +7 -1
- package/src/app/(backend)/webapi/plugin/store/route.ts +6 -1
- package/src/app/[variants]/(main)/(mobile)/me/(home)/__tests__/UserBanner.test.tsx +1 -0
- package/src/app/[variants]/(main)/_layout/Desktop/SideBar/TopActions.tsx +3 -2
- package/src/app/[variants]/(main)/_layout/Desktop/SideBar/index.tsx +24 -5
- package/src/app/[variants]/(main)/_layout/Desktop/Titlebar.tsx +27 -0
- package/src/app/[variants]/(main)/_layout/Desktop/index.tsx +32 -3
- package/src/app/[variants]/(main)/chat/(workspace)/@conversation/features/ChatList/WelcomeChatItem/InboxWelcome/AgentsSuggest.tsx +3 -0
- package/src/app/[variants]/(main)/chat/(workspace)/_layout/Desktop/ChatHeader/index.tsx +12 -2
- package/src/app/[variants]/(main)/chat/(workspace)/page.tsx +2 -1
- package/src/app/[variants]/(main)/settings/hotkey/features/HotkeySetting.tsx +15 -6
- package/src/app/[variants]/(main)/settings/hotkey/page.tsx +10 -1
- package/src/app/[variants]/(main)/settings/provider/features/ProviderConfig/index.tsx +3 -1
- package/src/app/desktop/devtools/page.tsx +37 -38
- package/src/config/aiModels/google.ts +25 -1
- package/src/config/aiModels/novita.ts +1 -32
- package/src/config/aiModels/ollama.ts +12 -1
- package/src/config/aiModels/qwen.ts +1 -1
- package/src/const/hotkeys.ts +10 -0
- package/src/features/Conversation/Error/index.tsx +10 -0
- package/src/features/DevPanel/PostgresViewer/SchemaSidebar/index.tsx +1 -1
- package/src/features/User/UserPanel/index.tsx +1 -1
- package/src/features/User/__tests__/PanelContent.test.tsx +1 -0
- package/src/libs/agent-runtime/error.ts +1 -0
- package/src/libs/agent-runtime/ollama/index.ts +16 -12
- package/src/libs/trpc/index.ts +8 -1
- package/src/libs/trpc/middleware/userAuth.ts +2 -3
- package/src/locales/default/hotkey.ts +4 -0
- package/src/server/globalConfig/genServerAiProviderConfig.ts +12 -3
- package/src/server/globalConfig/index.ts +6 -1
- package/src/server/globalConfig/parseSystemAgent.test.ts +56 -0
- package/src/server/globalConfig/parseSystemAgent.ts +25 -0
- package/src/server/modules/AssistantStore/index.ts +8 -0
- package/src/server/services/changelog/index.test.ts +2 -0
- package/src/server/services/changelog/index.ts +6 -2
- package/src/services/__tests__/chat.test.ts +3 -0
- package/src/services/electron/devtools.ts +1 -1
- package/src/store/user/slices/modelList/selectors/modelConfig.ts +4 -0
- package/src/styles/global.ts +14 -0
- package/src/types/hotkey.ts +1 -0
- package/src/utils/errorResponse.ts +2 -0
- package/tsconfig.json +1 -6
- package/packages/electron-client-ipc/src/events/devtools.ts +0 -6
- package/src/app/[variants]/(main)/settings/hotkey/index.tsx +0 -9
- package/src/types/electron.ts +0 -11
- package/src/utils/electron/dispatch.ts +0 -10
@@ -1,58 +1,69 @@
|
|
1
1
|
/**
|
2
2
|
* Generate PR comment with download links for desktop builds
|
3
|
+
* and handle comment creation/update logic
|
3
4
|
*/
|
4
5
|
module.exports = async ({ github, context, releaseUrl, version, tag }) => {
|
5
|
-
|
6
|
-
|
7
|
-
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
(
|
22
|
-
asset
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
27
|
-
|
28
|
-
|
29
|
-
(asset
|
30
|
-
|
31
|
-
|
32
|
-
|
33
|
-
|
34
|
-
|
35
|
-
|
36
|
-
|
37
|
-
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
42
|
-
|
43
|
-
|
44
|
-
|
45
|
-
|
46
|
-
|
47
|
-
|
48
|
-
|
49
|
-
|
50
|
-
|
51
|
-
|
52
|
-
|
53
|
-
|
6
|
+
// 用于识别构建评论的标识符
|
7
|
+
const COMMENT_IDENTIFIER = '<!-- DESKTOP-BUILD-COMMENT -->';
|
8
|
+
|
9
|
+
/**
|
10
|
+
* 生成评论内容
|
11
|
+
*/
|
12
|
+
const generateCommentBody = async () => {
|
13
|
+
try {
|
14
|
+
// Get release assets to create download links
|
15
|
+
const release = await github.rest.repos.getReleaseByTag({
|
16
|
+
owner: context.repo.owner,
|
17
|
+
repo: context.repo.repo,
|
18
|
+
tag,
|
19
|
+
});
|
20
|
+
|
21
|
+
// Organize assets by platform
|
22
|
+
const macAssets = release.data.assets.filter(
|
23
|
+
(asset) =>
|
24
|
+
((asset.name.includes('.dmg') || asset.name.includes('.zip')) &&
|
25
|
+
!asset.name.includes('.blockmap')) ||
|
26
|
+
(asset.name.includes('latest-mac') && asset.name.endsWith('.yml')),
|
27
|
+
);
|
28
|
+
|
29
|
+
const winAssets = release.data.assets.filter(
|
30
|
+
(asset) =>
|
31
|
+
(asset.name.includes('.exe') && !asset.name.includes('.blockmap')) ||
|
32
|
+
(asset.name.includes('latest-win') && asset.name.endsWith('.yml')),
|
33
|
+
);
|
34
|
+
|
35
|
+
const linuxAssets = release.data.assets.filter(
|
36
|
+
(asset) =>
|
37
|
+
(asset.name.includes('.AppImage') && !asset.name.includes('.blockmap')) ||
|
38
|
+
(asset.name.includes('latest-linux') && asset.name.endsWith('.yml')),
|
39
|
+
);
|
40
|
+
|
41
|
+
// Generate combined download table
|
42
|
+
let assetTable = '| Platform | File | Size |\n| --- | --- | --- |\n';
|
43
|
+
|
44
|
+
// Add macOS assets
|
45
|
+
macAssets.forEach((asset) => {
|
46
|
+
const sizeInMB = (asset.size / (1024 * 1024)).toFixed(2);
|
47
|
+
assetTable += `| macOS | [${asset.name}](${asset.browser_download_url}) | ${sizeInMB} MB |\n`;
|
48
|
+
});
|
49
|
+
|
50
|
+
// Add Windows assets
|
51
|
+
winAssets.forEach((asset) => {
|
52
|
+
const sizeInMB = (asset.size / (1024 * 1024)).toFixed(2);
|
53
|
+
assetTable += `| Windows | [${asset.name}](${asset.browser_download_url}) | ${sizeInMB} MB |\n`;
|
54
|
+
});
|
55
|
+
|
56
|
+
// Add Linux assets
|
57
|
+
linuxAssets.forEach((asset) => {
|
58
|
+
const sizeInMB = (asset.size / (1024 * 1024)).toFixed(2);
|
59
|
+
assetTable += `| Linux | [${asset.name}](${asset.browser_download_url}) | ${sizeInMB} MB |\n`;
|
60
|
+
});
|
61
|
+
|
62
|
+
return `${COMMENT_IDENTIFIER}
|
63
|
+
### 🚀 Desktop App Build Completed!
|
54
64
|
|
55
65
|
**Version**: \`${version}\`
|
66
|
+
**Build Time**: \`${new Date().toISOString()}\`
|
56
67
|
|
57
68
|
📦 [View All Build Artifacts](${releaseUrl})
|
58
69
|
|
@@ -64,17 +75,62 @@ ${assetTable}
|
|
64
75
|
> [!Warning]
|
65
76
|
>
|
66
77
|
> Note: This is a temporary build for testing purposes only.`;
|
67
|
-
|
68
|
-
|
69
|
-
|
70
|
-
|
78
|
+
} catch (error) {
|
79
|
+
console.error('Error generating PR comment:', error);
|
80
|
+
// Fallback to a simple comment if error occurs
|
81
|
+
return `${COMMENT_IDENTIFIER}
|
71
82
|
### 🚀 Desktop App Build Completed!
|
72
83
|
|
73
84
|
**Version**: \`${version}\`
|
85
|
+
**Build Time**: \`${new Date().toISOString()}\`
|
74
86
|
|
75
87
|
## 📦 [View All Build Artifacts](${releaseUrl})
|
76
88
|
|
77
89
|
> Note: This is a temporary build for testing purposes only.
|
78
|
-
|
79
|
-
|
90
|
+
`;
|
91
|
+
}
|
92
|
+
};
|
93
|
+
|
94
|
+
/**
|
95
|
+
* 查找并更新或创建PR评论
|
96
|
+
*/
|
97
|
+
const updateOrCreateComment = async () => {
|
98
|
+
// 生成评论内容
|
99
|
+
const body = await generateCommentBody();
|
100
|
+
|
101
|
+
// 查找我们之前可能创建的评论
|
102
|
+
const { data: comments } = await github.rest.issues.listComments({
|
103
|
+
issue_number: context.issue.number,
|
104
|
+
owner: context.repo.owner,
|
105
|
+
repo: context.repo.repo,
|
106
|
+
});
|
107
|
+
|
108
|
+
// 查找包含我们标识符的评论
|
109
|
+
const buildComment = comments.find((comment) => comment.body.includes(COMMENT_IDENTIFIER));
|
110
|
+
|
111
|
+
if (buildComment) {
|
112
|
+
// 如果找到现有评论,则更新它
|
113
|
+
await github.rest.issues.updateComment({
|
114
|
+
comment_id: buildComment.id,
|
115
|
+
owner: context.repo.owner,
|
116
|
+
repo: context.repo.repo,
|
117
|
+
body: body,
|
118
|
+
});
|
119
|
+
console.log(`已更新现有评论 ID: ${buildComment.id}`);
|
120
|
+
return { updated: true, id: buildComment.id };
|
121
|
+
} else {
|
122
|
+
// 如果没有找到现有评论,则创建新评论
|
123
|
+
const result = await github.rest.issues.createComment({
|
124
|
+
issue_number: context.issue.number,
|
125
|
+
owner: context.repo.owner,
|
126
|
+
repo: context.repo.repo,
|
127
|
+
body: body,
|
128
|
+
});
|
129
|
+
console.log(`已创建新评论 ID: ${result.data.id}`);
|
130
|
+
return { updated: false, id: result.data.id };
|
131
|
+
}
|
132
|
+
};
|
133
|
+
|
134
|
+
// 执行评论更新或创建
|
135
|
+
return await updateOrCreateComment();
|
80
136
|
};
|
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,57 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.77.12](https://github.com/lobehub/lobe-chat/compare/v1.77.11...v1.77.12)
|
6
|
+
|
7
|
+
<sup>Released on **2025-04-06**</sup>
|
8
|
+
|
9
|
+
#### 💄 Styles
|
10
|
+
|
11
|
+
- **misc**: Fix QVQ Max model, support default config for system agent and pre-merge some desktop code.
|
12
|
+
|
13
|
+
<br/>
|
14
|
+
|
15
|
+
<details>
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
17
|
+
|
18
|
+
#### Styles
|
19
|
+
|
20
|
+
- **misc**: Fix QVQ Max model, closes [#7289](https://github.com/lobehub/lobe-chat/issues/7289) ([540aaf6](https://github.com/lobehub/lobe-chat/commit/540aaf6))
|
21
|
+
- **misc**: Support default config for system agent and pre-merge some desktop code, closes [#7296](https://github.com/lobehub/lobe-chat/issues/7296) ([addea48](https://github.com/lobehub/lobe-chat/commit/addea48))
|
22
|
+
|
23
|
+
</details>
|
24
|
+
|
25
|
+
<div align="right">
|
26
|
+
|
27
|
+
[](#readme-top)
|
28
|
+
|
29
|
+
</div>
|
30
|
+
|
31
|
+
### [Version 1.77.11](https://github.com/lobehub/lobe-chat/compare/v1.77.10...v1.77.11)
|
32
|
+
|
33
|
+
<sup>Released on **2025-04-04**</sup>
|
34
|
+
|
35
|
+
#### 🐛 Bug Fixes
|
36
|
+
|
37
|
+
- **misc**: Fix fetch issue in offline mode and make jina crawler first.
|
38
|
+
|
39
|
+
<br/>
|
40
|
+
|
41
|
+
<details>
|
42
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
43
|
+
|
44
|
+
#### What's fixed
|
45
|
+
|
46
|
+
- **misc**: Fix fetch issue in offline mode and make jina crawler first, closes [#7288](https://github.com/lobehub/lobe-chat/issues/7288) ([6bed7a3](https://github.com/lobehub/lobe-chat/commit/6bed7a3))
|
47
|
+
|
48
|
+
</details>
|
49
|
+
|
50
|
+
<div align="right">
|
51
|
+
|
52
|
+
[](#readme-top)
|
53
|
+
|
54
|
+
</div>
|
55
|
+
|
5
56
|
### [Version 1.77.10](https://github.com/lobehub/lobe-chat/compare/v1.77.9...v1.77.10)
|
6
57
|
|
7
58
|
<sup>Released on **2025-04-03**</sup>
|
package/changelog/v1.json
CHANGED
@@ -1,4 +1,22 @@
|
|
1
1
|
[
|
2
|
+
{
|
3
|
+
"children": {
|
4
|
+
"improvements": [
|
5
|
+
"Fix QVQ Max model, support default config for system agent and pre-merge some desktop code."
|
6
|
+
]
|
7
|
+
},
|
8
|
+
"date": "2025-04-06",
|
9
|
+
"version": "1.77.12"
|
10
|
+
},
|
11
|
+
{
|
12
|
+
"children": {
|
13
|
+
"fixes": [
|
14
|
+
"Fix fetch issue in offline mode and make jina crawler first."
|
15
|
+
]
|
16
|
+
},
|
17
|
+
"date": "2025-04-04",
|
18
|
+
"version": "1.77.11"
|
19
|
+
},
|
2
20
|
{
|
3
21
|
"children": {},
|
4
22
|
"date": "2025-04-03",
|
@@ -71,6 +71,37 @@ Further reading:
|
|
71
71
|
|
72
72
|
- [\[RFC\] 022 - Default Assistant Parameters Configuration via Environment Variables](https://github.com/lobehub/lobe-chat/discussions/913)
|
73
73
|
|
74
|
+
### `SYSTEM_AGENT`
|
75
|
+
|
76
|
+
- Type: Optional
|
77
|
+
- Description: Used to configure models and providers for LobeChat system agents (such as topic generation, translation, etc.).
|
78
|
+
- Default value: `-`
|
79
|
+
- Example: `default=ollama/deepseek-v3` or `topic=openai/gpt-4,translation=anthropic/claude-1`
|
80
|
+
|
81
|
+
The `SYSTEM_AGENT` environment variable supports two configuration methods:
|
82
|
+
|
83
|
+
1. Use `default=provider/model` to set the same default configuration for all system agents
|
84
|
+
2. Configure specific system agents individually using the format `agent-name=provider/model`
|
85
|
+
|
86
|
+
Configuration details:
|
87
|
+
|
88
|
+
| Config Type | Format | Explanation |
|
89
|
+
| ------------------- | ----------------------------------------------- | ---------------------------------------------------------------------- |
|
90
|
+
| Default setting | `default=ollama/deepseek-v3` | Set deepseek-v3 from ollama as the default model for all system agents |
|
91
|
+
| Specific setting | `topic=openai/gpt-4` | Set a specific provider and model for topic generation |
|
92
|
+
| Mixed configuration | `default=ollama/deepseek-v3,topic=openai/gpt-4` | First set default values for all agents, then override specific agents |
|
93
|
+
|
94
|
+
Available system agents and their functions:
|
95
|
+
|
96
|
+
| System Agent | Key Name | Function Description |
|
97
|
+
| ------------------- | ----------------- | --------------------------------------------------------------------------------------------------- |
|
98
|
+
| Topic Generation | `topic` | Automatically generates topic names and summaries based on chat content |
|
99
|
+
| Translation | `translation` | Handles text translation between multiple languages |
|
100
|
+
| Metadata Generation | `agentMeta` | Generates descriptive information and metadata for assistants |
|
101
|
+
| History Compression | `historyCompress` | Compresses and organizes history for long conversations, optimizing context management |
|
102
|
+
| Query Rewrite | `queryRewrite` | Rewrites follow-up questions as standalone questions with context, improving conversation coherence |
|
103
|
+
| Thread Management | `thread` | Handles the creation and management of conversation threads |
|
104
|
+
|
74
105
|
### `FEATURE_FLAGS`
|
75
106
|
|
76
107
|
- Type: Optional
|
@@ -91,8 +122,16 @@ For specific content, please refer to the [Feature Flags](/docs/self-hosting/adv
|
|
91
122
|
If you're using Docker Desktop on Windows or macOS, it relies on a virtual machine. In this setup,
|
92
123
|
`localhost` / `127.0.0.1` refers to the localhost of the container itself. In such cases, please
|
93
124
|
try using `host.docker.internal` instead of `localhost`.
|
125
|
+
Use `http://user:password@127.0.0.1:7890` to connect to an authenticated proxy server.
|
94
126
|
</Callout>
|
95
127
|
|
128
|
+
### `ENABLE_PROXY_DNS`
|
129
|
+
|
130
|
+
- Type: Optional
|
131
|
+
- Description: Used to control whether to send DNS queries to the proxy server. When configured as `0`, all DNS queries are resolved locally. If you encounter API access failures or timeouts in your network environment, try setting this option to `1`.
|
132
|
+
- Default: `0`
|
133
|
+
- Example: `1` or `0`
|
134
|
+
|
96
135
|
### `SSRF_ALLOW_PRIVATE_IP_ADDRESS`
|
97
136
|
|
98
137
|
- Type: Optional
|
@@ -67,6 +67,37 @@ LobeChat 在部署时提供了一些额外的配置项,你可以使用环境
|
|
67
67
|
|
68
68
|
- [\[RFC\] 022 - 环境变量配置默认助手参数](https://github.com/lobehub/lobe-chat/discussions/913)
|
69
69
|
|
70
|
+
### `SYSTEM_AGENT`
|
71
|
+
|
72
|
+
- 类型:可选
|
73
|
+
- 描述:用于配置 LobeChat 系统助手(如主题生成、翻译等功能)的模型和供应商。
|
74
|
+
- 默认值:`-`
|
75
|
+
- 示例:`default=ollama/deepseek-v3` 或 `topic=openai/gpt-4,translation=anthropic/claude-1`
|
76
|
+
|
77
|
+
`SYSTEM_AGENT` 环境变量支持两种配置方式:
|
78
|
+
|
79
|
+
1. 使用 `default=供应商/模型` 为所有系统助手设置相同的默认配置
|
80
|
+
2. 针对特定的系统助手进行单独配置,格式为 `助手名称=供应商/模型`
|
81
|
+
|
82
|
+
配置项说明:
|
83
|
+
|
84
|
+
| 配置项 | 格式 | 解释 |
|
85
|
+
| ---- | ----------------------------------------------- | ----------------------------------- |
|
86
|
+
| 默认设置 | `default=ollama/deepseek-v3` | 为所有系统助手设置默认模型为 ollama 的 deepseek-v3 |
|
87
|
+
| 特定设置 | `topic=openai/gpt-4` | 为主题生成设置特定的供应商和模型 |
|
88
|
+
| 混合配置 | `default=ollama/deepseek-v3,topic=openai/gpt-4` | 先为所有助手设置默认值,然后针对特定助手进行覆盖 |
|
89
|
+
|
90
|
+
可配置的系统助手及其作用:
|
91
|
+
|
92
|
+
| 系统助手 | 键名 | 作用描述 |
|
93
|
+
| ------- | ----------------- | --------------------------- |
|
94
|
+
| 主题生成 | `topic` | 根据聊天内容自动生成主题名称和摘要 |
|
95
|
+
| 翻译 | `translation` | 文本翻译使用的助手 |
|
96
|
+
| 元数据生成 | `agentMeta` | 为助手生成描述性信息和元数据 |
|
97
|
+
| 历史记录压缩 | `historyCompress` | 压缩和整理长对话的历史记录,优化上下文管理 |
|
98
|
+
| 知识库查询重写 | `queryRewrite` | 将后续问题改写为包含上下文的独立问题,提升对话的连贯性 |
|
99
|
+
| 分支对话 | `thread` | 自定生成分支对话的标题 |
|
100
|
+
|
70
101
|
### `FEATURE_FLAGS`
|
71
102
|
|
72
103
|
- 类型:可选
|
@@ -85,9 +116,17 @@ LobeChat 在部署时提供了一些额外的配置项,你可以使用环境
|
|
85
116
|
|
86
117
|
<Callout type="info">
|
87
118
|
`Docker Desktop` 在 `Windows `和 `macOS `上走的是虚拟机方案,如果是 `localhost` / `127.0.0.1`
|
88
|
-
是走到自身容器的 `localhost`,此时请尝试用 `host.docker.internal` 替代 `localhost
|
119
|
+
是走到自身容器的 `localhost`,此时请尝试用 `host.docker.internal` 替代 `localhost`。
|
120
|
+
使用 `http://user:password@127.0.0.1:7890` 来连接到带认证的代理服务器。
|
89
121
|
</Callout>
|
90
122
|
|
123
|
+
### `ENABLE_PROXY_DNS`
|
124
|
+
|
125
|
+
- 类型:可选
|
126
|
+
- 描述:用于控制是否将DNS发送到代理服务器,配置为 `0` 时所有 DNS 查询在本地完成,当你的网络环境无法访问 API 或访问超时,请尝试将该项配置为 `1`。
|
127
|
+
- 默认值:`0`
|
128
|
+
- 示例:`1` or `0`
|
129
|
+
|
91
130
|
### `SSRF_ALLOW_PRIVATE_IP_ADDRESS`
|
92
131
|
|
93
132
|
- 类型:可选
|
package/locales/ar/hotkey.json
CHANGED
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "عرض جميع تعليمات استخدام الاختصارات",
|
16
16
|
"title": "فتح مساعدة الاختصارات"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "فتح صفحة إعدادات التطبيق",
|
20
|
+
"title": "إعدادات التطبيق"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "إعادة توليد آخر رسالة",
|
20
24
|
"title": "إعادة توليد الرسالة"
|
package/locales/ar/models.json
CHANGED
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) يوفر قدرة أساسية على معالجة التعليمات، مناسب للتطبيقات الخفيفة."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "جيمّا 3 27B هو نموذج لغوي مفتوح المصدر من جوجل، وقد وضع معايير جديدة من حيث الكفاءة والأداء."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "نموذج GPT 3.5 Turbo، مناسب لمجموعة متنوعة من مهام توليد وفهم النصوص، يشير حاليًا إلى gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "نموذج QVQ هو نموذج بحث تجريبي تم تطويره بواسطة فريق Qwen، يركز على تعزيز قدرات الاستدلال البصري، خاصة في مجال الاستدلال الرياضي."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "نموذج الاستدلال البصري QVQ، يدعم الإدخال البصري وإخراج سلسلة التفكير، وقد أظهر قدرات أقوى في الرياضيات، البرمجة، التحليل البصري، الإبداع، والمهام العامة."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "نموذج كود Qwen الشامل."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "Прегледайте инструкциите за използване на всички клавишни комбинации",
|
16
16
|
"title": "Отворете помощта за клавишни комбинации"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Отворете страницата с настройки на приложението",
|
20
|
+
"title": "Настройки на приложението"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Прегенерирайте последното съобщение",
|
20
24
|
"title": "Прегенериране на съобщение"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) предлага основни способности за обработка на инструкции, подходящи за леки приложения."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B е отворен езиков модел на Google, който поставя нови стандарти за ефективност и производителност."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo, подходящ за различни задачи по генериране и разбиране на текст, в момента сочи към gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "QVQ моделът е експериментален изследователски модел, разработен от екипа на Qwen, фокусиран върху повишаване на визуалните способности за разсъждение, особено в областта на математическото разсъждение."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "Моделът за визуално разсъждение QVQ на Tongyi Qianwen поддържа визуален вход и изход на мисловни вериги, демонстрирайки по-силни способности в математика, програмиране, визуален анализ, творчество и общи задачи."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Модел за кодиране Qwen с общо предназначение."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "Anleitung zur Verwendung aller Tastenkombinationen anzeigen",
|
16
16
|
"title": "Tastenkombinationshilfe öffnen"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Öffnen Sie die Anwendungseinstellungen",
|
20
|
+
"title": "Anwendungseinstellungen"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Die letzte Nachricht neu generieren",
|
20
24
|
"title": "Nachricht neu generieren"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) bietet grundlegende Anweisungsverarbeitungsfähigkeiten und eignet sich für leichte Anwendungen."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B ist ein Open-Source-Sprachmodell von Google, das neue Maßstäbe in Bezug auf Effizienz und Leistung setzt."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo eignet sich für eine Vielzahl von Textgenerierungs- und Verständnisaufgaben. Derzeit verweist es auf gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "Das QVQ-Modell ist ein experimentelles Forschungsmodell, das vom Qwen-Team entwickelt wurde und sich auf die Verbesserung der visuellen Schlussfolgerungsfähigkeiten konzentriert, insbesondere im Bereich der mathematischen Schlussfolgerungen."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "Das QVQ-Vision-Reasoning-Modell von Tongyi Qianwen unterstützt visuelle Eingaben und Denkketten-Ausgaben und zeigt in Mathematik, Programmierung, visueller Analyse, Kreativität und allgemeinen Aufgaben eine stärkere Leistungsfähigkeit."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Tongyi Qianwen Code-Modell."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "View instructions for all keyboard shortcuts",
|
16
16
|
"title": "Open Hotkey Help"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Open the application settings page",
|
20
|
+
"title": "Application Settings"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Regenerate the last message",
|
20
24
|
"title": "Regenerate Message"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) provides basic instruction processing capabilities, suitable for lightweight applications."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B is an open-source language model from Google that sets new standards in efficiency and performance."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo is suitable for various text generation and understanding tasks. Currently points to gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "The QVQ model is an experimental research model developed by the Qwen team, focusing on enhancing visual reasoning capabilities, particularly in the field of mathematical reasoning."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "The Tongyi Qianwen QVQ visual reasoning model supports visual input and chain-of-thought output, demonstrating enhanced capabilities in mathematics, programming, visual analysis, creative tasks, and general tasks."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Tongyi Qianwen code model."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "Ver las instrucciones de uso de todos los atajos de teclado",
|
16
16
|
"title": "Abrir ayuda de atajos de teclado"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Abrir la página de configuración de la aplicación",
|
20
|
+
"title": "Configuración de la aplicación"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Regenerar el último mensaje",
|
20
24
|
"title": "Regenerar mensaje"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) ofrece capacidades básicas de procesamiento de instrucciones, adecuado para aplicaciones ligeras."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B es un modelo de lenguaje de código abierto de Google, que establece nuevos estándares en eficiencia y rendimiento."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo, adecuado para diversas tareas de generación y comprensión de texto, actualmente apunta a gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "El modelo QVQ es un modelo de investigación experimental desarrollado por el equipo de Qwen, enfocado en mejorar la capacidad de razonamiento visual, especialmente en el ámbito del razonamiento matemático."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "El modelo de razonamiento visual QVQ de Tongyi Qianwen admite entradas visuales y salidas de cadena de pensamiento, mostrando una mayor capacidad en matemáticas, programación, análisis visual, creación y tareas generales."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Modelo de código Qwen de Tongyi."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "مشاهده تمام توضیحات استفاده از کلیدهای میانبر",
|
16
16
|
"title": "باز کردن راهنمای کلیدهای میانبر"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "صفحه تنظیمات برنامه را باز کنید",
|
20
|
+
"title": "تنظیمات برنامه"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "آخرین پیام را دوباره تولید کنید",
|
20
24
|
"title": "تولید مجدد پیام"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) توانایی پردازش دستورات پایه را فراهم میکند و برای برنامههای سبک مناسب است."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B یک مدل زبان متن باز از گوگل است که استانداردهای جدیدی را در زمینه کارایی و عملکرد تعیین کرده است."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 توربو، مناسب برای انواع وظایف تولید و درک متن، در حال حاضر به gpt-3.5-turbo-0125 اشاره میکند"
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "مدل QVQ یک مدل تحقیقاتی تجربی است که توسط تیم Qwen توسعه یافته و بر بهبود توانایی استدلال بصری، بهویژه در زمینه استدلال ریاضی تمرکز دارد."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "مدل استدلال بصری QVQ با قابلیت ورودی بصری و خروجی زنجیره تفکر، در ریاضیات، برنامهنویسی، تحلیل بصری، خلاقیت و وظایف عمومی عملکرد بهتری از خود نشان داده است."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "مدل کد Qwen با قابلیتهای جامع."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "Voir les instructions d'utilisation de tous les raccourcis",
|
16
16
|
"title": "Ouvrir l'aide des raccourcis"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Ouvrir la page des paramètres de l'application",
|
20
|
+
"title": "Paramètres de l'application"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Régénérer le dernier message",
|
20
24
|
"title": "Régénérer le message"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) offre des capacités de traitement d'instructions de base, adapté aux applications légères."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B est un modèle de langage open source de Google, qui a établi de nouvelles normes en matière d'efficacité et de performance."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo, adapté à diverses tâches de génération et de compréhension de texte, pointe actuellement vers gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "Le modèle QVQ est un modèle de recherche expérimental développé par l'équipe Qwen, axé sur l'amélioration des capacités de raisonnement visuel, en particulier dans le domaine du raisonnement mathématique."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "Le modèle de raisonnement visuel QVQ de Tongyi Qianwen prend en charge les entrées visuelles et les sorties de chaînes de pensée, montrant de meilleures capacités dans les domaines des mathématiques, de la programmation, de l'analyse visuelle, de la création et des tâches générales."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Modèle de code Qwen universel."
|
1597
1603
|
},
|
@@ -15,6 +15,10 @@
|
|
15
15
|
"desc": "Visualizza le istruzioni per l'uso di tutte le scorciatoie da tastiera",
|
16
16
|
"title": "Apri aiuto scorciatoie"
|
17
17
|
},
|
18
|
+
"openSettings": {
|
19
|
+
"desc": "Apri la pagina delle impostazioni dell'app",
|
20
|
+
"title": "Impostazioni dell'app"
|
21
|
+
},
|
18
22
|
"regenerateMessage": {
|
19
23
|
"desc": "Rigenera l'ultimo messaggio",
|
20
24
|
"title": "Rigenera messaggio"
|
@@ -1010,6 +1010,9 @@
|
|
1010
1010
|
"google/gemma-2b-it": {
|
1011
1011
|
"description": "Gemma Instruct (2B) offre capacità di elaborazione di istruzioni di base, adatta per applicazioni leggere."
|
1012
1012
|
},
|
1013
|
+
"google/gemma-3-27b-it": {
|
1014
|
+
"description": "Gemma 3 27B è un modello linguistico open source di Google, che ha stabilito nuovi standard in termini di efficienza e prestazioni."
|
1015
|
+
},
|
1013
1016
|
"gpt-3.5-turbo": {
|
1014
1017
|
"description": "GPT 3.5 Turbo, adatto a una varietà di compiti di generazione e comprensione del testo, attualmente punta a gpt-3.5-turbo-0125."
|
1015
1018
|
},
|
@@ -1592,6 +1595,9 @@
|
|
1592
1595
|
"qvq-72b-preview": {
|
1593
1596
|
"description": "Il modello QVQ è un modello di ricerca sperimentale sviluppato dal team Qwen, focalizzato sul miglioramento delle capacità di ragionamento visivo, in particolare nel campo del ragionamento matematico."
|
1594
1597
|
},
|
1598
|
+
"qvq-max-latest": {
|
1599
|
+
"description": "Il modello di ragionamento visivo QVQ di Tongyi Qianwen supporta input visivi e output di catene di pensiero, dimostrando capacità superiori in matematica, programmazione, analisi visiva, creazione e compiti generali."
|
1600
|
+
},
|
1595
1601
|
"qwen-coder-plus-latest": {
|
1596
1602
|
"description": "Modello di codice Qwen di Tongyi."
|
1597
1603
|
},
|