@lobehub/chat 0.147.3 → 0.147.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (35) hide show
  1. package/CHANGELOG.md +33 -0
  2. package/docs/self-hosting/advanced/sso-providers/auth0.mdx +2 -2
  3. package/docs/self-hosting/advanced/sso-providers/auth0.zh-CN.mdx +2 -2
  4. package/docs/self-hosting/advanced/sso-providers/authentik.mdx +2 -2
  5. package/docs/self-hosting/advanced/sso-providers/authentik.zh-CN.mdx +2 -2
  6. package/docs/self-hosting/advanced/sso-providers/github.mdx +1 -1
  7. package/docs/self-hosting/advanced/sso-providers/github.zh-CN.mdx +2 -2
  8. package/docs/self-hosting/advanced/sso-providers/microsoft-entra-id.mdx +1 -1
  9. package/docs/self-hosting/advanced/sso-providers/microsoft-entra-id.zh-CN.mdx +2 -2
  10. package/docs/self-hosting/advanced/sso-providers/zitadel.mdx +2 -2
  11. package/docs/self-hosting/advanced/sso-providers/zitadel.zh-CN.mdx +2 -2
  12. package/docs/self-hosting/examples/ollama.mdx +2 -0
  13. package/docs/self-hosting/platform/docker.mdx +1 -1
  14. package/docs/self-hosting/platform/docker.zh-CN.mdx +1 -1
  15. package/docs/self-hosting/platform/netlify.mdx +1 -1
  16. package/docs/self-hosting/platform/netlify.zh-CN.mdx +1 -1
  17. package/docs/self-hosting/platform/vercel.mdx +1 -1
  18. package/docs/self-hosting/platform/vercel.zh-CN.mdx +1 -1
  19. package/docs/usage/agents/custom-agent.mdx +2 -2
  20. package/docs/usage/agents/custom-agent.zh-CN.mdx +2 -2
  21. package/docs/usage/features/multi-ai-providers.mdx +1 -1
  22. package/docs/usage/features/multi-ai-providers.zh-CN.mdx +1 -1
  23. package/docs/usage/features/plugin-system.mdx +1 -1
  24. package/docs/usage/features/plugin-system.zh-CN.mdx +2 -2
  25. package/docs/usage/plugins/custom-plugin.mdx +1 -1
  26. package/docs/usage/plugins/custom-plugin.zh-CN.mdx +1 -1
  27. package/docs/usage/providers/ollama/gemma.mdx +3 -3
  28. package/docs/usage/providers/ollama/gemma.zh-CN.mdx +3 -3
  29. package/docs/usage/providers/ollama/qwen.mdx +3 -3
  30. package/docs/usage/providers/ollama/qwen.zh-CN.mdx +3 -3
  31. package/docs/usage/providers/ollama.mdx +3 -3
  32. package/docs/usage/providers/ollama.zh-CN.mdx +3 -3
  33. package/package.json +1 -1
  34. package/src/config/modelProviders/openrouter.ts +64 -0
  35. package/src/const/layoutTokens.ts +1 -1
package/CHANGELOG.md CHANGED
@@ -2,6 +2,39 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 0.147.4](https://github.com/lobehub/lobe-chat/compare/v0.147.3...v0.147.4)
6
+
7
+ <sup>Released on **2024-04-11**</sup>
8
+
9
+ #### 🐛 Bug Fixes
10
+
11
+ - **misc**: Add more builtin OpenRouter models.
12
+
13
+ #### 💄 Styles
14
+
15
+ - **misc**: Adjust minimum width value for DraggablePanel component.
16
+
17
+ <br/>
18
+
19
+ <details>
20
+ <summary><kbd>Improvements and Fixes</kbd></summary>
21
+
22
+ #### What's fixed
23
+
24
+ - **misc**: Add more builtin OpenRouter models, closes [#1973](https://github.com/lobehub/lobe-chat/issues/1973) ([0000b1a](https://github.com/lobehub/lobe-chat/commit/0000b1a))
25
+
26
+ #### Styles
27
+
28
+ - **misc**: Adjust minimum width value for DraggablePanel component, closes [#1901](https://github.com/lobehub/lobe-chat/issues/1901) ([a696d37](https://github.com/lobehub/lobe-chat/commit/a696d37))
29
+
30
+ </details>
31
+
32
+ <div align="right">
33
+
34
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
35
+
36
+ </div>
37
+
5
38
  ### [Version 0.147.3](https://github.com/lobehub/lobe-chat/compare/v0.147.2...v0.147.3)
6
39
 
7
40
  <sup>Released on **2024-04-11**</sup>
@@ -86,8 +86,8 @@ When deploying LobeChat, you need to configure the following environment variabl
86
86
  | `NEXTAUTH_URL` | Optional | The URL is used to specify the callback address for the execution of OAuth authentication in Auth.js. It needs to be set only when the default address is incorrect. `https://example.com/api/auth` |
87
87
 
88
88
  <Callout type={'tip'}>
89
- You can refer to the related variable details at [📘Environment Variables](/en/self-hosting/environment-variable#auth0).
90
-
89
+ You can refer to the related variable details at [📘Environment Variables](/en/docs/self-hosting/environment-variable#auth0).
90
+
91
91
  </Callout>
92
92
  </Steps>
93
93
 
@@ -81,8 +81,8 @@ http(s)://your-domain/api/auth/callback/auth0
81
81
  | `NEXTAUTH_URL` | 可选 | 该 URL 用于指定 Auth.js 在执行 OAuth 验证时的回调地址,当默认生成的重定向地址发生不正确时才需要设置。`https://example.com/api/auth` |
82
82
 
83
83
  <Callout type={'tip'}>
84
- 前往 [📘 环境变量](/zh/self-hosting/environment-variable#auth0) 可查阅相关变量详情。
85
-
84
+ 前往 [📘 环境变量](/zh/docs/self-hosting/environment-variable#auth0) 可查阅相关变量详情。
85
+
86
86
  </Callout>
87
87
  </Steps>
88
88
 
@@ -64,8 +64,8 @@ When deploying LobeChat, you need to configure the following environment variabl
64
64
  | `NEXTAUTH_URL` | Optional | This URL is used to specify the callback address for Auth.js when performing OAuth authentication. It only needs to be set when the default generated redirect address is incorrect. `https://example.com/api/auth` |
65
65
 
66
66
  <Callout type={'tip'}>
67
- Go to Environment Variables: /en/self-hosting/environment-variable#Authentik for details about the variables.
68
-
67
+ Go to Environment Variables: /en/docs/self-hosting/environment-variable#Authentik for details about the variables.
68
+
69
69
  </Callout>
70
70
  </Steps>
71
71
 
@@ -59,8 +59,8 @@ https://your-domain/api/auth/callback/authentik
59
59
  | `NEXTAUTH_URL` | 可选 | 该 URL 用于指定 Auth.js 在执行 OAuth 验证时的回调地址,当默认生成的重定向地址发生不正确时才需要设置。`https://example.com/api/auth` |
60
60
 
61
61
  <Callout type={'tip'}>
62
- 前往 [📘 环境变量](/zh/self-hosting/environment-variable#Authentik) 可查阅相关变量详情。
63
-
62
+ 前往 [📘 环境变量](/zh/docs/self-hosting/environment-variable#Authentik) 可查阅相关变量详情。
63
+
64
64
  </Callout>
65
65
  </Steps>
66
66
 
@@ -91,7 +91,7 @@ When deploying LobeChat, you need to configure the following environment variabl
91
91
  | `NEXTAUTH_URL` | Optional | This URL is used to specify the callback address for Auth.js when performing OAuth authentication. Only set it if the default generated redirect address is incorrect. `https://example.com/api/auth` |
92
92
 
93
93
  <Callout type={'tip'}>
94
- Go to [📘 Environment Variables](/en/self-hosting/environment-variable#Github) for detailed
94
+ Go to [📘 Environment Variables](/en/docs/self-hosting/environment-variable#Github) for detailed
95
95
  information on these variables.
96
96
  </Callout>
97
97
  </Steps>
@@ -87,8 +87,8 @@ tags:
87
87
  | `NEXTAUTH_URL` | 可选 | 该 URL 用于指定 Auth.js 在执行 OAuth 验证时的回调地址,当默认生成的重定向地址发生不正确时才需要设置。`https://example.com/api/auth` |
88
88
 
89
89
  <Callout type={'tip'}>
90
- 前往 [📘 环境变量](/zh/self-hosting/environment-variable#Github) 可查阅相关变量详情。
91
-
90
+ 前往 [📘 环境变量](/zh/docs/self-hosting/environment-variable#Github) 可查阅相关变量详情。
91
+
92
92
  </Callout>
93
93
  </Steps>
94
94
 
@@ -82,7 +82,7 @@ When deploying LobeChat, you need to configure the following environment variabl
82
82
 
83
83
  <Callout type={'tip'}>
84
84
  You can refer to [📘 environment
85
- variables](/en/self-hosting/environment-variable#microsoft-entra-id) for details on related
85
+ variables](/en/docs/self-hosting/environment-variable#microsoft-entra-id) for details on related
86
86
  variables.
87
87
  </Callout>
88
88
 
@@ -78,8 +78,8 @@ https://your-domain/api/auth/callback/azure-ad
78
78
  | `NEXTAUTH_URL` | 可选 | 该 URL 用于指定 Auth.js 在执行 OAuth 验证时的回调地址,当默认生成的重定向地址发生不正确时才需要设置。`https://example.com/api/auth` |
79
79
 
80
80
  <Callout type={'tip'}>
81
- 前往 [📘 环境变量](/zh/self-hosting/environment-variable#microsoft-entra-id) 可查阅相关变量详情。
82
-
81
+ 前往 [📘 环境变量](/zh/docs/self-hosting/environment-variable#microsoft-entra-id) 可查阅相关变量详情。
82
+
83
83
  </Callout>
84
84
  </Steps>
85
85
 
@@ -99,8 +99,8 @@ When deploying LobeChat, you need to configure the following environment variabl
99
99
  | `NEXTAUTH_URL` | Optional | The URL is used to specify the callback address for the execution of OAuth authentication in Auth.js. It needs to be set only when the default address is incorrect. `https://example.com/api/auth` |
100
100
 
101
101
  <Callout type={'tip'}>
102
- You can refer to the related variable details at [📘Environment Variables](/en/self-hosting/environment-variables/auth#zitadel).
103
-
102
+ You can refer to the related variable details at [📘Environment Variables](/en/docs/self-hosting/environment-variables/auth#zitadel).
103
+
104
104
  </Callout>
105
105
  </Steps>
106
106
 
@@ -95,8 +95,8 @@ http(s)://your-domain/api/auth/callback/zitadel
95
95
  | `NEXTAUTH_URL` | 可选 | 该 URL 用于指定 Auth.js 中执行 OAuth 认证的回调地址。仅当默认地址不正确时才需要设置。`https://example.com/api/auth` |
96
96
 
97
97
  <Callout type={'tip'}>
98
- 您可以在 [📘 环境变量](/zh/self-hosting/environment-variables/auth#zitadel) 中查看相关变量的详细信息。
99
-
98
+ 您可以在 [📘 环境变量](/zh/docs/self-hosting/environment-variables/auth#zitadel) 中查看相关变量的详细信息。
99
+
100
100
  </Callout>
101
101
  </Steps>
102
102
 
@@ -32,8 +32,10 @@ docker run -d -p 3210:3210 -e OLLAMA_PROXY_URL=http://host.docker.internal:11434
32
32
 
33
33
  Now, you can use LobeChat to converse with the local LLM.
34
34
 
35
+
35
36
  For more information on using Ollama in LobeChat, please refer to [Ollama Usage](/docs/usage/providers/ollama).
36
37
 
38
+
37
39
  ## Accessing Ollama from Non-Local Locations
38
40
 
39
41
  When you first initiate Ollama, it is configured to allow access only from the local machine. To enable access from other domains and set up port listening, you will need to adjust the environment variables accordingly.
@@ -64,7 +64,7 @@ Command explanation:
64
64
 
65
65
  - Replace `sk-xxxx` in the above command with your OpenAI API Key.
66
66
 
67
- - For the complete list of environment variables supported by LobeChat, please refer to the [Environment Variables](/zh/self-hosting/environment-ariable) section.
67
+ - For the complete list of environment variables supported by LobeChat, please refer to the [Environment Variables](/zh/docs/self-hosting/environment-ariable) section.
68
68
 
69
69
  <Callout type="tip">
70
70
  Since the official Docker image build takes about half an hour, if you see the "update available"
@@ -63,7 +63,7 @@ $ docker run -d -p 3210:3210 \
63
63
  - 使用你的 OpenAI API Key 替换上述命令中的 `sk-xxxx`
64
64
 
65
65
  <Callout type={'tip'}>
66
- LobeChat 支持的完整环境变量列表请参考 [📘 环境变量](/zh/self-hosting/environment-variables) 部分
66
+ LobeChat 支持的完整环境变量列表请参考 [📘 环境变量](/zh/docs/self-hosting/environment-variables) 部分
67
67
  </Callout>
68
68
 
69
69
  <Callout>
@@ -92,7 +92,7 @@ Taking OpenAI as an example, the environment variables you need to add are as fo
92
92
 
93
93
  <Callout type={'tip'}>
94
94
  For a complete list of environment variables supported by LobeChat, please refer to the [📘
95
- Environment Variables](/en/self-hosting/environment-variables)
95
+ Environment Variables](/en/docs/self-hosting/environment-variables)
96
96
  </Callout>
97
97
 
98
98
  Afteradding the variables, finally click "Deploy lobe-chat" to enter the deployment phase
@@ -86,7 +86,7 @@ tags:
86
86
  | `OPENAI_PROXY_URL` | 可选 | 如果你手动配置了 OpenAI 接口代理,可以使用此配置项来覆盖默认的 OpenAI API 请求基础 URL | `https://aihubmix.com/v1` ,默认值:`https://api.openai.com/v1` |
87
87
 
88
88
  <Callout type={'tip'}>
89
- LobeChat 支持的完整环境变量列表请参考 [📘 环境变量](/zh/self-hosting/environment-variables) 部分
89
+ LobeChat 支持的完整环境变量列表请参考 [📘 环境变量](/zh/docs/self-hosting/environment-variables) 部分
90
90
  </Callout>
91
91
 
92
92
  添加完成后,最后点击「Deploy lobe-chat」 进入部署阶段。
@@ -41,7 +41,7 @@ Vercel's assigned domain DNS may be polluted in some regions, so binding a custo
41
41
  If you have deployed your project using the one-click deployment steps mentioned above, you may find that you are always prompted with "updates available." This is because Vercel creates a new project for you by default instead of forking this project, which causes the inability to accurately detect updates.
42
42
 
43
43
  <Callout>
44
- We recommend following the [Self-Hosting Upstream Sync](/zh/self-hosting/upstream-sync) steps to
44
+ We recommend following the [Self-Hosting Upstream Sync](/zh/docs/self-hosting/upstream-sync) steps to
45
45
  Redeploy.
46
46
  </Callout>
47
47
 
@@ -40,7 +40,7 @@ Vercel 分配的域名 DNS 在某些区域被污染了,绑定自定义域名
40
40
  如果你根据上述中的一键部署步骤部署了自己的项目,你可能会发现总是被提示 “有可用更新”。这是因为 Vercel 默认为你创建新项目而非 fork 本项目,这将导致无法准确检测更新。
41
41
 
42
42
  <Callout>
43
- 我们建议按照 [📘 LobeChat 自部署保持更新](/zh/self-hosting/advanced/upstream-sync) 步骤重新部署。
43
+ 我们建议按照 [📘 LobeChat 自部署保持更新](/zh/docs/self-hosting/advanced/upstream-sync) 步骤重新部署。
44
44
  </Callout>
45
45
 
46
46
  [deploy-button-image]: https://vercel.com/button
@@ -52,7 +52,7 @@ When you need to handle specific tasks, you need to consider creating a custom a
52
52
  If you want to understand Prompt writing tips and common model parameter settings, you can continue to view:
53
53
 
54
54
  <Cards>
55
- <Card href={'/en/usage/agents/prompt'} title={'Prompt User Guide'} />
55
+ <Card href={'/en/docs/usage/agents/prompt'} title={'Prompt User Guide'} />
56
56
 
57
- <Card href={'/en/usage/agents/model'} title={'Large Language Model User Guide'} />
57
+ <Card href={'/en/docs/usage/agents/model'} title={'Large Language Model User Guide'} />
58
58
  </Cards>
@@ -47,7 +47,7 @@ tags:
47
47
  如果你希望理解 Prompt 编写技巧和常见的模型参数设置,可以继续查看:
48
48
 
49
49
  <Cards>
50
- <Card href={'/zh/usage/agents/prompt'} title={'Prompt 使用指南'} />
50
+ <Card href={'/zh/docs/usage/agents/prompt'} title={'Prompt 使用指南'} />
51
51
 
52
- <Card href={'/zh/usage/agents/model'} title={'大语言模型使用指南'} />
52
+ <Card href={'/zh/docs/usage/agents/model'} title={'大语言模型使用指南'} />
53
53
  </Cards>
@@ -55,4 +55,4 @@ At the same time, we are also planning to support more model service providers,
55
55
  src={'https://github.com/lobehub/lobe-chat/assets/28616219/ca9a21bc-ea6c-4c90-bf4a-fa53b4fb2b5c'}
56
56
  />
57
57
 
58
- To meet the specific needs of users, LobeChat also supports the use of local models based on [Ollama](https://ollama.ai), allowing users to flexibly use their own or third-party models. For more details, see [Local Model Support](/en/usage/features/local-llm).
58
+ To meet the specific needs of users, LobeChat also supports the use of local models based on [Ollama](https://ollama.ai), allowing users to flexibly use their own or third-party models. For more details, see [Local Model Support](/en/docs/usage/features/local-llm).
@@ -55,4 +55,4 @@ tags:
55
55
  src={'https://github.com/lobehub/lobe-chat/assets/28616219/ca9a21bc-ea6c-4c90-bf4a-fa53b4fb2b5c'}
56
56
  />
57
57
 
58
- 为了满足特定用户的需求,LobeChat 还基于 [Ollama](https://ollama.ai) 支持了本地模型的使用,让用户能够更灵活地使用自己的或第三方的模型,详见 [本地模型支持](/zh/usage/features/local-llm)。
58
+ 为了满足特定用户的需求,LobeChat 还基于 [Ollama](https://ollama.ai) 支持了本地模型的使用,让用户能够更灵活地使用自己的或第三方的模型,详见 [本地模型支持](/zh/docs/usage/features/local-llm)。
@@ -52,7 +52,7 @@ Learn more about [plugin usage](/en-US/usage/plugins/basic) by checking it out.
52
52
 
53
53
  <Callout>
54
54
  If you are interested in plugin development, please refer to our [📘 Plugin Development
55
- Guide](/en/usage/plugins/development) in the Wiki.
55
+ Guide](/en/docs/usage/plugins/development) in the Wiki.
56
56
  </Callout>
57
57
 
58
58
  - [lobe-chat-plugins][lobe-chat-plugins]: This is the plugin index for LobeChat. It retrieves the list of plugins from the index.json of this repository and displays them to the users.
@@ -27,7 +27,7 @@ LobeChat 的插件生态系统是其核心功能的重要扩展,它极大地
27
27
 
28
28
  此外,这些插件不仅局限于新闻聚合,还可以扩展到其他实用的功能,如快速检索文档、生成图片、获取 Bilibili 、Steam 等各种平台数据,以及与其他各式各样的第三方服务交互。
29
29
 
30
- 通过查看 [插件使用](/zh/usage/plugins/basic) 了解更多。
30
+ 通过查看 [插件使用](/zh/docs/usage/plugins/basic) 了解更多。
31
31
 
32
32
  <Callout type={'tip'}>
33
33
  为了帮助开发者更好地参与到这个生态中来,我们在提供了全面的开发资源。这包括详尽的组件开发文档、功能齐全的软件开发工具包(SDK),以及样板示例,这些都是为了简化开发过程,降低开发者的入门门槛。
@@ -40,7 +40,7 @@ LobeChat 的插件生态系统是其核心功能的重要扩展,它极大地
40
40
  ## 插件生态体系
41
41
 
42
42
  <Callout>
43
- 如果你对插件开发感兴趣,请在 Wiki 中查阅我们的 [📘 插件开发指南](/zh/usage/plugins/development)。
43
+ 如果你对插件开发感兴趣,请在 Wiki 中查阅我们的 [📘 插件开发指南](/zh/docs/usage/plugins/development)。
44
44
  </Callout>
45
45
 
46
46
  - [lobe-chat-plugins][lobe-chat-plugins]:这是 LobeChat 的插件索引。它从该仓库的 index.json 中获取插件列表并显示给用户。
@@ -34,4 +34,4 @@ If you want to try installing custom plugins on your own, you can use the follow
34
34
 
35
35
  ## Developing Custom Plugins
36
36
 
37
- If you wish to develop a LobeChat plugin on your own, feel free to refer to the [Plugin Development Guide](/en/usage/plugins/development) to expand the possibilities of your AI assistant!
37
+ If you wish to develop a LobeChat plugin on your own, feel free to refer to the [Plugin Development Guide](/en/docs/usage/plugins/development) to expand the possibilities of your AI assistant!
@@ -32,4 +32,4 @@ tags:
32
32
 
33
33
  ## 开发自定义插件
34
34
 
35
- 如果你希望自行开发一个 LobeChat 的插件,欢迎查阅 [插件开发指南](/zh/usage/plugins/development) 以扩展你的 AI 智能助手的可能性边界!
35
+ 如果你希望自行开发一个 LobeChat 的插件,欢迎查阅 [插件开发指南](/zh/docs/usage/plugins/development) 以扩展你的 AI 智能助手的可能性边界!
@@ -27,7 +27,7 @@ This document will guide you on how to use Google Gemma in LobeChat:
27
27
  <Steps>
28
28
  ### Install Ollama locally
29
29
 
30
- First, you need to install Ollama. For the installation process, please refer to the [Ollama usage documentation](/en/usage/providers/ollama).
30
+ First, you need to install Ollama. For the installation process, please refer to the [Ollama usage documentation](/en/docs/usage/providers/ollama).
31
31
 
32
32
  ### Pull Google Gemma model to local using Ollama
33
33
 
@@ -59,9 +59,9 @@ In the session page, open the model panel and then select the Gemma model.
59
59
 
60
60
  <Callout type={'info'}>
61
61
  If you do not see the Ollama provider in the model selection panel, please refer to [Integrating
62
- with Ollama](/en/self-hosting/examples/ollama) to learn how to enable the Ollama provider in
62
+ with Ollama](/en/docs/self-hosting/examples/ollama) to learn how to enable the Ollama provider in
63
63
  LobeChat.
64
-
64
+
65
65
  </Callout>
66
66
  </Steps>
67
67
 
@@ -27,7 +27,7 @@ tags:
27
27
  <Steps>
28
28
  ### 本地安装 Ollama
29
29
 
30
- 首先,你需要安装 Ollama,安装过程请查阅 [Ollama 使用文件](/zh/usage/providers/ollama)。
30
+ 首先,你需要安装 Ollama,安装过程请查阅 [Ollama 使用文件](/zh/docs/usage/providers/ollama)。
31
31
 
32
32
  ### 用 Ollama 拉取 Google Gemma 模型到本地
33
33
 
@@ -59,8 +59,8 @@ ollama pull gemma
59
59
 
60
60
  <Callout type={'info'}>
61
61
  如果你没有在模型选择面板中看到 Ollama 服务商,请查阅 [与 Ollama
62
- 集成](/zh/self-hosting/examples/ollama) 了解如何在 LobeChat 中开启 Ollama 服务商。
63
-
62
+ 集成](/zh/docs/self-hosting/examples/ollama) 了解如何在 LobeChat 中开启 Ollama 服务商。
63
+
64
64
  </Callout>
65
65
  </Steps>
66
66
 
@@ -26,7 +26,7 @@ Now, through the integration of LobeChat and [Ollama](https://ollama.com/), you
26
26
  <Steps>
27
27
  ## Local Installation of Ollama
28
28
 
29
- First, you need to install Ollama. For the installation process, please refer to the [Ollama Usage Document](/en/usage/providers/ollama).
29
+ First, you need to install Ollama. For the installation process, please refer to the [Ollama Usage Document](/en/docs/usage/providers/ollama).
30
30
 
31
31
  ## Pull the Qwen Model to Local with Ollama
32
32
 
@@ -61,8 +61,8 @@ In the LobeChat conversation page, open the model selection panel, and then sele
61
61
  />
62
62
 
63
63
  <Callout type={'info'}>
64
- If you do not see the Ollama provider in the model selection panel, please refer to [Integration with Ollama](/en/self-hosting/examples/ollama) to learn how to enable the Ollama provider in LobeChat.
65
-
64
+ If you do not see the Ollama provider in the model selection panel, please refer to [Integration with Ollama](/en/docs/self-hosting/examples/ollama) to learn how to enable the Ollama provider in LobeChat.
65
+
66
66
  </Callout>
67
67
  </Steps>
68
68
 
@@ -28,7 +28,7 @@ tags:
28
28
  <Steps>
29
29
  ### 本地安装 Ollama
30
30
 
31
- 首先,你需要安装 Ollama,安装过程请查阅 [Ollama 使用文件](/zh/usage/providers/ollama)。
31
+ 首先,你需要安装 Ollama,安装过程请查阅 [Ollama 使用文件](/zh/docs/usage/providers/ollama)。
32
32
 
33
33
  ### 用 Ollama 拉取 Qwen 模型到本地
34
34
 
@@ -58,8 +58,8 @@ ollama pull qwen:14b
58
58
 
59
59
  <Callout type={'info'}>
60
60
  如果你没有在模型选择面板中看到 Ollama 服务商,请查阅 [与 Ollama
61
- 集成](/zh/self-hosting/examples/ollama) 了解如何在 LobeChat 中开启 Ollama 服务商。
62
-
61
+ 集成](/zh/docs/self-hosting/examples/ollama) 了解如何在 LobeChat 中开启 Ollama 服务商。
62
+
63
63
  </Callout>
64
64
  </Steps>
65
65
 
@@ -52,7 +52,7 @@ First, you need to install Ollama, which supports macOS, Windows, and Linux syst
52
52
  ```
53
53
 
54
54
  </Tab>
55
-
55
+
56
56
  </Tabs>
57
57
 
58
58
  ### Pulling Models to Local with Ollama
@@ -76,9 +76,9 @@ Next, you can start conversing with the local LLM using LobeChat.
76
76
  />
77
77
 
78
78
  <Callout type={'info'}>
79
- You can visit [Integrating with Ollama](/en/self-hosting/examples/ollama) to learn how to deploy
79
+ You can visit [Integrating with Ollama](/en/docs/self-hosting/examples/ollama) to learn how to deploy
80
80
  LobeChat to meet the integration requirements with Ollama.
81
-
81
+
82
82
  </Callout>
83
83
  </Steps>
84
84
 
@@ -52,7 +52,7 @@ Ollama 是一款强大的本地运行大型语言模型(LLM)的框架,支
52
52
  ```
53
53
 
54
54
  </Tab>
55
-
55
+
56
56
  </Tabs>
57
57
 
58
58
  ### 用 Ollama 拉取模型到本地
@@ -76,8 +76,8 @@ Ollama 支持多种模型,你可以在 [Ollama Library](https://ollama.com/lib
76
76
  />
77
77
 
78
78
  <Callout type={'info'}>
79
- 你可以前往 [与 Ollama 集成](/zh/self-hosting/examples/ollama) 了解如何部署 LobeChat ,以满足与 Ollama 的集成需求。
80
-
79
+ 你可以前往 [与 Ollama 集成](/zh/docs/self-hosting/examples/ollama) 了解如何部署 LobeChat ,以满足与 Ollama 的集成需求。
80
+
81
81
  </Callout>
82
82
  </Steps>
83
83
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "0.147.3",
3
+ "version": "0.147.4",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -2,6 +2,14 @@ import { ModelProviderCard } from '@/types/llm';
2
2
 
3
3
  const OpenRouter: ModelProviderCard = {
4
4
  chatModels: [
5
+ {
6
+ displayName: 'Auto (best for prompt)',
7
+ enabled: true,
8
+ functionCall: false,
9
+ id: 'openrouter/auto',
10
+ tokens: 128_000,
11
+ vision: false,
12
+ },
5
13
  {
6
14
  displayName: 'Mistral 7B Instruct (free)',
7
15
  enabled: true,
@@ -42,6 +50,62 @@ const OpenRouter: ModelProviderCard = {
42
50
  tokens: 4096,
43
51
  vision: false,
44
52
  },
53
+ {
54
+ displayName: 'MythoMist 7B (free)',
55
+ enabled: true,
56
+ functionCall: false,
57
+ id: 'gryphe/mythomist-7b:free',
58
+ tokens: 32_768,
59
+ vision: false,
60
+ },
61
+ {
62
+ displayName: 'Toppy M 7B (free)',
63
+ enabled: true,
64
+ functionCall: false,
65
+ id: 'undi95/toppy-m-7b:free',
66
+ tokens: 4096,
67
+ vision: false,
68
+ },
69
+ {
70
+ displayName: 'Cinematika 7B (alpha) (free)',
71
+ enabled: true,
72
+ functionCall: false,
73
+ id: 'openrouter/cinematika-7b:free',
74
+ tokens: 8000,
75
+ vision: false,
76
+ },
77
+ {
78
+ displayName: 'RWKV v5 World 3B (free)',
79
+ enabled: true,
80
+ functionCall: false,
81
+ id: 'rwkv/rwkv-5-world-3b',
82
+ tokens: 10_000,
83
+ vision: false,
84
+ },
85
+ {
86
+ displayName: 'RWKV v5 3B AI Town (free)',
87
+ enabled: true,
88
+ functionCall: false,
89
+ id: 'recursal/rwkv-5-3b-ai-town',
90
+ tokens: 10_000,
91
+ vision: false,
92
+ },
93
+ {
94
+ displayName: 'RWKV v5: Eagle 7B (free)',
95
+ enabled: true,
96
+ functionCall: false,
97
+ id: 'recursal/eagle-7b',
98
+ tokens: 10_000,
99
+ vision: false,
100
+ },
101
+ {
102
+ displayName: 'Mistral: Mixtral 8x22B (base) (free)',
103
+ enabled: true,
104
+ functionCall: false,
105
+ id: 'mistralai/mixtral-8x22b',
106
+ tokens: 64_000,
107
+ vision: false,
108
+ },
45
109
  ],
46
110
  id: 'openrouter',
47
111
  };
@@ -8,7 +8,7 @@ export const CHAT_TEXTAREA_HEIGHT = 230;
8
8
  export const CHAT_TEXTAREA_HEIGHT_MOBILE = 108;
9
9
  export const CHAT_SIDEBAR_WIDTH = 280;
10
10
  export const MARKET_SIDEBAR_WIDTH = 400;
11
- export const FOLDER_WIDTH = 256;
11
+ export const FOLDER_WIDTH = 270;
12
12
  export const MAX_WIDTH = 1024;
13
13
  export const FORM_STYLE: FormProps = {
14
14
  itemMinWidth: 'max(30%,240px)',