@lobehub/chat 0.145.4 → 0.145.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +25 -0
- package/package.json +1 -1
- package/src/config/modelProviders/ollama.ts +49 -0
package/CHANGELOG.md
CHANGED
|
@@ -2,6 +2,31 @@
|
|
|
2
2
|
|
|
3
3
|
# Changelog
|
|
4
4
|
|
|
5
|
+
### [Version 0.145.5](https://github.com/lobehub/lobe-chat/compare/v0.145.4...v0.145.5)
|
|
6
|
+
|
|
7
|
+
<sup>Released on **2024-03-30**</sup>
|
|
8
|
+
|
|
9
|
+
#### 🐛 Bug Fixes
|
|
10
|
+
|
|
11
|
+
- **misc**: Add qwen api models patch in ollama.
|
|
12
|
+
|
|
13
|
+
<br/>
|
|
14
|
+
|
|
15
|
+
<details>
|
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
17
|
+
|
|
18
|
+
#### What's fixed
|
|
19
|
+
|
|
20
|
+
- **misc**: Add qwen api models patch in ollama, closes [#1630](https://github.com/lobehub/lobe-chat/issues/1630) ([a1e754c](https://github.com/lobehub/lobe-chat/commit/a1e754c))
|
|
21
|
+
|
|
22
|
+
</details>
|
|
23
|
+
|
|
24
|
+
<div align="right">
|
|
25
|
+
|
|
26
|
+
[](#readme-top)
|
|
27
|
+
|
|
28
|
+
</div>
|
|
29
|
+
|
|
5
30
|
### [Version 0.145.4](https://github.com/lobehub/lobe-chat/compare/v0.145.3...v0.145.4)
|
|
6
31
|
|
|
7
32
|
<sup>Released on **2024-03-29**</sup>
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@lobehub/chat",
|
|
3
|
-
"version": "0.145.
|
|
3
|
+
"version": "0.145.5",
|
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"framework",
|
|
@@ -153,6 +153,55 @@ const Ollama: ModelProviderCard = {
|
|
|
153
153
|
tokens: 4000,
|
|
154
154
|
vision: true,
|
|
155
155
|
},
|
|
156
|
+
// TODO: 在单独支持千问之后这些 Qwen 模型需要移动到千问的配置中
|
|
157
|
+
{
|
|
158
|
+
displayName: 'Qwen Plus',
|
|
159
|
+
functionCall: true,
|
|
160
|
+
hidden: true,
|
|
161
|
+
id: 'qwen-plus',
|
|
162
|
+
tokens: 30_000,
|
|
163
|
+
vision: false,
|
|
164
|
+
},
|
|
165
|
+
{
|
|
166
|
+
displayName: 'Qwen Turbo',
|
|
167
|
+
functionCall: true,
|
|
168
|
+
hidden: true,
|
|
169
|
+
id: 'qwen-turbo',
|
|
170
|
+
tokens: 6000,
|
|
171
|
+
vision: false,
|
|
172
|
+
},
|
|
173
|
+
{
|
|
174
|
+
displayName: 'Qwen Max',
|
|
175
|
+
functionCall: true,
|
|
176
|
+
hidden: true,
|
|
177
|
+
id: 'qwen-max',
|
|
178
|
+
tokens: 6000,
|
|
179
|
+
vision: false,
|
|
180
|
+
},
|
|
181
|
+
{
|
|
182
|
+
displayName: 'Qwen Max Long',
|
|
183
|
+
functionCall: true,
|
|
184
|
+
hidden: true,
|
|
185
|
+
id: 'qwen-max-longcontext',
|
|
186
|
+
tokens: 28_000,
|
|
187
|
+
vision: false,
|
|
188
|
+
},
|
|
189
|
+
{
|
|
190
|
+
displayName: 'Qwen VL Max',
|
|
191
|
+
functionCall: false,
|
|
192
|
+
hidden: true,
|
|
193
|
+
id: 'qwen-vl-max',
|
|
194
|
+
tokens: 6000,
|
|
195
|
+
vision: true,
|
|
196
|
+
},
|
|
197
|
+
{
|
|
198
|
+
displayName: 'Qwen VL Plus',
|
|
199
|
+
functionCall: false,
|
|
200
|
+
hidden: true,
|
|
201
|
+
id: 'qwen-vl-plus',
|
|
202
|
+
tokens: 30_000,
|
|
203
|
+
vision: true,
|
|
204
|
+
},
|
|
156
205
|
],
|
|
157
206
|
id: 'ollama',
|
|
158
207
|
};
|