@lobehub/chat 1.26.2 → 1.26.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md
CHANGED
@@ -2,6 +2,39 @@
|
|
2
2
|
|
3
3
|
# Changelog
|
4
4
|
|
5
|
+
### [Version 1.26.3](https://github.com/lobehub/lobe-chat/compare/v1.26.2...v1.26.3)
|
6
|
+
|
7
|
+
<sup>Released on **2024-10-28**</sup>
|
8
|
+
|
9
|
+
#### 🐛 Bug Fixes
|
10
|
+
|
11
|
+
- **misc**: Fix Huggingface API interrupting when the output exceeds 140 tokens.
|
12
|
+
|
13
|
+
#### 💄 Styles
|
14
|
+
|
15
|
+
- **misc**: Remove SenseChat-Vision model, due to model limitation.
|
16
|
+
|
17
|
+
<br/>
|
18
|
+
|
19
|
+
<details>
|
20
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
21
|
+
|
22
|
+
#### What's fixed
|
23
|
+
|
24
|
+
- **misc**: Fix Huggingface API interrupting when the output exceeds 140 tokens, closes [#4524](https://github.com/lobehub/lobe-chat/issues/4524) ([49acdd7](https://github.com/lobehub/lobe-chat/commit/49acdd7))
|
25
|
+
|
26
|
+
#### Styles
|
27
|
+
|
28
|
+
- **misc**: Remove SenseChat-Vision model, due to model limitation, closes [#4493](https://github.com/lobehub/lobe-chat/issues/4493) ([2aea033](https://github.com/lobehub/lobe-chat/commit/2aea033))
|
29
|
+
|
30
|
+
</details>
|
31
|
+
|
32
|
+
<div align="right">
|
33
|
+
|
34
|
+
[](#readme-top)
|
35
|
+
|
36
|
+
</div>
|
37
|
+
|
5
38
|
### [Version 1.26.2](https://github.com/lobehub/lobe-chat/compare/v1.26.1...v1.26.2)
|
6
39
|
|
7
40
|
<sup>Released on **2024-10-28**</sup>
|
@@ -17,7 +17,7 @@ tags:
|
|
17
17
|
## ChatGPT
|
18
18
|
|
19
19
|
- **gpt-3.5-turbo**: Currently the fastest generating ChatGPT model, it is faster but may sacrifice some text quality, with a context length of 4k.
|
20
|
-
- **gpt-4**: ChatGPT 4.0 has improved language understanding and generation capabilities compared to 3.5. It can better understand context and
|
20
|
+
- **gpt-4**: ChatGPT 4.0 has improved language understanding and generation capabilities compared to 3.5. It can better understand context and generate more accurate and natural responses. This is thanks to improvements in the GPT-4 model, including better language modeling and deeper semantic understanding, but it may be slower than other models, with a context length of 8k.
|
21
21
|
- **gpt-4-32k**: Similar to gpt-4, the context limit is increased to 32k tokens, with a higher cost.
|
22
22
|
|
23
23
|
## Concept of Model Parameters
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.26.
|
3
|
+
"version": "1.26.3",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -17,6 +17,8 @@ const SenseNova: ModelProviderCard = {
|
|
17
17
|
},
|
18
18
|
tokens: 131_072,
|
19
19
|
},
|
20
|
+
/*
|
21
|
+
// Not compatible with local mode (Not support Base64 Image)
|
20
22
|
{
|
21
23
|
description: '最新版本模型 (V5.5),16K上下文长度,支持多图的输入,全面实现模型基础能力优化,在对象属性识别、空间关系、动作事件识别、场景理解、情感识别、逻辑常识推理和文本理解生成上都实现了较大提升。',
|
22
24
|
displayName: 'SenseChat 5.5 Vision',
|
@@ -30,6 +32,7 @@ const SenseNova: ModelProviderCard = {
|
|
30
32
|
tokens: 16_384,
|
31
33
|
vision: true,
|
32
34
|
},
|
35
|
+
*/
|
33
36
|
{
|
34
37
|
description: '适用于快速问答、模型微调场景',
|
35
38
|
displayName: 'SenseChat 5.0 Turbo',
|
@@ -21,8 +21,10 @@ export const LobeHuggingFaceAI = LobeOpenAICompatibleFactory({
|
|
21
21
|
},
|
22
22
|
customClient: {
|
23
23
|
createChatCompletionStream: (client: HfInference, payload, instance) => {
|
24
|
+
const { max_tokens = 4096} = payload;
|
24
25
|
const hfRes = client.chatCompletionStream({
|
25
26
|
endpointUrl: instance.baseURL,
|
27
|
+
max_tokens: max_tokens,
|
26
28
|
messages: payload.messages,
|
27
29
|
model: payload.model,
|
28
30
|
stream: true,
|