@lobehub/chat 1.1.17 → 1.1.18
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +25 -0
- package/package.json +1 -1
- package/src/config/modelProviders/google.ts +44 -61
package/CHANGELOG.md
CHANGED
|
@@ -2,6 +2,31 @@
|
|
|
2
2
|
|
|
3
3
|
# Changelog
|
|
4
4
|
|
|
5
|
+
### [Version 1.1.18](https://github.com/lobehub/lobe-chat/compare/v1.1.17...v1.1.18)
|
|
6
|
+
|
|
7
|
+
<sup>Released on **2024-07-01**</sup>
|
|
8
|
+
|
|
9
|
+
#### 💄 Styles
|
|
10
|
+
|
|
11
|
+
- **misc**: Add Gemini 1.5 stable version model & 2M context window.
|
|
12
|
+
|
|
13
|
+
<br/>
|
|
14
|
+
|
|
15
|
+
<details>
|
|
16
|
+
<summary><kbd>Improvements and Fixes</kbd></summary>
|
|
17
|
+
|
|
18
|
+
#### Styles
|
|
19
|
+
|
|
20
|
+
- **misc**: Add Gemini 1.5 stable version model & 2M context window, closes [#2744](https://github.com/lobehub/lobe-chat/issues/2744) ([a35143b](https://github.com/lobehub/lobe-chat/commit/a35143b))
|
|
21
|
+
|
|
22
|
+
</details>
|
|
23
|
+
|
|
24
|
+
<div align="right">
|
|
25
|
+
|
|
26
|
+
[](#readme-top)
|
|
27
|
+
|
|
28
|
+
</div>
|
|
29
|
+
|
|
5
30
|
### [Version 1.1.17](https://github.com/lobehub/lobe-chat/compare/v1.1.16...v1.1.17)
|
|
6
31
|
|
|
7
32
|
<sup>Released on **2024-07-01**</sup>
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@lobehub/chat",
|
|
3
|
-
"version": "1.1.
|
|
3
|
+
"version": "1.1.18",
|
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
|
5
5
|
"keywords": [
|
|
6
6
|
"framework",
|
|
@@ -5,48 +5,52 @@ import { ModelProviderCard } from '@/types/llm';
|
|
|
5
5
|
const Google: ModelProviderCard = {
|
|
6
6
|
chatModels: [
|
|
7
7
|
{
|
|
8
|
-
description: '
|
|
9
|
-
displayName: '
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
8
|
+
description: 'Fast and versatile multimodal model for scaling across diverse tasks',
|
|
9
|
+
displayName: 'Gemini 1.5 Flash',
|
|
10
|
+
enabled: true,
|
|
11
|
+
functionCall: true,
|
|
12
|
+
id: 'gemini-1.5-flash-latest',
|
|
13
|
+
maxOutput: 8192,
|
|
14
|
+
tokens: 1_048_576 + 8192,
|
|
15
|
+
vision: true,
|
|
14
16
|
},
|
|
15
17
|
{
|
|
16
|
-
description: '
|
|
17
|
-
displayName: '
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
maxOutput:
|
|
21
|
-
tokens:
|
|
18
|
+
description: 'Fast and versatile multimodal model for scaling across diverse tasks',
|
|
19
|
+
displayName: 'Gemini 1.5 Flash 001',
|
|
20
|
+
functionCall: true,
|
|
21
|
+
id: 'gemini-1.5-flash-001',
|
|
22
|
+
maxOutput: 8192,
|
|
23
|
+
tokens: 1_048_576 + 8192,
|
|
24
|
+
vision: true,
|
|
22
25
|
},
|
|
23
26
|
{
|
|
24
|
-
description: '
|
|
25
|
-
displayName: 'Gemini 1.
|
|
27
|
+
description: 'Mid-size multimodal model that supports up to 1 million tokens',
|
|
28
|
+
displayName: 'Gemini 1.5 Pro',
|
|
29
|
+
enabled: true,
|
|
26
30
|
functionCall: true,
|
|
27
|
-
id: 'gemini-pro',
|
|
28
|
-
maxOutput:
|
|
29
|
-
tokens:
|
|
31
|
+
id: 'gemini-1.5-pro-latest',
|
|
32
|
+
maxOutput: 8192,
|
|
33
|
+
tokens: 2_097_152 + 8192,
|
|
34
|
+
vision: true,
|
|
30
35
|
},
|
|
31
36
|
{
|
|
32
|
-
description: '
|
|
33
|
-
displayName: 'Gemini 1.
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
+
description: 'Mid-size multimodal model that supports up to 1 million tokens',
|
|
38
|
+
displayName: 'Gemini 1.5 Pro 001',
|
|
39
|
+
functionCall: true,
|
|
40
|
+
id: 'gemini-1.5-pro-001',
|
|
41
|
+
maxOutput: 8192,
|
|
42
|
+
tokens: 2_097_152 + 8192,
|
|
37
43
|
vision: true,
|
|
38
44
|
},
|
|
39
45
|
{
|
|
40
|
-
description: 'The best
|
|
41
|
-
displayName: 'Gemini 1.0 Pro
|
|
42
|
-
id: 'gemini-pro-
|
|
43
|
-
maxOutput:
|
|
44
|
-
tokens:
|
|
45
|
-
vision: true,
|
|
46
|
+
description: 'The best model for scaling across a wide range of tasks. This is the latest model.',
|
|
47
|
+
displayName: 'Gemini 1.0 Pro',
|
|
48
|
+
id: 'gemini-1.0-pro-latest',
|
|
49
|
+
maxOutput: 2048,
|
|
50
|
+
tokens: 30_720 + 2048,
|
|
46
51
|
},
|
|
47
52
|
{
|
|
48
|
-
description:
|
|
49
|
-
'The best model for scaling across a wide range of tasks. This is a stable model that supports tuning.',
|
|
53
|
+
description: 'The best model for scaling across a wide range of tasks. This is a stable model that supports tuning.',
|
|
50
54
|
displayName: 'Gemini 1.0 Pro 001 (Tuning)',
|
|
51
55
|
functionCall: true,
|
|
52
56
|
id: 'gemini-1.0-pro-001',
|
|
@@ -54,41 +58,12 @@ const Google: ModelProviderCard = {
|
|
|
54
58
|
tokens: 30_720 + 2048,
|
|
55
59
|
},
|
|
56
60
|
{
|
|
57
|
-
description:
|
|
58
|
-
'The best model for scaling across a wide range of tasks. Released April 9, 2024.',
|
|
61
|
+
description: 'The best model for scaling across a wide range of tasks. Released April 9, 2024.',
|
|
59
62
|
displayName: 'Gemini 1.0 Pro 002 (Tuning)',
|
|
60
63
|
id: 'gemini-1.0-pro-002',
|
|
61
64
|
maxOutput: 2048,
|
|
62
65
|
tokens: 30_720 + 2048,
|
|
63
66
|
},
|
|
64
|
-
{
|
|
65
|
-
description:
|
|
66
|
-
'The best model for scaling across a wide range of tasks. This is the latest model.',
|
|
67
|
-
displayName: 'Gemini 1.0 Pro Latest',
|
|
68
|
-
id: 'gemini-1.0-pro-latest',
|
|
69
|
-
maxOutput: 2048,
|
|
70
|
-
tokens: 30_720 + 2048,
|
|
71
|
-
},
|
|
72
|
-
{
|
|
73
|
-
description: 'Fast and versatile multimodal model for scaling across diverse tasks',
|
|
74
|
-
displayName: 'Gemini 1.5 Flash',
|
|
75
|
-
enabled: true,
|
|
76
|
-
functionCall: true,
|
|
77
|
-
id: 'gemini-1.5-flash-latest',
|
|
78
|
-
maxOutput: 8192,
|
|
79
|
-
tokens: 1_048_576 + 8192,
|
|
80
|
-
vision: true,
|
|
81
|
-
},
|
|
82
|
-
{
|
|
83
|
-
description: 'Mid-size multimodal model that supports up to 1 million tokens',
|
|
84
|
-
displayName: 'Gemini 1.5 Pro',
|
|
85
|
-
enabled: true,
|
|
86
|
-
functionCall: true,
|
|
87
|
-
id: 'gemini-1.5-pro-latest',
|
|
88
|
-
maxOutput: 8192,
|
|
89
|
-
tokens: 1_048_576 + 8192,
|
|
90
|
-
vision: true,
|
|
91
|
-
},
|
|
92
67
|
{
|
|
93
68
|
description: 'The most capable model for highly complex tasks',
|
|
94
69
|
displayName: 'Gemini 1.0 Ultra',
|
|
@@ -96,8 +71,16 @@ const Google: ModelProviderCard = {
|
|
|
96
71
|
maxOutput: 2048,
|
|
97
72
|
tokens: 32_768,
|
|
98
73
|
},
|
|
74
|
+
{
|
|
75
|
+
description: 'A legacy text-only model optimized for chat conversations',
|
|
76
|
+
displayName: 'PaLM 2 Chat (Legacy)',
|
|
77
|
+
id: 'chat-bison-001',
|
|
78
|
+
legacy: true,
|
|
79
|
+
maxOutput: 1024,
|
|
80
|
+
// tokens: 4096 + 1024, // none tokens test
|
|
81
|
+
}
|
|
99
82
|
],
|
|
100
|
-
checkModel: 'gemini-
|
|
83
|
+
checkModel: 'gemini-1.5-flash',
|
|
101
84
|
id: 'google',
|
|
102
85
|
name: 'Google',
|
|
103
86
|
proxyUrl: {
|