@lobehub/chat 1.1.17 → 1.1.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,31 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ### [Version 1.1.18](https://github.com/lobehub/lobe-chat/compare/v1.1.17...v1.1.18)
6
+
7
+ <sup>Released on **2024-07-01**</sup>
8
+
9
+ #### 💄 Styles
10
+
11
+ - **misc**: Add Gemini 1.5 stable version model & 2M context window.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### Styles
19
+
20
+ - **misc**: Add Gemini 1.5 stable version model & 2M context window, closes [#2744](https://github.com/lobehub/lobe-chat/issues/2744) ([a35143b](https://github.com/lobehub/lobe-chat/commit/a35143b))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
5
30
  ### [Version 1.1.17](https://github.com/lobehub/lobe-chat/compare/v1.1.16...v1.1.17)
6
31
 
7
32
  <sup>Released on **2024-07-01**</sup>
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lobehub/chat",
3
- "version": "1.1.17",
3
+ "version": "1.1.18",
4
4
  "description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
5
5
  "keywords": [
6
6
  "framework",
@@ -5,48 +5,52 @@ import { ModelProviderCard } from '@/types/llm';
5
5
  const Google: ModelProviderCard = {
6
6
  chatModels: [
7
7
  {
8
- description: 'A legacy text-only model optimized for chat conversations',
9
- displayName: 'PaLM 2 Chat (Legacy)',
10
- id: 'chat-bison-001',
11
- legacy: true,
12
- maxOutput: 1024,
13
- // tokens: 4096 + 1024, // none tokens test
8
+ description: 'Fast and versatile multimodal model for scaling across diverse tasks',
9
+ displayName: 'Gemini 1.5 Flash',
10
+ enabled: true,
11
+ functionCall: true,
12
+ id: 'gemini-1.5-flash-latest',
13
+ maxOutput: 8192,
14
+ tokens: 1_048_576 + 8192,
15
+ vision: true,
14
16
  },
15
17
  {
16
- description: 'A legacy model that understands text and generates text as an output',
17
- displayName: 'PaLM 2 (Legacy)',
18
- id: 'text-bison-001',
19
- legacy: true,
20
- maxOutput: 1024,
21
- tokens: 8196 + 1024,
18
+ description: 'Fast and versatile multimodal model for scaling across diverse tasks',
19
+ displayName: 'Gemini 1.5 Flash 001',
20
+ functionCall: true,
21
+ id: 'gemini-1.5-flash-001',
22
+ maxOutput: 8192,
23
+ tokens: 1_048_576 + 8192,
24
+ vision: true,
22
25
  },
23
26
  {
24
- description: 'The best model for scaling across a wide range of tasks',
25
- displayName: 'Gemini 1.0 Pro',
27
+ description: 'Mid-size multimodal model that supports up to 1 million tokens',
28
+ displayName: 'Gemini 1.5 Pro',
29
+ enabled: true,
26
30
  functionCall: true,
27
- id: 'gemini-pro',
28
- maxOutput: 2048,
29
- tokens: 30_720 + 2048,
31
+ id: 'gemini-1.5-pro-latest',
32
+ maxOutput: 8192,
33
+ tokens: 2_097_152 + 8192,
34
+ vision: true,
30
35
  },
31
36
  {
32
- description: 'The best image understanding model to handle a broad range of applications',
33
- displayName: 'Gemini 1.0 Pro Vision',
34
- id: 'gemini-1.0-pro-vision-latest',
35
- maxOutput: 4096,
36
- tokens: 12_288 + 4096,
37
+ description: 'Mid-size multimodal model that supports up to 1 million tokens',
38
+ displayName: 'Gemini 1.5 Pro 001',
39
+ functionCall: true,
40
+ id: 'gemini-1.5-pro-001',
41
+ maxOutput: 8192,
42
+ tokens: 2_097_152 + 8192,
37
43
  vision: true,
38
44
  },
39
45
  {
40
- description: 'The best image understanding model to handle a broad range of applications',
41
- displayName: 'Gemini 1.0 Pro Vision',
42
- id: 'gemini-pro-vision',
43
- maxOutput: 4096,
44
- tokens: 12_288 + 4096,
45
- vision: true,
46
+ description: 'The best model for scaling across a wide range of tasks. This is the latest model.',
47
+ displayName: 'Gemini 1.0 Pro',
48
+ id: 'gemini-1.0-pro-latest',
49
+ maxOutput: 2048,
50
+ tokens: 30_720 + 2048,
46
51
  },
47
52
  {
48
- description:
49
- 'The best model for scaling across a wide range of tasks. This is a stable model that supports tuning.',
53
+ description: 'The best model for scaling across a wide range of tasks. This is a stable model that supports tuning.',
50
54
  displayName: 'Gemini 1.0 Pro 001 (Tuning)',
51
55
  functionCall: true,
52
56
  id: 'gemini-1.0-pro-001',
@@ -54,41 +58,12 @@ const Google: ModelProviderCard = {
54
58
  tokens: 30_720 + 2048,
55
59
  },
56
60
  {
57
- description:
58
- 'The best model for scaling across a wide range of tasks. Released April 9, 2024.',
61
+ description: 'The best model for scaling across a wide range of tasks. Released April 9, 2024.',
59
62
  displayName: 'Gemini 1.0 Pro 002 (Tuning)',
60
63
  id: 'gemini-1.0-pro-002',
61
64
  maxOutput: 2048,
62
65
  tokens: 30_720 + 2048,
63
66
  },
64
- {
65
- description:
66
- 'The best model for scaling across a wide range of tasks. This is the latest model.',
67
- displayName: 'Gemini 1.0 Pro Latest',
68
- id: 'gemini-1.0-pro-latest',
69
- maxOutput: 2048,
70
- tokens: 30_720 + 2048,
71
- },
72
- {
73
- description: 'Fast and versatile multimodal model for scaling across diverse tasks',
74
- displayName: 'Gemini 1.5 Flash',
75
- enabled: true,
76
- functionCall: true,
77
- id: 'gemini-1.5-flash-latest',
78
- maxOutput: 8192,
79
- tokens: 1_048_576 + 8192,
80
- vision: true,
81
- },
82
- {
83
- description: 'Mid-size multimodal model that supports up to 1 million tokens',
84
- displayName: 'Gemini 1.5 Pro',
85
- enabled: true,
86
- functionCall: true,
87
- id: 'gemini-1.5-pro-latest',
88
- maxOutput: 8192,
89
- tokens: 1_048_576 + 8192,
90
- vision: true,
91
- },
92
67
  {
93
68
  description: 'The most capable model for highly complex tasks',
94
69
  displayName: 'Gemini 1.0 Ultra',
@@ -96,8 +71,16 @@ const Google: ModelProviderCard = {
96
71
  maxOutput: 2048,
97
72
  tokens: 32_768,
98
73
  },
74
+ {
75
+ description: 'A legacy text-only model optimized for chat conversations',
76
+ displayName: 'PaLM 2 Chat (Legacy)',
77
+ id: 'chat-bison-001',
78
+ legacy: true,
79
+ maxOutput: 1024,
80
+ // tokens: 4096 + 1024, // none tokens test
81
+ }
99
82
  ],
100
- checkModel: 'gemini-pro',
83
+ checkModel: 'gemini-1.5-flash',
101
84
  id: 'google',
102
85
  name: 'Google',
103
86
  proxyUrl: {