@lobehub/chat 1.36.33 → 1.36.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/README.ja-JP.md +150 -115
- package/README.md +65 -30
- package/README.zh-CN.md +65 -30
- package/changelog/v1.json +18 -0
- package/locales/ar/common.json +27 -7
- package/locales/ar/models.json +0 -24
- package/locales/ar/providers.json +10 -4
- package/locales/bg-BG/common.json +27 -7
- package/locales/bg-BG/models.json +0 -24
- package/locales/bg-BG/providers.json +10 -4
- package/locales/de-DE/common.json +27 -7
- package/locales/de-DE/models.json +0 -24
- package/locales/de-DE/providers.json +10 -4
- package/locales/en-US/common.json +27 -7
- package/locales/en-US/models.json +0 -24
- package/locales/en-US/providers.json +10 -4
- package/locales/es-ES/common.json +27 -7
- package/locales/es-ES/models.json +0 -24
- package/locales/es-ES/providers.json +10 -4
- package/locales/fa-IR/common.json +27 -7
- package/locales/fa-IR/models.json +0 -24
- package/locales/fa-IR/providers.json +10 -4
- package/locales/fr-FR/common.json +27 -7
- package/locales/fr-FR/models.json +0 -24
- package/locales/fr-FR/providers.json +10 -4
- package/locales/it-IT/common.json +27 -7
- package/locales/it-IT/models.json +0 -24
- package/locales/it-IT/providers.json +10 -4
- package/locales/ja-JP/common.json +27 -7
- package/locales/ja-JP/models.json +0 -24
- package/locales/ja-JP/providers.json +10 -4
- package/locales/ko-KR/common.json +27 -7
- package/locales/ko-KR/models.json +0 -24
- package/locales/ko-KR/providers.json +10 -4
- package/locales/nl-NL/common.json +27 -7
- package/locales/nl-NL/models.json +0 -24
- package/locales/nl-NL/providers.json +10 -4
- package/locales/pl-PL/common.json +27 -7
- package/locales/pl-PL/models.json +0 -24
- package/locales/pl-PL/providers.json +10 -4
- package/locales/pt-BR/common.json +27 -7
- package/locales/pt-BR/models.json +0 -24
- package/locales/pt-BR/providers.json +10 -4
- package/locales/ru-RU/common.json +27 -7
- package/locales/ru-RU/models.json +0 -24
- package/locales/ru-RU/providers.json +10 -4
- package/locales/tr-TR/common.json +27 -7
- package/locales/tr-TR/models.json +0 -24
- package/locales/tr-TR/providers.json +10 -4
- package/locales/vi-VN/common.json +27 -7
- package/locales/vi-VN/models.json +0 -24
- package/locales/vi-VN/providers.json +10 -4
- package/locales/zh-CN/common.json +28 -8
- package/locales/zh-CN/models.json +4 -28
- package/locales/zh-CN/providers.json +10 -4
- package/locales/zh-TW/common.json +27 -7
- package/locales/zh-TW/models.json +0 -24
- package/locales/zh-TW/providers.json +10 -4
- package/package.json +1 -1
- package/scripts/readmeWorkflow/const.ts +22 -8
- package/scripts/readmeWorkflow/index.ts +2 -0
- package/scripts/readmeWorkflow/syncAgentIndex.ts +36 -28
- package/scripts/readmeWorkflow/syncPluginIndex.ts +28 -15
- package/scripts/readmeWorkflow/syncProviderIndex.ts +51 -0
- package/scripts/readmeWorkflow/utlis.ts +23 -12
- package/src/app/loading/Client/Content.tsx +38 -0
- package/src/app/loading/Client/Redirect.tsx +47 -0
- package/src/app/loading/Client/index.tsx +22 -0
- package/src/app/loading/{Content.tsx → Server/Content.tsx} +6 -2
- package/src/app/loading/{Redirect.tsx → Server/Redirect.tsx} +3 -12
- package/src/app/loading/Server/index.tsx +22 -0
- package/src/app/loading/index.tsx +4 -16
- package/src/app/loading/stage.ts +22 -0
- package/src/components/FullscreenLoading/index.tsx +9 -8
- package/src/components/InitProgress/index.tsx +42 -0
- package/src/config/modelProviders/ai21.ts +1 -0
- package/src/config/modelProviders/cloudflare.ts +1 -2
- package/src/config/modelProviders/higress.ts +2 -1
- package/src/config/modelProviders/sensenova.ts +6 -3
- package/src/features/FileViewer/index.tsx +1 -1
- package/src/features/MobileSwitchLoading/index.tsx +20 -7
- package/src/layout/AuthProvider/NextAuth/UserUpdater.tsx +3 -1
- package/src/locales/default/common.ts +29 -8
- package/src/locales/default/models.ts +1 -0
- package/src/locales/default/providers.ts +1 -0
- package/src/app/loading/type.ts +0 -6
@@ -10,10 +10,16 @@
|
|
10
10
|
}
|
11
11
|
},
|
12
12
|
"appLoading": {
|
13
|
-
"
|
14
|
-
"
|
15
|
-
"
|
16
|
-
"
|
13
|
+
"appInitializing": "应用启动中...",
|
14
|
+
"finished": "数据库初始化完成",
|
15
|
+
"goToChat": "对话页面加载中...",
|
16
|
+
"initAuth": "鉴权服务初始化...",
|
17
|
+
"initUser": "用户状态初始化...",
|
18
|
+
"initializing": "PGlite 数据库初始化...",
|
19
|
+
"loadingDependencies": "初始化依赖...",
|
20
|
+
"loadingWasm": "加载 WASM 模块...",
|
21
|
+
"migrating": "执行数据表迁移...",
|
22
|
+
"ready": "数据库已就绪"
|
17
23
|
},
|
18
24
|
"autoGenerate": "自动补全",
|
19
25
|
"autoGenerateTooltip": "基于提示词自动补全助手描述",
|
@@ -38,14 +44,28 @@
|
|
38
44
|
"error": "发生错误,请重试",
|
39
45
|
"idle": "等待初始化...",
|
40
46
|
"initializing": "正在初始化...",
|
41
|
-
"loadingDependencies": "
|
42
|
-
"loadingWasmModule": "加载 WASM
|
43
|
-
"migrating": "
|
47
|
+
"loadingDependencies": "加载依赖中...",
|
48
|
+
"loadingWasmModule": "加载 WASM 模块中...",
|
49
|
+
"migrating": "执行数据表迁移...",
|
44
50
|
"ready": "数据库已就绪"
|
45
51
|
},
|
46
52
|
"modal": {
|
47
|
-
"desc": "
|
53
|
+
"desc": "立即启用下一代客户端数据库。在你的浏览器中持久存储聊天数据,并使用知识库等进阶特性。",
|
48
54
|
"enable": "立即启用",
|
55
|
+
"features": {
|
56
|
+
"knowledgeBase": {
|
57
|
+
"desc": "沉淀你的个人知识库,并与你的助手轻松开启知识库对话(即将上线)",
|
58
|
+
"title": "支持知识库对话,开启第二大脑"
|
59
|
+
},
|
60
|
+
"localFirst": {
|
61
|
+
"desc": "聊天数据完全存储在浏览器中,你的数据始终在你的掌握。",
|
62
|
+
"title": "本地优先,隐私至上"
|
63
|
+
},
|
64
|
+
"pglite": {
|
65
|
+
"desc": "基于 PGlite 构建,原生支持 AI Native 高阶特性(向量检索)",
|
66
|
+
"title": "新一代客户端存储架构"
|
67
|
+
}
|
68
|
+
},
|
49
69
|
"init": {
|
50
70
|
"desc": "正在初始化数据库,视网络差异可能会用时 5~30 秒不等",
|
51
71
|
"title": "正在初始化 PGlite 数据库"
|
@@ -23,20 +23,6 @@
|
|
23
23
|
"4.0Ultra": {
|
24
24
|
"description": "Spark Ultra 是星火大模型系列中最为强大的版本,在升级联网搜索链路同时,提升对文本内容的理解和总结能力。它是用于提升办公生产力和准确响应需求的全方位解决方案,是引领行业的智能产品。"
|
25
25
|
},
|
26
|
-
"@cf/meta/llama-3-8b-instruct-awq": {},
|
27
|
-
"@cf/openchat/openchat-3.5-0106": {},
|
28
|
-
"@cf/qwen/qwen1.5-14b-chat-awq": {},
|
29
|
-
"@hf/google/gemma-7b-it": {},
|
30
|
-
"@hf/meta-llama/meta-llama-3-8b-instruct": {
|
31
|
-
"description": "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t"
|
32
|
-
},
|
33
|
-
"@hf/mistral/mistral-7b-instruct-v0.2": {},
|
34
|
-
"@hf/nexusflow/starling-lm-7b-beta": {},
|
35
|
-
"@hf/nousresearch/hermes-2-pro-mistral-7b": {},
|
36
|
-
"@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
|
37
|
-
"@hf/thebloke/neural-chat-7b-v3-1-awq": {},
|
38
|
-
"@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
|
39
|
-
"@hf/thebloke/zephyr-7b-beta-awq": {},
|
40
26
|
"Baichuan2-Turbo": {
|
41
27
|
"description": "采用搜索增强技术实现大模型与领域知识、全网知识的全面链接。支持PDF、Word等多种文档上传及网址输入,信息获取及时、全面,输出结果准确、专业。"
|
42
28
|
},
|
@@ -50,10 +36,10 @@
|
|
50
36
|
"description": "模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。"
|
51
37
|
},
|
52
38
|
"Baichuan4-Air": {
|
53
|
-
"description": ""
|
39
|
+
"description": "模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。"
|
54
40
|
},
|
55
41
|
"Baichuan4-Turbo": {
|
56
|
-
"description": ""
|
42
|
+
"description": "模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。"
|
57
43
|
},
|
58
44
|
"Doubao-lite-128k": {
|
59
45
|
"description": "Doubao-lite 拥有极致的响应速度,更好的性价比,为客户不同场景提供更灵活的选择。支持128k上下文窗口的推理和精调。"
|
@@ -127,7 +113,6 @@
|
|
127
113
|
"Nous-Hermes-2-Mixtral-8x7B-DPO": {
|
128
114
|
"description": "Hermes 2 Mixtral 8x7B DPO 是一款高度灵活的多模型合并,旨在提供卓越的创造性体验。"
|
129
115
|
},
|
130
|
-
"NousResearch/Hermes-3-Llama-3.1-8B": {},
|
131
116
|
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
|
132
117
|
"description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) 是高精度的指令模型,适用于复杂计算。"
|
133
118
|
},
|
@@ -530,14 +515,11 @@
|
|
530
515
|
"cohere-command-r-plus": {
|
531
516
|
"description": "Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。"
|
532
517
|
},
|
533
|
-
"command-light": {
|
534
|
-
"description": ""
|
535
|
-
},
|
536
518
|
"command-r": {
|
537
|
-
"description": ""
|
519
|
+
"description": "Command R 是优化用于对话和长上下文任务的LLM,特别适合动态交互与知识管理。"
|
538
520
|
},
|
539
521
|
"command-r-plus": {
|
540
|
-
"description": ""
|
522
|
+
"description": "Command R+ 是一款高性能的大型语言模型,专为真实企业场景和复杂应用而设计。"
|
541
523
|
},
|
542
524
|
"databricks/dbrx-instruct": {
|
543
525
|
"description": "DBRX Instruct 提供高可靠性的指令处理能力,支持多行业应用。"
|
@@ -794,9 +776,6 @@
|
|
794
776
|
"hunyuan-functioncall": {
|
795
777
|
"description": "混元最新 MOE 架构 FunctionCall 模型,经过高质量的 FunctionCall 数据训练,上下文窗口达 32K,在多个维度的评测指标上处于领先。"
|
796
778
|
},
|
797
|
-
"hunyuan-large": {
|
798
|
-
"description": ""
|
799
|
-
},
|
800
779
|
"hunyuan-lite": {
|
801
780
|
"description": "升级为 MOE 结构,上下文窗口为 256k ,在 NLP,代码,数学,行业等多项评测集上领先众多开源模型。"
|
802
781
|
},
|
@@ -830,8 +809,6 @@
|
|
830
809
|
"internlm2.5-latest": {
|
831
810
|
"description": "我们最新的模型系列,有着卓越的推理性能,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。"
|
832
811
|
},
|
833
|
-
"jamba-1.5-large": {},
|
834
|
-
"jamba-1.5-mini": {},
|
835
812
|
"learnlm-1.5-pro-experimental": {
|
836
813
|
"description": "LearnLM 是一个实验性的、特定于任务的语言模型,经过训练以符合学习科学原则,可在教学和学习场景中遵循系统指令,充当专家导师等。"
|
837
814
|
},
|
@@ -1027,7 +1004,6 @@
|
|
1027
1004
|
"meta.llama3-8b-instruct-v1:0": {
|
1028
1005
|
"description": "Meta Llama 3 是一款面向开发者、研究人员和企业的开放大型语言模型 (LLM),旨在帮助他们构建、实验并负责任地扩展他们的生成 AI 想法。作为全球社区创新的基础系统的一部分,它非常适合计算能力和资源有限、边缘设备和更快的训练时间。"
|
1029
1006
|
},
|
1030
|
-
"microsoft/Phi-3.5-mini-instruct": {},
|
1031
1007
|
"microsoft/WizardLM-2-8x22B": {
|
1032
1008
|
"description": "WizardLM 2 是微软AI提供的语言模型,在复杂对话、多语言、推理和智能助手领域表现尤为出色。"
|
1033
1009
|
},
|
@@ -1,5 +1,7 @@
|
|
1
1
|
{
|
2
|
-
"ai21": {
|
2
|
+
"ai21": {
|
3
|
+
"description": "AI21 Labs 为企业构建基础模型和人工智能系统,加速生成性人工智能在生产中的应用。"
|
4
|
+
},
|
3
5
|
"ai360": {
|
4
6
|
"description": "360 AI 是 360 公司推出的 AI 模型和服务平台,提供多种先进的自然语言处理模型,包括 360GPT2 Pro、360GPT Pro、360GPT Turbo 和 360GPT Turbo Responsibility 8K。这些模型结合了大规模参数和多模态能力,广泛应用于文本生成、语义理解、对话系统与代码生成等领域。通过灵活的定价策略,360 AI 满足多样化用户需求,支持开发者集成,推动智能化应用的革新和发展。"
|
5
7
|
},
|
@@ -15,7 +17,9 @@
|
|
15
17
|
"bedrock": {
|
16
18
|
"description": "Bedrock 是亚马逊 AWS 提供的一项服务,专注于为企业提供先进的 AI 语言模型和视觉模型。其模型家族包括 Anthropic 的 Claude 系列、Meta 的 Llama 3.1 系列等,涵盖从轻量级到高性能的多种选择,支持文本生成、对话、图像处理等多种任务,适用于不同规模和需求的企业应用。"
|
17
19
|
},
|
18
|
-
"cloudflare": {
|
20
|
+
"cloudflare": {
|
21
|
+
"description": "在 Cloudflare 的全球网络上运行由无服务器 GPU 驱动的机器学习模型。"
|
22
|
+
},
|
19
23
|
"deepseek": {
|
20
24
|
"description": "DeepSeek 是一家专注于人工智能技术研究和应用的公司,其最新模型 DeepSeek-V2.5 融合了通用对话和代码处理能力,并在人类偏好对齐、写作任务和指令跟随等方面实现了显著提升。"
|
21
25
|
},
|
@@ -35,7 +39,7 @@
|
|
35
39
|
"description": "Groq 的 LPU 推理引擎在最新的独立大语言模型(LLM)基准测试中表现卓越,以其惊人的速度和效率重新定义了 AI 解决方案的标准。Groq 是一种即时推理速度的代表,在基于云的部署中展现了良好的性能。"
|
36
40
|
},
|
37
41
|
"higress": {
|
38
|
-
"description": ""
|
42
|
+
"description": "Higress 是一款云原生 API 网关,在阿里内部为解决 Tengine reload 对长连接业务有损,以及 gRPC/Dubbo 负载均衡能力不足而诞生。"
|
39
43
|
},
|
40
44
|
"huggingface": {
|
41
45
|
"description": "HuggingFace Inference API 提供了一种快速且免费的方式,让您可以探索成千上万种模型,适用于各种任务。无论您是在为新应用程序进行原型设计,还是在尝试机器学习的功能,这个 API 都能让您即时访问多个领域的高性能模型。"
|
@@ -73,7 +77,9 @@
|
|
73
77
|
"qwen": {
|
74
78
|
"description": "通义千问是阿里云自主研发的超大规模语言模型,具有强大的自然语言理解和生成能力。它可以回答各种问题、创作文字内容、表达观点看法、撰写代码等,在多个领域发挥作用。"
|
75
79
|
},
|
76
|
-
"sensenova": {
|
80
|
+
"sensenova": {
|
81
|
+
"description": "商汤日日新,依托商汤大装置的强大的基础支撑,提供高效易用的全栈大模型服务。"
|
82
|
+
},
|
77
83
|
"siliconcloud": {
|
78
84
|
"description": "SiliconCloud,基于优秀开源基础模型的高性价比 GenAI 云服务"
|
79
85
|
},
|
@@ -10,10 +10,16 @@
|
|
10
10
|
}
|
11
11
|
},
|
12
12
|
"appLoading": {
|
13
|
-
"
|
14
|
-
"
|
15
|
-
"
|
16
|
-
"
|
13
|
+
"appInitializing": "應用啟動中...",
|
14
|
+
"finished": "資料庫初始化完成",
|
15
|
+
"goToChat": "對話頁面加載中...",
|
16
|
+
"initAuth": "鑑權服務初始化...",
|
17
|
+
"initUser": "用戶狀態初始化...",
|
18
|
+
"initializing": "PGlite 資料庫初始化...",
|
19
|
+
"loadingDependencies": "初始化依賴...",
|
20
|
+
"loadingWasm": "加載 WASM 模組...",
|
21
|
+
"migrating": "執行資料表遷移...",
|
22
|
+
"ready": "資料庫已就緒"
|
17
23
|
},
|
18
24
|
"autoGenerate": "自動生成",
|
19
25
|
"autoGenerateTooltip": "基於提示詞自動生成助手描述",
|
@@ -38,14 +44,28 @@
|
|
38
44
|
"error": "發生錯誤,請重試",
|
39
45
|
"idle": "等待初始化...",
|
40
46
|
"initializing": "正在初始化...",
|
41
|
-
"loadingDependencies": "
|
42
|
-
"loadingWasmModule": "
|
43
|
-
"migrating": "
|
47
|
+
"loadingDependencies": "載入依賴中...",
|
48
|
+
"loadingWasmModule": "載入 WASM 模組中...",
|
49
|
+
"migrating": "執行資料表遷移...",
|
44
50
|
"ready": "數據庫已就緒"
|
45
51
|
},
|
46
52
|
"modal": {
|
47
53
|
"desc": "啟用 PGlite 客戶端數據庫,在你的瀏覽器中持久存儲聊天數據,並使用知識庫等進階特性",
|
48
54
|
"enable": "立即啟用",
|
55
|
+
"features": {
|
56
|
+
"knowledgeBase": {
|
57
|
+
"desc": "沉澱你的個人知識庫,並與你的助手輕鬆開啟知識庫對話(即將上線)",
|
58
|
+
"title": "支持知識庫對話,開啟第二大腦"
|
59
|
+
},
|
60
|
+
"localFirst": {
|
61
|
+
"desc": "聊天數據完全存儲在瀏覽器中,你的數據始終在你的掌握。",
|
62
|
+
"title": "本地優先,隱私至上"
|
63
|
+
},
|
64
|
+
"pglite": {
|
65
|
+
"desc": "基於 PGlite 構建,原生支持 AI Native 高階特性(向量檢索)",
|
66
|
+
"title": "新一代客戶端存儲架構"
|
67
|
+
}
|
68
|
+
},
|
49
69
|
"init": {
|
50
70
|
"desc": "正在初始化數據庫,視網絡差異可能會用時 5~30 秒不等",
|
51
71
|
"title": "正在初始化 PGlite 數據庫"
|
@@ -23,20 +23,6 @@
|
|
23
23
|
"4.0Ultra": {
|
24
24
|
"description": "Spark4.0 Ultra 是星火大模型系列中最為強大的版本,在升級聯網搜索鏈路同時,提升對文本內容的理解和總結能力。它是用於提升辦公生產力和準確響應需求的全方位解決方案,是引領行業的智能產品。"
|
25
25
|
},
|
26
|
-
"@cf/meta/llama-3-8b-instruct-awq": {},
|
27
|
-
"@cf/openchat/openchat-3.5-0106": {},
|
28
|
-
"@cf/qwen/qwen1.5-14b-chat-awq": {},
|
29
|
-
"@hf/google/gemma-7b-it": {},
|
30
|
-
"@hf/meta-llama/meta-llama-3-8b-instruct": {
|
31
|
-
"description": "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning."
|
32
|
-
},
|
33
|
-
"@hf/mistral/mistral-7b-instruct-v0.2": {},
|
34
|
-
"@hf/nexusflow/starling-lm-7b-beta": {},
|
35
|
-
"@hf/nousresearch/hermes-2-pro-mistral-7b": {},
|
36
|
-
"@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
|
37
|
-
"@hf/thebloke/neural-chat-7b-v3-1-awq": {},
|
38
|
-
"@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
|
39
|
-
"@hf/thebloke/zephyr-7b-beta-awq": {},
|
40
26
|
"Baichuan2-Turbo": {
|
41
27
|
"description": "採用搜索增強技術實現大模型與領域知識、全網知識的全面連結。支持PDF、Word等多種文檔上傳及網址輸入,信息獲取及時、全面,輸出結果準確、專業。"
|
42
28
|
},
|
@@ -127,7 +113,6 @@
|
|
127
113
|
"Nous-Hermes-2-Mixtral-8x7B-DPO": {
|
128
114
|
"description": "Hermes 2 Mixtral 8x7B DPO 是一款高度靈活的多模型合併,旨在提供卓越的創造性體驗。"
|
129
115
|
},
|
130
|
-
"NousResearch/Hermes-3-Llama-3.1-8B": {},
|
131
116
|
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
|
132
117
|
"description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) 是高精度的指令模型,適用於複雜計算。"
|
133
118
|
},
|
@@ -530,9 +515,6 @@
|
|
530
515
|
"cohere-command-r-plus": {
|
531
516
|
"description": "Command R+是一個最先進的RAG優化模型,旨在應對企業級工作負載。"
|
532
517
|
},
|
533
|
-
"command-light": {
|
534
|
-
"description": ""
|
535
|
-
},
|
536
518
|
"command-r": {
|
537
519
|
"description": "Command R 是優化用於對話和長上下文任務的 LLM,特別適合動態交互與知識管理。"
|
538
520
|
},
|
@@ -794,9 +776,6 @@
|
|
794
776
|
"hunyuan-functioncall": {
|
795
777
|
"description": "混元最新 MOE 架構 FunctionCall 模型,經過高質量的 FunctionCall 數據訓練,上下文窗口達 32K,在多個維度的評測指標上處於領先。"
|
796
778
|
},
|
797
|
-
"hunyuan-large": {
|
798
|
-
"description": ""
|
799
|
-
},
|
800
779
|
"hunyuan-lite": {
|
801
780
|
"description": "升級為 MOE 結構,上下文窗口為 256k,在 NLP、代碼、數學、行業等多項評測集上領先眾多開源模型。"
|
802
781
|
},
|
@@ -830,8 +809,6 @@
|
|
830
809
|
"internlm2.5-latest": {
|
831
810
|
"description": "我們最新的模型系列,有著卓越的推理性能,支持 1M 的上下文長度以及更強的指令跟隨和工具調用能力。"
|
832
811
|
},
|
833
|
-
"jamba-1.5-large": {},
|
834
|
-
"jamba-1.5-mini": {},
|
835
812
|
"learnlm-1.5-pro-experimental": {
|
836
813
|
"description": "LearnLM 是一個實驗性的、特定於任務的語言模型,經過訓練以符合學習科學原則,可在教學和學習場景中遵循系統指令,充當專家導師等。"
|
837
814
|
},
|
@@ -1027,7 +1004,6 @@
|
|
1027
1004
|
"meta.llama3-8b-instruct-v1:0": {
|
1028
1005
|
"description": "Meta Llama 3 是一款面向開發者、研究人員和企業的開放大型語言模型 (LLM),旨在幫助他們構建、實驗並負責任地擴展他們的生成 AI 想法。作為全球社區創新的基礎系統的一部分,它非常適合計算能力和資源有限、邊緣設備和更快的訓練時間。"
|
1029
1006
|
},
|
1030
|
-
"microsoft/Phi-3.5-mini-instruct": {},
|
1031
1007
|
"microsoft/WizardLM-2-8x22B": {
|
1032
1008
|
"description": "WizardLM 2 是微軟AI提供的語言模型,在複雜對話、多語言、推理和智能助手領域表現尤為出色。"
|
1033
1009
|
},
|
@@ -1,5 +1,7 @@
|
|
1
1
|
{
|
2
|
-
"ai21": {
|
2
|
+
"ai21": {
|
3
|
+
"description": "AI21 Labs 為企業構建基礎模型和人工智慧系統,加速生成性人工智慧在生產中的應用。"
|
4
|
+
},
|
3
5
|
"ai360": {
|
4
6
|
"description": "360 AI 是 360 公司推出的 AI 模型和服務平台,提供多種先進的自然語言處理模型,包括 360GPT2 Pro、360GPT Pro、360GPT Turbo 和 360GPT Turbo Responsibility 8K。這些模型結合了大規模參數和多模態能力,廣泛應用於文本生成、語義理解、對話系統與代碼生成等領域。通過靈活的定價策略,360 AI 滿足多樣化用戶需求,支持開發者集成,推動智能化應用的革新和發展。"
|
5
7
|
},
|
@@ -15,7 +17,9 @@
|
|
15
17
|
"bedrock": {
|
16
18
|
"description": "Bedrock 是亞馬遜 AWS 提供的一項服務,專注於為企業提供先進的 AI 語言模型和視覺模型。其模型家族包括 Anthropic 的 Claude 系列、Meta 的 Llama 3.1 系列等,涵蓋從輕量級到高性能的多種選擇,支持文本生成、對話、圖像處理等多種任務,適用於不同規模和需求的企業應用。"
|
17
19
|
},
|
18
|
-
"cloudflare": {
|
20
|
+
"cloudflare": {
|
21
|
+
"description": "在 Cloudflare 的全球網絡上運行由無伺服器 GPU 驅動的機器學習模型。"
|
22
|
+
},
|
19
23
|
"deepseek": {
|
20
24
|
"description": "DeepSeek 是一家專注於人工智慧技術研究和應用的公司,其最新模型 DeepSeek-V2.5 融合了通用對話和代碼處理能力,並在人類偏好對齊、寫作任務和指令跟隨等方面實現了顯著提升。"
|
21
25
|
},
|
@@ -35,7 +39,7 @@
|
|
35
39
|
"description": "Groq 的 LPU 推理引擎在最新的獨立大語言模型(LLM)基準測試中表現卓越,以其驚人的速度和效率重新定義了 AI 解決方案的標準。Groq 是一種即時推理速度的代表,在基於雲的部署中展現了良好的性能。"
|
36
40
|
},
|
37
41
|
"higress": {
|
38
|
-
"description": ""
|
42
|
+
"description": "Higress 是一款雲原生 API 網關,為了解決 Tengine reload 對長連接業務的影響,以及 gRPC/Dubbo 負載均衡能力不足而在阿里內部誕生。"
|
39
43
|
},
|
40
44
|
"huggingface": {
|
41
45
|
"description": "HuggingFace Inference API 提供了一種快速且免費的方式,讓您可以探索成千上萬種模型,適用於各種任務。無論您是在為新應用程式進行原型設計,還是在嘗試機器學習的功能,這個 API 都能讓您即時訪問多個領域的高性能模型。"
|
@@ -73,7 +77,9 @@
|
|
73
77
|
"qwen": {
|
74
78
|
"description": "通義千問是阿里雲自主研發的超大規模語言模型,具有強大的自然語言理解和生成能力。它可以回答各種問題、創作文字內容、表達觀點看法、撰寫代碼等,在多個領域發揮作用。"
|
75
79
|
},
|
76
|
-
"sensenova": {
|
80
|
+
"sensenova": {
|
81
|
+
"description": "商湯日日新,依托商湯大裝置的強大基礎支撐,提供高效易用的全棧大模型服務。"
|
82
|
+
},
|
77
83
|
"siliconcloud": {
|
78
84
|
"description": "SiliconFlow 致力於加速 AGI,以惠及人類,通過易用與成本低的 GenAI 堆疊提升大規模 AI 效率。"
|
79
85
|
},
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@lobehub/chat",
|
3
|
-
"version": "1.36.
|
3
|
+
"version": "1.36.35",
|
4
4
|
"description": "Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.",
|
5
5
|
"keywords": [
|
6
6
|
"framework",
|
@@ -1,3 +1,9 @@
|
|
1
|
+
import { resolve } from 'node:path';
|
2
|
+
|
3
|
+
import { DEFAULT_MODEL_PROVIDER_LIST } from '@/config/modelProviders';
|
4
|
+
|
5
|
+
export const root = resolve(__dirname, '../..');
|
6
|
+
|
1
7
|
export interface DataItem {
|
2
8
|
author: string;
|
3
9
|
createdAt: string;
|
@@ -6,14 +12,22 @@ export interface DataItem {
|
|
6
12
|
meta: { avatar: string; description: string; tags: string[]; title: string };
|
7
13
|
}
|
8
14
|
|
9
|
-
export const
|
10
|
-
export const
|
11
|
-
|
12
|
-
export const
|
13
|
-
export const
|
14
|
-
|
15
|
-
export const PLUGIN_CN_URL = 'https://chat-plugins.lobehub.com/index.zh-CN.json';
|
16
|
-
export const PLUGIN_REPO = 'https://github.com/lobehub/lobe-chat-plugins';
|
15
|
+
export const AGENT_URL = 'https://chat-agents.lobehub.com/index.json';
|
16
|
+
export const AGENT_I18N_URL = (lang: string) =>
|
17
|
+
`https://chat-agents.lobehub.com/index.${lang}.json`;
|
18
|
+
export const PLUGIN_URL = 'https://chat-plugins.lobehub.com/index.json';
|
19
|
+
export const PLUGIN_I18N_URL = (lang: string) =>
|
20
|
+
`https://chat-plugins.lobehub.com/index.${lang}.json`;
|
17
21
|
|
18
22
|
export const AGENT_SPLIT = '<!-- AGENT LIST -->';
|
19
23
|
export const PLUGIN_SPLIT = '<!-- PLUGIN LIST -->';
|
24
|
+
export const PROVIDER_SPLIT = '<!-- PROVIDER LIST -->';
|
25
|
+
|
26
|
+
export const PROVIDER_LIST = DEFAULT_MODEL_PROVIDER_LIST.filter(
|
27
|
+
(item) => item.chatModels.length > 0 && item.id !== 'lobehub',
|
28
|
+
).map((item) => {
|
29
|
+
return {
|
30
|
+
id: item.id,
|
31
|
+
name: item.name,
|
32
|
+
};
|
33
|
+
});
|
@@ -2,9 +2,11 @@ import { consola } from 'consola';
|
|
2
2
|
|
3
3
|
import syncAgentIndex from './syncAgentIndex';
|
4
4
|
import syncPluginIndex from './syncPluginIndex';
|
5
|
+
import syncProviderIndex from './syncProviderIndex';
|
5
6
|
|
6
7
|
const runSync = async () => {
|
7
8
|
consola.start('Start sync readme workflow...');
|
9
|
+
await syncProviderIndex();
|
8
10
|
await syncAgentIndex();
|
9
11
|
await syncPluginIndex();
|
10
12
|
};
|
@@ -1,47 +1,55 @@
|
|
1
1
|
import { consola } from 'consola';
|
2
2
|
import { markdownTable } from 'markdown-table';
|
3
|
-
import
|
3
|
+
import urlJoin from 'url-join';
|
4
4
|
|
5
|
-
import {
|
6
|
-
import {
|
5
|
+
import { AGENT_SPLIT, DataItem } from './const';
|
6
|
+
import {
|
7
|
+
fetchAgentIndex,
|
8
|
+
genLink,
|
9
|
+
genTags,
|
10
|
+
getTitle,
|
11
|
+
readReadme,
|
12
|
+
updateReadme,
|
13
|
+
writeReadme,
|
14
|
+
} from './utlis';
|
7
15
|
|
8
|
-
const genAgentTable = (data: DataItem[], lang
|
9
|
-
const
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
...content,
|
26
|
-
]);
|
16
|
+
const genAgentTable = (data: DataItem[], lang?: string) => {
|
17
|
+
const title = getTitle(lang);
|
18
|
+
|
19
|
+
const content = data
|
20
|
+
.slice(0, 4)
|
21
|
+
.map((item) => [
|
22
|
+
[
|
23
|
+
genLink(
|
24
|
+
item.meta.title.replaceAll('|', ','),
|
25
|
+
urlJoin('https://lobechat.com/discover/assistant', item.identifier),
|
26
|
+
),
|
27
|
+
`<sup>By **${genLink(item.author, item.homepage)}** on **${(item as any).createdAt}**</sup>`,
|
28
|
+
].join('<br/>'),
|
29
|
+
[item.meta.description.replaceAll('|', ','), genTags(item.meta.tags)].join('<br/>'),
|
30
|
+
]);
|
31
|
+
|
32
|
+
return markdownTable([title, ...content]);
|
27
33
|
};
|
28
34
|
|
29
|
-
const runAgentTable = async (lang
|
35
|
+
const runAgentTable = async (lang?: string) => {
|
30
36
|
const data = await fetchAgentIndex(lang);
|
31
37
|
const md = readReadme(lang);
|
32
38
|
const mdTable = genAgentTable(data, lang);
|
33
39
|
const newMd = updateReadme(
|
34
40
|
AGENT_SPLIT,
|
35
41
|
md,
|
36
|
-
[
|
37
|
-
|
38
|
-
|
42
|
+
[
|
43
|
+
mdTable,
|
44
|
+
`> 📊 Total agents: ${genLink(`<kbd>**${data.length}**</kbd> `, 'https://lobechat.com/discover/assistants')}`,
|
45
|
+
].join('\n\n'),
|
39
46
|
);
|
40
47
|
writeReadme(newMd, lang);
|
41
|
-
consola.success(
|
48
|
+
consola.success(`Sync ${lang || 'en-US'} agent index success!`);
|
42
49
|
};
|
43
50
|
|
44
51
|
export default async () => {
|
45
|
-
await runAgentTable(
|
52
|
+
await runAgentTable();
|
46
53
|
await runAgentTable('zh-CN');
|
54
|
+
await runAgentTable('ja-JP');
|
47
55
|
};
|
@@ -1,42 +1,55 @@
|
|
1
1
|
import { consola } from 'consola';
|
2
2
|
import { markdownTable } from 'markdown-table';
|
3
|
+
import urlJoin from 'url-join';
|
3
4
|
|
4
|
-
import { DataItem,
|
5
|
-
import {
|
5
|
+
import { DataItem, PLUGIN_SPLIT } from './const';
|
6
|
+
import {
|
7
|
+
fetchPluginIndex,
|
8
|
+
genLink,
|
9
|
+
genTags,
|
10
|
+
getTitle,
|
11
|
+
readReadme,
|
12
|
+
updateReadme,
|
13
|
+
writeReadme,
|
14
|
+
} from './utlis';
|
15
|
+
|
16
|
+
const genPluginTable = (data: DataItem[], lang?: string) => {
|
17
|
+
const title = getTitle(lang);
|
6
18
|
|
7
|
-
const genPluginTable = (data: DataItem[], lang: string) => {
|
8
|
-
const isCN = lang === 'zh-CN';
|
9
19
|
const content = data
|
10
20
|
.slice(0, 4)
|
11
21
|
.map((item) => [
|
12
22
|
[
|
13
|
-
genLink(
|
23
|
+
genLink(
|
24
|
+
item.meta.title.replaceAll('|', ','),
|
25
|
+
urlJoin('https://lobechat.com/discover/plugin', item.identifier),
|
26
|
+
),
|
14
27
|
`<sup>By **${item.author}** on **${item.createdAt}**</sup>`,
|
15
28
|
].join('<br/>'),
|
16
29
|
[item.meta.description.replaceAll('|', ','), genTags(item.meta.tags)].join('<br/>'),
|
17
30
|
]);
|
18
|
-
|
19
|
-
|
20
|
-
...content,
|
21
|
-
]);
|
31
|
+
|
32
|
+
return markdownTable([title, ...content]);
|
22
33
|
};
|
23
34
|
|
24
|
-
const runPluginTable = async (lang
|
35
|
+
const runPluginTable = async (lang?: string) => {
|
25
36
|
const data = await fetchPluginIndex(lang);
|
26
37
|
const md = readReadme(lang);
|
27
38
|
const mdTable = genPluginTable(data, lang);
|
28
39
|
const newMd = updateReadme(
|
29
40
|
PLUGIN_SPLIT,
|
30
41
|
md,
|
31
|
-
[
|
32
|
-
|
33
|
-
|
42
|
+
[
|
43
|
+
mdTable,
|
44
|
+
`> 📊 Total plugins: ${genLink(`<kbd>**${data.length}**</kbd>`, 'https://lobechat.com/discover/plugins')}`,
|
45
|
+
].join('\n\n'),
|
34
46
|
);
|
35
47
|
writeReadme(newMd, lang);
|
36
|
-
consola.success(
|
48
|
+
consola.success(`Sync ${lang || 'en-US'} plugin index success!`);
|
37
49
|
};
|
38
50
|
|
39
51
|
export default async () => {
|
40
|
-
await runPluginTable(
|
52
|
+
await runPluginTable();
|
41
53
|
await runPluginTable('zh-CN');
|
54
|
+
await runPluginTable('ja-JP');
|
42
55
|
};
|
@@ -0,0 +1,51 @@
|
|
1
|
+
import { consola } from 'consola';
|
2
|
+
import { readJSONSync } from 'fs-extra';
|
3
|
+
import { resolve } from 'node:path';
|
4
|
+
import urlJoin from 'url-join';
|
5
|
+
|
6
|
+
import { PROVIDER_LIST, PROVIDER_SPLIT, root } from './const';
|
7
|
+
import { genLink, readReadme, updateReadme, writeReadme } from './utlis';
|
8
|
+
|
9
|
+
const genProviderTable = (data: { desc?: string; id: string; name: string }) => {
|
10
|
+
const title = genLink(data.name, urlJoin('https://lobechat.com/discover/provider', data.id));
|
11
|
+
|
12
|
+
return ['-', `**${title}**:`, data.desc].join(' ');
|
13
|
+
};
|
14
|
+
|
15
|
+
const runProviderTable = async (lang?: string) => {
|
16
|
+
const md = readReadme(lang);
|
17
|
+
const desc = readJSONSync(resolve(root, 'locales', lang || 'en-US', 'providers.json'));
|
18
|
+
const newMd = updateReadme(
|
19
|
+
PROVIDER_SPLIT,
|
20
|
+
md,
|
21
|
+
[
|
22
|
+
PROVIDER_LIST.slice(0, 10)
|
23
|
+
.map((item) =>
|
24
|
+
genProviderTable({
|
25
|
+
...item,
|
26
|
+
desc: desc?.[item.id]?.description,
|
27
|
+
}),
|
28
|
+
)
|
29
|
+
.join('\n'),
|
30
|
+
`<details><summary><kbd>See more providers (+${PROVIDER_LIST.length - 10})</kbd></summary>`,
|
31
|
+
PROVIDER_LIST.slice(10, PROVIDER_LIST.length)
|
32
|
+
.map((item) =>
|
33
|
+
genProviderTable({
|
34
|
+
...item,
|
35
|
+
desc: desc?.[item.id]?.description,
|
36
|
+
}),
|
37
|
+
)
|
38
|
+
.join('\n'),
|
39
|
+
'</details>',
|
40
|
+
`> 📊 Total providers: ${genLink(`<kbd>**${PROVIDER_LIST.length}**</kbd>`, 'https://lobechat.com/discover/providers')}`,
|
41
|
+
].join('\n\n'),
|
42
|
+
);
|
43
|
+
writeReadme(newMd, lang);
|
44
|
+
consola.success(`Sync ${lang || 'en-US'} provider index success!`);
|
45
|
+
};
|
46
|
+
|
47
|
+
export default async () => {
|
48
|
+
await runProviderTable();
|
49
|
+
await runProviderTable('zh-CN');
|
50
|
+
await runProviderTable('ja-JP');
|
51
|
+
};
|