@lobehub/chat 1.16.8 → 1.16.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of @lobehub/chat might be problematic. Click here for more details.

Files changed (51) hide show
  1. package/CHANGELOG.md +58 -0
  2. package/package.json +3 -3
  3. package/src/config/__tests__/auth.test.ts +200 -0
  4. package/src/config/auth.ts +97 -2
  5. package/src/config/modelProviders/ai360.ts +11 -0
  6. package/src/config/modelProviders/anthropic.ts +27 -18
  7. package/src/config/modelProviders/azure.ts +12 -3
  8. package/src/config/modelProviders/baichuan.ts +3 -1
  9. package/src/config/modelProviders/bedrock.ts +20 -18
  10. package/src/config/modelProviders/deepseek.ts +3 -1
  11. package/src/config/modelProviders/fireworksai.ts +33 -5
  12. package/src/config/modelProviders/google.ts +16 -13
  13. package/src/config/modelProviders/groq.ts +19 -8
  14. package/src/config/modelProviders/minimax.ts +8 -6
  15. package/src/config/modelProviders/mistral.ts +19 -3
  16. package/src/config/modelProviders/moonshot.ts +11 -1
  17. package/src/config/modelProviders/novita.ts +24 -0
  18. package/src/config/modelProviders/ollama.ts +58 -1
  19. package/src/config/modelProviders/openai.ts +52 -18
  20. package/src/config/modelProviders/openrouter.ts +21 -1
  21. package/src/config/modelProviders/perplexity.ts +19 -3
  22. package/src/config/modelProviders/qwen.ts +11 -8
  23. package/src/config/modelProviders/siliconcloud.ts +34 -1
  24. package/src/config/modelProviders/spark.ts +16 -7
  25. package/src/config/modelProviders/stepfun.ts +13 -1
  26. package/src/config/modelProviders/taichu.ts +7 -2
  27. package/src/config/modelProviders/togetherai.ts +38 -2
  28. package/src/config/modelProviders/upstage.ts +11 -4
  29. package/src/config/modelProviders/zeroone.ts +5 -1
  30. package/src/config/modelProviders/zhipu.ts +20 -18
  31. package/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap +13 -6
  32. package/src/libs/agent-runtime/qwen/index.test.ts +20 -4
  33. package/src/libs/agent-runtime/qwen/index.ts +1 -1
  34. package/src/libs/next-auth/auth.config.ts +3 -1
  35. package/src/libs/next-auth/sso-providers/auth0.ts +5 -3
  36. package/src/libs/next-auth/sso-providers/authelia.ts +6 -6
  37. package/src/libs/next-auth/sso-providers/authentik.ts +5 -3
  38. package/src/libs/next-auth/sso-providers/azure-ad.ts +5 -3
  39. package/src/libs/next-auth/sso-providers/cloudflare-zero-trust.ts +4 -3
  40. package/src/libs/next-auth/sso-providers/generic-oidc.ts +3 -3
  41. package/src/libs/next-auth/sso-providers/github.ts +4 -2
  42. package/src/libs/next-auth/sso-providers/logto.ts +3 -3
  43. package/src/libs/next-auth/sso-providers/zitadel.ts +5 -3
  44. package/src/migrations/FromV3ToV4/fixtures/ollama-output-v4.json +1 -0
  45. package/src/server/routers/edge/config/__snapshots__/index.test.ts.snap +24 -4
  46. package/src/server/routers/edge/config/index.test.ts +3 -11
  47. package/src/store/user/slices/modelList/__snapshots__/action.test.ts.snap +12 -0
  48. package/src/store/user/slices/modelList/action.test.ts +3 -7
  49. package/src/types/llm.ts +1 -0
  50. package/src/utils/__snapshots__/parseModels.test.ts.snap +32 -0
  51. package/src/utils/parseModels.test.ts +1 -28
@@ -1,26 +1,26 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
- // ref https://docs.aws.amazon.com/bedrock/latest/userguide/conversation-inference.html
4
- // ref https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/models
5
- // ref https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/models
3
+ // ref :https://docs.aws.amazon.com/bedrock/latest/userguide/conversation-inference.html
4
+ // ref :https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/models
5
+ // ref :https://us-west-2.console.aws.amazon.com/bedrock/home?region=us-west-2#/models
6
6
  const Bedrock: ModelProviderCard = {
7
7
  chatModels: [
8
8
  /*
9
9
  // TODO: Not support for now
10
10
  {
11
- description: 'Amazon Titan Text Lite is a light weight efficient model ideal for fine-tuning for English-language tasks, including like summarization and copywriting, where customers want a smaller, more cost-effective model that is also highly customizable.',
11
+ description: '亚马逊 Titan Text Lite 是一款轻量级高效模型,非常适合对英语任务进行微调,包括总结和文案编写等,客户希望有一个更小、更经济的模型,同时也非常可定制。',
12
12
  displayName: 'Titan Text G1 - Lite',
13
13
  id: 'amazon.titan-text-lite-v1',
14
14
  tokens: 4000,
15
15
  },
16
16
  {
17
- description: 'Amazon Titan Text Express has a context length of up to 8,000 tokens, making it well-suited for a wide range of advanced, general language tasks such as open-ended text generation and conversational chat, as well as support within Retrieval Augmented Generation (RAG). At launch, the model is optimized for English, with multilingual support for more than 100 additional languages available in preview.',
17
+ description: '亚马逊 Titan Text Express 的上下文长度可达 8,000 个标记,非常适合广泛的高级通用语言任务,如开放式文本生成和对话聊天,以及在检索增强生成 (RAG) 中的支持。在推出时,该模型针对英语进行了优化,预览版还支持其他 100 多种语言。',
18
18
  displayName: 'Titan Text G1 - Express',
19
19
  id: 'amazon.titan-text-express-v1',
20
20
  tokens: 8000,
21
21
  },
22
22
  {
23
- description: 'Titan Text Premier is a powerful and advanced model within the Titan Text family, designed to deliver superior performance across a wide range of enterprise applications. With its cutting-edge capabilities, it offers enhanced accuracy and exceptional results, making it an excellent choice for organizations seeking top-notch text processing solutions.',
23
+ description: 'Titan Text Premier Titan Text 系列中一款强大的先进模型,旨在为广泛的企业应用提供卓越的性能。凭借其尖端能力,它提供了更高的准确性和卓越的结果,是寻求一流文本处理解决方案的组织的绝佳选择。',
24
24
  displayName: 'Titan Text G1 - Premier',
25
25
  id: 'amazon.titan-text-premier-v1:0',
26
26
  tokens: 32_000,
@@ -28,7 +28,7 @@ const Bedrock: ModelProviderCard = {
28
28
  */
29
29
  {
30
30
  description:
31
- 'Claude 3.5 Sonnet raises the industry bar for intelligence, outperforming competitor models and Claude 3 Opus on a wide range of evaluations, with the speed and cost of our mid-tier model, Claude 3 Sonnet.',
31
+ 'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。',
32
32
  displayName: 'Claude 3.5 Sonnet',
33
33
  enabled: true,
34
34
  functionCall: true,
@@ -42,7 +42,7 @@ const Bedrock: ModelProviderCard = {
42
42
  },
43
43
  {
44
44
  description:
45
- 'Claude 3 Sonnet by Anthropic strikes the ideal balance between intelligence and speed—particularly for enterprise workloads. It offers maximum utility at a lower price than competitors, and is engineered to be the dependable, high-endurance workhorse for scaled AI deployments. Claude 3 Sonnet can process images and return text outputs, and features a 200K context window.',
45
+ 'Anthropic 的 Claude 3 Sonnet 在智能和速度之间达到了理想的平衡——特别适合企业工作负载。它以低于竞争对手的价格提供最大的效用,并被设计成为可靠的、高耐用的主力机,适用于规模化的 AI 部署。Claude 3 Sonnet 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
46
46
  displayName: 'Claude 3 Sonnet',
47
47
  enabled: true,
48
48
  functionCall: true,
@@ -56,7 +56,7 @@ const Bedrock: ModelProviderCard = {
56
56
  },
57
57
  {
58
58
  description:
59
- 'Claude 3 Opus is Anthropic most powerful AI model, with state-of-the-art performance on highly complex tasks. It can navigate open-ended prompts and sight-unseen scenarios with remarkable fluency and human-like understanding. Claude 3 Opus shows us the frontier of what’s possible with generative AI. Claude 3 Opus can process images and return text outputs, and features a 200K context window.',
59
+ 'Claude 3 Opus Anthropic 最强大的 AI 模型,具有在高度复杂任务上的最先进性能。它可以处理开放式提示和未见过的场景,具有出色的流畅性和类人的理解能力。Claude 3 Opus 展示了生成 AI 可能性的前沿。Claude 3 Opus 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
60
60
  displayName: 'Claude 3 Opus',
61
61
  enabled: true,
62
62
  functionCall: true,
@@ -70,7 +70,7 @@ const Bedrock: ModelProviderCard = {
70
70
  },
71
71
  {
72
72
  description:
73
- 'Claude 3 Haiku is Anthropic fastest, most compact model for near-instant responsiveness. It answers simple queries and requests with speed. Customers will be able to build seamless AI experiences that mimic human interactions. Claude 3 Haiku can process images and return text outputs, and features a 200K context window.',
73
+ 'Claude 3 Haiku Anthropic 最快、最紧凑的模型,提供近乎即时的响应速度。它可以快速回答简单的查询和请求。客户将能够构建模仿人类互动的无缝 AI 体验。Claude 3 Haiku 可以处理图像并返回文本输出,具有 200K 的上下文窗口。',
74
74
  displayName: 'Claude 3 Haiku',
75
75
  enabled: true,
76
76
  functionCall: true,
@@ -84,7 +84,7 @@ const Bedrock: ModelProviderCard = {
84
84
  },
85
85
  {
86
86
  description:
87
- 'An update to Claude 2 that features double the context window, plus improvements across reliability, hallucination rates, and evidence-based accuracy in long document and RAG contexts.',
87
+ 'Claude 2 的更新版,具有双倍的上下文窗口,以及在长文档和 RAG 上下文中的可靠性、幻觉率和基于证据的准确性的改进。',
88
88
  displayName: 'Claude 2.1',
89
89
  id: 'anthropic.claude-v2:1',
90
90
  pricing: {
@@ -95,7 +95,7 @@ const Bedrock: ModelProviderCard = {
95
95
  },
96
96
  {
97
97
  description:
98
- 'Anthropic highly capable model across a wide range of tasks from sophisticated dialogue and creative content generation to detailed instruction following.',
98
+ 'Anthropic 在从复杂对话和创意内容生成到详细指令跟随的广泛任务中都表现出高度能力的模型。',
99
99
  displayName: 'Claude 2.0',
100
100
  id: 'anthropic.claude-v2',
101
101
  pricing: {
@@ -106,7 +106,7 @@ const Bedrock: ModelProviderCard = {
106
106
  },
107
107
  {
108
108
  description:
109
- 'A fast, affordable yet still very capable model, which can handle a range of tasks including casual dialogue, text analysis, summarization, and document question-answering.',
109
+ '一款快速、经济且仍然非常有能力的模型,可以处理包括日常对话、文本分析、总结和文档问答在内的一系列任务。',
110
110
  displayName: 'Claude Instant',
111
111
  id: 'anthropic.claude-instant-v1',
112
112
  pricing: {
@@ -117,7 +117,7 @@ const Bedrock: ModelProviderCard = {
117
117
  },
118
118
  {
119
119
  description:
120
- 'An update to Meta Llama 3 8B Instruct that includes an expanded 128K context length, multilinguality and improved reasoning capabilities. The Llama 3.1 offering of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction-tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Llama 3.1 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks. The Llama 3.1 models also support the ability to leverage the outputs of its models to improve other models including synthetic data generation and distillation. Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.',
120
+ 'Meta Llama 3.1 8B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B70B 405B 大小 (文本输入/输出)Llama 3.1 指令调整的文本模型 (8B70B405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
121
121
  displayName: 'Llama 3.1 8B Instruct',
122
122
  enabled: true,
123
123
  functionCall: true,
@@ -130,7 +130,7 @@ const Bedrock: ModelProviderCard = {
130
130
  },
131
131
  {
132
132
  description:
133
- 'An update to Meta Llama 3 70B Instruct that includes an expanded 128K context length, multilinguality and improved reasoning capabilities. The Llama 3.1 offering of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction-tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Llama 3.1 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks. The Llama 3.1 models also support the ability to leverage the outputs of its models to improve other models including synthetic data generation and distillation. Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.',
133
+ 'Meta Llama 3.1 70B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B70B 405B 大小 (文本输入/输出)Llama 3.1 指令调整的文本模型 (8B70B405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
134
134
  displayName: 'Llama 3.1 70B Instruct',
135
135
  enabled: true,
136
136
  functionCall: true,
@@ -143,7 +143,7 @@ const Bedrock: ModelProviderCard = {
143
143
  },
144
144
  {
145
145
  description:
146
- 'Meta Llama 3.1 405B Instruct is the largest and most powerful of the Llama 3.1 Instruct models that is a highly advanced model for conversational inference and reasoning, synthetic data generation, and a base to do specialized continual pre-training or fine-tuning on a specific domain. The Llama 3.1 offering of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 8B, 70B and 405B sizes (text in/text out). The Llama 3.1 instruction-tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Llama 3.1 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistant-like chat, whereas pretrained models can be adapted for a variety of natural language generation tasks. The Llama 3.1 models also support the ability to leverage the outputs of its models to improve other models including synthetic data generation and distillation. Llama 3.1 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.',
146
+ 'Meta Llama 3.1 405B Instruct Llama 3.1 Instruct 模型中最大、最强大的模型,是一款高度先进的对话推理和合成数据生成模型,也可以用作在特定领域进行专业持续预训练或微调的基础。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B70B 405B 大小 (文本输入/输出)Llama 3.1 指令调整的文本模型 (8B70B405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。',
147
147
  displayName: 'Llama 3.1 405B Instruct',
148
148
  enabled: true,
149
149
  functionCall: true,
@@ -156,7 +156,7 @@ const Bedrock: ModelProviderCard = {
156
156
  },
157
157
  {
158
158
  description:
159
- 'Meta Llama 3 is an accessible, open large language model (LLM) designed for developers, researchers, and businesses to build, experiment, and responsibly scale their generative AI ideas. Part of a foundational system, it serves as a bedrock for innovation in the global community. Ideal for limited computational power and resources, edge devices, and faster training times.',
159
+ 'Meta Llama 3 是一款面向开发者、研究人员和企业的开放大型语言模型 (LLM),旨在帮助他们构建、实验并负责任地扩展他们的生成 AI 想法。作为全球社区创新的基础系统的一部分,它非常适合计算能力和资源有限、边缘设备和更快的训练时间。',
160
160
  displayName: 'Llama 3 8B Instruct',
161
161
  id: 'meta.llama3-8b-instruct-v1:0',
162
162
  pricing: {
@@ -167,7 +167,7 @@ const Bedrock: ModelProviderCard = {
167
167
  },
168
168
  {
169
169
  description:
170
- 'Meta Llama 3 is an accessible, open large language model (LLM) designed for developers, researchers, and businesses to build, experiment, and responsibly scale their generative AI ideas. Part of a foundational system, it serves as a bedrock for innovation in the global community. Ideal for content creation, conversational AI, language understanding, R&D, and Enterprise applications.',
170
+ 'Meta Llama 3 是一款面向开发者、研究人员和企业的开放大型语言模型 (LLM),旨在帮助他们构建、实验并负责任地扩展他们的生成 AI 想法。作为全球社区创新的基础系统的一部分,它非常适合内容创建、对话 AI、语言理解、研发和企业应用。',
171
171
  displayName: 'Llama 3 70B Instruct',
172
172
  id: 'meta.llama3-70b-instruct-v1:0',
173
173
  pricing: {
@@ -276,6 +276,8 @@ const Bedrock: ModelProviderCard = {
276
276
  */
277
277
  ],
278
278
  checkModel: 'anthropic.claude-instant-v1',
279
+ description:
280
+ 'Bedrock 是亚马逊 AWS 提供的一项服务,专注于为企业提供先进的 AI 语言模型和视觉模型。其模型家族包括 Anthropic 的 Claude 系列、Meta 的 Llama 3.1 系列等,涵盖从轻量级到高性能的多种选择,支持文本生成、对话、图像处理等多种任务,适用于不同规模和需求的企业应用。',
279
281
  id: 'bedrock',
280
282
  modelsUrl: 'https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html',
281
283
  name: 'Bedrock',
@@ -1,6 +1,6 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
- // ref https://platform.deepseek.com/api-docs/pricing
3
+ // ref: https://platform.deepseek.com/api-docs/pricing
4
4
  const DeepSeek: ModelProviderCard = {
5
5
  chatModels: [
6
6
  {
@@ -20,6 +20,8 @@ const DeepSeek: ModelProviderCard = {
20
20
  },
21
21
  ],
22
22
  checkModel: 'deepseek-chat',
23
+ description:
24
+ 'DeepSeek 是一家专注于人工智能技术研究和应用的公司,其最新模型 DeepSeek-V2.5 融合了通用对话和代码处理能力,并在人类偏好对齐、写作任务和指令跟随等方面实现了显著提升。',
23
25
  id: 'deepseek',
24
26
  modelList: { showModelFetcher: true },
25
27
  modelsUrl: 'https://platform.deepseek.com/api-docs/zh-cn/quick_start/pricing',
@@ -1,12 +1,12 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
- // ref https://fireworks.ai/models?show=Serverless
4
- // ref https://fireworks.ai/pricing
3
+ // ref: https://fireworks.ai/models?show=Serverless
4
+ // ref: https://fireworks.ai/pricing
5
5
  const FireworksAI: ModelProviderCard = {
6
6
  chatModels: [
7
7
  {
8
8
  description:
9
- 'Fireworks latest and most performant function-calling model. Firefunction-v2 is based on Llama-3 and trained to excel at function-calling as well as chat and instruction-following. See blog post for more details https://fireworks.ai/blog/firefunction-v2-launch-post',
9
+ 'Fireworks 公司最新推出的 Firefunction-v2 是一款性能卓越的函数调用模型,基于 Llama-3 开发,并通过大量优化,特别适用于函数调用、对话及指令跟随等场景。',
10
10
  displayName: 'Firefunction V2',
11
11
  enabled: true,
12
12
  functionCall: true,
@@ -14,7 +14,7 @@ const FireworksAI: ModelProviderCard = {
14
14
  tokens: 8192,
15
15
  },
16
16
  {
17
- description: 'Fireworks open-source function calling model.',
17
+ description: 'Fireworks 开源函数调用模型,提供卓越的指令执行能力和开放可定制的特性。',
18
18
  displayName: 'Firefunction V1',
19
19
  functionCall: true,
20
20
  id: 'accounts/fireworks/models/firefunction-v1',
@@ -22,7 +22,7 @@ const FireworksAI: ModelProviderCard = {
22
22
  },
23
23
  {
24
24
  description:
25
- 'Vision-language model allowing both image and text as inputs (single image is recommended), trained on OSS model generated training data and open sourced on huggingface at fireworks-ai/FireLLaVA-13b',
25
+ 'fireworks-ai/FireLLaVA-13b 是一款视觉语言模型,可以同时接收图像和文本输入,经过高质量数据训练,适合多模态任务。',
26
26
  displayName: 'FireLLaVA-13B',
27
27
  enabled: true,
28
28
  functionCall: false,
@@ -31,6 +31,8 @@ const FireworksAI: ModelProviderCard = {
31
31
  vision: true,
32
32
  },
33
33
  {
34
+ description:
35
+ 'Llama 3.1 8B 指令模型,专为多语言对话优化,能够在常见行业基准上超越多数开源及闭源模型。',
34
36
  displayName: 'Llama 3.1 8B Instruct',
35
37
  enabled: true,
36
38
  functionCall: false,
@@ -38,6 +40,8 @@ const FireworksAI: ModelProviderCard = {
38
40
  tokens: 131_072,
39
41
  },
40
42
  {
43
+ description:
44
+ 'Llama 3.1 70B 指令模型,提供卓越的自然语言理解和生成能力,是对话及分析任务的理想选择。',
41
45
  displayName: 'Llama 3.1 70B Instruct',
42
46
  enabled: true,
43
47
  functionCall: false,
@@ -45,6 +49,8 @@ const FireworksAI: ModelProviderCard = {
45
49
  tokens: 131_072,
46
50
  },
47
51
  {
52
+ description:
53
+ 'Llama 3.1 405B 指令模型,具备超大规模参数,适合复杂任务和高负载场景下的指令跟随。',
48
54
  displayName: 'Llama 3.1 405B Instruct',
49
55
  enabled: true,
50
56
  functionCall: false,
@@ -52,30 +58,38 @@ const FireworksAI: ModelProviderCard = {
52
58
  tokens: 131_072,
53
59
  },
54
60
  {
61
+ description: 'Llama 3 8B 指令模型,优化用于对话及多语言任务,表现卓越且高效。',
55
62
  displayName: 'Llama 3 8B Instruct',
56
63
  functionCall: false,
57
64
  id: 'accounts/fireworks/models/llama-v3-8b-instruct',
58
65
  tokens: 8192,
59
66
  },
60
67
  {
68
+ description: 'Llama 3 70B 指令模型,专为多语言对话和自然语言理解优化,性能优于多数竞争模型。',
61
69
  displayName: 'Llama 3 70B Instruct',
62
70
  functionCall: false,
63
71
  id: 'accounts/fireworks/models/llama-v3-70b-instruct',
64
72
  tokens: 8192,
65
73
  },
66
74
  {
75
+ description:
76
+ 'Llama 3 8B 指令模型(HF 版本),与官方实现结果一致,具备高度一致性和跨平台兼容性。',
67
77
  displayName: 'Llama 3 8B Instruct (HF version)',
68
78
  functionCall: false,
69
79
  id: 'accounts/fireworks/models/llama-v3-8b-instruct-hf',
70
80
  tokens: 8192,
71
81
  },
72
82
  {
83
+ description:
84
+ 'Llama 3 70B 指令模型(HF 版本),与官方实现结果保持一致,适合高质量的指令跟随任务。',
73
85
  displayName: 'Llama 3 70B Instruct (HF version)',
74
86
  functionCall: false,
75
87
  id: 'accounts/fireworks/models/llama-v3-70b-instruct-hf',
76
88
  tokens: 8192,
77
89
  },
78
90
  {
91
+ description:
92
+ 'Gemma 2 9B 指令模型,基于之前的Google技术,适合回答问题、总结和推理等多种文本生成任务。',
79
93
  displayName: 'Gemma 2 9B Instruct',
80
94
  enabled: true,
81
95
  functionCall: false,
@@ -83,6 +97,7 @@ const FireworksAI: ModelProviderCard = {
83
97
  tokens: 8192,
84
98
  },
85
99
  {
100
+ description: 'Mixtral MoE 8x7B 指令模型,多专家架构提供高效的指令跟随及执行。',
86
101
  displayName: 'Mixtral MoE 8x7B Instruct',
87
102
  enabled: true,
88
103
  functionCall: false,
@@ -90,6 +105,8 @@ const FireworksAI: ModelProviderCard = {
90
105
  tokens: 32_768,
91
106
  },
92
107
  {
108
+ description:
109
+ 'Mixtral MoE 8x22B 指令模型,大规模参数和多专家架构,全方位支持复杂任务的高效处理。',
93
110
  displayName: 'Mixtral MoE 8x22B Instruct',
94
111
  enabled: true,
95
112
  functionCall: false,
@@ -97,12 +114,16 @@ const FireworksAI: ModelProviderCard = {
97
114
  tokens: 65_536,
98
115
  },
99
116
  {
117
+ description:
118
+ 'Mixtral MoE 8x7B 指令模型(HF 版本),性能与官方实现一致,适合多种高效任务场景。',
100
119
  displayName: 'Mixtral MoE 8x7B Instruct (HF version)',
101
120
  functionCall: false,
102
121
  id: 'accounts/fireworks/models/mixtral-8x7b-instruct-hf',
103
122
  tokens: 32_768,
104
123
  },
105
124
  {
125
+ description:
126
+ 'Phi 3 Vision 指令模型,轻量级多模态模型,能够处理复杂的视觉和文本信息,具备较强的推理能力。',
106
127
  displayName: 'Phi 3 Vision Instruct',
107
128
  enabled: true,
108
129
  functionCall: false,
@@ -111,6 +132,7 @@ const FireworksAI: ModelProviderCard = {
111
132
  vision: true,
112
133
  },
113
134
  {
135
+ description: 'Yi-Large 模型,具备卓越的多语言处理能力,可用于各类语言生成和理解任务。',
114
136
  displayName: 'Yi-Large',
115
137
  enabled: true,
116
138
  functionCall: false,
@@ -118,18 +140,22 @@ const FireworksAI: ModelProviderCard = {
118
140
  tokens: 32_768,
119
141
  },
120
142
  {
143
+ description: 'StarCoder 7B 模型,针对80多种编程语言训练,拥有出色的编程填充能力和语境理解。',
121
144
  displayName: 'StarCoder 7B',
122
145
  functionCall: false,
123
146
  id: 'accounts/fireworks/models/starcoder-7b',
124
147
  tokens: 8192,
125
148
  },
126
149
  {
150
+ description:
151
+ 'StarCoder 15.5B 模型,支持高级编程任务,多语言能力增强,适合复杂代码生成和理解。',
127
152
  displayName: 'StarCoder 15.5B',
128
153
  functionCall: false,
129
154
  id: 'accounts/fireworks/models/starcoder-16b',
130
155
  tokens: 8192,
131
156
  },
132
157
  {
158
+ description: 'MythoMax L2 13B 模型,结合新颖的合并技术,擅长叙事和角色扮演。',
133
159
  displayName: 'MythoMax L2 13b',
134
160
  functionCall: false,
135
161
  id: 'accounts/fireworks/models/mythomax-l2-13b',
@@ -137,6 +163,8 @@ const FireworksAI: ModelProviderCard = {
137
163
  },
138
164
  ],
139
165
  checkModel: 'accounts/fireworks/models/firefunction-v2',
166
+ description:
167
+ 'Fireworks AI 是一家领先的高级语言模型服务商,专注于功能调用和多模态处理。其最新模型 Firefunction V2 基于 Llama-3,优化用于函数调用、对话及指令跟随。视觉语言模型 FireLLaVA-13B 支持图像和文本混合输入。其他 notable 模型包括 Llama 系列和 Mixtral 系列,提供高效的多语言指令跟随与生成支持。',
140
168
  id: 'fireworksai',
141
169
  modelList: { showModelFetcher: true },
142
170
  modelsUrl: 'https://fireworks.ai/models?show=Serverless',
@@ -1,9 +1,11 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
+ // ref: https://ai.google.dev/gemini-api/docs/models/gemini
3
4
  const Google: ModelProviderCard = {
4
5
  chatModels: [
5
6
  {
6
- description: 'Fast and versatile multimodal model for scaling across diverse tasks',
7
+ description:
8
+ 'Gemini 1.5 Flash 是Google最新的多模态AI模型,具备快速处理能力,支持文本、图像和视频输入,适用于多种任务的高效扩展。',
7
9
  displayName: 'Gemini 1.5 Flash',
8
10
  enabled: true,
9
11
  functionCall: true,
@@ -18,7 +20,7 @@ const Google: ModelProviderCard = {
18
20
  vision: true,
19
21
  },
20
22
  {
21
- description: 'Fast and versatile multimodal model for scaling across diverse tasks',
23
+ description: 'Gemini 1.5 Flash 0827 提供了优化后的多模态处理能力,适用多种复杂任务场景。',
22
24
  displayName: 'Gemini 1.5 Flash 0827',
23
25
  enabled: true,
24
26
  functionCall: true,
@@ -34,7 +36,7 @@ const Google: ModelProviderCard = {
34
36
  vision: true,
35
37
  },
36
38
  {
37
- description: 'Fast and versatile multimodal model for scaling across diverse tasks',
39
+ description: 'Gemini 1.5 Flash 8B 0827 专为处理大规模任务场景设计,提供无与伦比的处理速度。',
38
40
  displayName: 'Gemini 1.5 Flash 8B 0827',
39
41
  functionCall: true,
40
42
  id: 'gemini-1.5-flash-8b-exp-0827',
@@ -44,7 +46,7 @@ const Google: ModelProviderCard = {
44
46
  vision: true,
45
47
  },
46
48
  {
47
- description: 'Fast and versatile multimodal model for scaling across diverse tasks',
49
+ description: 'Gemini 1.5 Flash 001 是一款高效的多模态模型,支持广泛应用的扩展。',
48
50
  displayName: 'Gemini 1.5 Flash 001',
49
51
  functionCall: true,
50
52
  id: 'gemini-1.5-flash-001',
@@ -53,7 +55,8 @@ const Google: ModelProviderCard = {
53
55
  vision: true,
54
56
  },
55
57
  {
56
- description: 'Mid-size multimodal model that supports up to 2 million tokens',
58
+ description:
59
+ 'Gemini 1.5 Pro 支持高达200万个tokens,是中型多模态模型的理想选择,适用于复杂任务的多方面支持。',
57
60
  displayName: 'Gemini 1.5 Pro',
58
61
  enabled: true,
59
62
  functionCall: true,
@@ -69,7 +72,7 @@ const Google: ModelProviderCard = {
69
72
  vision: true,
70
73
  },
71
74
  {
72
- description: 'Mid-size multimodal model that supports up to 2 million tokens',
75
+ description: 'Gemini 1.5 Pro 0827 结合最新优化技术,带来更高效的多模态数据处理能力。',
73
76
  displayName: 'Gemini 1.5 Pro 0827',
74
77
  enabled: true,
75
78
  functionCall: true,
@@ -85,7 +88,7 @@ const Google: ModelProviderCard = {
85
88
  vision: true,
86
89
  },
87
90
  {
88
- description: 'Mid-size multimodal model that supports up to 2 million tokens',
91
+ description: 'Gemini 1.5 Pro 0801 提供出色的多模态处理能力,为应用开发带来更大灵活性。',
89
92
  displayName: 'Gemini 1.5 Pro 0801',
90
93
  functionCall: true,
91
94
  id: 'gemini-1.5-pro-exp-0801',
@@ -100,7 +103,7 @@ const Google: ModelProviderCard = {
100
103
  vision: true,
101
104
  },
102
105
  {
103
- description: 'Mid-size multimodal model that supports up to 2 million tokens',
106
+ description: 'Gemini 1.5 Pro 001 是可扩展的多模态AI解决方案,支持广泛的复杂任务。',
104
107
  displayName: 'Gemini 1.5 Pro 001',
105
108
  functionCall: true,
106
109
  id: 'gemini-1.5-pro-001',
@@ -115,8 +118,7 @@ const Google: ModelProviderCard = {
115
118
  vision: true,
116
119
  },
117
120
  {
118
- description:
119
- 'The best model for scaling across a wide range of tasks. This is the latest model.',
121
+ description: 'Gemini 1.0 Pro 是Google的高性能AI模型,专为广泛任务扩展而设计。',
120
122
  displayName: 'Gemini 1.0 Pro',
121
123
  id: 'gemini-1.0-pro-latest',
122
124
  maxOutput: 2048,
@@ -129,7 +131,7 @@ const Google: ModelProviderCard = {
129
131
  },
130
132
  {
131
133
  description:
132
- 'The best model for scaling across a wide range of tasks. This is a stable model that supports tuning.',
134
+ 'Gemini 1.0 Pro 001 (Tuning) 提供稳定并可调优的性能,是复杂任务解决方案的理想选择。',
133
135
  displayName: 'Gemini 1.0 Pro 001 (Tuning)',
134
136
  functionCall: true,
135
137
  id: 'gemini-1.0-pro-001',
@@ -142,8 +144,7 @@ const Google: ModelProviderCard = {
142
144
  tokens: 30_720 + 2048,
143
145
  },
144
146
  {
145
- description:
146
- 'The best model for scaling across a wide range of tasks. Released April 9, 2024.',
147
+ description: 'Gemini 1.0 Pro 002 (Tuning) 提供出色的多模态支持,专注于复杂任务的有效解决。',
147
148
  displayName: 'Gemini 1.0 Pro 002 (Tuning)',
148
149
  id: 'gemini-1.0-pro-002',
149
150
  maxOutput: 2048,
@@ -156,6 +157,8 @@ const Google: ModelProviderCard = {
156
157
  },
157
158
  ],
158
159
  checkModel: 'gemini-1.5-flash-latest',
160
+ description:
161
+ 'Google 的 Gemini 系列是其最先进、通用的 A I模型,由 Google DeepMind 打造,专为多模态设计,支持文本、代码、图像、音频和视频的无缝理解与处理。适用于从数据中心到移动设备的多种环境,极大提升了AI模型的效率与应用广泛性。',
159
162
  id: 'google',
160
163
  modelsUrl: 'https://ai.google.dev/gemini-api/docs/models/gemini',
161
164
  name: 'Google',
@@ -5,6 +5,8 @@ const Groq: ModelProviderCard = {
5
5
  chatModels: [
6
6
  // TODO: During preview launch, Groq is limiting 3.1 models to max_tokens of 8k.
7
7
  {
8
+ description:
9
+ 'Llama 3.1 8B 是一款高效能模型,提供了快速的文本生成能力,非常适合需要大规模效率和成本效益的应用场景。',
8
10
  displayName: 'Llama 3.1 8B (Preview)',
9
11
  enabled: true,
10
12
  functionCall: true,
@@ -17,6 +19,8 @@ const Groq: ModelProviderCard = {
17
19
  tokens: 131_072,
18
20
  },
19
21
  {
22
+ description:
23
+ 'Llama 3.1 70B 提供更强大的AI推理能力,适合复杂应用,支持超多的计算处理并保证高效和准确率。',
20
24
  displayName: 'Llama 3.1 70B (Preview)',
21
25
  enabled: true,
22
26
  functionCall: true,
@@ -38,7 +42,8 @@ const Groq: ModelProviderCard = {
38
42
  },
39
43
  */
40
44
  {
41
- displayName: 'Llama 3 Groq 8B Tool Use Preview 8K',
45
+ description: 'Llama 3 Groq 8B Tool Use 是针对高效工具使用优化的模型,支持快速并行计算。',
46
+ displayName: 'Llama 3 Groq 8B Tool Use (Preview)',
42
47
  enabled: true,
43
48
  functionCall: true,
44
49
  id: 'llama3-groq-8b-8192-tool-use-preview',
@@ -49,7 +54,8 @@ const Groq: ModelProviderCard = {
49
54
  tokens: 8192,
50
55
  },
51
56
  {
52
- displayName: 'Llama 3 Groq 70B Tool Use Preview 8K',
57
+ description: 'Llama 3 Groq 70B Tool Use 提供强大的工具调用能力,支持复杂任务的高效处理。',
58
+ displayName: 'Llama 3 Groq 70B Tool Use (Preview)',
53
59
  enabled: true,
54
60
  functionCall: true,
55
61
  id: 'llama3-groq-70b-8192-tool-use-preview',
@@ -60,8 +66,8 @@ const Groq: ModelProviderCard = {
60
66
  tokens: 8192,
61
67
  },
62
68
  {
69
+ description: 'Meta Llama 3 8B 带来优质的推理效能,适合多场景应用需求。',
63
70
  displayName: 'Meta Llama 3 8B',
64
- enabled: true,
65
71
  functionCall: true,
66
72
  id: 'llama3-8b-8192',
67
73
  pricing: {
@@ -71,8 +77,8 @@ const Groq: ModelProviderCard = {
71
77
  tokens: 8192,
72
78
  },
73
79
  {
80
+ description: 'Meta Llama 3 70B 提供无与伦比的复杂性处理能力,为高要求项目量身定制。',
74
81
  displayName: 'Meta Llama 3 70B',
75
- enabled: true,
76
82
  functionCall: true,
77
83
  id: 'llama3-70b-8192',
78
84
  pricing: {
@@ -82,7 +88,8 @@ const Groq: ModelProviderCard = {
82
88
  tokens: 8192,
83
89
  },
84
90
  {
85
- displayName: 'Gemma 2 9B 8k',
91
+ description: 'Gemma 2 9B 是一款优化用于特定任务和工具整合的模型。',
92
+ displayName: 'Gemma 2 9B',
86
93
  enabled: true,
87
94
  functionCall: true,
88
95
  id: 'gemma2-9b-it',
@@ -93,7 +100,8 @@ const Groq: ModelProviderCard = {
93
100
  tokens: 8192,
94
101
  },
95
102
  {
96
- displayName: 'Gemma 7B 8k Instruct',
103
+ description: 'Gemma 7B 适合中小规模任务处理,兼具成本效益。',
104
+ displayName: 'Gemma 7B',
97
105
  functionCall: true,
98
106
  id: 'gemma-7b-it',
99
107
  pricing: {
@@ -103,8 +111,8 @@ const Groq: ModelProviderCard = {
103
111
  tokens: 8192,
104
112
  },
105
113
  {
106
- displayName: 'Mixtral 8x7B Instruct 32k',
107
- enabled: true,
114
+ description: 'Mixtral 8x7B 提供高容错的并行计算能力,适合复杂任务。',
115
+ displayName: 'Mixtral 8x7B',
108
116
  functionCall: true,
109
117
  id: 'mixtral-8x7b-32768',
110
118
  pricing: {
@@ -114,6 +122,7 @@ const Groq: ModelProviderCard = {
114
122
  tokens: 32_768,
115
123
  },
116
124
  {
125
+ description: 'LLaVA 1.5 7B 提供视觉处理能力融合,通过视觉信息输入生成复杂输出。',
117
126
  displayName: 'LLaVA 1.5 7B',
118
127
  enabled: true,
119
128
  id: 'llava-v1.5-7b-4096-preview',
@@ -122,6 +131,8 @@ const Groq: ModelProviderCard = {
122
131
  },
123
132
  ],
124
133
  checkModel: 'gemma2-9b-it',
134
+ description:
135
+ 'Groq 的 LPU 推理引擎在最新的独立大语言模型(LLM)基准测试中表现卓越,以其惊人的速度和效率重新定义了 AI 解决方案的标准。Groq 是一种即时推理速度的代表,在基于云的部署中展现了良好的性能。',
125
136
  id: 'groq',
126
137
  modelsUrl: 'https://console.groq.com/docs/models',
127
138
  name: 'Groq',
@@ -1,10 +1,10 @@
1
1
  import { ModelProviderCard } from '@/types/llm';
2
2
 
3
- // ref https://www.minimaxi.com/document/guides/chat-model/pro/api
3
+ // ref: https://platform.minimaxi.com/document/Models
4
4
  const Minimax: ModelProviderCard = {
5
5
  chatModels: [
6
6
  {
7
- description: '通用场景',
7
+ description: '适用于广泛的自然语言处理任务,包括文本生成、对话系统等。',
8
8
  displayName: 'abab6.5s',
9
9
  enabled: true,
10
10
  functionCall: true,
@@ -12,7 +12,7 @@ const Minimax: ModelProviderCard = {
12
12
  tokens: 245_760,
13
13
  },
14
14
  {
15
- description: '英文等多语种人设对话场景',
15
+ description: '专为多语种人设对话设计,支持英文及其他多种语言的高质量对话生成。',
16
16
  displayName: 'abab6.5g',
17
17
  enabled: true,
18
18
  functionCall: true,
@@ -20,7 +20,7 @@ const Minimax: ModelProviderCard = {
20
20
  tokens: 8192,
21
21
  },
22
22
  {
23
- description: '中文人设对话场景',
23
+ description: '针对中文人设对话场景优化,提供流畅且符合中文表达习惯的对话生成能力。',
24
24
  displayName: 'abab6.5t',
25
25
  enabled: true,
26
26
  functionCall: true,
@@ -28,19 +28,21 @@ const Minimax: ModelProviderCard = {
28
28
  tokens: 8192,
29
29
  },
30
30
  {
31
- description: '生产力场景',
31
+ description: '面向生产力场景,支持复杂任务处理和高效文本生成,适用于专业领域应用。',
32
32
  displayName: 'abab5.5',
33
33
  id: 'abab5.5-chat',
34
34
  tokens: 16_384,
35
35
  },
36
36
  {
37
- description: '中文人设对话场景',
37
+ description: '专为中文人设对话场景设计,提供高质量的中文对话生成能力,适用于多种应用场景。',
38
38
  displayName: 'abab5.5s',
39
39
  id: 'abab5.5s-chat',
40
40
  tokens: 8192,
41
41
  },
42
42
  ],
43
43
  checkModel: 'abab6.5s-chat',
44
+ description:
45
+ 'MiniMax 是 2021 年成立的通用人工智能科技公司,致力于与用户共创智能。MiniMax 自主研发了不同模态的通用大模型,其中包括万亿参数的 MoE 文本大模型、语音大模型以及图像大模型。并推出了海螺 AI 等应用。',
44
46
  id: 'minimax',
45
47
  modelsUrl: 'https://platform.minimaxi.com/document/Models',
46
48
  name: 'Minimax',