@lobehub/chat 1.33.4 → 1.34.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.i18nrc.js +8 -2
- package/.releaserc.js +10 -1
- package/CHANGELOG.md +51 -16624
- package/changelog/CHANGELOG.v0.md +16621 -0
- package/changelog/v0.json +6064 -0
- package/changelog/v1.json +3356 -0
- package/docs/changelog/2024-11-25-november-providers.mdx +19 -0
- package/docs/changelog/2024-11-25-november-providers.zh-CN.mdx +17 -0
- package/docs/changelog/index.json +12 -0
- package/docs/changelog/schema.json +70 -0
- package/docs/self-hosting/advanced/auth/clerk.mdx +19 -23
- package/docs/self-hosting/advanced/auth/clerk.zh-CN.mdx +5 -4
- package/docs/self-hosting/advanced/auth/next-auth/authelia.mdx +2 -4
- package/docs/self-hosting/advanced/auth/next-auth/authelia.zh-CN.mdx +3 -5
- package/docs/self-hosting/advanced/auth/next-auth/authentik.zh-CN.mdx +2 -2
- package/docs/self-hosting/advanced/auth/next-auth/casdoor.mdx +49 -44
- package/docs/self-hosting/advanced/auth/next-auth/casdoor.zh-CN.mdx +42 -41
- package/docs/self-hosting/advanced/auth/next-auth/logto.mdx +29 -21
- package/docs/self-hosting/advanced/auth/next-auth/logto.zh-CN.mdx +2 -1
- package/docs/self-hosting/advanced/auth.mdx +10 -10
- package/docs/self-hosting/advanced/auth.zh-CN.mdx +10 -10
- package/docs/self-hosting/advanced/feature-flags.zh-CN.mdx +1 -1
- package/docs/self-hosting/advanced/model-list.mdx +1 -1
- package/docs/self-hosting/advanced/s3/cloudflare-r2.mdx +17 -12
- package/docs/self-hosting/advanced/s3/tencent-cloud.mdx +33 -20
- package/docs/self-hosting/advanced/s3.mdx +31 -28
- package/docs/self-hosting/advanced/s3.zh-CN.mdx +1 -0
- package/docs/self-hosting/advanced/webrtc.mdx +1 -0
- package/docs/self-hosting/environment-variables/s3.mdx +4 -3
- package/docs/self-hosting/environment-variables/s3.zh-CN.mdx +1 -0
- package/docs/self-hosting/environment-variables.mdx +6 -9
- package/docs/self-hosting/platform/alibaba-cloud.mdx +1 -2
- package/docs/self-hosting/platform/alibaba-cloud.zh-CN.mdx +1 -2
- package/docs/self-hosting/platform/btpanel.mdx +7 -13
- package/docs/self-hosting/platform/btpanel.zh-CN.mdx +8 -13
- package/docs/self-hosting/platform/docker.zh-CN.mdx +2 -1
- package/docs/self-hosting/platform/netlify.zh-CN.mdx +2 -1
- package/docs/self-hosting/server-database/docker.mdx +18 -5
- package/docs/self-hosting/server-database/docker.zh-CN.mdx +9 -6
- package/docs/self-hosting/server-database/netlify.mdx +0 -1
- package/docs/self-hosting/server-database/railway.mdx +0 -1
- package/docs/self-hosting/server-database/repocloud.mdx +5 -2
- package/docs/self-hosting/server-database/repocloud.zh-CN.mdx +23 -3
- package/docs/self-hosting/server-database/sealos.mdx +3 -0
- package/docs/self-hosting/server-database/vercel.mdx +35 -32
- package/docs/self-hosting/server-database/vercel.zh-CN.mdx +25 -25
- package/docs/self-hosting/server-database/zeabur.mdx +2 -2
- package/docs/self-hosting/server-database/zeabur.zh-CN.mdx +3 -4
- package/docs/self-hosting/server-database.mdx +23 -8
- package/docs/self-hosting/start.mdx +8 -2
- package/docs/usage/features/database.zh-CN.mdx +1 -1
- package/docs/usage/foundation/text2image.mdx +1 -2
- package/docs/usage/foundation/text2image.zh-CN.mdx +1 -3
- package/docs/usage/providers/ai21.mdx +14 -15
- package/docs/usage/providers/ai21.zh-CN.mdx +1 -3
- package/docs/usage/providers/ai360.mdx +14 -15
- package/docs/usage/providers/ai360.zh-CN.mdx +1 -3
- package/docs/usage/providers/cloudflare.mdx +1 -1
- package/docs/usage/providers/fireworksai.mdx +19 -19
- package/docs/usage/providers/fireworksai.zh-CN.mdx +1 -3
- package/docs/usage/providers/github.mdx +22 -21
- package/docs/usage/providers/github.zh-CN.mdx +6 -6
- package/docs/usage/providers/hunyuan.mdx +17 -18
- package/docs/usage/providers/hunyuan.zh-CN.mdx +1 -3
- package/docs/usage/providers/siliconcloud.mdx +14 -15
- package/docs/usage/providers/siliconcloud.zh-CN.mdx +3 -5
- package/docs/usage/providers/spark.mdx +17 -18
- package/docs/usage/providers/spark.zh-CN.mdx +1 -3
- package/docs/usage/providers/upstage.mdx +14 -15
- package/docs/usage/providers/upstage.zh-CN.mdx +1 -3
- package/docs/usage/providers/wenxin.mdx +17 -18
- package/docs/usage/providers/wenxin.zh-CN.mdx +1 -3
- package/docs/usage/providers/zeroone.mdx +2 -2
- package/locales/ar/chat.json +7 -0
- package/locales/ar/common.json +2 -0
- package/locales/ar/models.json +24 -0
- package/locales/ar/providers.json +3 -0
- package/locales/ar/setting.json +5 -0
- package/locales/ar/thread.json +5 -0
- package/locales/bg-BG/chat.json +7 -0
- package/locales/bg-BG/common.json +2 -0
- package/locales/bg-BG/models.json +24 -0
- package/locales/bg-BG/providers.json +3 -0
- package/locales/bg-BG/setting.json +5 -0
- package/locales/bg-BG/thread.json +5 -0
- package/locales/de-DE/chat.json +7 -0
- package/locales/de-DE/common.json +2 -0
- package/locales/de-DE/models.json +24 -0
- package/locales/de-DE/providers.json +3 -0
- package/locales/de-DE/setting.json +5 -0
- package/locales/de-DE/thread.json +5 -0
- package/locales/en-US/chat.json +7 -0
- package/locales/en-US/common.json +2 -0
- package/locales/en-US/models.json +24 -0
- package/locales/en-US/providers.json +3 -0
- package/locales/en-US/setting.json +5 -0
- package/locales/en-US/thread.json +5 -0
- package/locales/es-ES/chat.json +7 -0
- package/locales/es-ES/common.json +2 -0
- package/locales/es-ES/models.json +24 -0
- package/locales/es-ES/providers.json +3 -0
- package/locales/es-ES/setting.json +5 -0
- package/locales/es-ES/thread.json +5 -0
- package/locales/fa-IR/chat.json +7 -0
- package/locales/fa-IR/common.json +2 -0
- package/locales/fa-IR/models.json +24 -0
- package/locales/fa-IR/providers.json +3 -0
- package/locales/fa-IR/setting.json +5 -0
- package/locales/fa-IR/thread.json +5 -0
- package/locales/fr-FR/chat.json +7 -0
- package/locales/fr-FR/common.json +2 -0
- package/locales/fr-FR/models.json +24 -0
- package/locales/fr-FR/providers.json +3 -0
- package/locales/fr-FR/setting.json +5 -0
- package/locales/fr-FR/thread.json +5 -0
- package/locales/it-IT/chat.json +7 -0
- package/locales/it-IT/common.json +2 -0
- package/locales/it-IT/models.json +24 -0
- package/locales/it-IT/providers.json +3 -0
- package/locales/it-IT/setting.json +5 -0
- package/locales/it-IT/thread.json +5 -0
- package/locales/ja-JP/chat.json +7 -0
- package/locales/ja-JP/common.json +2 -0
- package/locales/ja-JP/models.json +24 -0
- package/locales/ja-JP/providers.json +3 -0
- package/locales/ja-JP/setting.json +5 -0
- package/locales/ja-JP/thread.json +5 -0
- package/locales/ko-KR/chat.json +7 -0
- package/locales/ko-KR/common.json +2 -0
- package/locales/ko-KR/models.json +24 -0
- package/locales/ko-KR/providers.json +3 -0
- package/locales/ko-KR/setting.json +5 -0
- package/locales/ko-KR/thread.json +5 -0
- package/locales/nl-NL/chat.json +7 -0
- package/locales/nl-NL/common.json +2 -0
- package/locales/nl-NL/models.json +24 -0
- package/locales/nl-NL/providers.json +3 -0
- package/locales/nl-NL/setting.json +5 -0
- package/locales/nl-NL/thread.json +5 -0
- package/locales/pl-PL/chat.json +7 -0
- package/locales/pl-PL/common.json +2 -0
- package/locales/pl-PL/models.json +24 -0
- package/locales/pl-PL/providers.json +3 -0
- package/locales/pl-PL/setting.json +5 -0
- package/locales/pl-PL/thread.json +5 -0
- package/locales/pt-BR/chat.json +7 -0
- package/locales/pt-BR/common.json +2 -0
- package/locales/pt-BR/models.json +24 -0
- package/locales/pt-BR/providers.json +3 -0
- package/locales/pt-BR/setting.json +5 -0
- package/locales/pt-BR/thread.json +5 -0
- package/locales/ru-RU/chat.json +7 -0
- package/locales/ru-RU/common.json +2 -0
- package/locales/ru-RU/models.json +24 -0
- package/locales/ru-RU/providers.json +3 -0
- package/locales/ru-RU/setting.json +5 -0
- package/locales/ru-RU/thread.json +5 -0
- package/locales/tr-TR/chat.json +7 -0
- package/locales/tr-TR/common.json +2 -0
- package/locales/tr-TR/models.json +24 -0
- package/locales/tr-TR/providers.json +3 -0
- package/locales/tr-TR/setting.json +5 -0
- package/locales/tr-TR/thread.json +5 -0
- package/locales/vi-VN/chat.json +7 -0
- package/locales/vi-VN/common.json +2 -0
- package/locales/vi-VN/models.json +24 -0
- package/locales/vi-VN/providers.json +3 -0
- package/locales/vi-VN/setting.json +5 -0
- package/locales/vi-VN/thread.json +5 -0
- package/locales/zh-CN/chat.json +7 -0
- package/locales/zh-CN/common.json +2 -0
- package/locales/zh-CN/models.json +24 -0
- package/locales/zh-CN/providers.json +3 -0
- package/locales/zh-CN/setting.json +5 -0
- package/locales/zh-CN/thread.json +5 -0
- package/locales/zh-TW/chat.json +7 -0
- package/locales/zh-TW/common.json +2 -0
- package/locales/zh-TW/models.json +24 -0
- package/locales/zh-TW/providers.json +3 -0
- package/locales/zh-TW/setting.json +5 -0
- package/locales/zh-TW/thread.json +5 -0
- package/package.json +6 -1
- package/scripts/changelogWorkflow/buildStaticChangelog.ts +135 -0
- package/scripts/changelogWorkflow/const.ts +11 -0
- package/scripts/changelogWorkflow/index.ts +10 -0
- package/src/app/(main)/chat/(workspace)/@conversation/default.tsx +2 -0
- package/src/app/(main)/chat/(workspace)/@conversation/features/ChatHydration/index.tsx +11 -2
- package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/index.tsx +7 -9
- package/src/app/(main)/chat/(workspace)/@conversation/features/ChatInput/Desktop/index.tsx +7 -2
- package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/Thread.tsx +62 -0
- package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/ThreadItem.tsx +68 -0
- package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/index.tsx +62 -2
- package/src/app/(main)/chat/(workspace)/@conversation/features/ThreadHydration.tsx +47 -0
- package/src/app/(main)/chat/(workspace)/@portal/_layout/Desktop.tsx +3 -2
- package/src/app/(main)/chat/(workspace)/@portal/_layout/Mobile.tsx +47 -6
- package/src/app/(main)/chat/(workspace)/@topic/features/SkeletonList.tsx +3 -2
- package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ByTimeMode/index.tsx +10 -3
- package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/FlatMode/index.tsx +1 -1
- package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ThreadItem/Content.tsx +164 -0
- package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ThreadItem/index.tsx +98 -0
- package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{TopicItem.tsx → TopicItem/index.tsx} +33 -22
- package/src/app/(main)/chat/(workspace)/_layout/Desktop/Portal.tsx +12 -5
- package/src/app/(main)/chat/(workspace)/_layout/Mobile/index.tsx +1 -2
- package/src/const/message.ts +2 -0
- package/src/const/settings/systemAgent.ts +1 -0
- package/src/database/server/migrations/0012_add_thread.sql +39 -0
- package/src/database/server/migrations/meta/0012_snapshot.json +3671 -0
- package/src/database/server/migrations/meta/_journal.json +7 -0
- package/src/database/server/models/_template.ts +2 -2
- package/src/database/server/models/message.ts +1 -0
- package/src/database/server/models/thread.ts +79 -0
- package/src/database/server/schemas/lobechat/message.ts +2 -1
- package/src/database/server/schemas/lobechat/relations.ts +13 -1
- package/src/database/server/schemas/lobechat/topic.ts +30 -1
- package/src/database/server/utils/idGenerator.ts +1 -0
- package/src/features/ChatInput/ActionBar/Token/TokenTag.tsx +6 -4
- package/src/features/ChatInput/ActionBar/Token/index.tsx +24 -5
- package/src/features/ChatInput/ActionBar/config.ts +3 -2
- package/src/features/ChatInput/Desktop/index.tsx +15 -7
- package/src/features/ChatInput/Mobile/index.tsx +4 -4
- package/src/features/Conversation/Actions/Assistant.tsx +24 -5
- package/src/features/Conversation/Actions/User.tsx +21 -4
- package/src/features/Conversation/Actions/index.ts +1 -66
- package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/index.tsx +3 -1
- package/src/features/Conversation/Messages/{Tool/index.tsx → Assistant/ToolCallItem/Tool.tsx} +10 -11
- package/src/features/Conversation/Messages/Assistant/ToolCallItem/index.tsx +5 -3
- package/src/features/Conversation/Messages/Assistant/index.tsx +22 -14
- package/src/features/Conversation/Messages/index.ts +0 -2
- package/src/features/Conversation/components/AutoScroll.tsx +1 -1
- package/src/features/Conversation/components/ChatItem/ActionsBar.tsx +79 -5
- package/src/features/Conversation/components/ChatItem/InPortalThreadContext.ts +3 -0
- package/src/features/Conversation/components/ChatItem/index.tsx +16 -5
- package/src/features/Conversation/components/MarkdownElements/LobeArtifact/Render/index.tsx +9 -1
- package/src/features/Conversation/components/ThreadDivider/index.tsx +19 -0
- package/src/features/Conversation/hooks/useChatListActionsBar.tsx +19 -4
- package/src/features/Portal/Thread/Chat/ChatInput/Footer.tsx +90 -0
- package/src/features/Portal/Thread/Chat/ChatInput/TextArea.tsx +30 -0
- package/src/features/Portal/Thread/Chat/ChatInput/index.tsx +66 -0
- package/src/features/Portal/Thread/Chat/ChatInput/useSend.ts +50 -0
- package/src/features/Portal/Thread/Chat/ChatItem.tsx +62 -0
- package/src/features/Portal/Thread/Chat/ChatList.tsx +49 -0
- package/src/features/Portal/Thread/Chat/ThreadDivider/index.tsx +19 -0
- package/src/features/Portal/Thread/Chat/index.tsx +28 -0
- package/src/features/Portal/Thread/Header/Active.tsx +35 -0
- package/src/features/Portal/Thread/Header/New.tsx +37 -0
- package/src/features/Portal/Thread/Header/Title.tsx +18 -0
- package/src/features/Portal/Thread/Header/index.tsx +20 -0
- package/src/features/Portal/Thread/hook.ts +8 -0
- package/src/features/Portal/Thread/index.ts +12 -0
- package/src/features/Portal/router.tsx +2 -1
- package/src/hooks/useFetchTopics.ts +7 -1
- package/src/locales/default/chat.ts +8 -1
- package/src/locales/default/common.ts +3 -0
- package/src/locales/default/index.ts +2 -0
- package/src/locales/default/setting.ts +5 -0
- package/src/locales/default/thread.ts +5 -0
- package/src/server/routers/lambda/index.ts +2 -0
- package/src/server/routers/lambda/thread.ts +83 -0
- package/src/services/thread.ts +54 -0
- package/src/store/chat/initialState.ts +3 -0
- package/src/store/chat/selectors.ts +2 -1
- package/src/store/chat/slices/aiChat/actions/__tests__/generateAIChat.test.ts +1 -1
- package/src/store/chat/slices/aiChat/actions/__tests__/rag.test.ts +1 -1
- package/src/store/chat/slices/aiChat/actions/generateAIChat.ts +31 -8
- package/src/store/chat/slices/aiChat/actions/rag.ts +1 -1
- package/src/store/chat/slices/message/selectors.test.ts +3 -3
- package/src/store/chat/slices/message/selectors.ts +50 -29
- package/src/store/chat/slices/plugin/action.ts +26 -8
- package/src/store/chat/slices/portal/action.ts +1 -0
- package/src/store/chat/slices/portal/initialState.ts +1 -0
- package/src/store/chat/slices/portal/selectors/thread.ts +17 -0
- package/src/store/chat/slices/portal/selectors.ts +2 -0
- package/src/store/chat/slices/thread/action.ts +326 -0
- package/src/store/chat/slices/thread/initialState.ts +34 -0
- package/src/store/chat/slices/thread/reducer.ts +48 -0
- package/src/store/chat/slices/thread/selectors/index.ts +202 -0
- package/src/store/chat/slices/thread/selectors/util.ts +22 -0
- package/src/store/chat/slices/topic/action.ts +5 -1
- package/src/store/chat/store.ts +5 -2
- package/src/store/global/initialState.ts +4 -0
- package/src/store/global/selectors.ts +4 -0
- package/src/store/user/slices/settings/selectors/systemAgent.ts +2 -0
- package/src/types/message/index.ts +17 -1
- package/src/types/topic/index.ts +1 -0
- package/src/types/topic/thread.ts +42 -0
- package/src/types/user/settings/systemAgent.ts +1 -0
- package/src/app/(main)/chat/(workspace)/@portal/features/Header.tsx +0 -11
- package/src/app/(main)/chat/(workspace)/_layout/Mobile/PortalModal.tsx +0 -35
- /package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/SendMore.tsx +0 -0
- /package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/ShortcutHint.tsx +0 -0
- /package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{DefaultContent.tsx → TopicItem/DefaultContent.tsx} +0 -0
- /package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{TopicContent.tsx → TopicItem/TopicContent.tsx} +0 -0
- /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/PluginResultJSON.tsx +0 -0
- /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/Settings.tsx +0 -0
- /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/style.ts +0 -0
@@ -176,6 +176,15 @@
|
|
176
176
|
"Qwen/Qwen2.5-Math-72B-Instruct": {
|
177
177
|
"description": "Qwen2.5-Math si concentra sulla risoluzione di problemi nel campo della matematica, fornendo risposte professionali a domande di alta difficoltà."
|
178
178
|
},
|
179
|
+
"Qwen2-72B-Instruct": {
|
180
|
+
"description": "Qwen2 è l'ultima serie del modello Qwen, supporta un contesto di 128k, e rispetto ai modelli open source attualmente migliori, Qwen2-72B supera significativamente i modelli leader attuali in comprensione del linguaggio naturale, conoscenza, codice, matematica e capacità multilingue."
|
181
|
+
},
|
182
|
+
"Qwen2-7B-Instruct": {
|
183
|
+
"description": "Qwen2 è l'ultima serie del modello Qwen, in grado di superare i modelli open source ottimali di dimensioni simili e anche modelli di dimensioni maggiori. Qwen2 7B ha ottenuto vantaggi significativi in vari test, in particolare nella comprensione del codice e del cinese."
|
184
|
+
},
|
185
|
+
"Qwen2.5-72B-Instruct": {
|
186
|
+
"description": "Qwen2.5-72B-Instruct supporta un contesto di 16k, generando testi lunghi oltre 8K. Supporta chiamate di funzione e interazioni senza soluzione di continuità con sistemi esterni, aumentando notevolmente flessibilità e scalabilità. La conoscenza del modello è notevolmente aumentata e ha migliorato significativamente le capacità di codifica e matematica, con supporto per oltre 29 lingue."
|
187
|
+
},
|
179
188
|
"SenseChat": {
|
180
189
|
"description": "Modello di base (V4), lunghezza del contesto di 4K, con potenti capacità generali."
|
181
190
|
},
|
@@ -206,6 +215,9 @@
|
|
206
215
|
"Tencent/Hunyuan-A52B-Instruct": {
|
207
216
|
"description": "Hunyuan-Large è il modello MoE con architettura Transformer open source più grande del settore, con un totale di 389 miliardi di parametri e 52 miliardi di parametri attivi."
|
208
217
|
},
|
218
|
+
"Yi-34B-Chat": {
|
219
|
+
"description": "Yi-1.5-34B, mantenendo le eccellenti capacità linguistiche generali della serie originale, ha notevolmente migliorato la logica matematica e le capacità di codifica attraverso un addestramento incrementale su 500 miliardi di token di alta qualità."
|
220
|
+
},
|
209
221
|
"abab5.5-chat": {
|
210
222
|
"description": "Focalizzato su scenari di produttività, supporta l'elaborazione di compiti complessi e la generazione di testo efficiente, adatto per applicazioni professionali."
|
211
223
|
},
|
@@ -368,6 +380,9 @@
|
|
368
380
|
"codegeex-4": {
|
369
381
|
"description": "CodeGeeX-4 è un potente assistente di programmazione AI, supporta domande intelligenti e completamento del codice in vari linguaggi di programmazione, migliorando l'efficienza dello sviluppo."
|
370
382
|
},
|
383
|
+
"codegeex4-all-9b": {
|
384
|
+
"description": "CodeGeeX4-ALL-9B è un modello di generazione di codice multilingue, supporta funzionalità complete tra cui completamento e generazione di codice, interprete di codice, ricerca in rete, chiamate di funzione e domande e risposte sul codice a livello di repository, coprendo vari scenari di sviluppo software. È un modello di generazione di codice di punta con meno di 10B di parametri."
|
385
|
+
},
|
371
386
|
"codegemma": {
|
372
387
|
"description": "CodeGemma è un modello linguistico leggero dedicato a vari compiti di programmazione, supporta iterazioni rapide e integrazione."
|
373
388
|
},
|
@@ -422,6 +437,9 @@
|
|
422
437
|
"deepseek-chat": {
|
423
438
|
"description": "Un nuovo modello open source che integra capacità generali e di codifica, mantenendo non solo le capacità conversazionali generali del modello Chat originale, ma anche la potente capacità di elaborazione del codice del modello Coder, allineandosi meglio alle preferenze umane. Inoltre, DeepSeek-V2.5 ha ottenuto notevoli miglioramenti in vari aspetti, come i compiti di scrittura e il rispetto delle istruzioni."
|
424
439
|
},
|
440
|
+
"deepseek-coder-33B-instruct": {
|
441
|
+
"description": "DeepSeek Coder 33B è un modello di linguaggio per codice, addestrato su 20 trilioni di dati, di cui l'87% è codice e il 13% è in cinese e inglese. Il modello introduce una dimensione della finestra di 16K e compiti di completamento, fornendo funzionalità di completamento del codice e riempimento di frammenti a livello di progetto."
|
442
|
+
},
|
425
443
|
"deepseek-coder-v2": {
|
426
444
|
"description": "DeepSeek Coder V2 è un modello di codice open source di esperti misti, eccelle nei compiti di codice, paragonabile a GPT4-Turbo."
|
427
445
|
},
|
@@ -476,6 +494,9 @@
|
|
476
494
|
"gemini-exp-1114": {
|
477
495
|
"description": "Gemini Exp 1114 è il modello AI multimodale sperimentale più recente di Google, dotato di capacità di elaborazione rapida, supporta input testuali, visivi e video, ed è adatto per un'ampia gamma di compiti con un'efficiente scalabilità."
|
478
496
|
},
|
497
|
+
"gemini-exp-1121": {
|
498
|
+
"description": "Gemini Exp 1121 è il più recente modello AI multimodale sperimentale di Google, dotato di capacità di elaborazione rapida, supporta input di testo, immagini e video, ed è adatto per un'ampia gamma di compiti con un'efficace scalabilità."
|
499
|
+
},
|
479
500
|
"gemma-7b-it": {
|
480
501
|
"description": "Gemma 7B è adatto per l'elaborazione di compiti di piccole e medie dimensioni, combinando efficienza dei costi."
|
481
502
|
},
|
@@ -503,6 +524,9 @@
|
|
503
524
|
"glm-4-0520": {
|
504
525
|
"description": "GLM-4-0520 è l'ultima versione del modello, progettata per compiti altamente complessi e diversificati, con prestazioni eccezionali."
|
505
526
|
},
|
527
|
+
"glm-4-9b-chat": {
|
528
|
+
"description": "GLM-4-9B-Chat mostra elevate prestazioni in vari aspetti come semantica, matematica, ragionamento, codice e conoscenza. Ha anche la capacità di navigare in rete, eseguire codice, chiamare strumenti personalizzati e inferire testi lunghi. Supporta 26 lingue, tra cui giapponese, coreano e tedesco."
|
529
|
+
},
|
506
530
|
"glm-4-air": {
|
507
531
|
"description": "GLM-4-Air è una versione economica, con prestazioni simili a GLM-4, che offre velocità elevate a un prezzo accessibile."
|
508
532
|
},
|
@@ -22,6 +22,9 @@
|
|
22
22
|
"fireworksai": {
|
23
23
|
"description": "Fireworks AI è un fornitore leader di servizi di modelli linguistici avanzati, focalizzato su chiamate funzionali e elaborazione multimodale. Il suo ultimo modello Firefunction V2, basato su Llama-3, è ottimizzato per chiamate di funzione, dialogo e rispetto delle istruzioni. Il modello di linguaggio visivo FireLLaVA-13B supporta input misti di immagini e testo. Altri modelli notevoli includono la serie Llama e la serie Mixtral, offrendo supporto efficiente per il rispetto e la generazione di istruzioni multilingue."
|
24
24
|
},
|
25
|
+
"giteeai": {
|
26
|
+
"description": "L'API Serverless di Gitee AI fornisce agli sviluppatori di AI un servizio API di inferenza di modelli di grandi dimensioni fuori dagli schemi."
|
27
|
+
},
|
25
28
|
"github": {
|
26
29
|
"description": "Con i modelli di GitHub, gli sviluppatori possono diventare ingegneri AI e costruire con i modelli AI leader del settore."
|
27
30
|
},
|
@@ -389,6 +389,11 @@
|
|
389
389
|
"modelDesc": "Modello specificato per ottimizzare le domande degli utenti",
|
390
390
|
"title": "Riscrittura delle domande del knowledge base"
|
391
391
|
},
|
392
|
+
"thread": {
|
393
|
+
"label": "Modello di denominazione dei sottoargomenti",
|
394
|
+
"modelDesc": "Modello designato per la rinominazione automatica dei sottoargomenti",
|
395
|
+
"title": "Rinominazione automatica dei sottoargomenti"
|
396
|
+
},
|
392
397
|
"title": "Assistente di sistema",
|
393
398
|
"topic": {
|
394
399
|
"label": "Modello di denominazione degli argomenti",
|
package/locales/ja-JP/chat.json
CHANGED
@@ -8,6 +8,7 @@
|
|
8
8
|
"agents": "エージェント",
|
9
9
|
"artifact": {
|
10
10
|
"generating": "生成中",
|
11
|
+
"inThread": "サブトピックでは表示できません。メインの対話エリアに切り替えてください。",
|
11
12
|
"thinking": "思考中",
|
12
13
|
"thought": "思考過程",
|
13
14
|
"unknownTitle": "未命名の作品"
|
@@ -65,6 +66,7 @@
|
|
65
66
|
},
|
66
67
|
"messageAction": {
|
67
68
|
"delAndRegenerate": "削除して再生成",
|
69
|
+
"deleteDisabledByThreads": "サブトピックが存在するため、削除できません。",
|
68
70
|
"regenerate": "再生成"
|
69
71
|
},
|
70
72
|
"newAgent": "新しいエージェント",
|
@@ -121,6 +123,11 @@
|
|
121
123
|
"loading": "認識中...",
|
122
124
|
"prettifying": "整形中..."
|
123
125
|
},
|
126
|
+
"thread": {
|
127
|
+
"divider": "サブトピック",
|
128
|
+
"threadMessageCount": "{{messageCount}} 件のメッセージ",
|
129
|
+
"title": "サブトピック"
|
130
|
+
},
|
124
131
|
"tokenDetails": {
|
125
132
|
"chats": "チャットメッセージ",
|
126
133
|
"historySummary": "履歴の要約",
|
@@ -16,6 +16,8 @@
|
|
16
16
|
"back": "戻る",
|
17
17
|
"batchDelete": "バッチ削除",
|
18
18
|
"blog": "製品ブログ",
|
19
|
+
"branching": "サブトピックを作成",
|
20
|
+
"branchingDisable": "「サブトピック」機能はサーバー版のみで利用可能です。この機能が必要な場合は、サーバー展開モードに切り替えるか、LobeChat Cloudを使用してください。",
|
19
21
|
"cancel": "キャンセル",
|
20
22
|
"changelog": "変更履歴",
|
21
23
|
"close": "閉じる",
|
@@ -176,6 +176,15 @@
|
|
176
176
|
"Qwen/Qwen2.5-Math-72B-Instruct": {
|
177
177
|
"description": "Qwen2.5-Mathは、数学分野の問題解決に特化しており、高難度の問題に対して専門的な解答を提供します。"
|
178
178
|
},
|
179
|
+
"Qwen2-72B-Instruct": {
|
180
|
+
"description": "Qwen2はQwenモデルの最新シリーズで、128kのコンテキストをサポートしています。現在の最適なオープンソースモデルと比較して、Qwen2-72Bは自然言語理解、知識、コード、数学、そして多言語などの能力において、現在のリーディングモデルを大幅に上回っています。"
|
181
|
+
},
|
182
|
+
"Qwen2-7B-Instruct": {
|
183
|
+
"description": "Qwen2はQwenモデルの最新シリーズで、同等の規模の最適なオープンソースモデルやそれ以上の規模のモデルを超えることができ、Qwen2 7Bは複数の評価で顕著な優位性を示し、特にコードと中国語理解において優れています。"
|
184
|
+
},
|
185
|
+
"Qwen2.5-72B-Instruct": {
|
186
|
+
"description": "Qwen2.5-72B-Instructは、16kのコンテキストをサポートし、8Kを超える長文を生成します。関数呼び出しと外部システムとのシームレスなインタラクションをサポートし、柔軟性と拡張性を大幅に向上させました。モデルの知識は明らかに増加し、コーディングと数学の能力が大幅に向上し、29以上の言語をサポートしています。"
|
187
|
+
},
|
179
188
|
"SenseChat": {
|
180
189
|
"description": "基本バージョンのモデル (V4)、4Kのコンテキスト長で、汎用能力が強力です。"
|
181
190
|
},
|
@@ -206,6 +215,9 @@
|
|
206
215
|
"Tencent/Hunyuan-A52B-Instruct": {
|
207
216
|
"description": "Hunyuan-Largeは業界最大のオープンソースTransformerアーキテクチャMoEモデルで、3890億の総パラメータ数と520億のアクティブパラメータ数を持っています。"
|
208
217
|
},
|
218
|
+
"Yi-34B-Chat": {
|
219
|
+
"description": "Yi-1.5-34Bは、元のシリーズモデルの優れた汎用言語能力を維持しつつ、5000億の高品質トークンを増分トレーニングすることで、数学的論理とコーディング能力を大幅に向上させました。"
|
220
|
+
},
|
209
221
|
"abab5.5-chat": {
|
210
222
|
"description": "生産性シーン向けであり、複雑なタスク処理と効率的なテキスト生成をサポートし、専門分野のアプリケーションに適しています。"
|
211
223
|
},
|
@@ -368,6 +380,9 @@
|
|
368
380
|
"codegeex-4": {
|
369
381
|
"description": "CodeGeeX-4は強力なAIプログラミングアシスタントで、さまざまなプログラミング言語のインテリジェントな質問応答とコード補完をサポートし、開発効率を向上させます。"
|
370
382
|
},
|
383
|
+
"codegeex4-all-9b": {
|
384
|
+
"description": "CodeGeeX4-ALL-9Bは、多言語コード生成モデルで、コード補完と生成、コードインタープリター、ウェブ検索、関数呼び出し、リポジトリレベルのコードQ&Aを含む包括的な機能をサポートし、ソフトウェア開発のさまざまなシーンをカバーしています。パラメータが10B未満のトップクラスのコード生成モデルです。"
|
385
|
+
},
|
371
386
|
"codegemma": {
|
372
387
|
"description": "CodeGemmaは、さまざまなプログラミングタスクに特化した軽量言語モデルであり、迅速な反復と統合をサポートします。"
|
373
388
|
},
|
@@ -422,6 +437,9 @@
|
|
422
437
|
"deepseek-chat": {
|
423
438
|
"description": "一般的な対話能力と強力なコード処理能力を兼ね備えた新しいオープンソースモデルであり、元のChatモデルの対話能力とCoderモデルのコード処理能力を保持しつつ、人間の好みにより良く整合しています。さらに、DeepSeek-V2.5は、執筆タスクや指示に従う能力など、さまざまな面で大幅な向上を実現しました。"
|
424
439
|
},
|
440
|
+
"deepseek-coder-33B-instruct": {
|
441
|
+
"description": "DeepSeek Coder 33Bは、2兆のデータを基にトレーニングされたコード言語モデルで、そのうち87%がコード、13%が中英語です。モデルは16Kのウィンドウサイズと穴埋めタスクを導入し、プロジェクトレベルのコード補完とスニペット埋め機能を提供します。"
|
442
|
+
},
|
425
443
|
"deepseek-coder-v2": {
|
426
444
|
"description": "DeepSeek Coder V2は、オープンソースの混合エキスパートコードモデルであり、コードタスクにおいて優れた性能を発揮し、GPT4-Turboに匹敵します。"
|
427
445
|
},
|
@@ -476,6 +494,9 @@
|
|
476
494
|
"gemini-exp-1114": {
|
477
495
|
"description": "Gemini Exp 1114は、Googleの最新の実験的なマルチモーダルAIモデルで、迅速な処理能力を備え、テキスト、画像、動画の入力をサポートし、様々なタスクに効率的に対応できます。"
|
478
496
|
},
|
497
|
+
"gemini-exp-1121": {
|
498
|
+
"description": "Gemini Exp 1121は、Googleの最新の実験的なマルチモーダルAIモデルで、高速処理能力を備え、テキスト、画像、動画の入力をサポートし、さまざまなタスクに対して効率的に拡張できます。"
|
499
|
+
},
|
479
500
|
"gemma-7b-it": {
|
480
501
|
"description": "Gemma 7Bは、中小規模のタスク処理に適しており、コスト効果を兼ね備えています。"
|
481
502
|
},
|
@@ -503,6 +524,9 @@
|
|
503
524
|
"glm-4-0520": {
|
504
525
|
"description": "GLM-4-0520は最新のモデルバージョンで、高度に複雑で多様なタスクのために設計され、優れたパフォーマンスを発揮します。"
|
505
526
|
},
|
527
|
+
"glm-4-9b-chat": {
|
528
|
+
"description": "GLM-4-9B-Chatは、意味、数学、推論、コード、知識などの多方面で高い性能を示しています。また、ウェブブラウジング、コード実行、カスタムツール呼び出し、長文推論を備えています。日本語、韓国語、ドイツ語を含む26の言語をサポートしています。"
|
529
|
+
},
|
506
530
|
"glm-4-air": {
|
507
531
|
"description": "GLM-4-Airはコストパフォーマンスが高いバージョンで、GLM-4に近い性能を提供し、高速かつ手頃な価格です。"
|
508
532
|
},
|
@@ -22,6 +22,9 @@
|
|
22
22
|
"fireworksai": {
|
23
23
|
"description": "Fireworks AIは、先進的な言語モデルサービスのリーダーであり、機能呼び出しと多モーダル処理に特化しています。最新のモデルFirefunction V2はLlama-3に基づいており、関数呼び出し、対話、指示の遵守に最適化されています。視覚言語モデルFireLLaVA-13Bは、画像とテキストの混合入力をサポートしています。他の注目すべきモデルには、LlamaシリーズやMixtralシリーズがあり、高効率の多言語指示遵守と生成サポートを提供しています。"
|
24
24
|
},
|
25
|
+
"giteeai": {
|
26
|
+
"description": "Gitee AIのServerless APIは、AI開発者に開梱即使用の大モデル推論APIサービスを提供する。"
|
27
|
+
},
|
25
28
|
"github": {
|
26
29
|
"description": "GitHubモデルを使用することで、開発者はAIエンジニアになり、業界をリードするAIモデルを使って構築できます。"
|
27
30
|
},
|
@@ -389,6 +389,11 @@
|
|
389
389
|
"modelDesc": "ユーザーの質問を最適化するために指定されたモデル",
|
390
390
|
"title": "知識ベースの質問の書き換え"
|
391
391
|
},
|
392
|
+
"thread": {
|
393
|
+
"label": "サブトピック命名モデル",
|
394
|
+
"modelDesc": "サブトピックの自動命名に使用されるモデルを指定します",
|
395
|
+
"title": "サブトピックの自動命名"
|
396
|
+
},
|
392
397
|
"title": "システムアシスタント",
|
393
398
|
"topic": {
|
394
399
|
"label": "トピックネーミングモデル",
|
package/locales/ko-KR/chat.json
CHANGED
@@ -8,6 +8,7 @@
|
|
8
8
|
"agents": "도우미",
|
9
9
|
"artifact": {
|
10
10
|
"generating": "생성 중",
|
11
|
+
"inThread": "하위 주제에서는 볼 수 없습니다. 주 대화 영역으로 전환하여 열어주세요.",
|
11
12
|
"thinking": "생각 중",
|
12
13
|
"thought": "사고 과정",
|
13
14
|
"unknownTitle": "제목 없음"
|
@@ -65,6 +66,7 @@
|
|
65
66
|
},
|
66
67
|
"messageAction": {
|
67
68
|
"delAndRegenerate": "삭제하고 다시 생성",
|
69
|
+
"deleteDisabledByThreads": "하위 주제가 존재하여 삭제할 수 없습니다.",
|
68
70
|
"regenerate": "다시 생성"
|
69
71
|
},
|
70
72
|
"newAgent": "새 도우미",
|
@@ -121,6 +123,11 @@
|
|
121
123
|
"loading": "인식 중...",
|
122
124
|
"prettifying": "정제 중..."
|
123
125
|
},
|
126
|
+
"thread": {
|
127
|
+
"divider": "하위 주제",
|
128
|
+
"threadMessageCount": "{{messageCount}}개의 메시지",
|
129
|
+
"title": "하위 주제"
|
130
|
+
},
|
124
131
|
"tokenDetails": {
|
125
132
|
"chats": "채팅 메시지",
|
126
133
|
"historySummary": "역사 요약",
|
@@ -16,6 +16,8 @@
|
|
16
16
|
"back": "뒤로",
|
17
17
|
"batchDelete": "일괄 삭제",
|
18
18
|
"blog": "제품 블로그",
|
19
|
+
"branching": "하위 주제 만들기",
|
20
|
+
"branchingDisable": "「하위 주제」 기능은 서버 버전에서만 사용할 수 있습니다. 이 기능이 필요하시면 서버 배포 모드로 전환하거나 LobeChat Cloud를 사용하세요.",
|
19
21
|
"cancel": "취소",
|
20
22
|
"changelog": "변경 로그",
|
21
23
|
"close": "닫기",
|
@@ -176,6 +176,15 @@
|
|
176
176
|
"Qwen/Qwen2.5-Math-72B-Instruct": {
|
177
177
|
"description": "Qwen2.5-Math는 수학 분야의 문제 해결에 중점을 두고 있으며, 고난이도 문제에 대한 전문적인 해답을 제공합니다."
|
178
178
|
},
|
179
|
+
"Qwen2-72B-Instruct": {
|
180
|
+
"description": "Qwen2는 Qwen 모델의 최신 시리즈로, 128k 컨텍스트를 지원합니다. 현재 최상의 오픈 소스 모델과 비교할 때, Qwen2-72B는 자연어 이해, 지식, 코드, 수학 및 다국어 등 여러 능력에서 현재 선도하는 모델을 현저히 초월합니다."
|
181
|
+
},
|
182
|
+
"Qwen2-7B-Instruct": {
|
183
|
+
"description": "Qwen2는 Qwen 모델의 최신 시리즈로, 동등한 규모의 최적 오픈 소스 모델은 물론 더 큰 규모의 모델을 초월할 수 있습니다. Qwen2 7B는 여러 평가에서 현저한 우위를 차지하였으며, 특히 코드 및 중국어 이해에서 두드러진 성과를 보였습니다."
|
184
|
+
},
|
185
|
+
"Qwen2.5-72B-Instruct": {
|
186
|
+
"description": "Qwen2.5-72B-Instruct는 16k 컨텍스트를 지원하며, 8K를 초과하는 긴 텍스트를 생성할 수 있습니다. 함수 호출 및 외부 시스템과의 원활한 상호작용을 지원하여 유연성과 확장성을 크게 향상시킵니다. 모델의 지식이 현저히 증가하였고, 인코딩 및 수학 능력이 크게 향상되었으며, 29개 이상의 언어를 지원합니다."
|
187
|
+
},
|
179
188
|
"SenseChat": {
|
180
189
|
"description": "기본 버전 모델(V4), 4K 컨텍스트 길이, 일반적인 능력이 강력합니다."
|
181
190
|
},
|
@@ -206,6 +215,9 @@
|
|
206
215
|
"Tencent/Hunyuan-A52B-Instruct": {
|
207
216
|
"description": "Hunyuan-Large는 업계에서 가장 큰 오픈 소스 Transformer 아키텍처 MoE 모델로, 총 3890억 개의 매개변수와 520억 개의 활성 매개변수를 가지고 있습니다."
|
208
217
|
},
|
218
|
+
"Yi-34B-Chat": {
|
219
|
+
"description": "Yi-1.5-34B는 원래 시리즈 모델의 뛰어난 일반 언어 능력을 유지하면서, 5000억 개의 고품질 토큰을 통해 점진적으로 훈련하여 수학적 논리 및 코드 능력을 크게 향상시켰습니다."
|
220
|
+
},
|
209
221
|
"abab5.5-chat": {
|
210
222
|
"description": "생산성 시나리오를 위해 설계되었으며, 복잡한 작업 처리 및 효율적인 텍스트 생성을 지원하여 전문 분야 응용에 적합합니다."
|
211
223
|
},
|
@@ -368,6 +380,9 @@
|
|
368
380
|
"codegeex-4": {
|
369
381
|
"description": "CodeGeeX-4는 강력한 AI 프로그래밍 도우미로, 다양한 프로그래밍 언어에 대한 스마트 Q&A 및 코드 완성을 지원하여 개발 효율성을 높입니다."
|
370
382
|
},
|
383
|
+
"codegeex4-all-9b": {
|
384
|
+
"description": "CodeGeeX4-ALL-9B는 다국어 코드 생성 모델로, 코드 완성 및 생성, 코드 해석기, 웹 검색, 함수 호출, 저장소 수준의 코드 질문 응답 등 다양한 기능을 지원하여 소프트웨어 개발의 여러 시나리오를 포괄합니다. 10B 미만의 매개변수를 가진 최고의 코드 생성 모델입니다."
|
385
|
+
},
|
371
386
|
"codegemma": {
|
372
387
|
"description": "CodeGemma는 다양한 프로그래밍 작업을 위한 경량 언어 모델로, 빠른 반복 및 통합을 지원합니다."
|
373
388
|
},
|
@@ -422,6 +437,9 @@
|
|
422
437
|
"deepseek-chat": {
|
423
438
|
"description": "일반 및 코드 능력을 융합한 새로운 오픈 소스 모델로, 기존 Chat 모델의 일반 대화 능력과 Coder 모델의 강력한 코드 처리 능력을 유지하면서 인간의 선호에 더 잘 맞춰졌습니다. 또한, DeepSeek-V2.5는 작문 작업, 지시 따르기 등 여러 측면에서 큰 향상을 이루었습니다."
|
424
439
|
},
|
440
|
+
"deepseek-coder-33B-instruct": {
|
441
|
+
"description": "DeepSeek Coder 33B는 코드 언어 모델로, 20조 개의 데이터로 훈련되었으며, 그 중 87%는 코드, 13%는 중문 및 영문입니다. 모델은 16K 창 크기와 빈칸 채우기 작업을 도입하여 프로젝트 수준의 코드 완성과 코드 조각 채우기 기능을 제공합니다."
|
442
|
+
},
|
425
443
|
"deepseek-coder-v2": {
|
426
444
|
"description": "DeepSeek Coder V2는 오픈 소스 혼합 전문가 코드 모델로, 코드 작업에서 뛰어난 성능을 발휘하며, GPT4-Turbo와 경쟁할 수 있습니다."
|
427
445
|
},
|
@@ -476,6 +494,9 @@
|
|
476
494
|
"gemini-exp-1114": {
|
477
495
|
"description": "Gemini Exp 1114는 Google의 최신 실험적 다중 모달 AI 모델로, 빠른 처리 능력을 갖추고 있으며, 텍스트, 이미지 및 비디오 입력을 지원하여 다양한 작업에 효율적으로 확장할 수 있습니다."
|
478
496
|
},
|
497
|
+
"gemini-exp-1121": {
|
498
|
+
"description": "Gemini Exp 1121은 Google의 최신 실험적 다중 모달 AI 모델로, 빠른 처리 능력을 갖추고 있으며 텍스트, 이미지 및 비디오 입력을 지원하여 다양한 작업에 효율적으로 확장할 수 있습니다."
|
499
|
+
},
|
479
500
|
"gemma-7b-it": {
|
480
501
|
"description": "Gemma 7B는 중소 규모 작업 처리에 적합하며, 비용 효과성을 갖추고 있습니다."
|
481
502
|
},
|
@@ -503,6 +524,9 @@
|
|
503
524
|
"glm-4-0520": {
|
504
525
|
"description": "GLM-4-0520은 최신 모델 버전으로, 매우 복잡하고 다양한 작업을 위해 설계되어 뛰어난 성능을 발휘합니다."
|
505
526
|
},
|
527
|
+
"glm-4-9b-chat": {
|
528
|
+
"description": "GLM-4-9B-Chat은 의미, 수학, 추론, 코드 및 지식 등 여러 분야에서 높은 성능을 보입니다. 웹 브라우징, 코드 실행, 사용자 정의 도구 호출 및 긴 텍스트 추론 기능도 갖추고 있습니다. 일본어, 한국어, 독일어를 포함한 26개 언어를 지원합니다."
|
529
|
+
},
|
506
530
|
"glm-4-air": {
|
507
531
|
"description": "GLM-4-Air는 가성비가 높은 버전으로, GLM-4에 가까운 성능을 제공하며 빠른 속도와 저렴한 가격을 자랑합니다."
|
508
532
|
},
|
@@ -22,6 +22,9 @@
|
|
22
22
|
"fireworksai": {
|
23
23
|
"description": "Fireworks AI는 기능 호출 및 다중 모드 처리를 전문으로 하는 선도적인 고급 언어 모델 서비스 제공업체입니다. 최신 모델인 Firefunction V2는 Llama-3를 기반으로 하며, 함수 호출, 대화 및 지시 따르기에 최적화되어 있습니다. 비주얼 언어 모델인 FireLLaVA-13B는 이미지와 텍스트 혼합 입력을 지원합니다. 기타 주목할 만한 모델로는 Llama 시리즈와 Mixtral 시리즈가 있으며, 효율적인 다국어 지시 따르기 및 생성 지원을 제공합니다."
|
24
24
|
},
|
25
|
+
"giteeai": {
|
26
|
+
"description": "Gitee AI의 Serverless API는 AI 개발자에게 즉시 사용할 수 있는 대형 모델 추론 API 서비스를 제공한다."
|
27
|
+
},
|
25
28
|
"github": {
|
26
29
|
"description": "GitHub 모델을 통해 개발자는 AI 엔지니어가 되어 업계 최고의 AI 모델로 구축할 수 있습니다."
|
27
30
|
},
|
@@ -389,6 +389,11 @@
|
|
389
389
|
"modelDesc": "사용자의 질문을 최적화하는 데 사용되는 모델 지정",
|
390
390
|
"title": "지식 베이스 질문 재작성"
|
391
391
|
},
|
392
|
+
"thread": {
|
393
|
+
"label": "하위 주제 명명 모델",
|
394
|
+
"modelDesc": "하위 주제 자동 이름 변경에 사용되는 모델 지정",
|
395
|
+
"title": "하위 주제 자동 명명"
|
396
|
+
},
|
392
397
|
"title": "시스템 도우미",
|
393
398
|
"topic": {
|
394
399
|
"label": "주제 명명 모델",
|
package/locales/nl-NL/chat.json
CHANGED
@@ -8,6 +8,7 @@
|
|
8
8
|
"agents": "Assistent",
|
9
9
|
"artifact": {
|
10
10
|
"generating": "Genereren",
|
11
|
+
"inThread": "Je kunt het niet bekijken in het subonderwerp, schakel over naar het hoofdgesprek om het te openen.",
|
11
12
|
"thinking": "Denken",
|
12
13
|
"thought": "Denken proces",
|
13
14
|
"unknownTitle": "Onbenoemd werk"
|
@@ -65,6 +66,7 @@
|
|
65
66
|
},
|
66
67
|
"messageAction": {
|
67
68
|
"delAndRegenerate": "Verwijderen en opnieuw genereren",
|
69
|
+
"deleteDisabledByThreads": "Er zijn subonderwerpen, verwijderen is niet mogelijk.",
|
68
70
|
"regenerate": "Opnieuw genereren"
|
69
71
|
},
|
70
72
|
"newAgent": "Nieuwe assistent",
|
@@ -121,6 +123,11 @@
|
|
121
123
|
"loading": "Bezig met herkennen...",
|
122
124
|
"prettifying": "Aan het verfraaien..."
|
123
125
|
},
|
126
|
+
"thread": {
|
127
|
+
"divider": "Subonderwerp",
|
128
|
+
"threadMessageCount": "{{messageCount}} berichten",
|
129
|
+
"title": "Subonderwerp"
|
130
|
+
},
|
124
131
|
"tokenDetails": {
|
125
132
|
"chats": "Chats",
|
126
133
|
"historySummary": "Geschiedenis samenvatting",
|
@@ -16,6 +16,8 @@
|
|
16
16
|
"back": "Terug",
|
17
17
|
"batchDelete": "Batch verwijderen",
|
18
18
|
"blog": "Product Blog",
|
19
|
+
"branching": "Subonderwerp aanmaken",
|
20
|
+
"branchingDisable": "De functie 'Subonderwerp' is alleen beschikbaar in de serverversie. Als je deze functie wilt gebruiken, schakel dan over naar de serverimplementatiemodus of gebruik LobeChat Cloud.",
|
19
21
|
"cancel": "Annuleren",
|
20
22
|
"changelog": "Wijzigingslogboek",
|
21
23
|
"close": "Sluiten",
|
@@ -176,6 +176,15 @@
|
|
176
176
|
"Qwen/Qwen2.5-Math-72B-Instruct": {
|
177
177
|
"description": "Qwen2.5-Math richt zich op het oplossen van wiskundige vraagstukken en biedt professionele antwoorden op moeilijke vragen."
|
178
178
|
},
|
179
|
+
"Qwen2-72B-Instruct": {
|
180
|
+
"description": "Qwen2 is de nieuwste serie van het Qwen-model, dat 128k context ondersteunt. In vergelijking met de huidige beste open-source modellen, overtreft Qwen2-72B op het gebied van natuurlijke taalbegrip, kennis, code, wiskunde en meertaligheid aanzienlijk de huidige toonaangevende modellen."
|
181
|
+
},
|
182
|
+
"Qwen2-7B-Instruct": {
|
183
|
+
"description": "Qwen2 is de nieuwste serie van het Qwen-model, dat in staat is om de beste open-source modellen van gelijke grootte of zelfs grotere modellen te overtreffen. Qwen2 7B heeft aanzienlijke voordelen behaald in verschillende evaluaties, vooral op het gebied van code en begrip van het Chinees."
|
184
|
+
},
|
185
|
+
"Qwen2.5-72B-Instruct": {
|
186
|
+
"description": "Qwen2.5-72B-Instruct ondersteunt 16k context en genereert lange teksten van meer dan 8K. Het ondersteunt functieaanroepen en naadloze interactie met externe systemen, wat de flexibiliteit en schaalbaarheid aanzienlijk vergroot. De kennis van het model is duidelijk toegenomen en de coderings- en wiskundige vaardigheden zijn sterk verbeterd, met ondersteuning voor meer dan 29 talen."
|
187
|
+
},
|
179
188
|
"SenseChat": {
|
180
189
|
"description": "Basisversie van het model (V4), met een contextlengte van 4K, heeft sterke algemene capaciteiten."
|
181
190
|
},
|
@@ -206,6 +215,9 @@
|
|
206
215
|
"Tencent/Hunyuan-A52B-Instruct": {
|
207
216
|
"description": "Hunyuan-Large is het grootste open source Transformer-architectuur MoE-model in de industrie, met een totaal van 389 miljard parameters en 52 miljard geactiveerde parameters."
|
208
217
|
},
|
218
|
+
"Yi-34B-Chat": {
|
219
|
+
"description": "Yi-1.5-34B heeft de uitstekende algemene taalvaardigheden van de oorspronkelijke modelserie behouden en heeft door incrementele training van 500 miljard hoogwaardige tokens de wiskundige logica en codevaardigheden aanzienlijk verbeterd."
|
220
|
+
},
|
209
221
|
"abab5.5-chat": {
|
210
222
|
"description": "Gericht op productiviteitsscenario's, ondersteunt complexe taakverwerking en efficiënte tekstgeneratie, geschikt voor professionele toepassingen."
|
211
223
|
},
|
@@ -368,6 +380,9 @@
|
|
368
380
|
"codegeex-4": {
|
369
381
|
"description": "CodeGeeX-4 is een krachtige AI-programmeerassistent die slimme vraag- en antwoordmogelijkheden en code-aanvulling ondersteunt voor verschillende programmeertalen, waardoor de ontwikkelingssnelheid wordt verhoogd."
|
370
382
|
},
|
383
|
+
"codegeex4-all-9b": {
|
384
|
+
"description": "CodeGeeX4-ALL-9B is een meertalig codegeneratiemodel dat uitgebreide functionaliteit biedt, waaronder code-aanvulling en -generatie, code-interpreter, webzoekfunctie, functieaanroepen en repository-niveau codevragen, en dekt verschillende scenario's in softwareontwikkeling. Het is een top codegeneratiemodel met minder dan 10B parameters."
|
385
|
+
},
|
371
386
|
"codegemma": {
|
372
387
|
"description": "CodeGemma is een lichtgewicht taalmodel dat speciaal is ontworpen voor verschillende programmeertaken, ondersteunt snelle iteratie en integratie."
|
373
388
|
},
|
@@ -422,6 +437,9 @@
|
|
422
437
|
"deepseek-chat": {
|
423
438
|
"description": "Een nieuw open-source model dat algemene en code-capaciteiten combineert, behoudt niet alleen de algemene conversatiecapaciteiten van het oorspronkelijke Chat-model en de krachtige codeverwerkingscapaciteiten van het Coder-model, maar is ook beter afgestemd op menselijke voorkeuren. Bovendien heeft DeepSeek-V2.5 aanzienlijke verbeteringen gerealiseerd in schrijfopdrachten, instructievolging en andere gebieden."
|
424
439
|
},
|
440
|
+
"deepseek-coder-33B-instruct": {
|
441
|
+
"description": "DeepSeek Coder 33B is een code-taalmodel, getraind op 20 biljoen gegevens, waarvan 87% code en 13% in het Chinees en Engels. Het model introduceert een venstergrootte van 16K en invultaken, en biedt projectniveau code-aanvulling en fragmentinvulfunctionaliteit."
|
442
|
+
},
|
425
443
|
"deepseek-coder-v2": {
|
426
444
|
"description": "DeepSeek Coder V2 is een open-source hybride expertcode-model, presteert uitstekend in code-taken en is vergelijkbaar met GPT4-Turbo."
|
427
445
|
},
|
@@ -476,6 +494,9 @@
|
|
476
494
|
"gemini-exp-1114": {
|
477
495
|
"description": "Gemini Exp 1114 is Google's nieuwste experimentele multimodale AI-model, met een snelle verwerkingscapaciteit, ondersteunt tekst-, beeld- en video-invoer, geschikt voor efficiënte schaling van diverse taken."
|
478
496
|
},
|
497
|
+
"gemini-exp-1121": {
|
498
|
+
"description": "Gemini Exp 1121 is Google's nieuwste experimentele multimodale AI-model, dat snel kan verwerken en ondersteuning biedt voor tekst-, beeld- en video-invoer, geschikt voor efficiënte opschaling van verschillende taken."
|
499
|
+
},
|
479
500
|
"gemma-7b-it": {
|
480
501
|
"description": "Gemma 7B is geschikt voor het verwerken van middelgrote taken, met een goede kosteneffectiviteit."
|
481
502
|
},
|
@@ -503,6 +524,9 @@
|
|
503
524
|
"glm-4-0520": {
|
504
525
|
"description": "GLM-4-0520 is de nieuwste modelversie, speciaal ontworpen voor zeer complexe en diverse taken, met uitstekende prestaties."
|
505
526
|
},
|
527
|
+
"glm-4-9b-chat": {
|
528
|
+
"description": "GLM-4-9B-Chat presteert goed op het gebied van semantiek, wiskunde, redeneren, code en kennis. Het beschikt ook over webbrowserfunctionaliteit, code-uitvoering, aangepaste toolaanroepen en lange tekstredenering. Ondersteunt 26 talen, waaronder Japans, Koreaans en Duits."
|
529
|
+
},
|
506
530
|
"glm-4-air": {
|
507
531
|
"description": "GLM-4-Air is een kosteneffectieve versie met prestaties die dicht bij GLM-4 liggen, met snelle snelheid en een betaalbare prijs."
|
508
532
|
},
|
@@ -22,6 +22,9 @@
|
|
22
22
|
"fireworksai": {
|
23
23
|
"description": "Fireworks AI is een toonaangevende aanbieder van geavanceerde taalmodellen, met een focus op functionele aanroepen en multimodale verwerking. Hun nieuwste model Firefunction V2 is gebaseerd op Llama-3 en geoptimaliseerd voor functieaanroepen, dialogen en het volgen van instructies. Het visuele taalmodel FireLLaVA-13B ondersteunt gemengde invoer van afbeeldingen en tekst. Andere opmerkelijke modellen zijn de Llama-serie en de Mixtral-serie, die efficiënte ondersteuning bieden voor meertalig volgen van instructies en genereren."
|
24
24
|
},
|
25
|
+
"giteeai": {
|
26
|
+
"description": "Gitee AI's Serverless API biedt AI ontwikkelaars een out of the box grote model inference API service."
|
27
|
+
},
|
25
28
|
"github": {
|
26
29
|
"description": "Met GitHub-modellen kunnen ontwikkelaars AI-ingenieurs worden en bouwen met de toonaangevende AI-modellen in de industrie."
|
27
30
|
},
|
@@ -389,6 +389,11 @@
|
|
389
389
|
"modelDesc": "Model dat is opgegeven voor het optimaliseren van gebruikersvragen",
|
390
390
|
"title": "Herformuleren van vragen in de kennisbank"
|
391
391
|
},
|
392
|
+
"thread": {
|
393
|
+
"label": "Subtopic Naamgevingsmodel",
|
394
|
+
"modelDesc": "Model dat wordt gebruikt voor het automatisch hernoemen van subonderwerpen",
|
395
|
+
"title": "Automatische naamgeving van subonderwerpen"
|
396
|
+
},
|
392
397
|
"title": "Systeemassistent",
|
393
398
|
"topic": {
|
394
399
|
"label": "Onderwerp Naamgevingsmodel",
|
package/locales/pl-PL/chat.json
CHANGED
@@ -8,6 +8,7 @@
|
|
8
8
|
"agents": "Asystent",
|
9
9
|
"artifact": {
|
10
10
|
"generating": "Generowanie",
|
11
|
+
"inThread": "Nie można przeglądać w wątku, przełącz się na główny obszar rozmowy, aby otworzyć",
|
11
12
|
"thinking": "Myślenie",
|
12
13
|
"thought": "Proces myślenia",
|
13
14
|
"unknownTitle": "Nienazwane dzieło"
|
@@ -65,6 +66,7 @@
|
|
65
66
|
},
|
66
67
|
"messageAction": {
|
67
68
|
"delAndRegenerate": "Usuń i wygeneruj ponownie",
|
69
|
+
"deleteDisabledByThreads": "Istnieją podwątki, nie można usunąć",
|
68
70
|
"regenerate": "Wygeneruj ponownie"
|
69
71
|
},
|
70
72
|
"newAgent": "Nowy asystent",
|
@@ -121,6 +123,11 @@
|
|
121
123
|
"loading": "Rozpoznawanie...",
|
122
124
|
"prettifying": "Upiększanie..."
|
123
125
|
},
|
126
|
+
"thread": {
|
127
|
+
"divider": "Podwątek",
|
128
|
+
"threadMessageCount": "{{messageCount}} wiadomości",
|
129
|
+
"title": "Podwątek"
|
130
|
+
},
|
124
131
|
"tokenDetails": {
|
125
132
|
"chats": "Rozmowy",
|
126
133
|
"historySummary": "Podsumowanie historii",
|
@@ -16,6 +16,8 @@
|
|
16
16
|
"back": "Powrót",
|
17
17
|
"batchDelete": "Usuń wiele",
|
18
18
|
"blog": "Blog produktowy",
|
19
|
+
"branching": "Utwórz podtemat",
|
20
|
+
"branchingDisable": "Funkcja „podtemat” jest dostępna tylko w wersji serwerowej. Aby skorzystać z tej funkcji, przełącz się na tryb wdrożenia serwera lub użyj LobeChat Cloud.",
|
19
21
|
"cancel": "Anuluj",
|
20
22
|
"changelog": "Dziennik zmian",
|
21
23
|
"close": "Zamknij",
|
@@ -176,6 +176,15 @@
|
|
176
176
|
"Qwen/Qwen2.5-Math-72B-Instruct": {
|
177
177
|
"description": "Qwen2.5-Math koncentruje się na rozwiązywaniu problemów w dziedzinie matematyki, oferując profesjonalne odpowiedzi na trudne pytania."
|
178
178
|
},
|
179
|
+
"Qwen2-72B-Instruct": {
|
180
|
+
"description": "Qwen2 to najnowsza seria modeli Qwen, obsługująca kontekst 128k. W porównaniu do obecnie najlepszych modeli open source, Qwen2-72B znacznie przewyższa w zakresie rozumienia języka naturalnego, wiedzy, kodowania, matematyki i wielu języków."
|
181
|
+
},
|
182
|
+
"Qwen2-7B-Instruct": {
|
183
|
+
"description": "Qwen2 to najnowsza seria modeli Qwen, która przewyższa najlepsze modele open source o podobnej skali, a nawet większe. Qwen2 7B osiągnęła znaczną przewagę w wielu testach, szczególnie w zakresie kodowania i rozumienia języka chińskiego."
|
184
|
+
},
|
185
|
+
"Qwen2.5-72B-Instruct": {
|
186
|
+
"description": "Qwen2.5-72B-Instruct obsługuje kontekst 16k, generując długie teksty przekraczające 8K. Wspiera wywołania funkcji i bezproblemową interakcję z systemami zewnętrznymi, znacznie zwiększając elastyczność i skalowalność. Wiedza modelu znacznie wzrosła, a jego zdolności w zakresie kodowania i matematyki uległy znacznemu poprawieniu, z obsługą ponad 29 języków."
|
187
|
+
},
|
179
188
|
"SenseChat": {
|
180
189
|
"description": "Podstawowa wersja modelu (V4), długość kontekstu 4K, silne zdolności ogólne."
|
181
190
|
},
|
@@ -206,6 +215,9 @@
|
|
206
215
|
"Tencent/Hunyuan-A52B-Instruct": {
|
207
216
|
"description": "Hunyuan-Large to największy w branży otwarty model Transformer MoE, z 389 miliardami parametrów ogółem i 52 miliardami aktywowanych parametrów."
|
208
217
|
},
|
218
|
+
"Yi-34B-Chat": {
|
219
|
+
"description": "Yi-1.5-34B, zachowując doskonałe ogólne zdolności językowe oryginalnej serii modeli, znacznie poprawił zdolności logiczne i kodowania dzięki dodatkowym treningom na 500 miliardach wysokiej jakości tokenów."
|
220
|
+
},
|
209
221
|
"abab5.5-chat": {
|
210
222
|
"description": "Skierowany do scenariuszy produkcyjnych, wspierający przetwarzanie złożonych zadań i efektywne generowanie tekstu, odpowiedni do zastosowań w profesjonalnych dziedzinach."
|
211
223
|
},
|
@@ -368,6 +380,9 @@
|
|
368
380
|
"codegeex-4": {
|
369
381
|
"description": "CodeGeeX-4 to potężny asystent programowania AI, obsługujący inteligentne pytania i odpowiedzi oraz uzupełnianie kodu w różnych językach programowania, zwiększając wydajność programistów."
|
370
382
|
},
|
383
|
+
"codegeex4-all-9b": {
|
384
|
+
"description": "CodeGeeX4-ALL-9B to model generowania kodu w wielu językach, który obsługuje kompleksowe funkcje, w tym uzupełnianie i generowanie kodu, interpreter kodu, wyszukiwanie w sieci, wywołania funkcji oraz pytania i odpowiedzi na poziomie repozytoriów, obejmując różne scenariusze rozwoju oprogramowania. Jest to wiodący model generowania kodu z mniej niż 10B parametrów."
|
385
|
+
},
|
371
386
|
"codegemma": {
|
372
387
|
"description": "CodeGemma to lekki model językowy, specjalizujący się w różnych zadaniach programistycznych, wspierający szybkie iteracje i integrację."
|
373
388
|
},
|
@@ -422,6 +437,9 @@
|
|
422
437
|
"deepseek-chat": {
|
423
438
|
"description": "Nowy otwarty model łączący zdolności ogólne i kodowe, który nie tylko zachowuje ogólne zdolności dialogowe oryginalnego modelu czatu i potężne zdolności przetwarzania kodu modelu Coder, ale także lepiej dostosowuje się do ludzkich preferencji. Ponadto, DeepSeek-V2.5 osiągnął znaczne poprawy w zadaniach pisarskich, przestrzeganiu instrukcji i innych obszarach."
|
424
439
|
},
|
440
|
+
"deepseek-coder-33B-instruct": {
|
441
|
+
"description": "DeepSeek Coder 33B to model języka kodu, wytrenowany na 20 bilionach danych, z czego 87% to kod, a 13% to języki chiński i angielski. Model wprowadza okno o rozmiarze 16K oraz zadania uzupełniania, oferując funkcje uzupełniania kodu na poziomie projektu i wypełniania fragmentów."
|
442
|
+
},
|
425
443
|
"deepseek-coder-v2": {
|
426
444
|
"description": "DeepSeek Coder V2 to otwarty model kodowy Mixture-of-Experts, który doskonale radzi sobie z zadaniami kodowymi, porównywalny z GPT4-Turbo."
|
427
445
|
},
|
@@ -476,6 +494,9 @@
|
|
476
494
|
"gemini-exp-1114": {
|
477
495
|
"description": "Gemini Exp 1114 to najnowszy eksperymentalny model AI o wielu modalnościach od Google, który charakteryzuje się szybkim przetwarzaniem i obsługuje wejścia tekstowe, obrazowe oraz wideo, co umożliwia wydajne rozszerzanie go na różne zadania."
|
478
496
|
},
|
497
|
+
"gemini-exp-1121": {
|
498
|
+
"description": "Gemini Exp 1121 to najnowszy eksperymentalny model AI Google, który charakteryzuje się szybkim przetwarzaniem i obsługuje wejścia tekstowe, obrazowe i wideo, co czyni go wydajnym narzędziem do rozwiązywania różnych zadań."
|
499
|
+
},
|
479
500
|
"gemma-7b-it": {
|
480
501
|
"description": "Gemma 7B nadaje się do przetwarzania zadań średniej i małej skali, łącząc efektywność kosztową."
|
481
502
|
},
|
@@ -503,6 +524,9 @@
|
|
503
524
|
"glm-4-0520": {
|
504
525
|
"description": "GLM-4-0520 to najnowsza wersja modelu, zaprojektowana do wysoko złożonych i zróżnicowanych zadań, z doskonałymi wynikami."
|
505
526
|
},
|
527
|
+
"glm-4-9b-chat": {
|
528
|
+
"description": "GLM-4-9B-Chat wykazuje wysoką wydajność w wielu aspektach, takich jak semantyka, matematyka, wnioskowanie, kodowanie i wiedza. Posiada również funkcje przeglądania stron internetowych, wykonywania kodu, wywoływania niestandardowych narzędzi oraz wnioskowania z długich tekstów. Obsługuje 26 języków, w tym japoński, koreański i niemiecki."
|
529
|
+
},
|
506
530
|
"glm-4-air": {
|
507
531
|
"description": "GLM-4-Air to opłacalna wersja, której wydajność jest zbliżona do GLM-4, oferująca szybkie działanie i przystępną cenę."
|
508
532
|
},
|
@@ -22,6 +22,9 @@
|
|
22
22
|
"fireworksai": {
|
23
23
|
"description": "Fireworks AI to wiodący dostawca zaawansowanych modeli językowych, skoncentrowany na wywołaniach funkcji i przetwarzaniu multimodalnym. Jego najnowszy model Firefunction V2 oparty na Llama-3, zoptymalizowany do wywołań funkcji, dialogów i przestrzegania instrukcji. Model wizualny FireLLaVA-13B wspiera mieszane wejścia obrazów i tekstu. Inne znaczące modele to seria Llama i seria Mixtral, oferujące efektywne wsparcie dla wielojęzycznego przestrzegania instrukcji i generacji."
|
24
24
|
},
|
25
|
+
"giteeai": {
|
26
|
+
"description": "Serverless API Gitee AI zapewnia deweloperom sztucznej inteligencji gotową usługę interfejsu interfejsu interfejsu dużych modeli."
|
27
|
+
},
|
25
28
|
"github": {
|
26
29
|
"description": "Dzięki modelom GitHub, deweloperzy mogą stać się inżynierami AI i budować z wykorzystaniem wiodących modeli AI w branży."
|
27
30
|
},
|