@lobehub/chat 1.33.5 → 1.34.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (203) hide show
  1. package/CHANGELOG.md +25 -0
  2. package/changelog/v1.json +9 -0
  3. package/locales/ar/chat.json +7 -0
  4. package/locales/ar/common.json +2 -0
  5. package/locales/ar/models.json +24 -0
  6. package/locales/ar/setting.json +5 -0
  7. package/locales/ar/thread.json +5 -0
  8. package/locales/bg-BG/chat.json +7 -0
  9. package/locales/bg-BG/common.json +2 -0
  10. package/locales/bg-BG/models.json +24 -0
  11. package/locales/bg-BG/setting.json +5 -0
  12. package/locales/bg-BG/thread.json +5 -0
  13. package/locales/de-DE/chat.json +7 -0
  14. package/locales/de-DE/common.json +2 -0
  15. package/locales/de-DE/models.json +24 -0
  16. package/locales/de-DE/setting.json +5 -0
  17. package/locales/de-DE/thread.json +5 -0
  18. package/locales/en-US/chat.json +7 -0
  19. package/locales/en-US/common.json +2 -0
  20. package/locales/en-US/models.json +24 -0
  21. package/locales/en-US/setting.json +5 -0
  22. package/locales/en-US/thread.json +5 -0
  23. package/locales/es-ES/chat.json +7 -0
  24. package/locales/es-ES/common.json +2 -0
  25. package/locales/es-ES/models.json +24 -0
  26. package/locales/es-ES/setting.json +5 -0
  27. package/locales/es-ES/thread.json +5 -0
  28. package/locales/fa-IR/chat.json +7 -0
  29. package/locales/fa-IR/common.json +2 -0
  30. package/locales/fa-IR/models.json +24 -0
  31. package/locales/fa-IR/setting.json +5 -0
  32. package/locales/fa-IR/thread.json +5 -0
  33. package/locales/fr-FR/chat.json +7 -0
  34. package/locales/fr-FR/common.json +2 -0
  35. package/locales/fr-FR/models.json +24 -0
  36. package/locales/fr-FR/setting.json +5 -0
  37. package/locales/fr-FR/thread.json +5 -0
  38. package/locales/it-IT/chat.json +7 -0
  39. package/locales/it-IT/common.json +2 -0
  40. package/locales/it-IT/models.json +24 -0
  41. package/locales/it-IT/setting.json +5 -0
  42. package/locales/it-IT/thread.json +5 -0
  43. package/locales/ja-JP/chat.json +7 -0
  44. package/locales/ja-JP/common.json +2 -0
  45. package/locales/ja-JP/models.json +24 -0
  46. package/locales/ja-JP/setting.json +5 -0
  47. package/locales/ja-JP/thread.json +5 -0
  48. package/locales/ko-KR/chat.json +7 -0
  49. package/locales/ko-KR/common.json +2 -0
  50. package/locales/ko-KR/models.json +24 -0
  51. package/locales/ko-KR/setting.json +5 -0
  52. package/locales/ko-KR/thread.json +5 -0
  53. package/locales/nl-NL/chat.json +7 -0
  54. package/locales/nl-NL/common.json +2 -0
  55. package/locales/nl-NL/models.json +24 -0
  56. package/locales/nl-NL/setting.json +5 -0
  57. package/locales/nl-NL/thread.json +5 -0
  58. package/locales/pl-PL/chat.json +7 -0
  59. package/locales/pl-PL/common.json +2 -0
  60. package/locales/pl-PL/models.json +24 -0
  61. package/locales/pl-PL/setting.json +5 -0
  62. package/locales/pl-PL/thread.json +5 -0
  63. package/locales/pt-BR/chat.json +7 -0
  64. package/locales/pt-BR/common.json +2 -0
  65. package/locales/pt-BR/models.json +24 -0
  66. package/locales/pt-BR/setting.json +5 -0
  67. package/locales/pt-BR/thread.json +5 -0
  68. package/locales/ru-RU/chat.json +7 -0
  69. package/locales/ru-RU/common.json +2 -0
  70. package/locales/ru-RU/models.json +24 -0
  71. package/locales/ru-RU/setting.json +5 -0
  72. package/locales/ru-RU/thread.json +5 -0
  73. package/locales/tr-TR/chat.json +7 -0
  74. package/locales/tr-TR/common.json +2 -0
  75. package/locales/tr-TR/models.json +24 -0
  76. package/locales/tr-TR/setting.json +5 -0
  77. package/locales/tr-TR/thread.json +5 -0
  78. package/locales/vi-VN/chat.json +7 -0
  79. package/locales/vi-VN/common.json +2 -0
  80. package/locales/vi-VN/models.json +24 -0
  81. package/locales/vi-VN/setting.json +5 -0
  82. package/locales/vi-VN/thread.json +5 -0
  83. package/locales/zh-CN/chat.json +7 -0
  84. package/locales/zh-CN/common.json +2 -0
  85. package/locales/zh-CN/models.json +24 -0
  86. package/locales/zh-CN/setting.json +5 -0
  87. package/locales/zh-CN/thread.json +5 -0
  88. package/locales/zh-TW/chat.json +7 -0
  89. package/locales/zh-TW/common.json +2 -0
  90. package/locales/zh-TW/models.json +24 -0
  91. package/locales/zh-TW/setting.json +5 -0
  92. package/locales/zh-TW/thread.json +5 -0
  93. package/package.json +1 -1
  94. package/src/app/(main)/chat/(workspace)/@conversation/default.tsx +2 -0
  95. package/src/app/(main)/chat/(workspace)/@conversation/features/ChatHydration/index.tsx +11 -2
  96. package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/index.tsx +7 -9
  97. package/src/app/(main)/chat/(workspace)/@conversation/features/ChatInput/Desktop/index.tsx +7 -2
  98. package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/Thread.tsx +62 -0
  99. package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/ThreadItem.tsx +68 -0
  100. package/src/app/(main)/chat/(workspace)/@conversation/features/ChatList/ChatItem/index.tsx +62 -2
  101. package/src/app/(main)/chat/(workspace)/@conversation/features/ThreadHydration.tsx +47 -0
  102. package/src/app/(main)/chat/(workspace)/@portal/_layout/Desktop.tsx +3 -2
  103. package/src/app/(main)/chat/(workspace)/@portal/_layout/Mobile.tsx +47 -6
  104. package/src/app/(main)/chat/(workspace)/@topic/features/SkeletonList.tsx +3 -2
  105. package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ByTimeMode/index.tsx +10 -3
  106. package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/FlatMode/index.tsx +1 -1
  107. package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ThreadItem/Content.tsx +164 -0
  108. package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/ThreadItem/index.tsx +98 -0
  109. package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{TopicItem.tsx → TopicItem/index.tsx} +33 -22
  110. package/src/app/(main)/chat/(workspace)/_layout/Desktop/Portal.tsx +12 -5
  111. package/src/app/(main)/chat/(workspace)/_layout/Mobile/index.tsx +1 -2
  112. package/src/const/message.ts +2 -0
  113. package/src/const/settings/systemAgent.ts +1 -0
  114. package/src/database/server/migrations/0012_add_thread.sql +39 -0
  115. package/src/database/server/migrations/meta/0012_snapshot.json +3671 -0
  116. package/src/database/server/migrations/meta/_journal.json +7 -0
  117. package/src/database/server/models/_template.ts +2 -2
  118. package/src/database/server/models/message.ts +1 -0
  119. package/src/database/server/models/thread.ts +79 -0
  120. package/src/database/server/schemas/lobechat/message.ts +2 -1
  121. package/src/database/server/schemas/lobechat/relations.ts +13 -1
  122. package/src/database/server/schemas/lobechat/topic.ts +30 -1
  123. package/src/database/server/utils/idGenerator.ts +1 -0
  124. package/src/features/ChatInput/ActionBar/Token/TokenTag.tsx +6 -4
  125. package/src/features/ChatInput/ActionBar/Token/index.tsx +24 -5
  126. package/src/features/ChatInput/ActionBar/config.ts +3 -2
  127. package/src/features/ChatInput/Desktop/index.tsx +15 -7
  128. package/src/features/ChatInput/Mobile/index.tsx +4 -4
  129. package/src/features/Conversation/Actions/Assistant.tsx +24 -5
  130. package/src/features/Conversation/Actions/User.tsx +21 -4
  131. package/src/features/Conversation/Actions/index.ts +1 -66
  132. package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/index.tsx +3 -1
  133. package/src/features/Conversation/Messages/{Tool/index.tsx → Assistant/ToolCallItem/Tool.tsx} +10 -11
  134. package/src/features/Conversation/Messages/Assistant/ToolCallItem/index.tsx +5 -3
  135. package/src/features/Conversation/Messages/Assistant/index.tsx +22 -14
  136. package/src/features/Conversation/Messages/index.ts +0 -2
  137. package/src/features/Conversation/components/AutoScroll.tsx +1 -1
  138. package/src/features/Conversation/components/ChatItem/ActionsBar.tsx +79 -5
  139. package/src/features/Conversation/components/ChatItem/InPortalThreadContext.ts +3 -0
  140. package/src/features/Conversation/components/ChatItem/index.tsx +16 -5
  141. package/src/features/Conversation/components/MarkdownElements/LobeArtifact/Render/index.tsx +9 -1
  142. package/src/features/Conversation/components/ThreadDivider/index.tsx +19 -0
  143. package/src/features/Conversation/hooks/useChatListActionsBar.tsx +19 -4
  144. package/src/features/Portal/Thread/Chat/ChatInput/Footer.tsx +90 -0
  145. package/src/features/Portal/Thread/Chat/ChatInput/TextArea.tsx +30 -0
  146. package/src/features/Portal/Thread/Chat/ChatInput/index.tsx +66 -0
  147. package/src/features/Portal/Thread/Chat/ChatInput/useSend.ts +50 -0
  148. package/src/features/Portal/Thread/Chat/ChatItem.tsx +62 -0
  149. package/src/features/Portal/Thread/Chat/ChatList.tsx +49 -0
  150. package/src/features/Portal/Thread/Chat/ThreadDivider/index.tsx +19 -0
  151. package/src/features/Portal/Thread/Chat/index.tsx +28 -0
  152. package/src/features/Portal/Thread/Header/Active.tsx +35 -0
  153. package/src/features/Portal/Thread/Header/New.tsx +37 -0
  154. package/src/features/Portal/Thread/Header/Title.tsx +18 -0
  155. package/src/features/Portal/Thread/Header/index.tsx +20 -0
  156. package/src/features/Portal/Thread/hook.ts +8 -0
  157. package/src/features/Portal/Thread/index.ts +12 -0
  158. package/src/features/Portal/router.tsx +2 -1
  159. package/src/hooks/useFetchTopics.ts +7 -1
  160. package/src/locales/default/chat.ts +8 -1
  161. package/src/locales/default/common.ts +3 -0
  162. package/src/locales/default/index.ts +2 -0
  163. package/src/locales/default/setting.ts +5 -0
  164. package/src/locales/default/thread.ts +5 -0
  165. package/src/server/routers/lambda/index.ts +2 -0
  166. package/src/server/routers/lambda/thread.ts +83 -0
  167. package/src/services/thread.ts +54 -0
  168. package/src/store/chat/initialState.ts +3 -0
  169. package/src/store/chat/selectors.ts +2 -1
  170. package/src/store/chat/slices/aiChat/actions/__tests__/generateAIChat.test.ts +1 -1
  171. package/src/store/chat/slices/aiChat/actions/__tests__/rag.test.ts +1 -1
  172. package/src/store/chat/slices/aiChat/actions/generateAIChat.ts +31 -8
  173. package/src/store/chat/slices/aiChat/actions/rag.ts +1 -1
  174. package/src/store/chat/slices/message/selectors.test.ts +3 -3
  175. package/src/store/chat/slices/message/selectors.ts +50 -29
  176. package/src/store/chat/slices/plugin/action.ts +26 -8
  177. package/src/store/chat/slices/portal/action.ts +1 -0
  178. package/src/store/chat/slices/portal/initialState.ts +1 -0
  179. package/src/store/chat/slices/portal/selectors/thread.ts +17 -0
  180. package/src/store/chat/slices/portal/selectors.ts +2 -0
  181. package/src/store/chat/slices/thread/action.ts +326 -0
  182. package/src/store/chat/slices/thread/initialState.ts +34 -0
  183. package/src/store/chat/slices/thread/reducer.ts +48 -0
  184. package/src/store/chat/slices/thread/selectors/index.ts +202 -0
  185. package/src/store/chat/slices/thread/selectors/util.ts +22 -0
  186. package/src/store/chat/slices/topic/action.ts +5 -1
  187. package/src/store/chat/store.ts +5 -2
  188. package/src/store/global/initialState.ts +4 -0
  189. package/src/store/global/selectors.ts +4 -0
  190. package/src/store/user/slices/settings/selectors/systemAgent.ts +2 -0
  191. package/src/types/message/index.ts +17 -1
  192. package/src/types/topic/index.ts +1 -0
  193. package/src/types/topic/thread.ts +42 -0
  194. package/src/types/user/settings/systemAgent.ts +1 -0
  195. package/src/app/(main)/chat/(workspace)/@portal/features/Header.tsx +0 -11
  196. package/src/app/(main)/chat/(workspace)/_layout/Mobile/PortalModal.tsx +0 -35
  197. /package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/SendMore.tsx +0 -0
  198. /package/src/{features → app/(main)/chat/(workspace)/@conversation/features}/ChatInput/Desktop/Footer/ShortcutHint.tsx +0 -0
  199. /package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{DefaultContent.tsx → TopicItem/DefaultContent.tsx} +0 -0
  200. /package/src/app/(main)/chat/(workspace)/@topic/features/TopicListContent/{TopicContent.tsx → TopicItem/TopicContent.tsx} +0 -0
  201. /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/PluginResultJSON.tsx +0 -0
  202. /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/Settings.tsx +0 -0
  203. /package/src/features/Conversation/Messages/{Tool → Assistant/ToolCallItem}/Inspector/style.ts +0 -0
package/CHANGELOG.md CHANGED
@@ -2,6 +2,31 @@
2
2
 
3
3
  # Changelog
4
4
 
5
+ ## [Version 1.34.0](https://github.com/lobehub/lobe-chat/compare/v1.33.5...v1.34.0)
6
+
7
+ <sup>Released on **2024-11-26**</sup>
8
+
9
+ #### ✨ Features
10
+
11
+ - **misc**: Forkable Chat Mode.
12
+
13
+ <br/>
14
+
15
+ <details>
16
+ <summary><kbd>Improvements and Fixes</kbd></summary>
17
+
18
+ #### What's improved
19
+
20
+ - **misc**: Forkable Chat Mode, closes [#4632](https://github.com/lobehub/lobe-chat/issues/4632) ([832f0ce](https://github.com/lobehub/lobe-chat/commit/832f0ce))
21
+
22
+ </details>
23
+
24
+ <div align="right">
25
+
26
+ [![](https://img.shields.io/badge/-BACK_TO_TOP-151515?style=flat-square)](#readme-top)
27
+
28
+ </div>
29
+
5
30
  ### [Version 1.33.5](https://github.com/lobehub/lobe-chat/compare/v1.33.4...v1.33.5)
6
31
 
7
32
  <sup>Released on **2024-11-26**</sup>
package/changelog/v1.json CHANGED
@@ -1,4 +1,13 @@
1
1
  [
2
+ {
3
+ "children": {
4
+ "features": [
5
+ "Forkable Chat Mode."
6
+ ]
7
+ },
8
+ "date": "2024-11-26",
9
+ "version": "1.34.0"
10
+ },
2
11
  {
3
12
  "children": {
4
13
  "improvements": [
@@ -8,6 +8,7 @@
8
8
  "agents": "مساعد",
9
9
  "artifact": {
10
10
  "generating": "جاري الإنشاء",
11
+ "inThread": "لا يمكن عرض الموضوعات الفرعية، يرجى التبديل إلى منطقة المحادثة الرئيسية لفتحها",
11
12
  "thinking": "جاري التفكير",
12
13
  "thought": "عملية التفكير",
13
14
  "unknownTitle": "عمل غير مسمى"
@@ -65,6 +66,7 @@
65
66
  },
66
67
  "messageAction": {
67
68
  "delAndRegenerate": "حذف وإعادة الإنشاء",
69
+ "deleteDisabledByThreads": "يوجد موضوعات فرعية، لا يمكن الحذف",
68
70
  "regenerate": "إعادة الإنشاء"
69
71
  },
70
72
  "newAgent": "مساعد جديد",
@@ -121,6 +123,11 @@
121
123
  "loading": "جارٍ التعرف...",
122
124
  "prettifying": "جارٍ التجميل..."
123
125
  },
126
+ "thread": {
127
+ "divider": "موضوع فرعي",
128
+ "threadMessageCount": "{{messageCount}} رسالة",
129
+ "title": "موضوع فرعي"
130
+ },
124
131
  "tokenDetails": {
125
132
  "chats": "رسائل المحادثة",
126
133
  "historySummary": "ملخص التاريخ",
@@ -16,6 +16,8 @@
16
16
  "back": "عودة",
17
17
  "batchDelete": "حذف دفعة",
18
18
  "blog": "مدونة المنتجات",
19
+ "branching": "إنشاء موضوع فرعي",
20
+ "branchingDisable": "ميزة \"الموضوع الفرعي\" متاحة فقط في إصدار الخادم. إذا كنت بحاجة إلى هذه الميزة، يرجى التبديل إلى وضع نشر الخادم أو استخدام LobeChat Cloud",
19
21
  "cancel": "إلغاء",
20
22
  "changelog": "سجل التغييرات",
21
23
  "close": "إغلاق",
@@ -176,6 +176,15 @@
176
176
  "Qwen/Qwen2.5-Math-72B-Instruct": {
177
177
  "description": "Qwen2.5-Math يركز على حل المشكلات في مجال الرياضيات، ويقدم إجابات احترافية للأسئلة الصعبة."
178
178
  },
179
+ "Qwen2-72B-Instruct": {
180
+ "description": "Qwen2 هو أحدث سلسلة من نموذج Qwen، ويدعم سياقًا يصل إلى 128 ألف، مقارنةً بأفضل النماذج مفتوحة المصدر الحالية، يتفوق Qwen2-72B بشكل ملحوظ في فهم اللغة الطبيعية والمعرفة والترميز والرياضيات والقدرات متعددة اللغات."
181
+ },
182
+ "Qwen2-7B-Instruct": {
183
+ "description": "Qwen2 هو أحدث سلسلة من نموذج Qwen، قادر على التفوق على النماذج مفتوحة المصدر ذات الحجم المماثل أو حتى النماذج الأكبر حجمًا، حقق Qwen2 7B مزايا ملحوظة في عدة تقييمات، خاصة في فهم الترميز والصينية."
184
+ },
185
+ "Qwen2.5-72B-Instruct": {
186
+ "description": "يدعم Qwen2.5-72B-Instruct سياقًا يصل إلى 16 ألف، وينتج نصوصًا طويلة تتجاوز 8 آلاف. يدعم استدعاء الوظائف والتفاعل السلس مع الأنظمة الخارجية، مما يعزز بشكل كبير من المرونة وقابلية التوسع. لقد زادت معرفة النموذج بشكل ملحوظ، كما تحسنت قدراته في الترميز والرياضيات بشكل كبير، ويدعم أكثر من 29 لغة."
187
+ },
179
188
  "SenseChat": {
180
189
  "description": "نموذج الإصدار الأساسي (V4)، بطول سياق 4K، يمتلك قدرات قوية وعامة."
181
190
  },
@@ -206,6 +215,9 @@
206
215
  "Tencent/Hunyuan-A52B-Instruct": {
207
216
  "description": "Hunyuan-Large هو أكبر نموذج MoE مفتوح المصدر في الصناعة، مع 389 مليار إجمالي عدد المعلمات و52 مليار عدد المعلمات النشطة."
208
217
  },
218
+ "Yi-34B-Chat": {
219
+ "description": "Yi-1.5-34B، مع الحفاظ على القدرات اللغوية العامة الممتازة للنموذج الأصلي، تم تدريبه بشكل إضافي على 500 مليار توكن عالي الجودة، مما أدى إلى تحسين كبير في المنطق الرياضي وقدرات الترميز."
220
+ },
209
221
  "abab5.5-chat": {
210
222
  "description": "موجه لمشاهد الإنتاجية، يدعم معالجة المهام المعقدة وتوليد النصوص بكفاءة، مناسب للتطبيقات في المجالات المهنية."
211
223
  },
@@ -368,6 +380,9 @@
368
380
  "codegeex-4": {
369
381
  "description": "CodeGeeX-4 هو مساعد برمجي قوي، يدعم مجموعة متنوعة من لغات البرمجة في الإجابة الذكية وإكمال الشيفرة، مما يعزز من كفاءة التطوير."
370
382
  },
383
+ "codegeex4-all-9b": {
384
+ "description": "CodeGeeX4-ALL-9B هو نموذج توليد كود متعدد اللغات، يدعم مجموعة شاملة من الوظائف بما في ذلك إكمال الشيفرات والتوليد، ومفسر الشيفرات، والبحث عبر الإنترنت، واستدعاء الوظائف، وأسئلة وأجوبة على مستوى المستودع، مما يغطي جميع سيناريوهات تطوير البرمجيات. إنه أحد أفضل نماذج توليد الشيفرات بأقل من 10 مليار معلمة."
385
+ },
371
386
  "codegemma": {
372
387
  "description": "CodeGemma هو نموذج لغوي خفيف الوزن مخصص لمهام البرمجة المختلفة، يدعم التكرار السريع والتكامل."
373
388
  },
@@ -422,6 +437,9 @@
422
437
  "deepseek-chat": {
423
438
  "description": "نموذج مفتوح المصدر الجديد الذي يجمع بين القدرات العامة وقدرات البرمجة، لا يحتفظ فقط بالقدرات الحوارية العامة لنموذج الدردشة الأصلي وقدرات معالجة الشيفرة القوية لنموذج Coder، بل يتماشى أيضًا بشكل أفضل مع تفضيلات البشر. بالإضافة إلى ذلك، حقق DeepSeek-V2.5 تحسينات كبيرة في مهام الكتابة، واتباع التعليمات، وغيرها من المجالات."
424
439
  },
440
+ "deepseek-coder-33B-instruct": {
441
+ "description": "DeepSeek Coder 33B هو نموذج لغة برمجية، تم تدريبه على 20 تريليون بيانات، منها 87% كود و13% لغات صينية وإنجليزية. يقدم النموذج حجم نافذة 16K ومهام ملء الفراغ، مما يوفر إكمال الشيفرات على مستوى المشروع ووظائف ملء المقاطع."
442
+ },
425
443
  "deepseek-coder-v2": {
426
444
  "description": "DeepSeek Coder V2 هو نموذج شيفرة مفتوح المصدر من نوع خبير مختلط، يقدم أداءً ممتازًا في مهام الشيفرة، ويضاهي GPT4-Turbo."
427
445
  },
@@ -476,6 +494,9 @@
476
494
  "gemini-exp-1114": {
477
495
  "description": "جيمني إكسب 1114 هو أحدث نموذج ذكاء اصطناعي متعدد الوسائط تجريبي من Google، يتميز بقدرة معالجة سريعة، ويدعم إدخالات النصوص والصور والفيديو، مما يجعله مناسبًا للتوسع الفعال في مهام متعددة."
478
496
  },
497
+ "gemini-exp-1121": {
498
+ "description": "جمني إكسب 1121 هو أحدث نموذج تجريبي متعدد الوسائط من جوجل، يتمتع بقدرة معالجة سريعة، ويدعم إدخال النصوص والصور والفيديو، مما يجعله مناسبًا للتوسع الفعال في مجموعة متنوعة من المهام."
499
+ },
479
500
  "gemma-7b-it": {
480
501
  "description": "Gemma 7B مناسب لمعالجة المهام المتوسطة والصغيرة، ويجمع بين الكفاءة من حيث التكلفة."
481
502
  },
@@ -503,6 +524,9 @@
503
524
  "glm-4-0520": {
504
525
  "description": "GLM-4-0520 هو أحدث إصدار من النموذج، مصمم للمهام المعقدة والمتنوعة، ويظهر أداءً ممتازًا."
505
526
  },
527
+ "glm-4-9b-chat": {
528
+ "description": "يظهر GLM-4-9B-Chat أداءً عاليًا في مجالات متعددة مثل الدلالات والرياضيات والاستدلال والترميز والمعرفة. كما أنه مزود بقدرات تصفح الويب وتنفيذ الشيفرات واستدعاء الأدوات المخصصة واستدلال النصوص الطويلة. يدعم 26 لغة بما في ذلك اليابانية والكورية والألمانية."
529
+ },
506
530
  "glm-4-air": {
507
531
  "description": "GLM-4-Air هو إصدار ذو قيمة عالية، يتمتع بأداء قريب من GLM-4، ويقدم سرعة عالية وسعرًا معقولًا."
508
532
  },
@@ -389,6 +389,11 @@
389
389
  "modelDesc": "نموذج مخصص لتحسين أسئلة المستخدمين",
390
390
  "title": "إعادة صياغة سؤال قاعدة المعرفة"
391
391
  },
392
+ "thread": {
393
+ "label": "نموذج تسمية الموضوعات الفرعية",
394
+ "modelDesc": "نموذج مخصص لإعادة تسمية الموضوعات الفرعية تلقائيًا",
395
+ "title": "تسمية الموضوعات الفرعية تلقائيًا"
396
+ },
392
397
  "title": "مساعد النظام",
393
398
  "topic": {
394
399
  "label": "نموذج تسمية الموضوع",
@@ -0,0 +1,5 @@
1
+ {
2
+ "actions": {
3
+ "confirmRemoveThread": "سيتم حذف هذا الموضوع الفرعي، ولن يمكن استعادته بعد الحذف، يرجى توخي الحذر."
4
+ }
5
+ }
@@ -8,6 +8,7 @@
8
8
  "agents": "Асистент",
9
9
  "artifact": {
10
10
  "generating": "Генериране",
11
+ "inThread": "Не можете да видите в подтемата, моля, превключете към основната дискусия.",
11
12
  "thinking": "В процес на мислене",
12
13
  "thought": "Процес на мислене",
13
14
  "unknownTitle": "Неназован артефакт"
@@ -65,6 +66,7 @@
65
66
  },
66
67
  "messageAction": {
67
68
  "delAndRegenerate": "Изтрий и прегенерирай",
69
+ "deleteDisabledByThreads": "Съществуват подтеми, не можете да изтриете.",
68
70
  "regenerate": "Прегенерирай"
69
71
  },
70
72
  "newAgent": "Нов агент",
@@ -121,6 +123,11 @@
121
123
  "loading": "Разпознаване...",
122
124
  "prettifying": "Изглаждане..."
123
125
  },
126
+ "thread": {
127
+ "divider": "Подтема",
128
+ "threadMessageCount": "{{messageCount}} съобщения",
129
+ "title": "Подтема"
130
+ },
124
131
  "tokenDetails": {
125
132
  "chats": "Чат съобщения",
126
133
  "historySummary": "Историческо резюме",
@@ -16,6 +16,8 @@
16
16
  "back": "Назад",
17
17
  "batchDelete": "Пакетно изтриване",
18
18
  "blog": "Продуктов блог",
19
+ "branching": "Създаване на подтема",
20
+ "branchingDisable": "Функцията „подтема“ е налична само в сървърната версия. Ако искате да използвате тази функция, моля, превключете на режим на сървърно разполагане или използвайте LobeChat Cloud.",
19
21
  "cancel": "Отказ",
20
22
  "changelog": "Дневник на промените",
21
23
  "close": "Затвори",
@@ -176,6 +176,15 @@
176
176
  "Qwen/Qwen2.5-Math-72B-Instruct": {
177
177
  "description": "Qwen2.5-Math се фокусира върху решаването на математически проблеми, предоставяйки професионални отговори на трудни задачи."
178
178
  },
179
+ "Qwen2-72B-Instruct": {
180
+ "description": "Qwen2 е най-новата серия на модела Qwen, поддържаща 128k контекст. В сравнение с текущите най-добри отворени модели, Qwen2-72B значително надминава водещите модели в области като разбиране на естествен език, знания, код, математика и многоезичност."
181
+ },
182
+ "Qwen2-7B-Instruct": {
183
+ "description": "Qwen2 е най-новата серия на модела Qwen, способен да надмине оптималните отворени модели с равен размер или дори по-големи модели. Qwen2 7B постига значителни предимства в множество тестове, особено в разбирането на код и китайския език."
184
+ },
185
+ "Qwen2.5-72B-Instruct": {
186
+ "description": "Qwen2.5-72B-Instruct поддържа 16k контекст, генерира дълги текстове над 8K. Поддържа функция за извикване и безпроблемна интеграция с външни системи, значително увеличаваща гъвкавостта и разширяемостта. Моделът има значително увеличени знания и значително подобрени способности в кодиране и математика, с поддръжка на над 29 езика."
187
+ },
179
188
  "SenseChat": {
180
189
  "description": "Основна версия на модела (V4), с контекстна дължина 4K, с мощни общи способности."
181
190
  },
@@ -206,6 +215,9 @@
206
215
  "Tencent/Hunyuan-A52B-Instruct": {
207
216
  "description": "Hunyuan-Large е най-голямата отворена трансформаторна архитектура MoE в индустрията, с общо 3890 милиарда параметри и 52 милиарда активни параметри."
208
217
  },
218
+ "Yi-34B-Chat": {
219
+ "description": "Yi-1.5-34B значително подобрява математическата логика и способностите в кодирането, като запазва отличните общи езикови способности на оригиналната серия модели, чрез инкрементално обучение с 500 милиарда висококачествени токени."
220
+ },
209
221
  "abab5.5-chat": {
210
222
  "description": "Насочена към производствени сценарии, поддържаща обработка на сложни задачи и ефективно генериране на текст, подходяща за професионални приложения."
211
223
  },
@@ -368,6 +380,9 @@
368
380
  "codegeex-4": {
369
381
  "description": "CodeGeeX-4 е мощен AI помощник за програмиране, който поддържа интелигентни въпроси и отговори и автоматично допълване на код за различни програмни езици, повишавайки ефективността на разработката."
370
382
  },
383
+ "codegeex4-all-9b": {
384
+ "description": "CodeGeeX4-ALL-9B е многоезичен модел за генериране на код, който предлага пълни функции, включително попълване и генериране на код, интерпретатор на код, уеб търсене, извикване на функции и въпроси и отговори на ниво хранилище, обхващащ различни сценарии на софтуерна разработка. Това е водещ модел за генериране на код с по-малко от 10B параметри."
385
+ },
371
386
  "codegemma": {
372
387
  "description": "CodeGemma е лек езиков модел, специализиран в различни програмни задачи, поддържащ бърза итерация и интеграция."
373
388
  },
@@ -422,6 +437,9 @@
422
437
  "deepseek-chat": {
423
438
  "description": "Новооткритият отворен модел, който съчетава общи и кодови способности, не само запазва общата диалогова способност на оригиналния Chat модел и мощната способност за обработка на код на Coder модела, но също така по-добре се съгласува с човешките предпочитания. Освен това, DeepSeek-V2.5 постигна значителни подобрения в писателските задачи, следването на инструкции и много други области."
424
439
  },
440
+ "deepseek-coder-33B-instruct": {
441
+ "description": "DeepSeek Coder 33B е модел за кодови езици, обучен на 20 трилиона данни, от които 87% са код и 13% са на китайски и английски. Моделът въвежда размер на прозореца от 16K и задачи за попълване, предоставяйки функции за попълване на код на проектно ниво и попълване на фрагменти."
442
+ },
425
443
  "deepseek-coder-v2": {
426
444
  "description": "DeepSeek Coder V2 е отворен хибриден експертен кодов модел, който се представя отлично в кодовите задачи, сравним с GPT4-Turbo."
427
445
  },
@@ -476,6 +494,9 @@
476
494
  "gemini-exp-1114": {
477
495
  "description": "Gemini Exp 1114 е най-новият експериментален многомодален AI модел на Google, който предлага бърза обработка и поддържа вход от текст, изображения и видео, подходящ за ефективно разширение на множество задачи."
478
496
  },
497
+ "gemini-exp-1121": {
498
+ "description": "Gemini Exp 1121 е най-новият експериментален мултимодален AI модел на Google, който предлага бърза обработка и поддържа текстови, изображенчески и видео входове, подходящ за ефективно разширяване на множество задачи."
499
+ },
479
500
  "gemma-7b-it": {
480
501
  "description": "Gemma 7B е подходяща за обработка на средни и малки задачи, съчетаваща икономичност."
481
502
  },
@@ -503,6 +524,9 @@
503
524
  "glm-4-0520": {
504
525
  "description": "GLM-4-0520 е най-новата версия на модела, проектирана за високо сложни и разнообразни задачи, с отлични резултати."
505
526
  },
527
+ "glm-4-9b-chat": {
528
+ "description": "GLM-4-9B-Chat показва висока производителност в множество области, включително семантика, математика, логическо разсъждение, код и знания. Също така предлага уеб браузинг, изпълнение на код, извикване на персонализирани инструменти и разсъждение върху дълги текстове. Поддържа 26 езика, включително японски, корейски и немски."
529
+ },
506
530
  "glm-4-air": {
507
531
  "description": "GLM-4-Air е икономичен вариант, с производителност близка до GLM-4, предлагаща бързина и достъпна цена."
508
532
  },
@@ -389,6 +389,11 @@
389
389
  "modelDesc": "Определя модел за оптимизиране на запитванията на потребителите",
390
390
  "title": "Пренаписване на въпроси от базата данни"
391
391
  },
392
+ "thread": {
393
+ "label": "Модел за именуване на подтеми",
394
+ "modelDesc": "Модел, предназначен за автоматично преименуване на подтеми",
395
+ "title": "Автоматично именуване на подтеми"
396
+ },
392
397
  "title": "Системен асистент",
393
398
  "topic": {
394
399
  "label": "Модел за именуване на теми",
@@ -0,0 +1,5 @@
1
+ {
2
+ "actions": {
3
+ "confirmRemoveThread": "Ще изтриете тази подтема. След изтриването ѝ няма да може да бъде възстановена, моля, бъдете внимателни."
4
+ }
5
+ }
@@ -8,6 +8,7 @@
8
8
  "agents": "Assistent",
9
9
  "artifact": {
10
10
  "generating": "Wird generiert",
11
+ "inThread": "In Unterthemen kann nicht angezeigt werden, bitte wechseln Sie zum Hauptdiskussionsbereich.",
11
12
  "thinking": "Denken",
12
13
  "thought": "Denkenprozess",
13
14
  "unknownTitle": "Unbenanntes Werk"
@@ -65,6 +66,7 @@
65
66
  },
66
67
  "messageAction": {
67
68
  "delAndRegenerate": "Löschen und neu generieren",
69
+ "deleteDisabledByThreads": "Es gibt Unterthemen, die Löschung ist nicht möglich.",
68
70
  "regenerate": "Neu generieren"
69
71
  },
70
72
  "newAgent": "Neuer Assistent",
@@ -121,6 +123,11 @@
121
123
  "loading": "Erkenne...",
122
124
  "prettifying": "Verschönern..."
123
125
  },
126
+ "thread": {
127
+ "divider": "Unterthema",
128
+ "threadMessageCount": "{{messageCount}} Nachrichten",
129
+ "title": "Unterthema"
130
+ },
124
131
  "tokenDetails": {
125
132
  "chats": "Chats",
126
133
  "historySummary": "Historische Zusammenfassung",
@@ -16,6 +16,8 @@
16
16
  "back": "Zurück",
17
17
  "batchDelete": "Massenlöschung",
18
18
  "blog": "Produkt-Blog",
19
+ "branching": "Unterthema erstellen",
20
+ "branchingDisable": "Die Funktion „Unterthema“ ist nur in der Serverversion verfügbar. Wenn Sie diese Funktion benötigen, wechseln Sie bitte in den Serverbereitstellungsmodus oder verwenden Sie LobeChat Cloud.",
19
21
  "cancel": "Abbrechen",
20
22
  "changelog": "Änderungsprotokoll",
21
23
  "close": "Schließen",
@@ -176,6 +176,15 @@
176
176
  "Qwen/Qwen2.5-Math-72B-Instruct": {
177
177
  "description": "Qwen2.5-Math konzentriert sich auf die Problemlösung im Bereich Mathematik und bietet professionelle Lösungen für schwierige Aufgaben."
178
178
  },
179
+ "Qwen2-72B-Instruct": {
180
+ "description": "Qwen2 ist die neueste Reihe des Qwen-Modells, das 128k Kontext unterstützt. Im Vergleich zu den derzeit besten Open-Source-Modellen übertrifft Qwen2-72B in den Bereichen natürliche Sprachverständnis, Wissen, Code, Mathematik und Mehrsprachigkeit deutlich die führenden Modelle."
181
+ },
182
+ "Qwen2-7B-Instruct": {
183
+ "description": "Qwen2 ist die neueste Reihe des Qwen-Modells, das in der Lage ist, die besten Open-Source-Modelle ähnlicher Größe oder sogar größerer Modelle zu übertreffen. Qwen2 7B hat in mehreren Bewertungen signifikante Vorteile erzielt, insbesondere im Bereich Code und Verständnis der chinesischen Sprache."
184
+ },
185
+ "Qwen2.5-72B-Instruct": {
186
+ "description": "Qwen2.5-72B-Instruct unterstützt 16k Kontext und generiert lange Texte über 8K. Es unterstützt Funktionsaufrufe und nahtlose Interaktionen mit externen Systemen, was die Flexibilität und Skalierbarkeit erheblich verbessert. Das Wissen des Modells hat deutlich zugenommen, und die Codierungs- und mathematischen Fähigkeiten wurden erheblich verbessert, mit Unterstützung für über 29 Sprachen."
187
+ },
179
188
  "SenseChat": {
180
189
  "description": "Basisversion des Modells (V4) mit 4K Kontextlänge, die über starke allgemeine Fähigkeiten verfügt."
181
190
  },
@@ -206,6 +215,9 @@
206
215
  "Tencent/Hunyuan-A52B-Instruct": {
207
216
  "description": "Hunyuan-Large ist das größte Open-Source-Transformer-Architektur MoE-Modell der Branche mit insgesamt 389 Milliarden Parametern und 52 Milliarden aktiven Parametern."
208
217
  },
218
+ "Yi-34B-Chat": {
219
+ "description": "Yi-1.5-34B hat die hervorragenden allgemeinen Sprachfähigkeiten des ursprünglichen Modells beibehalten und durch inkrementelles Training von 500 Milliarden hochwertigen Tokens die mathematische Logik und Codierungsfähigkeiten erheblich verbessert."
220
+ },
209
221
  "abab5.5-chat": {
210
222
  "description": "Für produktivitätsorientierte Szenarien konzipiert, unterstützt es die Verarbeitung komplexer Aufgaben und die effiziente Textgenerierung, geeignet für professionelle Anwendungen."
211
223
  },
@@ -368,6 +380,9 @@
368
380
  "codegeex-4": {
369
381
  "description": "CodeGeeX-4 ist ein leistungsstarker AI-Programmierassistent, der intelligente Fragen und Codevervollständigung in verschiedenen Programmiersprachen unterstützt und die Entwicklungseffizienz steigert."
370
382
  },
383
+ "codegeex4-all-9b": {
384
+ "description": "CodeGeeX4-ALL-9B ist ein mehrsprachiges Code-Generierungsmodell, das umfassende Funktionen unterstützt, darunter Code-Vervollständigung und -Generierung, Code-Interpreter, Websuche, Funktionsaufrufe und repository-weite Codefragen und -antworten, und deckt verschiedene Szenarien der Softwareentwicklung ab. Es ist das führende Code-Generierungsmodell mit weniger als 10B Parametern."
385
+ },
371
386
  "codegemma": {
372
387
  "description": "CodeGemma ist ein leichtgewichtiges Sprachmodell, das speziell für verschiedene Programmieraufgaben entwickelt wurde und schnelle Iterationen und Integrationen unterstützt."
373
388
  },
@@ -422,6 +437,9 @@
422
437
  "deepseek-chat": {
423
438
  "description": "Ein neues Open-Source-Modell, das allgemeine und Codefähigkeiten kombiniert. Es bewahrt nicht nur die allgemeinen Dialogfähigkeiten des ursprünglichen Chat-Modells und die leistungsstarken Codeverarbeitungsfähigkeiten des Coder-Modells, sondern stimmt auch besser mit menschlichen Präferenzen überein. Darüber hinaus hat DeepSeek-V2.5 in mehreren Bereichen wie Schreibaufgaben und Befolgung von Anweisungen erhebliche Verbesserungen erzielt."
424
439
  },
440
+ "deepseek-coder-33B-instruct": {
441
+ "description": "DeepSeek Coder 33B ist ein Code-Sprachmodell, das auf 20 Billionen Daten trainiert wurde, von denen 87 % Code und 13 % in Chinesisch und Englisch sind. Das Modell führt eine Fenstergröße von 16K und Aufgaben zur Lückenergänzung ein und bietet projektbezogene Code-Vervollständigung und Fragmentfüllfunktionen."
442
+ },
425
443
  "deepseek-coder-v2": {
426
444
  "description": "DeepSeek Coder V2 ist ein Open-Source-Mischexperten-Code-Modell, das in Codeaufgaben hervorragende Leistungen erbringt und mit GPT4-Turbo vergleichbar ist."
427
445
  },
@@ -476,6 +494,9 @@
476
494
  "gemini-exp-1114": {
477
495
  "description": "Gemini Exp 1114 ist Googles neuestes experimentelles multimodales KI-Modell, das über eine schnelle Verarbeitungskapazität verfügt und Texte, Bilder und Videoeingaben unterstützt, um eine effiziente Skalierung für verschiedene Aufgaben zu ermöglichen."
478
496
  },
497
+ "gemini-exp-1121": {
498
+ "description": "Gemini Exp 1121 ist Googles neuestes experimentelles multimodales KI-Modell, das über eine schnelle Verarbeitungskapazität verfügt und Texte, Bilder und Videoeingaben unterstützt, um eine effiziente Skalierung für verschiedene Aufgaben zu ermöglichen."
499
+ },
479
500
  "gemma-7b-it": {
480
501
  "description": "Gemma 7B eignet sich für die Verarbeitung von mittelgroßen Aufgaben und bietet ein gutes Kosten-Nutzen-Verhältnis."
481
502
  },
@@ -503,6 +524,9 @@
503
524
  "glm-4-0520": {
504
525
  "description": "GLM-4-0520 ist die neueste Modellversion, die für hochkomplexe und vielfältige Aufgaben konzipiert wurde und hervorragende Leistungen zeigt."
505
526
  },
527
+ "glm-4-9b-chat": {
528
+ "description": "GLM-4-9B-Chat zeigt in den Bereichen Semantik, Mathematik, Schlussfolgerungen, Code und Wissen eine hohe Leistung. Es verfügt auch über Funktionen wie Web-Browsing, Code-Ausführung, benutzerdefinierte Toolaufrufe und langes Textverständnis. Es unterstützt 26 Sprachen, darunter Japanisch, Koreanisch und Deutsch."
529
+ },
506
530
  "glm-4-air": {
507
531
  "description": "GLM-4-Air ist eine kosteneffiziente Version, die in der Leistung nahe am GLM-4 liegt und schnelle Geschwindigkeiten zu einem erschwinglichen Preis bietet."
508
532
  },
@@ -389,6 +389,11 @@
389
389
  "modelDesc": "Modell zur Optimierung der Benutzeranfragen",
390
390
  "title": "Wiederformulierung von Fragen aus der Wissensdatenbank"
391
391
  },
392
+ "thread": {
393
+ "label": "Unterthema-Namensmodell",
394
+ "modelDesc": "Modell zur automatischen Umbenennung von Unterthemen",
395
+ "title": "Automatische Benennung von Unterthemen"
396
+ },
392
397
  "title": "Systemassistent",
393
398
  "topic": {
394
399
  "label": "Themenbenennungsmodell",
@@ -0,0 +1,5 @@
1
+ {
2
+ "actions": {
3
+ "confirmRemoveThread": "Sie sind dabei, dieses Unterthema zu löschen. Nach dem Löschen kann es nicht wiederhergestellt werden. Bitte seien Sie vorsichtig."
4
+ }
5
+ }
@@ -8,6 +8,7 @@
8
8
  "agents": "Assistants",
9
9
  "artifact": {
10
10
  "generating": "Generating",
11
+ "inThread": "Cannot view in subtopic, please switch to the main conversation area to open",
11
12
  "thinking": "Thinking",
12
13
  "thought": "Thought Process",
13
14
  "unknownTitle": "Untitled Work"
@@ -65,6 +66,7 @@
65
66
  },
66
67
  "messageAction": {
67
68
  "delAndRegenerate": "Delete and Regenerate",
69
+ "deleteDisabledByThreads": "There are subtopics, deletion is not allowed",
68
70
  "regenerate": "Regenerate"
69
71
  },
70
72
  "newAgent": "New Assistant",
@@ -121,6 +123,11 @@
121
123
  "loading": "Recognizing...",
122
124
  "prettifying": "Polishing..."
123
125
  },
126
+ "thread": {
127
+ "divider": "Subtopic",
128
+ "threadMessageCount": "{{messageCount}} messages",
129
+ "title": "Subtopic"
130
+ },
124
131
  "tokenDetails": {
125
132
  "chats": "Chat Messages",
126
133
  "historySummary": "History Summary",
@@ -16,6 +16,8 @@
16
16
  "back": "Back",
17
17
  "batchDelete": "Batch Delete",
18
18
  "blog": "Product Blog",
19
+ "branching": "Create Subtopic",
20
+ "branchingDisable": "The 'Subtopic' feature is only available in the server version. If you need this feature, please switch to server deployment mode or use LobeChat Cloud.",
19
21
  "cancel": "Cancel",
20
22
  "changelog": "Changelog",
21
23
  "close": "Close",
@@ -176,6 +176,15 @@
176
176
  "Qwen/Qwen2.5-Math-72B-Instruct": {
177
177
  "description": "Qwen2.5-Math focuses on problem-solving in the field of mathematics, providing expert solutions for challenging problems."
178
178
  },
179
+ "Qwen2-72B-Instruct": {
180
+ "description": "Qwen2 is the latest series of the Qwen model, supporting 128k context. Compared to the current best open-source models, Qwen2-72B significantly surpasses leading models in natural language understanding, knowledge, coding, mathematics, and multilingual capabilities."
181
+ },
182
+ "Qwen2-7B-Instruct": {
183
+ "description": "Qwen2 is the latest series of the Qwen model, capable of outperforming optimal open-source models of similar size and even larger models. Qwen2 7B has achieved significant advantages in multiple evaluations, especially in coding and Chinese comprehension."
184
+ },
185
+ "Qwen2.5-72B-Instruct": {
186
+ "description": "Qwen2.5-72B-Instruct supports 16k context and generates long texts exceeding 8K. It enables seamless interaction with external systems through function calls, greatly enhancing flexibility and scalability. The model's knowledge has significantly increased, and its coding and mathematical abilities have been greatly improved, with multilingual support for over 29 languages."
187
+ },
179
188
  "SenseChat": {
180
189
  "description": "Basic version model (V4) with a context length of 4K, featuring strong general capabilities."
181
190
  },
@@ -206,6 +215,9 @@
206
215
  "Tencent/Hunyuan-A52B-Instruct": {
207
216
  "description": "Hunyuan-Large is the industry's largest open-source Transformer architecture MoE model, with a total of 389 billion parameters and 52 billion active parameters."
208
217
  },
218
+ "Yi-34B-Chat": {
219
+ "description": "Yi-1.5-34B significantly enhances mathematical logic and coding abilities by incrementally training on 500 billion high-quality tokens while maintaining the excellent general language capabilities of the original series."
220
+ },
209
221
  "abab5.5-chat": {
210
222
  "description": "Targeted at productivity scenarios, supporting complex task processing and efficient text generation, suitable for professional applications."
211
223
  },
@@ -368,6 +380,9 @@
368
380
  "codegeex-4": {
369
381
  "description": "CodeGeeX-4 is a powerful AI programming assistant that supports intelligent Q&A and code completion in various programming languages, enhancing development efficiency."
370
382
  },
383
+ "codegeex4-all-9b": {
384
+ "description": "CodeGeeX4-ALL-9B is a multilingual code generation model that supports comprehensive functions including code completion and generation, code interpretation, web search, function calls, and repository-level code Q&A, covering various scenarios in software development. It is a top-tier code generation model with fewer than 10B parameters."
385
+ },
371
386
  "codegemma": {
372
387
  "description": "CodeGemma is a lightweight language model dedicated to various programming tasks, supporting rapid iteration and integration."
373
388
  },
@@ -422,6 +437,9 @@
422
437
  "deepseek-chat": {
423
438
  "description": "A new open-source model that integrates general and coding capabilities, retaining the general conversational abilities of the original Chat model and the powerful code handling capabilities of the Coder model, while better aligning with human preferences. Additionally, DeepSeek-V2.5 has achieved significant improvements in writing tasks, instruction following, and more."
424
439
  },
440
+ "deepseek-coder-33B-instruct": {
441
+ "description": "DeepSeek Coder 33B is a code language model trained on 20 trillion data points, of which 87% are code and 13% are in Chinese and English. The model introduces a 16K window size and fill-in-the-blank tasks, providing project-level code completion and snippet filling capabilities."
442
+ },
425
443
  "deepseek-coder-v2": {
426
444
  "description": "DeepSeek Coder V2 is an open-source hybrid expert code model that performs excellently in coding tasks, comparable to GPT4-Turbo."
427
445
  },
@@ -476,6 +494,9 @@
476
494
  "gemini-exp-1114": {
477
495
  "description": "Gemini Exp 1114 is Google's latest experimental multimodal AI model, featuring rapid processing capabilities and supporting text, image, and video inputs, making it suitable for efficient scaling across various tasks."
478
496
  },
497
+ "gemini-exp-1121": {
498
+ "description": "Gemini Exp 1121 is Google's latest experimental multimodal AI model, featuring rapid processing capabilities and supporting text, image, and video inputs, making it efficient for a variety of tasks."
499
+ },
479
500
  "gemma-7b-it": {
480
501
  "description": "Gemma 7B is suitable for medium to small-scale task processing, offering cost-effectiveness."
481
502
  },
@@ -503,6 +524,9 @@
503
524
  "glm-4-0520": {
504
525
  "description": "GLM-4-0520 is the latest model version designed for highly complex and diverse tasks, demonstrating outstanding performance."
505
526
  },
527
+ "glm-4-9b-chat": {
528
+ "description": "GLM-4-9B-Chat demonstrates high performance across various aspects, including semantics, mathematics, reasoning, coding, and knowledge. It also features web browsing, code execution, custom tool invocation, and long text reasoning, supporting 26 languages including Japanese, Korean, and German."
529
+ },
506
530
  "glm-4-air": {
507
531
  "description": "GLM-4-Air is a cost-effective version with performance close to GLM-4, offering fast speed at an affordable price."
508
532
  },
@@ -389,6 +389,11 @@
389
389
  "modelDesc": "Specify the model used to optimize user inquiries",
390
390
  "title": "Knowledge Base Question Rewrite"
391
391
  },
392
+ "thread": {
393
+ "label": "Subtopic Naming Model",
394
+ "modelDesc": "The model designated for automatic renaming of subtopics",
395
+ "title": "Automatic Subtopic Naming"
396
+ },
392
397
  "title": "System Assistants",
393
398
  "topic": {
394
399
  "label": "Topic Naming Model",
@@ -0,0 +1,5 @@
1
+ {
2
+ "actions": {
3
+ "confirmRemoveThread": "You are about to delete this subtopic. Once deleted, it cannot be recovered. Please proceed with caution."
4
+ }
5
+ }
@@ -8,6 +8,7 @@
8
8
  "agents": "Asistente",
9
9
  "artifact": {
10
10
  "generating": "Generando",
11
+ "inThread": "No se puede ver en el subtema, cambie a la zona de conversación principal para abrirlo",
11
12
  "thinking": "Pensando",
12
13
  "thought": "Proceso de pensamiento",
13
14
  "unknownTitle": "Obra sin título"
@@ -65,6 +66,7 @@
65
66
  },
66
67
  "messageAction": {
67
68
  "delAndRegenerate": "Eliminar y Regenerar",
69
+ "deleteDisabledByThreads": "Existen subtemas, no se puede eliminar",
68
70
  "regenerate": "Regenerar"
69
71
  },
70
72
  "newAgent": "Nuevo asistente",
@@ -121,6 +123,11 @@
121
123
  "loading": "Reconociendo...",
122
124
  "prettifying": "Embelleciendo..."
123
125
  },
126
+ "thread": {
127
+ "divider": "Subtema",
128
+ "threadMessageCount": "{{messageCount}} mensajes",
129
+ "title": "Subtema"
130
+ },
124
131
  "tokenDetails": {
125
132
  "chats": "Mensajes de chat",
126
133
  "historySummary": "Resumen histórico",
@@ -16,6 +16,8 @@
16
16
  "back": "Volver",
17
17
  "batchDelete": "Eliminar en lote",
18
18
  "blog": "Blog de productos",
19
+ "branching": "Crear subtemas",
20
+ "branchingDisable": "La función de «subtemas» solo está disponible en la versión del servidor. Si necesita esta función, cambie al modo de implementación del servidor o utilice LobeChat Cloud.",
19
21
  "cancel": "Cancelar",
20
22
  "changelog": "Registro de cambios",
21
23
  "close": "Cerrar",