@lobehub/chat 1.74.1 → 1.74.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +59 -0
- package/README.md +2 -2
- package/README.zh-CN.md +2 -2
- package/changelog/v1.json +18 -0
- package/docs/developer/database-schema.dbml +54 -2
- package/docs/self-hosting/environment-variables/model-provider.mdx +9 -7
- package/docs/self-hosting/environment-variables/model-provider.zh-CN.mdx +9 -7
- package/locales/ar/common.json +51 -0
- package/locales/ar/models.json +69 -3
- package/locales/ar/providers.json +6 -0
- package/locales/bg-BG/common.json +51 -0
- package/locales/bg-BG/models.json +69 -3
- package/locales/bg-BG/providers.json +6 -0
- package/locales/de-DE/common.json +51 -0
- package/locales/de-DE/models.json +69 -3
- package/locales/de-DE/providers.json +6 -0
- package/locales/en-US/common.json +51 -0
- package/locales/en-US/models.json +69 -3
- package/locales/en-US/providers.json +6 -3
- package/locales/es-ES/common.json +51 -0
- package/locales/es-ES/models.json +69 -3
- package/locales/es-ES/providers.json +6 -0
- package/locales/fa-IR/common.json +51 -0
- package/locales/fa-IR/models.json +69 -3
- package/locales/fa-IR/providers.json +6 -0
- package/locales/fr-FR/common.json +51 -0
- package/locales/fr-FR/models.json +69 -3
- package/locales/fr-FR/providers.json +6 -0
- package/locales/it-IT/common.json +51 -0
- package/locales/it-IT/models.json +69 -3
- package/locales/it-IT/providers.json +6 -0
- package/locales/ja-JP/common.json +51 -0
- package/locales/ja-JP/models.json +78 -4
- package/locales/ja-JP/providers.json +6 -0
- package/locales/ko-KR/common.json +51 -0
- package/locales/ko-KR/models.json +69 -3
- package/locales/ko-KR/providers.json +6 -0
- package/locales/nl-NL/common.json +51 -0
- package/locales/nl-NL/models.json +69 -3
- package/locales/nl-NL/providers.json +6 -0
- package/locales/pl-PL/common.json +51 -0
- package/locales/pl-PL/models.json +69 -3
- package/locales/pl-PL/providers.json +6 -0
- package/locales/pt-BR/common.json +51 -0
- package/locales/pt-BR/models.json +69 -3
- package/locales/pt-BR/providers.json +6 -0
- package/locales/ru-RU/common.json +51 -0
- package/locales/ru-RU/models.json +69 -3
- package/locales/ru-RU/providers.json +6 -0
- package/locales/tr-TR/common.json +51 -0
- package/locales/tr-TR/models.json +69 -3
- package/locales/tr-TR/providers.json +6 -0
- package/locales/vi-VN/common.json +51 -0
- package/locales/vi-VN/models.json +69 -3
- package/locales/vi-VN/providers.json +6 -0
- package/locales/zh-CN/common.json +53 -2
- package/locales/zh-CN/models.json +79 -13
- package/locales/zh-CN/providers.json +6 -4
- package/locales/zh-TW/common.json +51 -0
- package/locales/zh-TW/models.json +81 -4
- package/locales/zh-TW/providers.json +6 -0
- package/package.json +1 -1
- package/packages/web-crawler/src/utils/__tests__/withTimeout.test.ts +0 -1
- package/src/app/[variants]/(main)/settings/provider/features/ProviderConfig/Checker.tsx +9 -1
- package/src/config/aiModels/qwen.ts +4 -4
- package/src/config/aiModels/volcengine.ts +2 -2
- package/src/database/client/db.ts +102 -11
- package/src/database/client/migrations.json +38 -8
- package/src/database/migrations/0018_add_client_id_for_entities.sql +32 -0
- package/src/database/migrations/meta/0018_snapshot.json +4212 -0
- package/src/database/migrations/meta/_journal.json +7 -0
- package/src/database/models/drizzleMigration.ts +23 -0
- package/src/database/schemas/agent.ts +48 -31
- package/src/database/schemas/file.ts +32 -16
- package/src/database/schemas/message.ts +91 -54
- package/src/database/schemas/rag.ts +65 -32
- package/src/database/schemas/session.ts +6 -3
- package/src/database/schemas/topic.ts +31 -24
- package/src/features/InitClientDB/ErrorResult.tsx +53 -32
- package/src/features/InitClientDB/features/DatabaseRepair/Backup.tsx +77 -0
- package/src/features/InitClientDB/features/DatabaseRepair/Diagnosis.tsx +98 -0
- package/src/features/InitClientDB/features/DatabaseRepair/Repair.tsx +220 -0
- package/src/features/InitClientDB/features/DatabaseRepair/index.tsx +85 -0
- package/src/features/ModelSwitchPanel/index.tsx +13 -7
- package/src/locales/default/common.ts +53 -1
- package/src/store/global/actions/clientDb.ts +19 -3
- package/src/store/global/initialState.ts +6 -1
- package/src/store/global/selectors/clientDB.ts +43 -0
- package/src/store/global/selectors/index.ts +1 -0
- package/src/store/user/slices/settings/selectors/general.test.ts +90 -0
- package/src/types/clientDB.ts +13 -0
@@ -518,6 +518,18 @@
|
|
518
518
|
"baichuan/baichuan2-13b-chat": {
|
519
519
|
"description": "Baichuan-13B е отворен, комерсиален голям езиков модел, разработен от Baichuan Intelligence, с 13 милиарда параметри, който постига най-добрите резултати в своя размер на авторитетни бенчмаркове на китайски и английски."
|
520
520
|
},
|
521
|
+
"c4ai-aya-expanse-32b": {
|
522
|
+
"description": "Aya Expanse е високопроизводителен многоезичен модел с 32B, проектиран да предизвика представянето на едноезични модели чрез иновации в настройката на инструкции, арбитраж на данни, обучение на предпочитания и комбиниране на модели. Той поддържа 23 езика."
|
523
|
+
},
|
524
|
+
"c4ai-aya-expanse-8b": {
|
525
|
+
"description": "Aya Expanse е високопроизводителен многоезичен модел с 8B, проектиран да предизвика представянето на едноезични модели чрез иновации в настройката на инструкции, арбитраж на данни, обучение на предпочитания и комбиниране на модели. Той поддържа 23 езика."
|
526
|
+
},
|
527
|
+
"c4ai-aya-vision-32b": {
|
528
|
+
"description": "Aya Vision е авангарден много модален модел, който показва отлични резултати в множество ключови бенчмаркове за езикови, текстови и визуални способности. Той поддържа 23 езика. Тази версия с 32 милиарда параметри се фокусира върху авангарден многоезичен представител."
|
529
|
+
},
|
530
|
+
"c4ai-aya-vision-8b": {
|
531
|
+
"description": "Aya Vision е авангарден много модален модел, който показва отлични резултати в множество ключови бенчмаркове за езикови, текстови и визуални способности. Тази версия с 8 милиарда параметри се фокусира върху ниска латентност и оптимална производителност."
|
532
|
+
},
|
521
533
|
"charglm-3": {
|
522
534
|
"description": "CharGLM-3 е проектиран за ролеви игри и емоционално придружаване, поддържаща дълга многократна памет и персонализиран диалог, с широко приложение."
|
523
535
|
},
|
@@ -602,12 +614,39 @@
|
|
602
614
|
"cohere-command-r-plus": {
|
603
615
|
"description": "Command R+ е модел, оптимизиран за RAG, проектиран да се справя с натоварвания на ниво предприятие."
|
604
616
|
},
|
617
|
+
"command": {
|
618
|
+
"description": "Диалогов модел, следващ инструкции, който показва високо качество и надеждност в езиковите задачи, с по-дълга контекстна дължина в сравнение с нашия основен генеративен модел."
|
619
|
+
},
|
620
|
+
"command-a-03-2025": {
|
621
|
+
"description": "Команда A е нашият най-мощен модел до момента, който показва отлични резултати в използването на инструменти, агенти, подобрено генериране на информация (RAG) и многоезични приложения. Команда A разполага с контекстна дължина от 256K и може да работи само с две GPU, а производителността е увеличена с 150% в сравнение с Команда R+ 08-2024."
|
622
|
+
},
|
623
|
+
"command-light": {
|
624
|
+
"description": "По-малка и по-бърза версия на Команда, почти толкова мощна, но с по-бърза скорост."
|
625
|
+
},
|
626
|
+
"command-light-nightly": {
|
627
|
+
"description": "За да съкратим времевия интервал между основните версии, пуснахме нощна версия на модела Команда. За серията command-light, тази версия се нарича command-light-nightly. Обърнете внимание, че command-light-nightly е най-новата, най-експериментална и (възможно) нестабилна версия. Нощните версии се актуализират редовно и без предварително уведомление, затова не се препоръчва използването им в производствени среди."
|
628
|
+
},
|
629
|
+
"command-nightly": {
|
630
|
+
"description": "За да съкратим времевия интервал между основните версии, пуснахме нощна версия на модела Команда. За серията Команда, тази версия се нарича command-cightly. Обърнете внимание, че command-nightly е най-новата, най-експериментална и (възможно) нестабилна версия. Нощните версии се актуализират редовно и без предварително уведомление, затова не се препоръчва използването им в производствени среди."
|
631
|
+
},
|
605
632
|
"command-r": {
|
606
633
|
"description": "Command R е LLM, оптимизиран за диалогови и дълги контекстуални задачи, особено подходящ за динамично взаимодействие и управление на знания."
|
607
634
|
},
|
635
|
+
"command-r-03-2024": {
|
636
|
+
"description": "Команда R е диалогов модел, следващ инструкции, който показва по-високо качество и надеждност в езиковите задачи, с по-дълга контекстна дължина в сравнение с предишните модели. Той може да се използва за сложни работни потоци, като генериране на код, подобрено генериране на информация (RAG), използване на инструменти и агенти."
|
637
|
+
},
|
638
|
+
"command-r-08-2024": {
|
639
|
+
"description": "command-r-08-2024 е актуализирана версия на модела Команда R, пусната през август 2024 г."
|
640
|
+
},
|
608
641
|
"command-r-plus": {
|
609
642
|
"description": "Command R+ е високопроизводителен голям езиков модел, проектиран за реални бизнес сценарии и сложни приложения."
|
610
643
|
},
|
644
|
+
"command-r-plus-04-2024": {
|
645
|
+
"description": "Команда R+ е диалогов модел, следващ инструкции, който показва по-високо качество и надеждност в езиковите задачи, с по-дълга контекстна дължина в сравнение с предишните модели. Той е най-подходящ за сложни RAG работни потоци и многократна употреба на инструменти."
|
646
|
+
},
|
647
|
+
"command-r7b-12-2024": {
|
648
|
+
"description": "command-r7b-12-2024 е малка и ефективна актуализирана версия, пусната през декември 2024 г. Тя показва отлични резултати в задачи, изискващи сложни разсъждения и многократна обработка, като RAG, използване на инструменти и агенти."
|
649
|
+
},
|
611
650
|
"dall-e-2": {
|
612
651
|
"description": "Второ поколение модел DALL·E, поддържащ по-реалистично и точно генериране на изображения, с резолюция 4 пъти по-висока от първото поколение."
|
613
652
|
},
|
@@ -668,12 +707,24 @@
|
|
668
707
|
"deepseek-r1": {
|
669
708
|
"description": "DeepSeek-R1 е модел за извеждане, управляван от подсилено обучение (RL), който решава проблемите с повторяемостта и четимостта в модела. Преди RL, DeepSeek-R1 въвежда данни за студен старт, за да оптимизира допълнително производителността на извеждане. Той показва сравнима производителност с OpenAI-o1 в математически, кодови и извеждащи задачи и подобрява общите резултати чрез внимателно проектирани методи на обучение."
|
670
709
|
},
|
710
|
+
"deepseek-r1-70b-fast-online": {
|
711
|
+
"description": "DeepSeek R1 70B бърза версия, поддържаща търсене в реално време, предлагаща по-бърза скорост на отговор, без да компрометира производителността на модела."
|
712
|
+
},
|
713
|
+
"deepseek-r1-70b-online": {
|
714
|
+
"description": "DeepSeek R1 70B стандартна версия, поддържаща търсене в реално време, подходяща за диалози и текстови задачи, изискващи най-новата информация."
|
715
|
+
},
|
671
716
|
"deepseek-r1-distill-llama-70b": {
|
672
717
|
"description": "DeepSeek R1 - по-голям и по-интелигентен модел в комплекта DeepSeek - е дестилиран в архитектурата Llama 70B. На базата на бенчмаркове и човешка оценка, този модел е по-интелигентен от оригиналния Llama 70B, особено в задачи, изискващи математическа и фактическа точност."
|
673
718
|
},
|
674
719
|
"deepseek-r1-distill-llama-8b": {
|
675
720
|
"description": "Моделите от серията DeepSeek-R1-Distill са получени чрез техника на знание дестилация, като се фино настройват образците, генерирани от DeepSeek-R1, спрямо отворени модели като Qwen и Llama."
|
676
721
|
},
|
722
|
+
"deepseek-r1-distill-qianfan-llama-70b": {
|
723
|
+
"description": "Първоначално пуснат на 14 февруари 2025 г., дестилиран от екипа за разработка на модела Qianfan с базов модел Llama3_70B (създаден с Meta Llama), в дестилираните данни също е добавен корпус от Qianfan."
|
724
|
+
},
|
725
|
+
"deepseek-r1-distill-qianfan-llama-8b": {
|
726
|
+
"description": "Първоначално пуснат на 14 февруари 2025 г., дестилиран от екипа за разработка на модела Qianfan с базов модел Llama3_8B (създаден с Meta Llama), в дестилираните данни също е добавен корпус от Qianfan."
|
727
|
+
},
|
677
728
|
"deepseek-r1-distill-qwen-1.5b": {
|
678
729
|
"description": "Моделите от серията DeepSeek-R1-Distill са получени чрез техника на знание дестилация, като се фино настройват образците, генерирани от DeepSeek-R1, спрямо отворени модели като Qwen и Llama."
|
679
730
|
},
|
@@ -686,6 +737,12 @@
|
|
686
737
|
"deepseek-r1-distill-qwen-7b": {
|
687
738
|
"description": "Моделите от серията DeepSeek-R1-Distill са получени чрез техника на знание дестилация, като се фино настройват образците, генерирани от DeepSeek-R1, спрямо отворени модели като Qwen и Llama."
|
688
739
|
},
|
740
|
+
"deepseek-r1-fast-online": {
|
741
|
+
"description": "DeepSeek R1 пълна бърза версия, поддържаща търсене в реално време, комбинираща мощността на 671B параметри с по-бърза скорост на отговор."
|
742
|
+
},
|
743
|
+
"deepseek-r1-online": {
|
744
|
+
"description": "DeepSeek R1 пълна версия, с 671B параметри, поддържаща търсене в реално време, с по-силни способности за разбиране и генериране."
|
745
|
+
},
|
689
746
|
"deepseek-reasoner": {
|
690
747
|
"description": "Модел за извеждане, разработен от DeepSeek. Преди да предостави окончателния отговор, моделът първо извежда част от веригата на мислене, за да повиши точността на крайния отговор."
|
691
748
|
},
|
@@ -764,6 +821,9 @@
|
|
764
821
|
"ernie-4.0-turbo-8k-preview": {
|
765
822
|
"description": "Флагманският голям езиков модел, разработен от Baidu, с отлични общи резултати, широко приложим в сложни задачи в различни области; поддържа автоматично свързване с плъгина за търсене на Baidu, осигурявайки актуалност на информацията. В сравнение с ERNIE 4.0, показва по-добри резултати."
|
766
823
|
},
|
824
|
+
"ernie-4.5-8k-preview": {
|
825
|
+
"description": "Моделът Ernie 4.5 е ново поколение оригинален много модален основен модел, разработен от Baidu, който постига съвместна оптимизация чрез многомодално моделиране, с отлични способности за разбиране на много модалности; предлага усъвършенствани езикови способности, с подобрено разбиране, генериране, логика и памет, значително подобрени способности за избягване на халюцинации, логическо разсъждение и код."
|
826
|
+
},
|
767
827
|
"ernie-char-8k": {
|
768
828
|
"description": "Специализиран голям езиков модел, разработен от Baidu, подходящ за приложения като NPC в игри, диалози на клиентска поддръжка и ролеви игри, с по-изразителен и последователен стил на персонажите, по-силна способност за следване на инструкции и по-добра производителност на разсъжденията."
|
769
829
|
},
|
@@ -1097,9 +1157,6 @@
|
|
1097
1157
|
"hunyuan-turbo": {
|
1098
1158
|
"description": "Предварителна версия на новото поколение голям езиков модел на HunYuan, използваща нова структура на смесен експертен модел (MoE), с по-бърза скорост на извеждане и по-силни резултати в сравнение с hunyuan-pro."
|
1099
1159
|
},
|
1100
|
-
"hunyuan-turbo-20241120": {
|
1101
|
-
"description": "Фиксирана версия на hunyuan-turbo от 20 ноември 2024 г., която е между hunyuan-turbo и hunyuan-turbo-latest."
|
1102
|
-
},
|
1103
1160
|
"hunyuan-turbo-20241223": {
|
1104
1161
|
"description": "Оптимизация в тази версия: скалиране на данни и инструкции, значително повишаване на общата генерализационна способност на модела; значително повишаване на математическите, кодовите и логическите способности; оптимизиране на свързаните с разбирането на текста и думите способности; оптимизиране на качеството на генерираното съдържание при създаване на текст."
|
1105
1162
|
},
|
@@ -1109,6 +1166,15 @@
|
|
1109
1166
|
"hunyuan-turbo-vision": {
|
1110
1167
|
"description": "Новото поколение визуално езиково флагманско голямо модел на Hunyuan, използващо нова структура на смесен експертен модел (MoE), с цялостно подобрение на способностите за основно разпознаване, създаване на съдържание, отговори на въпроси и анализ и разсъждение в сравнение с предишното поколение модели."
|
1111
1168
|
},
|
1169
|
+
"hunyuan-turbos-20250226": {
|
1170
|
+
"description": "hunyuan-TurboS pv2.1.2 фиксирана версия, предтренировъчна база с увеличен брой токени; подобрени способности за разсъждение в математика/логика/код и др.; подобрено изживяване на китайски и английски, включително текстово творчество, разбиране на текст, въпроси и отговори, разговори и др."
|
1171
|
+
},
|
1172
|
+
"hunyuan-turbos-20250313": {
|
1173
|
+
"description": "Уеднаквяване на стила на математическите решения, засилване на многократните въпроси и отговори по математика. Оптимизация на стила на отговорите в текстовото творчество, премахване на AI привкус и добавяне на литературност."
|
1174
|
+
},
|
1175
|
+
"hunyuan-turbos-latest": {
|
1176
|
+
"description": "hunyuan-TurboS е последната версия на флагманския модел Hunyuan, с по-силни способности за разсъждение и по-добро потребителско изживяване."
|
1177
|
+
},
|
1112
1178
|
"hunyuan-vision": {
|
1113
1179
|
"description": "Най-новият мултимодален модел на HunYuan, поддържащ генериране на текстово съдържание от изображения и текстови входове."
|
1114
1180
|
},
|
@@ -23,6 +23,9 @@
|
|
23
23
|
"cloudflare": {
|
24
24
|
"description": "Работа с модели на машинно обучение, задвижвани от безсървърни GPU, в глобалната мрежа на Cloudflare."
|
25
25
|
},
|
26
|
+
"cohere": {
|
27
|
+
"description": "Cohere ви предлага най-съвременни многоезични модели, напреднали функции за търсене и AI работно пространство, проектирано специално за съвременните предприятия — всичко интегрирано в една сигурна платформа."
|
28
|
+
},
|
26
29
|
"deepseek": {
|
27
30
|
"description": "DeepSeek е компания, специализирана в изследвания и приложения на технологии за изкуствен интелект, чийто най-нов модел DeepSeek-V2.5 комбинира способности за общи диалози и обработка на код, постигайки значителни подобрения в съответствието с човешките предпочитания, писателските задачи и следването на инструкции."
|
28
31
|
},
|
@@ -101,6 +104,9 @@
|
|
101
104
|
"sambanova": {
|
102
105
|
"description": "SambaNova Cloud позволява на разработчиците лесно да използват най-добрите отворени модели и да се наслаждават на най-бързата скорост на извеждане."
|
103
106
|
},
|
107
|
+
"search1api": {
|
108
|
+
"description": "Search1API предоставя достъп до серията модели DeepSeek, които могат да се свързват в мрежа при нужда, включително стандартна и бърза версия, с поддръжка за избор на модели с различни параметри."
|
109
|
+
},
|
104
110
|
"sensenova": {
|
105
111
|
"description": "SenseNova, с мощната основа на SenseTime, предлага ефективни и лесни за използване услуги за големи модели с пълен стек."
|
106
112
|
},
|
@@ -41,7 +41,10 @@
|
|
41
41
|
"error": {
|
42
42
|
"desc": "Es tut uns leid, während des Initialisierungsprozesses der Pglite-Datenbank ist ein Fehler aufgetreten. Bitte klicken Sie auf die Schaltfläche, um es erneut zu versuchen. Wenn der Fehler nach mehreren Versuchen weiterhin auftritt, <1>reichen Sie bitte ein Problem ein</1>, und wir werden Ihnen umgehend bei der Untersuchung helfen.",
|
43
43
|
"detail": "Fehlerursache: [{{type}}] {{message}}. Einzelheiten sind wie folgt:",
|
44
|
+
"detailTitle": "Fehlerursache",
|
45
|
+
"report": "Problem melden",
|
44
46
|
"retry": "Erneut versuchen",
|
47
|
+
"selfSolve": "Selbsthilfe",
|
45
48
|
"title": "Datenbankinitialisierung fehlgeschlagen"
|
46
49
|
},
|
47
50
|
"initing": {
|
@@ -80,6 +83,54 @@
|
|
80
83
|
"button": "Jetzt verwenden",
|
81
84
|
"desc": "Jetzt verwenden",
|
82
85
|
"title": "PGlite-Datenbank ist bereit"
|
86
|
+
},
|
87
|
+
"solve": {
|
88
|
+
"backup": {
|
89
|
+
"backup": "Sicherung",
|
90
|
+
"backupSuccess": "Sicherung erfolgreich",
|
91
|
+
"desc": "Wichtige Daten aus der aktuellen Datenbank exportieren",
|
92
|
+
"export": "Alle Daten exportieren",
|
93
|
+
"exportDesc": "Die exportierten Daten werden im JSON-Format gespeichert und können für eine spätere Wiederherstellung oder Analyse verwendet werden.",
|
94
|
+
"reset": {
|
95
|
+
"alert": "Warnung",
|
96
|
+
"alertDesc": "Die folgenden Aktionen können zu Datenverlust führen. Bitte stellen Sie sicher, dass Sie wichtige Daten gesichert haben, bevor Sie fortfahren.",
|
97
|
+
"button": "Datenbank vollständig zurücksetzen (alle Daten löschen)",
|
98
|
+
"confirm": {
|
99
|
+
"desc": "Diese Aktion wird alle Daten löschen und ist nicht rückgängig zu machen. Möchten Sie fortfahren?",
|
100
|
+
"title": "Datenbank zurücksetzen bestätigen"
|
101
|
+
},
|
102
|
+
"desc": "Datenbank zurücksetzen, wenn eine Wiederherstellung nicht möglich ist",
|
103
|
+
"title": "Datenbank zurücksetzen"
|
104
|
+
},
|
105
|
+
"restore": "Wiederherstellen",
|
106
|
+
"restoreSuccess": "Wiederherstellung erfolgreich",
|
107
|
+
"title": "Datenbackup"
|
108
|
+
},
|
109
|
+
"diagnosis": {
|
110
|
+
"createdAt": "Erstellungszeit",
|
111
|
+
"migratedAt": "Migration abgeschlossen",
|
112
|
+
"sql": "Migration SQL",
|
113
|
+
"title": "Migrationsstatus"
|
114
|
+
},
|
115
|
+
"repair": {
|
116
|
+
"desc": "Manuelle Verwaltung des Migrationsstatus",
|
117
|
+
"runSQL": "Benutzerdefinierte Ausführung",
|
118
|
+
"sql": {
|
119
|
+
"clear": "Leeren",
|
120
|
+
"desc": "Benutzerdefinierte SQL-Anweisungen ausführen, um Datenbankprobleme zu beheben",
|
121
|
+
"markFinished": "Als abgeschlossen markieren",
|
122
|
+
"placeholder": "SQL-Anweisung eingeben...",
|
123
|
+
"result": "Ausführungsergebnis",
|
124
|
+
"run": "Ausführen",
|
125
|
+
"title": "SQL-Ausführungswerkzeug"
|
126
|
+
},
|
127
|
+
"title": "Migrationskontrolle"
|
128
|
+
},
|
129
|
+
"tabs": {
|
130
|
+
"backup": "Backup und Wiederherstellung",
|
131
|
+
"diagnosis": "Diagnose",
|
132
|
+
"repair": "Reparatur"
|
133
|
+
}
|
83
134
|
}
|
84
135
|
},
|
85
136
|
"close": "Schließen",
|
@@ -518,6 +518,18 @@
|
|
518
518
|
"baichuan/baichuan2-13b-chat": {
|
519
519
|
"description": "Baichuan-13B ist ein Open-Source-Sprachmodell mit 13 Milliarden Parametern, das von Baichuan Intelligence entwickelt wurde und in autorisierten chinesischen und englischen Benchmarks die besten Ergebnisse in seiner Größenordnung erzielt hat."
|
520
520
|
},
|
521
|
+
"c4ai-aya-expanse-32b": {
|
522
|
+
"description": "Aya Expanse ist ein leistungsstarkes 32B mehrsprachiges Modell, das darauf abzielt, die Leistung von einsprachigen Modellen durch innovative Ansätze wie Anweisungsoptimierung, Datenarbitrage, Präferenztraining und Modellfusion herauszufordern. Es unterstützt 23 Sprachen."
|
523
|
+
},
|
524
|
+
"c4ai-aya-expanse-8b": {
|
525
|
+
"description": "Aya Expanse ist ein leistungsstarkes 8B mehrsprachiges Modell, das darauf abzielt, die Leistung von einsprachigen Modellen durch innovative Ansätze wie Anweisungsoptimierung, Datenarbitrage, Präferenztraining und Modellfusion herauszufordern. Es unterstützt 23 Sprachen."
|
526
|
+
},
|
527
|
+
"c4ai-aya-vision-32b": {
|
528
|
+
"description": "Aya Vision ist ein hochmodernes multimodales Modell, das in mehreren wichtigen Benchmarks für Sprache, Text und Bild hervorragende Leistungen zeigt. Diese 32B-Version konzentriert sich auf die fortschrittlichste mehrsprachige Leistung und unterstützt 23 Sprachen."
|
529
|
+
},
|
530
|
+
"c4ai-aya-vision-8b": {
|
531
|
+
"description": "Aya Vision ist ein hochmodernes multimodales Modell, das in mehreren wichtigen Benchmarks für Sprache, Text und Bild hervorragende Leistungen zeigt. Diese 8B-Version konzentriert sich auf niedrige Latenz und optimale Leistung."
|
532
|
+
},
|
521
533
|
"charglm-3": {
|
522
534
|
"description": "CharGLM-3 ist für Rollenspiele und emotionale Begleitung konzipiert und unterstützt extrem lange Mehrfachgedächtnisse und personalisierte Dialoge, mit breiter Anwendung."
|
523
535
|
},
|
@@ -602,12 +614,39 @@
|
|
602
614
|
"cohere-command-r-plus": {
|
603
615
|
"description": "Command R+ ist ein hochmodernes, RAG-optimiertes Modell, das für unternehmensgerechte Arbeitslasten konzipiert ist."
|
604
616
|
},
|
617
|
+
"command": {
|
618
|
+
"description": "Ein dialogbasiertes Modell, das Anweisungen folgt und in sprachlichen Aufgaben hohe Qualität und Zuverlässigkeit bietet. Im Vergleich zu unserem grundlegenden Generierungsmodell hat es eine längere Kontextlänge."
|
619
|
+
},
|
620
|
+
"command-a-03-2025": {
|
621
|
+
"description": "Command A ist unser bisher leistungsstärkstes Modell, das in der Nutzung von Werkzeugen, Agenten, Retrieval-Enhanced Generation (RAG) und mehrsprachigen Anwendungsszenarien hervorragende Leistungen zeigt. Command A hat eine Kontextlänge von 256K, benötigt nur zwei GPUs zum Betrieb und bietet im Vergleich zu Command R+ 08-2024 eine Steigerung der Durchsatzrate um 150 %."
|
622
|
+
},
|
623
|
+
"command-light": {
|
624
|
+
"description": "Eine kleinere, schnellere Version von Command, die fast ebenso leistungsstark ist, aber schneller arbeitet."
|
625
|
+
},
|
626
|
+
"command-light-nightly": {
|
627
|
+
"description": "Um die Zeitspanne zwischen den Hauptversionsveröffentlichungen zu verkürzen, haben wir eine nächtliche Version des Command Modells eingeführt. Für die command-light-Serie wird diese Version als command-light-nightly bezeichnet. Bitte beachten Sie, dass command-light-nightly die neueste, experimentellste und (möglicherweise) instabilste Version ist. Die nächtlichen Versionen werden regelmäßig aktualisiert, ohne vorherige Ankündigung, daher wird die Verwendung in Produktionsumgebungen nicht empfohlen."
|
628
|
+
},
|
629
|
+
"command-nightly": {
|
630
|
+
"description": "Um die Zeitspanne zwischen den Hauptversionsveröffentlichungen zu verkürzen, haben wir eine nächtliche Version des Command Modells eingeführt. Für die Command-Serie wird diese Version als command-cightly bezeichnet. Bitte beachten Sie, dass command-nightly die neueste, experimentellste und (möglicherweise) instabilste Version ist. Die nächtlichen Versionen werden regelmäßig aktualisiert, ohne vorherige Ankündigung, daher wird die Verwendung in Produktionsumgebungen nicht empfohlen."
|
631
|
+
},
|
605
632
|
"command-r": {
|
606
633
|
"description": "Command R ist ein LLM, das für Dialoge und Aufgaben mit langen Kontexten optimiert ist und sich besonders gut für dynamische Interaktionen und Wissensmanagement eignet."
|
607
634
|
},
|
635
|
+
"command-r-03-2024": {
|
636
|
+
"description": "Command R ist ein dialogbasiertes Modell, das Anweisungen folgt und in sprachlichen Aufgaben eine höhere Qualität und Zuverlässigkeit bietet. Im Vergleich zu früheren Modellen hat es eine längere Kontextlänge. Es kann für komplexe Workflows wie Codegenerierung, Retrieval-Enhanced Generation (RAG), Werkzeugnutzung und Agenten verwendet werden."
|
637
|
+
},
|
638
|
+
"command-r-08-2024": {
|
639
|
+
"description": "command-r-08-2024 ist die aktualisierte Version des Command R Modells, das im August 2024 veröffentlicht wurde."
|
640
|
+
},
|
608
641
|
"command-r-plus": {
|
609
642
|
"description": "Command R+ ist ein leistungsstarkes großes Sprachmodell, das speziell für reale Unternehmensszenarien und komplexe Anwendungen entwickelt wurde."
|
610
643
|
},
|
644
|
+
"command-r-plus-04-2024": {
|
645
|
+
"description": "Command R+ ist ein dialogbasiertes Modell, das Anweisungen folgt und in sprachlichen Aufgaben eine höhere Qualität und Zuverlässigkeit bietet. Im Vergleich zu früheren Modellen hat es eine längere Kontextlänge. Es eignet sich am besten für komplexe RAG-Workflows und mehrstufige Werkzeugnutzung."
|
646
|
+
},
|
647
|
+
"command-r7b-12-2024": {
|
648
|
+
"description": "command-r7b-12-2024 ist eine kompakte und effiziente aktualisierte Version, die im Dezember 2024 veröffentlicht wurde. Es zeigt hervorragende Leistungen in Aufgaben, die komplexes Denken und mehrstufige Verarbeitung erfordern, wie RAG, Werkzeugnutzung und Agenten."
|
649
|
+
},
|
611
650
|
"dall-e-2": {
|
612
651
|
"description": "Zweite Generation des DALL·E-Modells, unterstützt realistischere und genauere Bildgenerierung, mit einer Auflösung, die viermal so hoch ist wie die der ersten Generation."
|
613
652
|
},
|
@@ -668,12 +707,24 @@
|
|
668
707
|
"deepseek-r1": {
|
669
708
|
"description": "DeepSeek-R1 ist ein durch verstärkendes Lernen (RL) gesteuertes Inferenzmodell, das die Probleme der Wiederholbarkeit und Lesbarkeit im Modell löst. Vor dem RL führte DeepSeek-R1 Kaltstartdaten ein, um die Inferenzleistung weiter zu optimieren. Es zeigt in mathematischen, programmierbezogenen und Inferenzaufgaben eine vergleichbare Leistung zu OpenAI-o1 und verbessert durch sorgfältig gestaltete Trainingsmethoden die Gesamteffizienz."
|
670
709
|
},
|
710
|
+
"deepseek-r1-70b-fast-online": {
|
711
|
+
"description": "DeepSeek R1 70B Schnellversion, die Echtzeit-Online-Suche unterstützt und eine schnellere Reaktionszeit bei gleichbleibender Modellleistung bietet."
|
712
|
+
},
|
713
|
+
"deepseek-r1-70b-online": {
|
714
|
+
"description": "DeepSeek R1 70B Standardversion, die Echtzeit-Online-Suche unterstützt und sich für Dialoge und Textverarbeitungsaufgaben eignet, die aktuelle Informationen benötigen."
|
715
|
+
},
|
671
716
|
"deepseek-r1-distill-llama-70b": {
|
672
717
|
"description": "DeepSeek R1 – das größere und intelligentere Modell im DeepSeek-Paket – wurde in die Llama 70B-Architektur destilliert. Basierend auf Benchmark-Tests und menschlicher Bewertung ist dieses Modell intelligenter als das ursprüngliche Llama 70B, insbesondere bei Aufgaben, die mathematische und faktische Genauigkeit erfordern."
|
673
718
|
},
|
674
719
|
"deepseek-r1-distill-llama-8b": {
|
675
720
|
"description": "Das DeepSeek-R1-Distill Modell wurde durch Wissensdistillationstechniken entwickelt, indem Proben, die von DeepSeek-R1 generiert wurden, auf Qwen, Llama und andere Open-Source-Modelle feinabgestimmt wurden."
|
676
721
|
},
|
722
|
+
"deepseek-r1-distill-qianfan-llama-70b": {
|
723
|
+
"description": "Erstmals veröffentlicht am 14. Februar 2025, destilliert vom Qianfan-Modellteam auf Basis des Llama3_70B Modells (gebaut mit Meta Llama), wobei auch die Qianfan-Korpora in die Destillationsdaten aufgenommen wurden."
|
724
|
+
},
|
725
|
+
"deepseek-r1-distill-qianfan-llama-8b": {
|
726
|
+
"description": "Erstmals veröffentlicht am 14. Februar 2025, destilliert vom Qianfan-Modellteam auf Basis des Llama3_8B Modells (gebaut mit Meta Llama), wobei auch die Qianfan-Korpora in die Destillationsdaten aufgenommen wurden."
|
727
|
+
},
|
677
728
|
"deepseek-r1-distill-qwen-1.5b": {
|
678
729
|
"description": "Das DeepSeek-R1-Distill Modell wurde durch Wissensdistillationstechniken entwickelt, indem Proben, die von DeepSeek-R1 generiert wurden, auf Qwen, Llama und andere Open-Source-Modelle feinabgestimmt wurden."
|
679
730
|
},
|
@@ -686,6 +737,12 @@
|
|
686
737
|
"deepseek-r1-distill-qwen-7b": {
|
687
738
|
"description": "Das DeepSeek-R1-Distill Modell wurde durch Wissensdistillationstechniken entwickelt, indem Proben, die von DeepSeek-R1 generiert wurden, auf Qwen, Llama und andere Open-Source-Modelle feinabgestimmt wurden."
|
688
739
|
},
|
740
|
+
"deepseek-r1-fast-online": {
|
741
|
+
"description": "DeepSeek R1 Vollschnellversion, die Echtzeit-Online-Suche unterstützt und die leistungsstarken Fähigkeiten von 671B Parametern mit einer schnelleren Reaktionszeit kombiniert."
|
742
|
+
},
|
743
|
+
"deepseek-r1-online": {
|
744
|
+
"description": "DeepSeek R1 Vollversion mit 671B Parametern, die Echtzeit-Online-Suche unterstützt und über verbesserte Verständnis- und Generierungsfähigkeiten verfügt."
|
745
|
+
},
|
689
746
|
"deepseek-reasoner": {
|
690
747
|
"description": "Das von DeepSeek entwickelte Inferenzmodell. Bevor das Modell die endgültige Antwort ausgibt, gibt es zunächst eine Denkprozesskette aus, um die Genauigkeit der endgültigen Antwort zu erhöhen."
|
691
748
|
},
|
@@ -764,6 +821,9 @@
|
|
764
821
|
"ernie-4.0-turbo-8k-preview": {
|
765
822
|
"description": "Das von Baidu entwickelte Flaggschiff-Modell für große Sprachmodelle zeigt hervorragende Gesamtergebnisse und ist weit verbreitet in komplexen Aufgabenbereichen anwendbar; es unterstützt die automatische Anbindung an das Baidu-Suchplugin, um die Aktualität der Antwortinformationen zu gewährleisten. Im Vergleich zu ERNIE 4.0 bietet es eine bessere Leistung."
|
766
823
|
},
|
824
|
+
"ernie-4.5-8k-preview": {
|
825
|
+
"description": "Das ERNIE 4.5 Modell ist ein neu entwickeltes, natives multimodales Basis-Modell von Baidu, das durch die gemeinsame Modellierung mehrerer Modalitäten eine synergistische Optimierung erreicht und über hervorragende multimodale Verständnisfähigkeiten verfügt; es bietet verbesserte Sprachfähigkeiten, umfassende Verbesserungen in Verständnis, Generierung, Logik und Gedächtnis, sowie signifikante Verbesserungen in der Vermeidung von Halluzinationen, logischen Schlussfolgerungen und Programmierfähigkeiten."
|
826
|
+
},
|
767
827
|
"ernie-char-8k": {
|
768
828
|
"description": "Das von Baidu entwickelte große Sprachmodell für vertikale Szenarien eignet sich für Anwendungen wie NPCs in Spielen, Kundenservice-Dialoge und Rollenspiele, mit einem klareren und konsistenteren Charakterstil, einer stärkeren Befolgung von Anweisungen und besserer Inferenzleistung."
|
769
829
|
},
|
@@ -1097,9 +1157,6 @@
|
|
1097
1157
|
"hunyuan-turbo": {
|
1098
1158
|
"description": "Die Vorschauversion des neuen großen Sprachmodells von Hunyuan verwendet eine neuartige hybride Expertenmodellstruktur (MoE) und bietet im Vergleich zu Hunyuan-Pro eine schnellere Inferenz und bessere Leistung."
|
1099
1159
|
},
|
1100
|
-
"hunyuan-turbo-20241120": {
|
1101
|
-
"description": "Hunyuan-turbo Version vom 20. November 2024, eine feste Version, die zwischen hunyuan-turbo und hunyuan-turbo-latest liegt."
|
1102
|
-
},
|
1103
1160
|
"hunyuan-turbo-20241223": {
|
1104
1161
|
"description": "Diese Version optimiert: Datenanweisungs-Skalierung, erhebliche Verbesserung der allgemeinen Generalisierungsfähigkeit des Modells; erhebliche Verbesserung der mathematischen, programmierbaren und logischen Denkfähigkeiten; Optimierung der Fähigkeiten im Textverständnis und der Wortverständnisfähigkeiten; Optimierung der Qualität der Inhaltserzeugung in der Texterstellung."
|
1105
1162
|
},
|
@@ -1109,6 +1166,15 @@
|
|
1109
1166
|
"hunyuan-turbo-vision": {
|
1110
1167
|
"description": "Das neue Flaggschiff-Modell der visuellen Sprache von Hunyuan, das eine brandneue Struktur des gemischten Expertenmodells (MoE) verwendet, bietet umfassende Verbesserungen in den Fähigkeiten zur grundlegenden Erkennung, Inhaltserstellung, Wissensfragen und Analyse sowie Schlussfolgerungen im Vergleich zum vorherigen Modell."
|
1111
1168
|
},
|
1169
|
+
"hunyuan-turbos-20250226": {
|
1170
|
+
"description": "hunyuan-TurboS pv2.1.2 ist eine feste Version mit aktualisierten Trainings-Tokens; verbesserte Denkfähigkeiten in Mathematik/Logik/Code; verbesserte allgemeine Erfahrung in Chinesisch und Englisch, einschließlich Textgenerierung, Textverständnis, Wissensfragen und Smalltalk."
|
1171
|
+
},
|
1172
|
+
"hunyuan-turbos-20250313": {
|
1173
|
+
"description": "Ein einheitlicher Stil für mathematische Problemlösungsprozesse, der die mehrstufige Beantwortung von mathematischen Fragen verbessert. Textgenerierung optimiert den Antwortstil, entfernt AI-typische Formulierungen und erhöht die sprachliche Eleganz."
|
1174
|
+
},
|
1175
|
+
"hunyuan-turbos-latest": {
|
1176
|
+
"description": "hunyuan-TurboS ist die neueste Version des Hunyuan-Flaggschiffmodells, das über verbesserte Denkfähigkeiten und ein besseres Nutzungserlebnis verfügt."
|
1177
|
+
},
|
1112
1178
|
"hunyuan-vision": {
|
1113
1179
|
"description": "Das neueste multimodale Modell von Hunyuan unterstützt die Eingabe von Bildern und Text zur Generierung von Textinhalten."
|
1114
1180
|
},
|
@@ -23,6 +23,9 @@
|
|
23
23
|
"cloudflare": {
|
24
24
|
"description": "Führen Sie von serverlosen GPUs betriebene Machine-Learning-Modelle im globalen Netzwerk von Cloudflare aus."
|
25
25
|
},
|
26
|
+
"cohere": {
|
27
|
+
"description": "Cohere bringt Ihnen die fortschrittlichsten mehrsprachigen Modelle, leistungsstarke Suchfunktionen und einen maßgeschneiderten KI-Arbeitsbereich für moderne Unternehmen – alles integriert in einer sicheren Plattform."
|
28
|
+
},
|
26
29
|
"deepseek": {
|
27
30
|
"description": "DeepSeek ist ein Unternehmen, das sich auf die Forschung und Anwendung von KI-Technologien spezialisiert hat. Ihr neuestes Modell, DeepSeek-V2.5, kombiniert allgemeine Dialog- und Codeverarbeitungsfähigkeiten und hat signifikante Fortschritte in den Bereichen menschliche Präferenzanpassung, Schreibaufgaben und Befehlsbefolgung erzielt."
|
28
31
|
},
|
@@ -101,6 +104,9 @@
|
|
101
104
|
"sambanova": {
|
102
105
|
"description": "SambaNova Cloud ermöglicht es Entwicklern, die besten Open-Source-Modelle einfach zu nutzen und von der schnellsten Inferenzgeschwindigkeit zu profitieren."
|
103
106
|
},
|
107
|
+
"search1api": {
|
108
|
+
"description": "Search1API bietet Zugriff auf die DeepSeek-Modellreihe, die bei Bedarf selbstständig online gehen kann, einschließlich der Standard- und Schnellversion, und unterstützt die Auswahl von Modellen in verschiedenen Parametergrößen."
|
109
|
+
},
|
104
110
|
"sensenova": {
|
105
111
|
"description": "SenseTime bietet mit der starken Basisunterstützung von SenseTimes großem Gerät effiziente und benutzerfreundliche Full-Stack-Modelldienste."
|
106
112
|
},
|
@@ -41,7 +41,10 @@
|
|
41
41
|
"error": {
|
42
42
|
"desc": "We apologize, an error occurred during the Pglite database initialization process. Please click the button to retry. If the error persists after multiple attempts, please <1>submit an issue</1>, and we will assist you as soon as possible.",
|
43
43
|
"detail": "Error reason: [{{type}}] {{message}}. Details are as follows:",
|
44
|
+
"detailTitle": "Error Reason",
|
45
|
+
"report": "Report Issue",
|
44
46
|
"retry": "Retry",
|
47
|
+
"selfSolve": "Self-Solve",
|
45
48
|
"title": "Database Initialization Failed"
|
46
49
|
},
|
47
50
|
"initing": {
|
@@ -80,6 +83,54 @@
|
|
80
83
|
"button": "Use Now",
|
81
84
|
"desc": "Ready to use",
|
82
85
|
"title": "PGlite Database is Ready"
|
86
|
+
},
|
87
|
+
"solve": {
|
88
|
+
"backup": {
|
89
|
+
"backup": "Backup",
|
90
|
+
"backupSuccess": "Backup Successful",
|
91
|
+
"desc": "Export key data from the current database",
|
92
|
+
"export": "Export All Data",
|
93
|
+
"exportDesc": "The exported data will be saved in JSON format, which can be used for future recovery or analysis.",
|
94
|
+
"reset": {
|
95
|
+
"alert": "Warning",
|
96
|
+
"alertDesc": "The following actions may result in data loss. Please ensure you have backed up important data before proceeding.",
|
97
|
+
"button": "Completely Reset Database (Delete All Data)",
|
98
|
+
"confirm": {
|
99
|
+
"desc": "This action will delete all data and cannot be undone. Do you confirm to continue?",
|
100
|
+
"title": "Confirm Database Reset"
|
101
|
+
},
|
102
|
+
"desc": "Reset the database in case of irreversible migration",
|
103
|
+
"title": "Database Reset"
|
104
|
+
},
|
105
|
+
"restore": "Restore",
|
106
|
+
"restoreSuccess": "Restore Successful",
|
107
|
+
"title": "Data Backup"
|
108
|
+
},
|
109
|
+
"diagnosis": {
|
110
|
+
"createdAt": "Creation Time",
|
111
|
+
"migratedAt": "Migration Completion Time",
|
112
|
+
"sql": "Migration SQL",
|
113
|
+
"title": "Migration Status"
|
114
|
+
},
|
115
|
+
"repair": {
|
116
|
+
"desc": "Manually manage migration status",
|
117
|
+
"runSQL": "Custom Execute",
|
118
|
+
"sql": {
|
119
|
+
"clear": "Clear",
|
120
|
+
"desc": "Execute custom SQL statements to fix database issues",
|
121
|
+
"markFinished": "Mark as Finished",
|
122
|
+
"placeholder": "Enter SQL statement...",
|
123
|
+
"result": "Execution Result",
|
124
|
+
"run": "Execute",
|
125
|
+
"title": "SQL Executor"
|
126
|
+
},
|
127
|
+
"title": "Migration Control"
|
128
|
+
},
|
129
|
+
"tabs": {
|
130
|
+
"backup": "Backup & Restore",
|
131
|
+
"diagnosis": "Diagnosis",
|
132
|
+
"repair": "Repair"
|
133
|
+
}
|
83
134
|
}
|
84
135
|
},
|
85
136
|
"close": "Close",
|
@@ -518,6 +518,18 @@
|
|
518
518
|
"baichuan/baichuan2-13b-chat": {
|
519
519
|
"description": "Baichuan-13B is an open-source, commercially usable large language model developed by Baichuan Intelligence, containing 13 billion parameters, achieving the best results in its size on authoritative Chinese and English benchmarks."
|
520
520
|
},
|
521
|
+
"c4ai-aya-expanse-32b": {
|
522
|
+
"description": "Aya Expanse is a high-performance 32B multilingual model designed to challenge the performance of single-language models through innovations in instruction tuning, data arbitrage, preference training, and model merging. It supports 23 languages."
|
523
|
+
},
|
524
|
+
"c4ai-aya-expanse-8b": {
|
525
|
+
"description": "Aya Expanse is a high-performance 8B multilingual model designed to challenge the performance of single-language models through innovations in instruction tuning, data arbitrage, preference training, and model merging. It supports 23 languages."
|
526
|
+
},
|
527
|
+
"c4ai-aya-vision-32b": {
|
528
|
+
"description": "Aya Vision is a state-of-the-art multimodal model that excels in multiple key benchmarks for language, text, and image capabilities. This 32 billion parameter version focuses on cutting-edge multilingual performance and supports 23 languages."
|
529
|
+
},
|
530
|
+
"c4ai-aya-vision-8b": {
|
531
|
+
"description": "Aya Vision is a state-of-the-art multimodal model that excels in multiple key benchmarks for language, text, and image capabilities. This 8 billion parameter version focuses on low latency and optimal performance."
|
532
|
+
},
|
521
533
|
"charglm-3": {
|
522
534
|
"description": "CharGLM-3 is designed for role-playing and emotional companionship, supporting ultra-long multi-turn memory and personalized dialogue, with wide applications."
|
523
535
|
},
|
@@ -602,12 +614,39 @@
|
|
602
614
|
"cohere-command-r-plus": {
|
603
615
|
"description": "Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads."
|
604
616
|
},
|
617
|
+
"command": {
|
618
|
+
"description": "An instruction-following dialogue model that delivers high quality and reliability in language tasks, with a longer context length compared to our base generation models."
|
619
|
+
},
|
620
|
+
"command-a-03-2025": {
|
621
|
+
"description": "Command A is our most powerful model to date, excelling in tool usage, agent tasks, retrieval-augmented generation (RAG), and multilingual applications. Command A features a context length of 256K and can run on just two GPUs, achieving a 150% increase in throughput compared to Command R+ 08-2024."
|
622
|
+
},
|
623
|
+
"command-light": {
|
624
|
+
"description": "A smaller, faster version of Command that is nearly as powerful but operates at a higher speed."
|
625
|
+
},
|
626
|
+
"command-light-nightly": {
|
627
|
+
"description": "To shorten the time interval between major version releases, we have launched nightly versions of the Command model. For the command-light series, this version is called command-light-nightly. Please note that command-light-nightly is the latest, most experimental, and (potentially) unstable version. Nightly versions are updated regularly without prior notice, so they are not recommended for production use."
|
628
|
+
},
|
629
|
+
"command-nightly": {
|
630
|
+
"description": "To shorten the time interval between major version releases, we have launched nightly versions of the Command model. For the Command series, this version is called command-cightly. Please note that command-nightly is the latest, most experimental, and (potentially) unstable version. Nightly versions are updated regularly without prior notice, so they are not recommended for production use."
|
631
|
+
},
|
605
632
|
"command-r": {
|
606
633
|
"description": "Command R is an LLM optimized for dialogue and long context tasks, particularly suitable for dynamic interactions and knowledge management."
|
607
634
|
},
|
635
|
+
"command-r-03-2024": {
|
636
|
+
"description": "Command R is an instruction-following dialogue model that provides higher quality and reliability in language tasks, with a longer context length than previous models. It can be used for complex workflows such as code generation, retrieval-augmented generation (RAG), tool usage, and agent tasks."
|
637
|
+
},
|
638
|
+
"command-r-08-2024": {
|
639
|
+
"description": "command-r-08-2024 is an updated version of the Command R model, released in August 2024."
|
640
|
+
},
|
608
641
|
"command-r-plus": {
|
609
642
|
"description": "Command R+ is a high-performance large language model designed for real enterprise scenarios and complex applications."
|
610
643
|
},
|
644
|
+
"command-r-plus-04-2024": {
|
645
|
+
"description": "Command R+ is an instruction-following dialogue model that delivers higher quality and reliability in language tasks, with a longer context length than previous models. It is best suited for complex RAG workflows and multi-step tool usage."
|
646
|
+
},
|
647
|
+
"command-r7b-12-2024": {
|
648
|
+
"description": "command-r7b-12-2024 is a compact and efficient updated version, released in December 2024. It excels in tasks requiring complex reasoning and multi-step processing, such as RAG, tool usage, and agent tasks."
|
649
|
+
},
|
611
650
|
"dall-e-2": {
|
612
651
|
"description": "The second generation DALL·E model, supporting more realistic and accurate image generation, with a resolution four times that of the first generation."
|
613
652
|
},
|
@@ -668,12 +707,24 @@
|
|
668
707
|
"deepseek-r1": {
|
669
708
|
"description": "DeepSeek-R1 is a reinforcement learning (RL) driven inference model that addresses issues of repetitiveness and readability within the model. Prior to RL, DeepSeek-R1 introduced cold start data to further optimize inference performance. It performs comparably to OpenAI-o1 in mathematical, coding, and reasoning tasks, and enhances overall effectiveness through meticulously designed training methods."
|
670
709
|
},
|
710
|
+
"deepseek-r1-70b-fast-online": {
|
711
|
+
"description": "DeepSeek R1 70B fast version, supporting real-time online search, providing faster response times while maintaining model performance."
|
712
|
+
},
|
713
|
+
"deepseek-r1-70b-online": {
|
714
|
+
"description": "DeepSeek R1 70B standard version, supporting real-time online search, suitable for dialogue and text processing tasks that require the latest information."
|
715
|
+
},
|
671
716
|
"deepseek-r1-distill-llama-70b": {
|
672
717
|
"description": "DeepSeek R1—the larger and smarter model in the DeepSeek suite—has been distilled into the Llama 70B architecture. Based on benchmark tests and human evaluations, this model is smarter than the original Llama 70B, especially excelling in tasks requiring mathematical and factual accuracy."
|
673
718
|
},
|
674
719
|
"deepseek-r1-distill-llama-8b": {
|
675
720
|
"description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
|
676
721
|
},
|
722
|
+
"deepseek-r1-distill-qianfan-llama-70b": {
|
723
|
+
"description": "First released on February 14, 2025, distilled by the Qianfan model development team using Llama3_70B as the base model (Built with Meta Llama), with Qianfan's corpus also added to the distilled data."
|
724
|
+
},
|
725
|
+
"deepseek-r1-distill-qianfan-llama-8b": {
|
726
|
+
"description": "First released on February 14, 2025, distilled by the Qianfan model development team using Llama3_8B as the base model (Built with Meta Llama), with Qianfan's corpus also added to the distilled data."
|
727
|
+
},
|
677
728
|
"deepseek-r1-distill-qwen-1.5b": {
|
678
729
|
"description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
|
679
730
|
},
|
@@ -686,6 +737,12 @@
|
|
686
737
|
"deepseek-r1-distill-qwen-7b": {
|
687
738
|
"description": "The DeepSeek-R1-Distill series models are fine-tuned versions of samples generated by DeepSeek-R1, using knowledge distillation techniques on open-source models like Qwen and Llama."
|
688
739
|
},
|
740
|
+
"deepseek-r1-fast-online": {
|
741
|
+
"description": "DeepSeek R1 full fast version, supporting real-time online search, combining the powerful capabilities of 671B parameters with faster response times."
|
742
|
+
},
|
743
|
+
"deepseek-r1-online": {
|
744
|
+
"description": "DeepSeek R1 full version, with 671B parameters, supporting real-time online search, offering enhanced understanding and generation capabilities."
|
745
|
+
},
|
689
746
|
"deepseek-reasoner": {
|
690
747
|
"description": "The reasoning model launched by DeepSeek. Before outputting the final answer, the model first provides a chain of thought to enhance the accuracy of the final response."
|
691
748
|
},
|
@@ -764,6 +821,9 @@
|
|
764
821
|
"ernie-4.0-turbo-8k-preview": {
|
765
822
|
"description": "Baidu's flagship ultra-large-scale language model, demonstrating outstanding overall performance, widely applicable to complex task scenarios across various fields; it supports automatic integration with Baidu search plugins to ensure the timeliness of Q&A information. It performs better than ERNIE 4.0 in terms of performance."
|
766
823
|
},
|
824
|
+
"ernie-4.5-8k-preview": {
|
825
|
+
"description": "ERNIE 4.5 is Baidu's self-developed next-generation native multimodal foundational model, achieving collaborative optimization through joint modeling of multiple modalities, with excellent multimodal understanding capabilities; it features enhanced language abilities, with significant improvements in understanding, generation, logic, and memory, as well as reduced hallucinations and improved logical reasoning and coding capabilities."
|
826
|
+
},
|
767
827
|
"ernie-char-8k": {
|
768
828
|
"description": "Baidu's vertical scene large language model, suitable for applications such as game NPCs, customer service dialogues, and role-playing conversations, with a more distinct and consistent character style, stronger instruction-following capabilities, and superior inference performance."
|
769
829
|
},
|
@@ -1097,9 +1157,6 @@
|
|
1097
1157
|
"hunyuan-turbo": {
|
1098
1158
|
"description": "The preview version of the next-generation Hunyuan large language model, featuring a brand-new mixed expert model (MoE) structure, which offers faster inference efficiency and stronger performance compared to Hunyuan Pro."
|
1099
1159
|
},
|
1100
|
-
"hunyuan-turbo-20241120": {
|
1101
|
-
"description": "Hunyuan-turbo fixed version as of November 20, 2024, a version that lies between hunyuan-turbo and hunyuan-turbo-latest."
|
1102
|
-
},
|
1103
1160
|
"hunyuan-turbo-20241223": {
|
1104
1161
|
"description": "This version optimizes: data instruction scaling, significantly enhancing the model's generalization capabilities; greatly improving mathematical, coding, and logical reasoning abilities; optimizing text understanding and word comprehension capabilities; enhancing the quality of content generation in text creation."
|
1105
1162
|
},
|
@@ -1109,6 +1166,15 @@
|
|
1109
1166
|
"hunyuan-turbo-vision": {
|
1110
1167
|
"description": "The next-generation flagship visual language model from Hunyuan, utilizing a new mixed expert model (MoE) structure, with comprehensive improvements in basic recognition, content creation, knowledge Q&A, and analytical reasoning capabilities compared to the previous generation model."
|
1111
1168
|
},
|
1169
|
+
"hunyuan-turbos-20250226": {
|
1170
|
+
"description": "hunyuan-TurboS pv2.1.2 fixed version with upgraded training tokens; enhanced reasoning capabilities in mathematics, logic, and coding; improved performance in both Chinese and English across text creation, comprehension, knowledge Q&A, and casual conversation."
|
1171
|
+
},
|
1172
|
+
"hunyuan-turbos-20250313": {
|
1173
|
+
"description": "Unifies the style of mathematical problem-solving steps and enhances multi-turn Q&A in mathematics. Optimizes the response style for text creation, removing AI-like characteristics and adding literary flair."
|
1174
|
+
},
|
1175
|
+
"hunyuan-turbos-latest": {
|
1176
|
+
"description": "The latest version of hunyuan-TurboS, the flagship model of Hunyuan, features enhanced reasoning capabilities and improved user experience."
|
1177
|
+
},
|
1112
1178
|
"hunyuan-vision": {
|
1113
1179
|
"description": "The latest multimodal model from Hunyuan, supporting image + text input to generate textual content."
|
1114
1180
|
},
|
@@ -23,6 +23,9 @@
|
|
23
23
|
"cloudflare": {
|
24
24
|
"description": "Run serverless GPU-powered machine learning models on Cloudflare's global network."
|
25
25
|
},
|
26
|
+
"cohere": {
|
27
|
+
"description": "Cohere brings you cutting-edge multilingual models, advanced retrieval capabilities, and an AI workspace tailored for modern enterprises—all integrated into a secure platform."
|
28
|
+
},
|
26
29
|
"deepseek": {
|
27
30
|
"description": "DeepSeek is a company focused on AI technology research and application, with its latest model DeepSeek-V2.5 integrating general dialogue and code processing capabilities, achieving significant improvements in human preference alignment, writing tasks, and instruction following."
|
28
31
|
},
|
@@ -101,6 +104,9 @@
|
|
101
104
|
"sambanova": {
|
102
105
|
"description": "SambaNova Cloud allows developers to easily utilize the best open-source models and enjoy the fastest inference speeds."
|
103
106
|
},
|
107
|
+
"search1api": {
|
108
|
+
"description": "Search1API provides access to the DeepSeek series of models that can connect to the internet as needed, including standard and fast versions, supporting a variety of model sizes."
|
109
|
+
},
|
104
110
|
"sensenova": {
|
105
111
|
"description": "SenseNova, backed by SenseTime's robust infrastructure, offers efficient and user-friendly full-stack large model services."
|
106
112
|
},
|
@@ -145,8 +151,5 @@
|
|
145
151
|
},
|
146
152
|
"zhipu": {
|
147
153
|
"description": "Zhipu AI offers an open platform for multimodal and language models, supporting a wide range of AI application scenarios, including text processing, image understanding, and programming assistance."
|
148
|
-
},
|
149
|
-
"ppio": {
|
150
|
-
"description": "PPIO supports stable and cost-efficient open-source LLM APIs, such as DeepSeek, Llama, Qwen etc."
|
151
154
|
}
|
152
155
|
}
|