@lobehub/chat 1.36.33 → 1.36.35

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (87) hide show
  1. package/CHANGELOG.md +50 -0
  2. package/README.ja-JP.md +150 -115
  3. package/README.md +65 -30
  4. package/README.zh-CN.md +65 -30
  5. package/changelog/v1.json +18 -0
  6. package/locales/ar/common.json +27 -7
  7. package/locales/ar/models.json +0 -24
  8. package/locales/ar/providers.json +10 -4
  9. package/locales/bg-BG/common.json +27 -7
  10. package/locales/bg-BG/models.json +0 -24
  11. package/locales/bg-BG/providers.json +10 -4
  12. package/locales/de-DE/common.json +27 -7
  13. package/locales/de-DE/models.json +0 -24
  14. package/locales/de-DE/providers.json +10 -4
  15. package/locales/en-US/common.json +27 -7
  16. package/locales/en-US/models.json +0 -24
  17. package/locales/en-US/providers.json +10 -4
  18. package/locales/es-ES/common.json +27 -7
  19. package/locales/es-ES/models.json +0 -24
  20. package/locales/es-ES/providers.json +10 -4
  21. package/locales/fa-IR/common.json +27 -7
  22. package/locales/fa-IR/models.json +0 -24
  23. package/locales/fa-IR/providers.json +10 -4
  24. package/locales/fr-FR/common.json +27 -7
  25. package/locales/fr-FR/models.json +0 -24
  26. package/locales/fr-FR/providers.json +10 -4
  27. package/locales/it-IT/common.json +27 -7
  28. package/locales/it-IT/models.json +0 -24
  29. package/locales/it-IT/providers.json +10 -4
  30. package/locales/ja-JP/common.json +27 -7
  31. package/locales/ja-JP/models.json +0 -24
  32. package/locales/ja-JP/providers.json +10 -4
  33. package/locales/ko-KR/common.json +27 -7
  34. package/locales/ko-KR/models.json +0 -24
  35. package/locales/ko-KR/providers.json +10 -4
  36. package/locales/nl-NL/common.json +27 -7
  37. package/locales/nl-NL/models.json +0 -24
  38. package/locales/nl-NL/providers.json +10 -4
  39. package/locales/pl-PL/common.json +27 -7
  40. package/locales/pl-PL/models.json +0 -24
  41. package/locales/pl-PL/providers.json +10 -4
  42. package/locales/pt-BR/common.json +27 -7
  43. package/locales/pt-BR/models.json +0 -24
  44. package/locales/pt-BR/providers.json +10 -4
  45. package/locales/ru-RU/common.json +27 -7
  46. package/locales/ru-RU/models.json +0 -24
  47. package/locales/ru-RU/providers.json +10 -4
  48. package/locales/tr-TR/common.json +27 -7
  49. package/locales/tr-TR/models.json +0 -24
  50. package/locales/tr-TR/providers.json +10 -4
  51. package/locales/vi-VN/common.json +27 -7
  52. package/locales/vi-VN/models.json +0 -24
  53. package/locales/vi-VN/providers.json +10 -4
  54. package/locales/zh-CN/common.json +28 -8
  55. package/locales/zh-CN/models.json +4 -28
  56. package/locales/zh-CN/providers.json +10 -4
  57. package/locales/zh-TW/common.json +27 -7
  58. package/locales/zh-TW/models.json +0 -24
  59. package/locales/zh-TW/providers.json +10 -4
  60. package/package.json +1 -1
  61. package/scripts/readmeWorkflow/const.ts +22 -8
  62. package/scripts/readmeWorkflow/index.ts +2 -0
  63. package/scripts/readmeWorkflow/syncAgentIndex.ts +36 -28
  64. package/scripts/readmeWorkflow/syncPluginIndex.ts +28 -15
  65. package/scripts/readmeWorkflow/syncProviderIndex.ts +51 -0
  66. package/scripts/readmeWorkflow/utlis.ts +23 -12
  67. package/src/app/loading/Client/Content.tsx +38 -0
  68. package/src/app/loading/Client/Redirect.tsx +47 -0
  69. package/src/app/loading/Client/index.tsx +22 -0
  70. package/src/app/loading/{Content.tsx → Server/Content.tsx} +6 -2
  71. package/src/app/loading/{Redirect.tsx → Server/Redirect.tsx} +3 -12
  72. package/src/app/loading/Server/index.tsx +22 -0
  73. package/src/app/loading/index.tsx +4 -16
  74. package/src/app/loading/stage.ts +22 -0
  75. package/src/components/FullscreenLoading/index.tsx +9 -8
  76. package/src/components/InitProgress/index.tsx +42 -0
  77. package/src/config/modelProviders/ai21.ts +1 -0
  78. package/src/config/modelProviders/cloudflare.ts +1 -2
  79. package/src/config/modelProviders/higress.ts +2 -1
  80. package/src/config/modelProviders/sensenova.ts +6 -3
  81. package/src/features/FileViewer/index.tsx +1 -1
  82. package/src/features/MobileSwitchLoading/index.tsx +20 -7
  83. package/src/layout/AuthProvider/NextAuth/UserUpdater.tsx +3 -1
  84. package/src/locales/default/common.ts +29 -8
  85. package/src/locales/default/models.ts +1 -0
  86. package/src/locales/default/providers.ts +1 -0
  87. package/src/app/loading/type.ts +0 -6
@@ -23,20 +23,6 @@
23
23
  "4.0Ultra": {
24
24
  "description": "Spark4.0 Ultra е най-мощната версия в серията Starfire, която подобрява разбирането и обобщаването на текстовото съдържание, докато надгражда свързаните търсения. Това е всестранно решение за повишаване на производителността в офиса и точно отговаряне на нуждите, водещо в индустрията интелигентно решение."
25
25
  },
26
- "@cf/meta/llama-3-8b-instruct-awq": {},
27
- "@cf/openchat/openchat-3.5-0106": {},
28
- "@cf/qwen/qwen1.5-14b-chat-awq": {},
29
- "@hf/google/gemma-7b-it": {},
30
- "@hf/meta-llama/meta-llama-3-8b-instruct": {
31
- "description": "Поколение след поколение, Meta Llama 3 демонстрира най-съвременна производителност в широк спектър от индустриални оценки и предлага нови възможности, включително подобрено разсъждение."
32
- },
33
- "@hf/mistral/mistral-7b-instruct-v0.2": {},
34
- "@hf/nexusflow/starling-lm-7b-beta": {},
35
- "@hf/nousresearch/hermes-2-pro-mistral-7b": {},
36
- "@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
37
- "@hf/thebloke/neural-chat-7b-v3-1-awq": {},
38
- "@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
39
- "@hf/thebloke/zephyr-7b-beta-awq": {},
40
26
  "Baichuan2-Turbo": {
41
27
  "description": "Използва технологии за подобряване на търсенето, за да свърже голям модел с областни знания и знания от интернет. Поддържа качване на различни документи като PDF, Word и вход на уебсайтове, с бърз и цялостен достъп до информация, предоставяйки точни и професионални резултати."
42
28
  },
@@ -127,7 +113,6 @@
127
113
  "Nous-Hermes-2-Mixtral-8x7B-DPO": {
128
114
  "description": "Hermes 2 Mixtral 8x7B DPO е високо гъвкава многомоделна комбинация, предназначена да предостави изключителен креативен опит."
129
115
  },
130
- "NousResearch/Hermes-3-Llama-3.1-8B": {},
131
116
  "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
132
117
  "description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) е модел с висока точност за инструкции, подходящ за сложни изчисления."
133
118
  },
@@ -530,9 +515,6 @@
530
515
  "cohere-command-r-plus": {
531
516
  "description": "Command R+ е модел, оптимизиран за RAG, проектиран да се справя с натоварвания на ниво предприятие."
532
517
  },
533
- "command-light": {
534
- "description": ""
535
- },
536
518
  "command-r": {
537
519
  "description": "Command R е LLM, оптимизиран за диалогови и дълги контекстуални задачи, особено подходящ за динамично взаимодействие и управление на знания."
538
520
  },
@@ -794,9 +776,6 @@
794
776
  "hunyuan-functioncall": {
795
777
  "description": "Най-новият модел на HunYuan с MOE архитектура за извикване на функции, обучен с висококачествени данни за извикване на функции, с контекстен прозорец от 32K, водещ в множество измерения на оценъчните показатели."
796
778
  },
797
- "hunyuan-large": {
798
- "description": ""
799
- },
800
779
  "hunyuan-lite": {
801
780
  "description": "Актуализиран до MOE структура, контекстният прозорец е 256k, водещ в множество оценъчни набори в NLP, код, математика и индустрия, пред много от отворените модели."
802
781
  },
@@ -830,8 +809,6 @@
830
809
  "internlm2.5-latest": {
831
810
  "description": "Нашата най-нова серия модели с изключителни способности за извеждане, поддържаща контекстна дължина от 1M и по-силни способности за следване на инструкции и извикване на инструменти."
832
811
  },
833
- "jamba-1.5-large": {},
834
- "jamba-1.5-mini": {},
835
812
  "learnlm-1.5-pro-experimental": {
836
813
  "description": "LearnLM е експериментален езиков модел, специфичен за задачи, обучен да отговаря на принципите на научното обучение, способен да следва системни инструкции в учебни и обучителни сценарии, да действа като експертен ментор и др."
837
814
  },
@@ -1027,7 +1004,6 @@
1027
1004
  "meta.llama3-8b-instruct-v1:0": {
1028
1005
  "description": "Meta Llama 3 е отворен голям езиков модел (LLM), насочен към разработчици, изследователи и предприятия, предназначен да им помогне да изградят, експериментират и отговорно разширят своите идеи за генеративен ИИ. Като част от основната система на глобалната общност за иновации, той е особено подходящ за устройства с ограничени изчислителни ресурси и по-бързо време за обучение."
1029
1006
  },
1030
- "microsoft/Phi-3.5-mini-instruct": {},
1031
1007
  "microsoft/WizardLM-2-8x22B": {
1032
1008
  "description": "WizardLM 2 е езиков модел, предоставен от Microsoft AI, който показва особени способности в сложни разговори, многоезичност, разсъждения и интелигентни асистенти."
1033
1009
  },
@@ -1,5 +1,7 @@
1
1
  {
2
- "ai21": {},
2
+ "ai21": {
3
+ "description": "AI21 Labs изгражда основни модели и системи за изкуствен интелект за предприятия, ускорявайки приложението на генеративния изкуствен интелект в производството."
4
+ },
3
5
  "ai360": {
4
6
  "description": "360 AI е платформа за AI модели и услуги, предлагана от компания 360, предлагаща множество напреднали модели за обработка на естествен език, включително 360GPT2 Pro, 360GPT Pro, 360GPT Turbo и 360GPT Turbo Responsibility 8K. Тези модели комбинират голям брой параметри и мултимодални способности, широко използвани в текстово генериране, семантично разбиране, диалогови системи и генериране на код. Чрез гъвкава ценова стратегия, 360 AI отговаря на разнообразни потребителски нужди, поддържайки интеграция за разработчици и насърчавайки иновации и развитие на интелигентни приложения."
5
7
  },
@@ -15,7 +17,9 @@
15
17
  "bedrock": {
16
18
  "description": "Bedrock е услуга, предоставяна от Amazon AWS, фокусирана върху предоставянето на напреднали AI езикови и визуални модели за предприятия. Семейството на моделите включва серията Claude на Anthropic, серията Llama 3.1 на Meta и други, обхващащи разнообразие от опции от леки до високо производителни, поддържащи текстово генериране, диалог, обработка на изображения и много други задачи, подходящи за различни мащаби и нужди на бизнес приложения."
17
19
  },
18
- "cloudflare": {},
20
+ "cloudflare": {
21
+ "description": "Работа с модели на машинно обучение, задвижвани от безсървърни GPU, в глобалната мрежа на Cloudflare."
22
+ },
19
23
  "deepseek": {
20
24
  "description": "DeepSeek е компания, специализирана в изследвания и приложения на технологии за изкуствен интелект, чийто най-нов модел DeepSeek-V2.5 комбинира способности за общи диалози и обработка на код, постигайки значителни подобрения в съответствието с човешките предпочитания, писателските задачи и следването на инструкции."
21
25
  },
@@ -35,7 +39,7 @@
35
39
  "description": "Инженерният двигател LPU на Groq показва изключителни резултати в последните независими тестове на големи езикови модели (LLM), преосмисляйки стандартите за AI решения с невероятната си скорост и ефективност. Groq е представител на мигновен скорост на изводите, демонстрирайки добро представяне в облачни внедрения."
36
40
  },
37
41
  "higress": {
38
- "description": ""
42
+ "description": "Higress е облачно роден API шлюз, създаден в Alibaba, за да реши проблемите с Tengine reload, които вредят на дългосрочните връзки, и недостатъчните възможности за баланс на натоварването на gRPC/Dubbo."
39
43
  },
40
44
  "huggingface": {
41
45
  "description": "HuggingFace Inference API предлагава бърз и безплатен начин да изследвате хиляди модели, подходящи за различни задачи. Независимо дали проектирате прототип за ново приложение, или опитвате функционалността на машинното обучение, този API ви предоставя незабавен достъп до високопроизводителни модели в множество области."
@@ -73,7 +77,9 @@
73
77
  "qwen": {
74
78
  "description": "Qwen е самостоятелно разработен свръхголям езиков модел на Alibaba Cloud, с мощни способности за разбиране и генериране на естествен език. Може да отговаря на различни въпроси, да създава текстово съдържание, да изразява мнения и да пише код, играейки роля в множество области."
75
79
  },
76
- "sensenova": {},
80
+ "sensenova": {
81
+ "description": "SenseNova, с мощната основа на SenseTime, предлага ефективни и лесни за използване услуги за големи модели с пълен стек."
82
+ },
77
83
  "siliconcloud": {
78
84
  "description": "SiliconFlow се стреми да ускори AGI, за да бъде от полза за човечеството, повишавайки ефективността на мащабния AI чрез лесен за използване и икономически изгоден GenAI стек."
79
85
  },
@@ -10,10 +10,16 @@
10
10
  }
11
11
  },
12
12
  "appLoading": {
13
- "goToChat": "(4/4) Lade die Chat-Seite...",
14
- "initAuth": "(2/4) Authentifizierungsdienst wird initialisiert...",
15
- "initUser": "(3/4) Benutzerstatus wird initialisiert...",
16
- "initializing": "(1/4) Anwendung wird gestartet..."
13
+ "appInitializing": "Anwendung wird gestartet...",
14
+ "finished": "Datenbankinitialisierung abgeschlossen",
15
+ "goToChat": "Lade die Chat-Seite...",
16
+ "initAuth": "Initialisiere den Authentifizierungsdienst...",
17
+ "initUser": "Initialisiere den Benutzerstatus...",
18
+ "initializing": "Initialisiere die PGlite-Datenbank...",
19
+ "loadingDependencies": "Abhängigkeiten werden initialisiert...",
20
+ "loadingWasm": "WASM-Module werden geladen...",
21
+ "migrating": "Datenbankmigration wird durchgeführt...",
22
+ "ready": "Datenbank ist bereit"
17
23
  },
18
24
  "autoGenerate": "Automatisch generieren",
19
25
  "autoGenerateTooltip": "Assistentenbeschreibung automatisch auf Basis von Vorschlägen vervollständigen",
@@ -38,14 +44,28 @@
38
44
  "error": "Ein Fehler ist aufgetreten, bitte versuchen Sie es erneut",
39
45
  "idle": "Warte auf die Initialisierung...",
40
46
  "initializing": "Wird initialisiert...",
41
- "loadingDependencies": "Lade Abhängigkeiten ({ {progress}}%)...",
42
- "loadingWasmModule": "Lade WASM-Modul ({ {progress}}%)...",
43
- "migrating": "Daten werden migriert...",
47
+ "loadingDependencies": "Abhängigkeiten werden geladen...",
48
+ "loadingWasmModule": "WASM-Modul wird geladen...",
49
+ "migrating": "Datenbankmigration wird durchgeführt...",
44
50
  "ready": "Datenbank ist bereit"
45
51
  },
46
52
  "modal": {
47
53
  "desc": "Aktivieren Sie die PGlite-Clientdatenbank, um Chatdaten in Ihrem Browser dauerhaft zu speichern und erweiterte Funktionen wie Wissensdatenbanken zu nutzen.",
48
54
  "enable": "Jetzt aktivieren",
55
+ "features": {
56
+ "knowledgeBase": {
57
+ "desc": "Bauen Sie Ihre persönliche Wissensdatenbank auf und führen Sie mühelos Gespräche mit Ihrem Assistenten (demnächst verfügbar)",
58
+ "title": "Unterstützung für Wissensdatenbankgespräche, aktivieren Sie Ihr zweites Gehirn"
59
+ },
60
+ "localFirst": {
61
+ "desc": "Chat-Daten werden vollständig im Browser gespeichert, Ihre Daten sind immer in Ihrer Kontrolle.",
62
+ "title": "Lokale Priorität, Datenschutz an erster Stelle"
63
+ },
64
+ "pglite": {
65
+ "desc": "Basierend auf PGlite, unterstützt nativ AI Native fortgeschrittene Funktionen (Vektorsuche)",
66
+ "title": "Neue Generation der Client-Speicherarchitektur"
67
+ }
68
+ },
49
69
  "init": {
50
70
  "desc": "Die Datenbank wird initialisiert, je nach Netzwerkbedingungen kann dies 5 bis 30 Sekunden dauern.",
51
71
  "title": "Initialisiere PGlite-Datenbank"
@@ -23,20 +23,6 @@
23
23
  "4.0Ultra": {
24
24
  "description": "Spark4.0 Ultra ist die leistungsstärkste Version der Spark-Großmodellreihe, die die Online-Suchverbindung aktualisiert und die Fähigkeit zur Textverständnis und -zusammenfassung verbessert. Es ist eine umfassende Lösung zur Steigerung der Büroproduktivität und zur genauen Reaktion auf Anforderungen und ein führendes intelligentes Produkt in der Branche."
25
25
  },
26
- "@cf/meta/llama-3-8b-instruct-awq": {},
27
- "@cf/openchat/openchat-3.5-0106": {},
28
- "@cf/qwen/qwen1.5-14b-chat-awq": {},
29
- "@hf/google/gemma-7b-it": {},
30
- "@hf/meta-llama/meta-llama-3-8b-instruct": {
31
- "description": "Generation über Generation zeigt Meta Llama 3 eine Spitzenleistung in einer Vielzahl von Branchenbenchmarks und bietet neue Fähigkeiten, einschließlich verbesserter Argumentation."
32
- },
33
- "@hf/mistral/mistral-7b-instruct-v0.2": {},
34
- "@hf/nexusflow/starling-lm-7b-beta": {},
35
- "@hf/nousresearch/hermes-2-pro-mistral-7b": {},
36
- "@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
37
- "@hf/thebloke/neural-chat-7b-v3-1-awq": {},
38
- "@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
39
- "@hf/thebloke/zephyr-7b-beta-awq": {},
40
26
  "Baichuan2-Turbo": {
41
27
  "description": "Verwendet Suchverbesserungstechnologie, um eine umfassende Verknüpfung zwischen großen Modellen und Fachwissen sowie Wissen aus dem gesamten Internet zu ermöglichen. Unterstützt das Hochladen von Dokumenten wie PDF, Word und die Eingabe von URLs, um Informationen zeitnah und umfassend zu erhalten, mit genauen und professionellen Ergebnissen."
42
28
  },
@@ -127,7 +113,6 @@
127
113
  "Nous-Hermes-2-Mixtral-8x7B-DPO": {
128
114
  "description": "Hermes 2 Mixtral 8x7B DPO ist eine hochflexible Multi-Modell-Kombination, die darauf abzielt, außergewöhnliche kreative Erlebnisse zu bieten."
129
115
  },
130
- "NousResearch/Hermes-3-Llama-3.1-8B": {},
131
116
  "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
132
117
  "description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) ist ein hochpräzises Anweisungsmodell, das für komplexe Berechnungen geeignet ist."
133
118
  },
@@ -530,9 +515,6 @@
530
515
  "cohere-command-r-plus": {
531
516
  "description": "Command R+ ist ein hochmodernes, RAG-optimiertes Modell, das für unternehmensgerechte Arbeitslasten konzipiert ist."
532
517
  },
533
- "command-light": {
534
- "description": ""
535
- },
536
518
  "command-r": {
537
519
  "description": "Command R ist ein LLM, das für Dialoge und Aufgaben mit langen Kontexten optimiert ist und sich besonders gut für dynamische Interaktionen und Wissensmanagement eignet."
538
520
  },
@@ -794,9 +776,6 @@
794
776
  "hunyuan-functioncall": {
795
777
  "description": "Das neueste MOE-Architektur-FunctionCall-Modell von Hunyuan, das mit hochwertigen FunctionCall-Daten trainiert wurde, hat ein Kontextfenster von 32K und führt in mehreren Bewertungsmetriken."
796
778
  },
797
- "hunyuan-large": {
798
- "description": ""
799
- },
800
779
  "hunyuan-lite": {
801
780
  "description": "Aufgerüstet auf eine MOE-Struktur mit einem Kontextfenster von 256k, führt es in mehreren Bewertungssets in NLP, Code, Mathematik und Industrie zahlreiche Open-Source-Modelle an."
802
781
  },
@@ -830,8 +809,6 @@
830
809
  "internlm2.5-latest": {
831
810
  "description": "Unsere neueste Modellreihe mit herausragender Schlussfolgerungsleistung, die eine Kontextlänge von 1M unterstützt und über verbesserte Anweisungsbefolgung und Toolaufrufmöglichkeiten verfügt."
832
811
  },
833
- "jamba-1.5-large": {},
834
- "jamba-1.5-mini": {},
835
812
  "learnlm-1.5-pro-experimental": {
836
813
  "description": "LearnLM ist ein experimentelles, aufgabenorientiertes Sprachmodell, das darauf trainiert wurde, den Prinzipien der Lernwissenschaft zu entsprechen und in Lehr- und Lernszenarien systematische Anweisungen zu befolgen, als Expertenmentor zu fungieren usw."
837
814
  },
@@ -1027,7 +1004,6 @@
1027
1004
  "meta.llama3-8b-instruct-v1:0": {
1028
1005
  "description": "Meta Llama 3 ist ein offenes großes Sprachmodell (LLM), das sich an Entwickler, Forscher und Unternehmen richtet und ihnen hilft, ihre Ideen für generative KI zu entwickeln, zu experimentieren und verantwortungsbewusst zu skalieren. Als Teil eines globalen Innovationssystems ist es besonders geeignet für Umgebungen mit begrenzter Rechenleistung und Ressourcen, für Edge-Geräte und schnellere Trainingszeiten."
1029
1006
  },
1030
- "microsoft/Phi-3.5-mini-instruct": {},
1031
1007
  "microsoft/WizardLM-2-8x22B": {
1032
1008
  "description": "WizardLM 2 ist ein Sprachmodell von Microsoft AI, das in komplexen Dialogen, Mehrsprachigkeit, Inferenz und intelligenten Assistenten besonders gut abschneidet."
1033
1009
  },
@@ -1,5 +1,7 @@
1
1
  {
2
- "ai21": {},
2
+ "ai21": {
3
+ "description": "AI21 Labs entwickelt Basis-Modelle und KI-Systeme für Unternehmen und beschleunigt die Anwendung generativer KI in der Produktion."
4
+ },
3
5
  "ai360": {
4
6
  "description": "360 AI ist die von der 360 Company eingeführte Plattform für KI-Modelle und -Dienste, die eine Vielzahl fortschrittlicher Modelle zur Verarbeitung natürlicher Sprache anbietet, darunter 360GPT2 Pro, 360GPT Pro, 360GPT Turbo und 360GPT Turbo Responsibility 8K. Diese Modelle kombinieren große Parameter mit multimodalen Fähigkeiten und finden breite Anwendung in den Bereichen Textgenerierung, semantisches Verständnis, Dialogsysteme und Codegenerierung. Durch flexible Preisstrategien erfüllt 360 AI die vielfältigen Bedürfnisse der Nutzer, unterstützt Entwickler bei der Integration und fördert die Innovation und Entwicklung intelligenter Anwendungen."
5
7
  },
@@ -15,7 +17,9 @@
15
17
  "bedrock": {
16
18
  "description": "Bedrock ist ein Service von Amazon AWS, der sich darauf konzentriert, Unternehmen fortschrittliche KI-Sprach- und visuelle Modelle bereitzustellen. Die Modellfamilie umfasst die Claude-Serie von Anthropic, die Llama 3.1-Serie von Meta und mehr, und bietet eine Vielzahl von Optionen von leichtgewichtig bis hochleistungsfähig, die Textgenerierung, Dialoge, Bildverarbeitung und andere Aufgaben unterstützen und für Unternehmensanwendungen unterschiedlicher Größen und Anforderungen geeignet sind."
17
19
  },
18
- "cloudflare": {},
20
+ "cloudflare": {
21
+ "description": "Führen Sie von serverlosen GPUs betriebene Machine-Learning-Modelle im globalen Netzwerk von Cloudflare aus."
22
+ },
19
23
  "deepseek": {
20
24
  "description": "DeepSeek ist ein Unternehmen, das sich auf die Forschung und Anwendung von KI-Technologien spezialisiert hat. Ihr neuestes Modell, DeepSeek-V2.5, kombiniert allgemeine Dialog- und Codeverarbeitungsfähigkeiten und hat signifikante Fortschritte in den Bereichen menschliche Präferenzanpassung, Schreibaufgaben und Befehlsbefolgung erzielt."
21
25
  },
@@ -35,7 +39,7 @@
35
39
  "description": "Der LPU-Inferenz-Engine von Groq hat in den neuesten unabhängigen Benchmark-Tests für große Sprachmodelle (LLM) hervorragende Leistungen gezeigt und definiert mit seiner erstaunlichen Geschwindigkeit und Effizienz die Standards für KI-Lösungen neu. Groq ist ein Beispiel für sofortige Inferenzgeschwindigkeit und zeigt in cloudbasierten Bereitstellungen eine gute Leistung."
36
40
  },
37
41
  "higress": {
38
- "description": ""
42
+ "description": "Higress ist ein cloud-natives API-Gateway, das intern bei Alibaba entwickelt wurde, um die Probleme von Tengine Reload bei langanhaltenden Verbindungen zu lösen und die unzureichenden Lastverteilungsmöglichkeiten von gRPC/Dubbo zu verbessern."
39
43
  },
40
44
  "huggingface": {
41
45
  "description": "Die HuggingFace Inference API bietet eine schnelle und kostenlose Möglichkeit, Tausende von Modellen für verschiedene Aufgaben zu erkunden. Egal, ob Sie Prototypen für neue Anwendungen erstellen oder die Funktionen des maschinellen Lernens ausprobieren, diese API ermöglicht Ihnen den sofortigen Zugriff auf leistungsstarke Modelle aus verschiedenen Bereichen."
@@ -73,7 +77,9 @@
73
77
  "qwen": {
74
78
  "description": "Tongyi Qianwen ist ein von Alibaba Cloud selbst entwickeltes, groß angelegtes Sprachmodell mit starken Fähigkeiten zur Verarbeitung und Generierung natürlicher Sprache. Es kann eine Vielzahl von Fragen beantworten, Texte erstellen, Meinungen äußern und Code schreiben und spielt in mehreren Bereichen eine Rolle."
75
79
  },
76
- "sensenova": {},
80
+ "sensenova": {
81
+ "description": "SenseTime bietet mit der starken Basisunterstützung von SenseTimes großem Gerät effiziente und benutzerfreundliche Full-Stack-Modelldienste."
82
+ },
77
83
  "siliconcloud": {
78
84
  "description": "SiliconFlow hat sich zum Ziel gesetzt, AGI zu beschleunigen, um der Menschheit zu dienen, und die Effizienz großangelegter KI durch eine benutzerfreundliche und kostengünstige GenAI-Stack zu steigern."
79
85
  },
@@ -10,10 +10,16 @@
10
10
  }
11
11
  },
12
12
  "appLoading": {
13
- "goToChat": "(4/4) Loading chat page...",
14
- "initAuth": "(2/4) Initializing authentication service...",
15
- "initUser": "(3/4) Initializing user status...",
16
- "initializing": "(1/4) Starting up application..."
13
+ "appInitializing": "Application is starting...",
14
+ "finished": "Database initialization completed",
15
+ "goToChat": "Loading chat page...",
16
+ "initAuth": "Initializing authentication service...",
17
+ "initUser": "Initializing user status...",
18
+ "initializing": "Initializing PGlite database...",
19
+ "loadingDependencies": "Initializing dependencies...",
20
+ "loadingWasm": "Loading WASM module...",
21
+ "migrating": "Performing database migration...",
22
+ "ready": "Database is ready"
17
23
  },
18
24
  "autoGenerate": "Auto Generate",
19
25
  "autoGenerateTooltip": "Auto-generate assistant description based on prompts",
@@ -38,14 +44,28 @@
38
44
  "error": "An error occurred, please try again",
39
45
  "idle": "Waiting for initialization...",
40
46
  "initializing": "Initializing...",
41
- "loadingDependencies": "Loading dependencies ({{progress}}%)...",
42
- "loadingWasmModule": "Loading WASM module ({{progress}}%)...",
43
- "migrating": "Migrating data...",
47
+ "loadingDependencies": "Loading dependencies...",
48
+ "loadingWasmModule": "Loading WASM module...",
49
+ "migrating": "Performing database migration...",
44
50
  "ready": "Database is ready"
45
51
  },
46
52
  "modal": {
47
53
  "desc": "Enable the PGlite client database to persistently store chat data in your browser and use advanced features like knowledge base.",
48
54
  "enable": "Enable Now",
55
+ "features": {
56
+ "knowledgeBase": {
57
+ "desc": "Build your personal knowledge base and easily start conversations with your assistant (coming soon)",
58
+ "title": "Support for knowledge base conversations, unlock your second brain"
59
+ },
60
+ "localFirst": {
61
+ "desc": "Chat data is stored entirely in the browser, keeping your data always under your control.",
62
+ "title": "Local first, privacy first"
63
+ },
64
+ "pglite": {
65
+ "desc": "Built on PGlite, natively supports AI Native advanced features (vector retrieval)",
66
+ "title": "Next-generation client storage architecture"
67
+ }
68
+ },
49
69
  "init": {
50
70
  "desc": "Initializing the database, which may take 5 to 30 seconds depending on network conditions.",
51
71
  "title": "Initializing PGlite Database"
@@ -23,20 +23,6 @@
23
23
  "4.0Ultra": {
24
24
  "description": "Spark4.0 Ultra is the most powerful version in the Spark large model series, enhancing text content understanding and summarization capabilities while upgrading online search links. It is a comprehensive solution for improving office productivity and accurately responding to demands, leading the industry as an intelligent product."
25
25
  },
26
- "@cf/meta/llama-3-8b-instruct-awq": {},
27
- "@cf/openchat/openchat-3.5-0106": {},
28
- "@cf/qwen/qwen1.5-14b-chat-awq": {},
29
- "@hf/google/gemma-7b-it": {},
30
- "@hf/meta-llama/meta-llama-3-8b-instruct": {
31
- "description": "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning."
32
- },
33
- "@hf/mistral/mistral-7b-instruct-v0.2": {},
34
- "@hf/nexusflow/starling-lm-7b-beta": {},
35
- "@hf/nousresearch/hermes-2-pro-mistral-7b": {},
36
- "@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
37
- "@hf/thebloke/neural-chat-7b-v3-1-awq": {},
38
- "@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
39
- "@hf/thebloke/zephyr-7b-beta-awq": {},
40
26
  "Baichuan2-Turbo": {
41
27
  "description": "Utilizes search enhancement technology to achieve comprehensive links between large models and domain knowledge, as well as knowledge from the entire web. Supports uploads of various documents such as PDF and Word, and URL input, providing timely and comprehensive information retrieval with accurate and professional output."
42
28
  },
@@ -127,7 +113,6 @@
127
113
  "Nous-Hermes-2-Mixtral-8x7B-DPO": {
128
114
  "description": "Hermes 2 Mixtral 8x7B DPO is a highly flexible multi-model fusion designed to provide an exceptional creative experience."
129
115
  },
130
- "NousResearch/Hermes-3-Llama-3.1-8B": {},
131
116
  "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
132
117
  "description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) is a high-precision instruction model suitable for complex computations."
133
118
  },
@@ -530,9 +515,6 @@
530
515
  "cohere-command-r-plus": {
531
516
  "description": "Command R+ is a state-of-the-art RAG-optimized model designed to tackle enterprise-grade workloads."
532
517
  },
533
- "command-light": {
534
- "description": ""
535
- },
536
518
  "command-r": {
537
519
  "description": "Command R is an LLM optimized for dialogue and long context tasks, particularly suitable for dynamic interactions and knowledge management."
538
520
  },
@@ -794,9 +776,6 @@
794
776
  "hunyuan-functioncall": {
795
777
  "description": "The latest MOE architecture FunctionCall model from Hunyuan, trained on high-quality FunctionCall data, with a context window of 32K, leading in multiple dimensions of evaluation metrics."
796
778
  },
797
- "hunyuan-large": {
798
- "description": ""
799
- },
800
779
  "hunyuan-lite": {
801
780
  "description": "Upgraded to a MOE structure with a context window of 256k, leading many open-source models in various NLP, coding, mathematics, and industry benchmarks."
802
781
  },
@@ -830,8 +809,6 @@
830
809
  "internlm2.5-latest": {
831
810
  "description": "Our latest model series, featuring exceptional reasoning performance, supporting a context length of 1M, and enhanced instruction following and tool invocation capabilities."
832
811
  },
833
- "jamba-1.5-large": {},
834
- "jamba-1.5-mini": {},
835
812
  "learnlm-1.5-pro-experimental": {
836
813
  "description": "LearnLM is an experimental, task-specific language model trained to align with learning science principles, capable of following systematic instructions in teaching and learning scenarios, acting as an expert tutor, among other roles."
837
814
  },
@@ -1027,7 +1004,6 @@
1027
1004
  "meta.llama3-8b-instruct-v1:0": {
1028
1005
  "description": "Meta Llama 3 is an open large language model (LLM) aimed at developers, researchers, and enterprises, designed to help them build, experiment, and responsibly scale their generative AI ideas. As part of a foundational system for global community innovation, it is particularly suitable for those with limited computational power and resources, edge devices, and faster training times."
1029
1006
  },
1030
- "microsoft/Phi-3.5-mini-instruct": {},
1031
1007
  "microsoft/WizardLM-2-8x22B": {
1032
1008
  "description": "WizardLM 2 is a language model provided by Microsoft AI, excelling in complex dialogues, multilingual capabilities, reasoning, and intelligent assistant tasks."
1033
1009
  },
@@ -1,5 +1,7 @@
1
1
  {
2
- "ai21": {},
2
+ "ai21": {
3
+ "description": "AI21 Labs builds foundational models and AI systems for enterprises, accelerating the application of generative AI in production."
4
+ },
3
5
  "ai360": {
4
6
  "description": "360 AI is an AI model and service platform launched by 360 Company, offering various advanced natural language processing models, including 360GPT2 Pro, 360GPT Pro, 360GPT Turbo, and 360GPT Turbo Responsibility 8K. These models combine large-scale parameters and multimodal capabilities, widely applied in text generation, semantic understanding, dialogue systems, and code generation. With flexible pricing strategies, 360 AI meets diverse user needs, supports developer integration, and promotes the innovation and development of intelligent applications."
5
7
  },
@@ -15,7 +17,9 @@
15
17
  "bedrock": {
16
18
  "description": "Bedrock is a service provided by Amazon AWS, focusing on delivering advanced AI language and visual models for enterprises. Its model family includes Anthropic's Claude series, Meta's Llama 3.1 series, and more, offering a range of options from lightweight to high-performance, supporting tasks such as text generation, conversation, and image processing for businesses of varying scales and needs."
17
19
  },
18
- "cloudflare": {},
20
+ "cloudflare": {
21
+ "description": "Run serverless GPU-powered machine learning models on Cloudflare's global network."
22
+ },
19
23
  "deepseek": {
20
24
  "description": "DeepSeek is a company focused on AI technology research and application, with its latest model DeepSeek-V2.5 integrating general dialogue and code processing capabilities, achieving significant improvements in human preference alignment, writing tasks, and instruction following."
21
25
  },
@@ -35,7 +39,7 @@
35
39
  "description": "Groq's LPU inference engine has excelled in the latest independent large language model (LLM) benchmarks, redefining the standards for AI solutions with its remarkable speed and efficiency. Groq represents instant inference speed, demonstrating strong performance in cloud-based deployments."
36
40
  },
37
41
  "higress": {
38
- "description": ""
42
+ "description": "Higress is a cloud-native API gateway that was developed internally at Alibaba to address the issues of Tengine reload affecting long-lived connections and the insufficient load balancing capabilities for gRPC/Dubbo."
39
43
  },
40
44
  "huggingface": {
41
45
  "description": "The HuggingFace Inference API provides a fast and free way for you to explore thousands of models for various tasks. Whether you are prototyping for a new application or experimenting with the capabilities of machine learning, this API gives you instant access to high-performance models across multiple domains."
@@ -73,7 +77,9 @@
73
77
  "qwen": {
74
78
  "description": "Tongyi Qianwen is a large-scale language model independently developed by Alibaba Cloud, featuring strong natural language understanding and generation capabilities. It can answer various questions, create written content, express opinions, and write code, playing a role in multiple fields."
75
79
  },
76
- "sensenova": {},
80
+ "sensenova": {
81
+ "description": "SenseNova, backed by SenseTime's robust infrastructure, offers efficient and user-friendly full-stack large model services."
82
+ },
77
83
  "siliconcloud": {
78
84
  "description": "SiliconFlow is dedicated to accelerating AGI for the benefit of humanity, enhancing large-scale AI efficiency through an easy-to-use and cost-effective GenAI stack."
79
85
  },
@@ -10,10 +10,16 @@
10
10
  }
11
11
  },
12
12
  "appLoading": {
13
- "goToChat": "(4/4) Cargando la página de chat...",
14
- "initAuth": "(2/4) Inicializando el servicio de autenticación...",
15
- "initUser": "(3/4) Inicializando el estado del usuario...",
16
- "initializing": "(1/4) Iniciando la aplicación..."
13
+ "appInitializing": "Iniciando la aplicación...",
14
+ "finished": "Inicialización de la base de datos completada",
15
+ "goToChat": "Cargando la página de chat...",
16
+ "initAuth": "Inicializando el servicio de autenticación...",
17
+ "initUser": "Inicializando el estado del usuario...",
18
+ "initializing": "Inicializando la base de datos PGlite...",
19
+ "loadingDependencies": "Inicializando dependencias...",
20
+ "loadingWasm": "Cargando módulo WASM...",
21
+ "migrating": "Ejecutando migración de tablas de datos...",
22
+ "ready": "Base de datos lista"
17
23
  },
18
24
  "autoGenerate": "Generación automática",
19
25
  "autoGenerateTooltip": "Completar automáticamente la descripción del asistente basándose en las sugerencias",
@@ -38,14 +44,28 @@
38
44
  "error": "Ha ocurrido un error, por favor reintente",
39
45
  "idle": "Esperando la inicialización...",
40
46
  "initializing": "Inicializando...",
41
- "loadingDependencies": "Cargando dependencias ({{progress}}%)...",
42
- "loadingWasmModule": "Cargando módulo WASM ({{progress}}%)...",
43
- "migrating": "Migrando datos...",
47
+ "loadingDependencies": "Cargando dependencias...",
48
+ "loadingWasmModule": "Cargando módulo WASM...",
49
+ "migrating": "Ejecutando migración de tablas de datos...",
44
50
  "ready": "La base de datos está lista"
45
51
  },
46
52
  "modal": {
47
53
  "desc": "Habilite la base de datos del cliente PGlite para almacenar de forma persistente los datos del chat en su navegador y utilizar características avanzadas como la base de conocimientos.",
48
54
  "enable": "Habilitar ahora",
55
+ "features": {
56
+ "knowledgeBase": {
57
+ "desc": "Consolida tu base de conocimientos personal y comienza conversaciones sobre ella con tu asistente fácilmente (próximamente)",
58
+ "title": "Soporte para conversaciones de base de conocimientos, activa tu segundo cerebro"
59
+ },
60
+ "localFirst": {
61
+ "desc": "Los datos de chat se almacenan completamente en el navegador, tus datos siempre están bajo tu control.",
62
+ "title": "Prioridad local, privacidad ante todo"
63
+ },
64
+ "pglite": {
65
+ "desc": "Construido sobre PGlite, soporte nativo para características avanzadas de AI Native (búsqueda vectorial)",
66
+ "title": "Nueva arquitectura de almacenamiento de cliente de próxima generación"
67
+ }
68
+ },
49
69
  "init": {
50
70
  "desc": "Inicializando la base de datos, el tiempo puede variar de 5 a 30 segundos dependiendo de la red.",
51
71
  "title": "Inicializando la base de datos PGlite"
@@ -23,20 +23,6 @@
23
23
  "4.0Ultra": {
24
24
  "description": "Spark4.0 Ultra es la versión más poderosa de la serie de modelos grandes de Xinghuo, mejorando la comprensión y capacidad de resumen de contenido textual al actualizar la conexión de búsqueda en línea. Es una solución integral para mejorar la productividad en la oficina y responder con precisión a las necesidades, siendo un producto inteligente líder en la industria."
25
25
  },
26
- "@cf/meta/llama-3-8b-instruct-awq": {},
27
- "@cf/openchat/openchat-3.5-0106": {},
28
- "@cf/qwen/qwen1.5-14b-chat-awq": {},
29
- "@hf/google/gemma-7b-it": {},
30
- "@hf/meta-llama/meta-llama-3-8b-instruct": {
31
- "description": "Generación tras generación, Meta Llama 3 demuestra un rendimiento de vanguardia en una amplia gama de benchmarks de la industria y ofrece nuevas capacidades, incluyendo un razonamiento mejorado."
32
- },
33
- "@hf/mistral/mistral-7b-instruct-v0.2": {},
34
- "@hf/nexusflow/starling-lm-7b-beta": {},
35
- "@hf/nousresearch/hermes-2-pro-mistral-7b": {},
36
- "@hf/thebloke/deepseek-coder-6.7b-instruct-awq": {},
37
- "@hf/thebloke/neural-chat-7b-v3-1-awq": {},
38
- "@hf/thebloke/openhermes-2.5-mistral-7b-awq": {},
39
- "@hf/thebloke/zephyr-7b-beta-awq": {},
40
26
  "Baichuan2-Turbo": {
41
27
  "description": "Utiliza tecnología de búsqueda mejorada para lograr un enlace completo entre el gran modelo y el conocimiento del dominio, así como el conocimiento de toda la red. Soporta la carga de documentos en PDF, Word y otros formatos, así como la entrada de URL, proporcionando información oportuna y completa, con resultados precisos y profesionales."
42
28
  },
@@ -127,7 +113,6 @@
127
113
  "Nous-Hermes-2-Mixtral-8x7B-DPO": {
128
114
  "description": "Hermes 2 Mixtral 8x7B DPO es una fusión de múltiples modelos altamente flexible, diseñada para ofrecer una experiencia creativa excepcional."
129
115
  },
130
- "NousResearch/Hermes-3-Llama-3.1-8B": {},
131
116
  "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
132
117
  "description": "Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) es un modelo de instrucciones de alta precisión, adecuado para cálculos complejos."
133
118
  },
@@ -530,9 +515,6 @@
530
515
  "cohere-command-r-plus": {
531
516
  "description": "Command R+ es un modelo optimizado para RAG de última generación diseñado para abordar cargas de trabajo de nivel empresarial."
532
517
  },
533
- "command-light": {
534
- "description": ""
535
- },
536
518
  "command-r": {
537
519
  "description": "Command R es un LLM optimizado para tareas de diálogo y contexto largo, especialmente adecuado para interacciones dinámicas y gestión del conocimiento."
538
520
  },
@@ -794,9 +776,6 @@
794
776
  "hunyuan-functioncall": {
795
777
  "description": "El último modelo FunctionCall de Hunyuan con arquitectura MOE, entrenado con datos de FunctionCall de alta calidad, con una ventana de contexto de 32K, liderando en múltiples dimensiones de métricas de evaluación."
796
778
  },
797
- "hunyuan-large": {
798
- "description": ""
799
- },
800
779
  "hunyuan-lite": {
801
780
  "description": "Actualizado a una estructura MOE, con una ventana de contexto de 256k, lidera en múltiples conjuntos de evaluación en NLP, código, matemáticas, industria y más, superando a muchos modelos de código abierto."
802
781
  },
@@ -830,8 +809,6 @@
830
809
  "internlm2.5-latest": {
831
810
  "description": "Nuestra última serie de modelos, con un rendimiento de inferencia excepcional, que admite una longitud de contexto de 1M y una mayor capacidad de seguimiento de instrucciones y llamadas a herramientas."
832
811
  },
833
- "jamba-1.5-large": {},
834
- "jamba-1.5-mini": {},
835
812
  "learnlm-1.5-pro-experimental": {
836
813
  "description": "LearnLM es un modelo de lenguaje experimental y específico para tareas, entrenado para cumplir con los principios de la ciencia del aprendizaje, capaz de seguir instrucciones sistemáticas en escenarios de enseñanza y aprendizaje, actuando como un tutor experto, entre otros."
837
814
  },
@@ -1027,7 +1004,6 @@
1027
1004
  "meta.llama3-8b-instruct-v1:0": {
1028
1005
  "description": "Meta Llama 3 es un modelo de lenguaje de gran tamaño (LLM) abierto dirigido a desarrolladores, investigadores y empresas, diseñado para ayudarles a construir, experimentar y escalar de manera responsable sus ideas de IA generativa. Como parte de un sistema base para la innovación de la comunidad global, es ideal para dispositivos de borde con recursos y capacidades computacionales limitadas, así como para tiempos de entrenamiento más rápidos."
1029
1006
  },
1030
- "microsoft/Phi-3.5-mini-instruct": {},
1031
1007
  "microsoft/WizardLM-2-8x22B": {
1032
1008
  "description": "WizardLM 2 es un modelo de lenguaje proporcionado por Microsoft AI, que destaca en diálogos complejos, multilingüismo, razonamiento y asistentes inteligentes."
1033
1009
  },
@@ -1,5 +1,7 @@
1
1
  {
2
- "ai21": {},
2
+ "ai21": {
3
+ "description": "AI21 Labs construye modelos fundamentales y sistemas de inteligencia artificial para empresas, acelerando la aplicación de la inteligencia artificial generativa en producción."
4
+ },
3
5
  "ai360": {
4
6
  "description": "360 AI es una plataforma de modelos y servicios de IA lanzada por la empresa 360, que ofrece una variedad de modelos avanzados de procesamiento del lenguaje natural, incluidos 360GPT2 Pro, 360GPT Pro, 360GPT Turbo y 360GPT Turbo Responsibility 8K. Estos modelos combinan parámetros a gran escala y capacidades multimodales, siendo ampliamente utilizados en generación de texto, comprensión semántica, sistemas de diálogo y generación de código. A través de una estrategia de precios flexible, 360 AI satisface diversas necesidades de los usuarios, apoyando la integración de desarrolladores y promoviendo la innovación y desarrollo de aplicaciones inteligentes."
5
7
  },
@@ -15,7 +17,9 @@
15
17
  "bedrock": {
16
18
  "description": "Bedrock es un servicio proporcionado por Amazon AWS, enfocado en ofrecer modelos de lenguaje y visuales avanzados para empresas. Su familia de modelos incluye la serie Claude de Anthropic, la serie Llama 3.1 de Meta, entre otros, abarcando una variedad de opciones desde ligeras hasta de alto rendimiento, apoyando tareas como generación de texto, diálogos y procesamiento de imágenes, adecuadas para aplicaciones empresariales de diferentes escalas y necesidades."
17
19
  },
18
- "cloudflare": {},
20
+ "cloudflare": {
21
+ "description": "Ejecuta modelos de aprendizaje automático impulsados por GPU sin servidor en la red global de Cloudflare."
22
+ },
19
23
  "deepseek": {
20
24
  "description": "DeepSeek es una empresa centrada en la investigación y aplicación de tecnologías de inteligencia artificial, cuyo modelo más reciente, DeepSeek-V2.5, combina capacidades de diálogo general y procesamiento de código, logrando mejoras significativas en alineación con preferencias humanas, tareas de escritura y seguimiento de instrucciones."
21
25
  },
@@ -35,7 +39,7 @@
35
39
  "description": "El motor de inferencia LPU de Groq ha demostrado un rendimiento excepcional en las pruebas de referencia de modelos de lenguaje de gran tamaño (LLM), redefiniendo los estándares de soluciones de IA con su asombrosa velocidad y eficiencia. Groq es un referente en velocidad de inferencia instantánea, mostrando un buen rendimiento en implementaciones basadas en la nube."
36
40
  },
37
41
  "higress": {
38
- "description": ""
42
+ "description": "Higress es una puerta de enlace API nativa de la nube, que nació en Alibaba para resolver los problemas que el recargado de Tengine causa en los negocios de conexiones largas, así como la insuficiencia de la capacidad de balanceo de carga de gRPC/Dubbo."
39
43
  },
40
44
  "huggingface": {
41
45
  "description": "La API de Inferencia de HuggingFace ofrece una forma rápida y gratuita de explorar miles de modelos para diversas tareas. Ya sea que esté prototipando una nueva aplicación o probando las capacidades del aprendizaje automático, esta API le brinda acceso instantáneo a modelos de alto rendimiento en múltiples dominios."
@@ -73,7 +77,9 @@
73
77
  "qwen": {
74
78
  "description": "Tongyi Qianwen es un modelo de lenguaje de gran escala desarrollado de forma independiente por Alibaba Cloud, con potentes capacidades de comprensión y generación de lenguaje natural. Puede responder a diversas preguntas, crear contenido escrito, expresar opiniones y redactar código, desempeñando un papel en múltiples campos."
75
79
  },
76
- "sensenova": {},
80
+ "sensenova": {
81
+ "description": "SenseTime ofrece servicios de modelos grandes de pila completa, aprovechando el sólido soporte de la gran infraestructura de SenseTime."
82
+ },
77
83
  "siliconcloud": {
78
84
  "description": "SiliconFlow se dedica a acelerar la AGI para beneficiar a la humanidad, mejorando la eficiencia de la IA a gran escala a través de un stack GenAI fácil de usar y de bajo costo."
79
85
  },