@lobehub/chat 1.77.16 → 1.77.18
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/changelog/v1.json +18 -0
- package/contributing/Basic/Architecture.md +1 -1
- package/contributing/Basic/Architecture.zh-CN.md +1 -1
- package/contributing/Basic/Chat-API.md +326 -108
- package/contributing/Basic/Chat-API.zh-CN.md +313 -133
- package/contributing/Basic/Contributing-Guidelines.md +7 -4
- package/contributing/Basic/Contributing-Guidelines.zh-CN.md +7 -6
- package/contributing/Home.md +5 -5
- package/contributing/State-Management/State-Management-Intro.md +1 -1
- package/contributing/State-Management/State-Management-Intro.zh-CN.md +1 -1
- package/docker-compose/local/docker-compose.yml +2 -1
- package/locales/ar/components.json +4 -0
- package/locales/ar/modelProvider.json +1 -0
- package/locales/ar/models.json +8 -5
- package/locales/ar/tool.json +21 -1
- package/locales/bg-BG/components.json +4 -0
- package/locales/bg-BG/modelProvider.json +1 -0
- package/locales/bg-BG/models.json +8 -5
- package/locales/bg-BG/tool.json +21 -1
- package/locales/de-DE/components.json +4 -0
- package/locales/de-DE/modelProvider.json +1 -0
- package/locales/de-DE/models.json +8 -5
- package/locales/de-DE/tool.json +21 -1
- package/locales/en-US/components.json +4 -0
- package/locales/en-US/modelProvider.json +1 -0
- package/locales/en-US/models.json +8 -5
- package/locales/en-US/tool.json +21 -1
- package/locales/es-ES/components.json +4 -0
- package/locales/es-ES/modelProvider.json +1 -0
- package/locales/es-ES/models.json +7 -4
- package/locales/es-ES/tool.json +21 -1
- package/locales/fa-IR/components.json +4 -0
- package/locales/fa-IR/modelProvider.json +1 -0
- package/locales/fa-IR/models.json +7 -4
- package/locales/fa-IR/tool.json +21 -1
- package/locales/fr-FR/components.json +4 -0
- package/locales/fr-FR/modelProvider.json +1 -0
- package/locales/fr-FR/models.json +8 -5
- package/locales/fr-FR/tool.json +21 -1
- package/locales/it-IT/components.json +4 -0
- package/locales/it-IT/modelProvider.json +1 -0
- package/locales/it-IT/models.json +7 -4
- package/locales/it-IT/tool.json +21 -1
- package/locales/ja-JP/components.json +4 -0
- package/locales/ja-JP/modelProvider.json +1 -0
- package/locales/ja-JP/models.json +8 -5
- package/locales/ja-JP/tool.json +21 -1
- package/locales/ko-KR/components.json +4 -0
- package/locales/ko-KR/modelProvider.json +1 -0
- package/locales/ko-KR/models.json +8 -5
- package/locales/ko-KR/tool.json +21 -1
- package/locales/nl-NL/components.json +4 -0
- package/locales/nl-NL/modelProvider.json +1 -0
- package/locales/nl-NL/models.json +8 -5
- package/locales/nl-NL/tool.json +21 -1
- package/locales/pl-PL/components.json +4 -0
- package/locales/pl-PL/modelProvider.json +1 -0
- package/locales/pl-PL/models.json +8 -5
- package/locales/pl-PL/tool.json +21 -1
- package/locales/pt-BR/components.json +4 -0
- package/locales/pt-BR/modelProvider.json +1 -0
- package/locales/pt-BR/models.json +7 -4
- package/locales/pt-BR/tool.json +21 -1
- package/locales/ru-RU/components.json +4 -0
- package/locales/ru-RU/modelProvider.json +1 -0
- package/locales/ru-RU/models.json +7 -4
- package/locales/ru-RU/tool.json +21 -1
- package/locales/tr-TR/components.json +4 -0
- package/locales/tr-TR/modelProvider.json +1 -0
- package/locales/tr-TR/models.json +8 -5
- package/locales/tr-TR/tool.json +21 -1
- package/locales/vi-VN/components.json +4 -0
- package/locales/vi-VN/modelProvider.json +1 -0
- package/locales/vi-VN/models.json +8 -5
- package/locales/vi-VN/tool.json +21 -1
- package/locales/zh-CN/components.json +4 -0
- package/locales/zh-CN/modelProvider.json +1 -0
- package/locales/zh-CN/models.json +9 -6
- package/locales/zh-CN/tool.json +30 -1
- package/locales/zh-TW/components.json +4 -0
- package/locales/zh-TW/modelProvider.json +1 -0
- package/locales/zh-TW/models.json +7 -4
- package/locales/zh-TW/tool.json +21 -1
- package/package.json +1 -1
- package/src/app/(backend)/webapi/models/[provider]/pull/route.ts +34 -0
- package/src/app/(backend)/webapi/{chat/models → models}/[provider]/route.ts +1 -2
- package/src/app/[variants]/(main)/settings/llm/ProviderList/Ollama/index.tsx +0 -7
- package/src/app/[variants]/(main)/settings/provider/(detail)/ollama/CheckError.tsx +1 -1
- package/src/components/FormAction/index.tsx +1 -1
- package/src/database/models/__tests__/aiProvider.test.ts +100 -0
- package/src/database/models/aiProvider.ts +11 -1
- package/src/features/Conversation/Error/OllamaBizError/InvalidOllamaModel.tsx +43 -0
- package/src/features/Conversation/Error/OllamaDesktopSetupGuide/index.tsx +61 -0
- package/src/features/Conversation/Error/index.tsx +7 -0
- package/src/features/DevPanel/SystemInspector/ServerConfig.tsx +18 -2
- package/src/features/DevPanel/SystemInspector/index.tsx +25 -6
- package/src/features/OllamaModelDownloader/index.tsx +149 -0
- package/src/libs/agent-runtime/AgentRuntime.ts +6 -0
- package/src/libs/agent-runtime/BaseAI.ts +7 -0
- package/src/libs/agent-runtime/ollama/index.ts +84 -2
- package/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap +24 -3263
- package/src/libs/agent-runtime/openrouter/fixtures/frontendModels.json +25 -0
- package/src/libs/agent-runtime/openrouter/fixtures/models.json +0 -3353
- package/src/libs/agent-runtime/openrouter/index.test.ts +56 -1
- package/src/libs/agent-runtime/openrouter/index.ts +9 -4
- package/src/libs/agent-runtime/types/index.ts +1 -0
- package/src/libs/agent-runtime/types/model.ts +44 -0
- package/src/libs/agent-runtime/utils/streams/index.ts +1 -0
- package/src/libs/agent-runtime/utils/streams/model.ts +110 -0
- package/src/locales/default/components.ts +4 -0
- package/src/locales/default/modelProvider.ts +1 -0
- package/src/locales/default/tool.ts +30 -1
- package/src/server/modules/SearXNG.ts +10 -2
- package/src/server/routers/tools/__test__/search.test.ts +3 -1
- package/src/server/routers/tools/search.ts +10 -2
- package/src/services/__tests__/models.test.ts +21 -0
- package/src/services/_url.ts +4 -1
- package/src/services/chat.ts +1 -1
- package/src/services/models.ts +153 -7
- package/src/services/search.ts +2 -2
- package/src/store/aiInfra/slices/aiModel/action.ts +1 -1
- package/src/store/aiInfra/slices/aiProvider/action.ts +2 -1
- package/src/store/chat/slices/builtinTool/actions/searXNG.test.ts +28 -8
- package/src/store/chat/slices/builtinTool/actions/searXNG.ts +22 -5
- package/src/store/user/slices/modelList/action.test.ts +2 -2
- package/src/store/user/slices/modelList/action.ts +1 -1
- package/src/tools/web-browsing/Portal/Search/index.tsx +1 -1
- package/src/tools/web-browsing/Render/Search/SearchQuery/SearchView.tsx +1 -1
- package/src/tools/web-browsing/Render/Search/SearchQuery/index.tsx +1 -1
- package/src/tools/web-browsing/Render/Search/SearchResult/index.tsx +1 -1
- package/src/tools/web-browsing/components/CategoryAvatar.tsx +27 -0
- package/src/tools/web-browsing/components/SearchBar.tsx +84 -4
- package/src/tools/web-browsing/const.ts +26 -0
- package/src/tools/web-browsing/index.ts +58 -28
- package/src/tools/web-browsing/systemRole.ts +62 -1
- package/src/types/tool/search.ts +10 -1
- package/src/app/[variants]/(main)/settings/llm/ProviderList/Ollama/Checker.tsx +0 -73
- package/src/app/[variants]/(main)/settings/provider/(detail)/ollama/OllamaModelDownloader/index.tsx +0 -127
- package/src/features/Conversation/Error/OllamaBizError/InvalidOllamaModel/index.tsx +0 -154
- package/src/features/Conversation/Error/OllamaBizError/InvalidOllamaModel/useDownloadMonitor.ts +0 -29
- package/src/helpers/url.ts +0 -17
- package/src/services/__tests__/ollama.test.ts +0 -28
- package/src/services/ollama.ts +0 -83
- /package/src/{app/[variants]/(main)/settings/provider/(detail)/ollama → features}/OllamaModelDownloader/useDownloadMonitor.ts +0 -0
package/locales/fr-FR/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "Mots-clés",
|
20
20
|
"tooltip": "Cela va récupérer à nouveau les résultats de recherche et créer un nouveau message de résumé"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "Catégorie de recherche",
|
24
|
+
"title": "Catégorie de recherche :",
|
25
|
+
"value": {
|
26
|
+
"files": "Fichiers",
|
27
|
+
"general": "Général",
|
28
|
+
"images": "Images",
|
29
|
+
"it": "Technologies de l'information",
|
30
|
+
"map": "Cartes",
|
31
|
+
"music": "Musique",
|
32
|
+
"news": "Actualités",
|
33
|
+
"science": "Science",
|
34
|
+
"social_media": "Médias sociaux",
|
35
|
+
"videos": "Vidéos"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "Moteur de recherche",
|
40
|
+
"title": "Moteur de recherche :"
|
41
|
+
},
|
23
42
|
"searchResult": "Nombre de recherches :",
|
43
|
+
"searchTimeRange": "Plage horaire :",
|
24
44
|
"summary": "Résumé",
|
25
45
|
"summaryTooltip": "Résumer le contenu actuel",
|
26
46
|
"viewMoreResults": "Voir {{results}} résultats supplémentaires"
|
@@ -91,6 +91,10 @@
|
|
91
91
|
"provider": "Provider"
|
92
92
|
},
|
93
93
|
"OllamaSetupGuide": {
|
94
|
+
"action": {
|
95
|
+
"close": "Chiudi avviso",
|
96
|
+
"start": "Installato e in esecuzione, inizia la conversazione"
|
97
|
+
},
|
94
98
|
"cors": {
|
95
99
|
"description": "A causa delle restrizioni di sicurezza del browser, è necessario configurare il cross-origin per utilizzare Ollama correttamente.",
|
96
100
|
"linux": {
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama sta scaricando questo modello, per favore non chiudere questa pagina. Il download verrà interrotto e riprenderà dal punto in cui si è interrotto in caso di riavvio.",
|
167
|
+
"failed": "Download del modello fallito, controlla la rete o le impostazioni di Ollama e riprova",
|
167
168
|
"remainingTime": "Tempo rimanente",
|
168
169
|
"speed": "Velocità di download",
|
169
170
|
"title": "Download del modello in corso {{model}}"
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tiny è un modello di linguaggio di grandi dimensioni ad alte prestazioni sviluppato internamente da Baidu, con i costi di distribuzione e messa a punto più bassi della serie Wencin."
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "Il grande modello Wénxīn X1 possiede una comprensione, pianificazione, riflessione e capacità evolutive superiori. Come modello di pensiero profondo più completo, Wénxīn X1 combina precisione, creatività e stile, eccellendo in domande e risposte in cinese, creazione letteraria, scrittura di documenti, conversazioni quotidiane, ragionamento logico, calcoli complessi e invocazione di strumenti."
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001 (Tuning) offre prestazioni stabili e ottimizzabili, è la scelta ideale per soluzioni a compiti complessi."
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "Gemini 2.5 Pro Experimental è il modello di pensiero più avanzato di Google, in grado di ragionare su problemi complessi in codice, matematica e nei campi STEM, e di analizzare grandi set di dati, codebase e documenti utilizzando contesti lunghi."
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "Gemini 2.5 Pro Preview è il modello di pensiero più avanzato di Google, in grado di ragionare su problemi complessi in codice, matematica e nei campi STEM, oltre a utilizzare analisi di lungo contesto per grandi set di dati, codici sorgente e documenti."
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7B è adatto per l'elaborazione di compiti di piccole e medie dimensioni, combinando efficienza dei costi."
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1 è il nuovo modello di inferenza di OpenAI, adatto a compiti complessi che richiedono una vasta conoscenza generale. Questo modello ha un contesto di 128K e una data di cutoff della conoscenza di ottobre 2023."
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7B è una libreria di modelli linguistici open source, ottimizzata tramite la strategia di 'C-RLFT (fine-tuning di rinforzo condizionato)'."
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "In base alla lunghezza del contesto, al tema e alla complessità, la tua richiesta verrà inviata a Llama 3 70B Instruct, Claude 3.5 Sonnet (auto-regolato) o GPT-4o."
|
1579
1582
|
},
|
@@ -1595,7 +1598,7 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "Il modello QVQ è un modello di ricerca sperimentale sviluppato dal team Qwen, focalizzato sul miglioramento delle capacità di ragionamento visivo, in particolare nel campo del ragionamento matematico."
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1601
|
+
"qvq-max": {
|
1599
1602
|
"description": "Il modello di ragionamento visivo QVQ di Tongyi Qianwen supporta input visivi e output di catene di pensiero, dimostrando capacità superiori in matematica, programmazione, analisi visiva, creazione e compiti generali."
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
package/locales/it-IT/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "Parole chiave",
|
20
20
|
"tooltip": "Ricaricherà i risultati di ricerca e creerà un nuovo messaggio di sintesi"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "Cerca categoria",
|
24
|
+
"title": "Categoria di ricerca:",
|
25
|
+
"value": {
|
26
|
+
"files": "File",
|
27
|
+
"general": "Generale",
|
28
|
+
"images": "Immagini",
|
29
|
+
"it": "Tecnologia dell'informazione",
|
30
|
+
"map": "Mappa",
|
31
|
+
"music": "Musica",
|
32
|
+
"news": "Notizie",
|
33
|
+
"science": "Scienza",
|
34
|
+
"social_media": "Social media",
|
35
|
+
"videos": "Video"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "Motore di ricerca",
|
40
|
+
"title": "Motore di ricerca:"
|
41
|
+
},
|
23
42
|
"searchResult": "Numero di ricerche:",
|
43
|
+
"searchTimeRange": "Intervallo di tempo:",
|
24
44
|
"summary": "Riepilogo",
|
25
45
|
"summaryTooltip": "Riepiloga il contenuto attuale",
|
26
46
|
"viewMoreResults": "Visualizza altri {{results}} risultati"
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollamaはこのモデルをダウンロードしています。このページを閉じないでください。再ダウンロードすると中断したところから再開されます。",
|
167
|
+
"failed": "モデルのダウンロードに失敗しました。ネットワークまたはOllamaの設定を確認して再試行してください",
|
167
168
|
"remainingTime": "残り時間",
|
168
169
|
"speed": "ダウンロード速度",
|
169
170
|
"title": "モデル{{model}}をダウンロード中"
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tinyは、百度が独自に開発した超高性能の大規模言語モデルで、文心シリーズモデルの中でデプロイと微調整コストが最も低いです。"
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "文心大モデルX1は、より強力な理解、計画、反省、進化の能力を備えています。より包括的な深い思考モデルとして、文心X1は正確さ、創造性、文才を兼ね備え、中国語の知識問答、文学創作、文書作成、日常会話、論理推論、複雑な計算、ツールの呼び出しなどの分野で特に優れたパフォーマンスを発揮します。"
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001(チューニング)は、安定した調整可能な性能を提供し、複雑なタスクのソリューションに理想的な選択肢です。"
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "Gemini 2.5 Pro Experimentalは、Googleの最先端の思考モデルであり、コード、数学、STEM分野の複雑な問題に対して推論を行うことができ、長いコンテキストを利用して大規模なデータセット、コードベース、文書を分析します。"
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "Gemini 2.5 Pro Previewは、Googleの最先端の思考モデルであり、コード、数学、STEM分野の複雑な問題に対して推論を行い、長いコンテキストを使用して大規模なデータセット、コードベース、文書を分析することができます。"
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7Bは、中小規模のタスク処理に適しており、コスト効果を兼ね備えています。"
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1はOpenAIの新しい推論モデルで、広範な一般知識を必要とする複雑なタスクに適しています。このモデルは128Kのコンテキストを持ち、2023年10月の知識のカットオフがあります。"
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7Bは「C-RLFT(条件強化学習微調整)」戦略で微調整されたオープンソース言語モデルライブラリです。"
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "コンテキストの長さ、テーマ、複雑さに応じて、あなたのリクエストはLlama 3 70B Instruct、Claude 3.5 Sonnet(自己調整)、またはGPT-4oに送信されます。"
|
1579
1582
|
},
|
@@ -1595,8 +1598,8 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "QVQモデルはQwenチームによって開発された実験的研究モデルで、視覚推論能力の向上に特化しており、特に数学推論の分野で優れた性能を発揮。"
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1599
|
-
"description": "通義千問QVQ
|
1601
|
+
"qvq-max": {
|
1602
|
+
"description": "通義千問QVQ視覚推論モデルは、視覚入力と思考の連鎖出力をサポートし、数学、プログラミング、視覚分析、創作、一般的なタスクにおいてより強力な能力を発揮します。"
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
1602
1605
|
"description": "通義千問コードモデル。"
|
package/locales/ja-JP/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "キーワード",
|
20
20
|
"tooltip": "検索結果を再取得し、新しい要約メッセージを作成します"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "カテゴリを検索",
|
24
|
+
"title": "カテゴリを検索:",
|
25
|
+
"value": {
|
26
|
+
"files": "ファイル",
|
27
|
+
"general": "一般",
|
28
|
+
"images": "画像",
|
29
|
+
"it": "情報技術",
|
30
|
+
"map": "地図",
|
31
|
+
"music": "音楽",
|
32
|
+
"news": "ニュース",
|
33
|
+
"science": "科学",
|
34
|
+
"social_media": "ソーシャルメディア",
|
35
|
+
"videos": "動画"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "検索エンジン",
|
40
|
+
"title": "検索エンジン:"
|
41
|
+
},
|
23
42
|
"searchResult": "検索結果の数:",
|
43
|
+
"searchTimeRange": "時間範囲:",
|
24
44
|
"summary": "要約",
|
25
45
|
"summaryTooltip": "現在の内容を要約",
|
26
46
|
"viewMoreResults": "さらに {{results}} 件の結果を見る"
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama가 모델을 다운로드하고 있습니다. 이 페이지를 닫지 마세요. 다시 다운로드할 경우 중단된 지점에서 계속됩니다.",
|
167
|
+
"failed": "모델 다운로드에 실패했습니다. 네트워크 또는 Ollama 설정을 확인한 후 다시 시도해 주세요.",
|
167
168
|
"remainingTime": "남은 시간",
|
168
169
|
"speed": "다운로드 속도",
|
169
170
|
"title": "모델 {{model}} 다운로드 중"
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tiny는 바이두가 자체 개발한 초고성능 대형 언어 모델로, 문신 시리즈 모델 중 배포 및 미세 조정 비용이 가장 낮습니다."
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "문심 대모델 X1은 더 강력한 이해, 계획, 반성 및 진화 능력을 갖추고 있습니다. 보다 포괄적인 심층 사고 모델로서, 문심 X1은 정확성, 창의성 및 문체를 겸비하여 중국어 지식 질문 응답, 문학 창작, 문서 작성, 일상 대화, 논리 추론, 복잡한 계산 및 도구 호출 등에서 특히 뛰어난 성과를 보입니다."
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001 (Tuning)은 안정적이고 조정 가능한 성능을 제공하며, 복잡한 작업 솔루션의 이상적인 선택입니다."
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "Gemini 2.5 Pro Experimental은 Google의 최첨단 사고 모델로, 코드, 수학 및 STEM 분야의 복잡한 문제를 추론할 수 있으며, 긴 문맥을 활용하여 대규모 데이터 세트, 코드베이스 및 문서를 분석할 수 있습니다."
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "Gemini 2.5 Pro Preview는 Google의 최첨단 사고 모델로, 코드, 수학 및 STEM 분야의 복잡한 문제를 추론하고 긴 맥락을 사용하여 대규모 데이터 세트, 코드베이스 및 문서를 분석할 수 있습니다."
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7B는 중소 규모 작업 처리에 적합하며, 비용 효과성을 갖추고 있습니다."
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1은 OpenAI의 새로운 추론 모델로, 광범위한 일반 지식이 필요한 복잡한 작업에 적합합니다. 이 모델은 128K의 컨텍스트와 2023년 10월의 지식 기준일을 가지고 있습니다."
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7B는 'C-RLFT(조건 강화 학습 미세 조정)' 전략으로 정교하게 조정된 오픈 소스 언어 모델 라이브러리입니다."
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "요청은 컨텍스트 길이, 주제 및 복잡성에 따라 Llama 3 70B Instruct, Claude 3.5 Sonnet(자기 조정) 또는 GPT-4o로 전송됩니다."
|
1579
1582
|
},
|
@@ -1595,8 +1598,8 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "QVQ 모델은 Qwen 팀이 개발한 실험적 연구 모델로, 시각적 추론 능력 향상에 중점을 두고 있으며, 특히 수학적 추론 분야에서 두드러진 성과를 보입니다."
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1599
|
-
"description": "통의천문 QVQ 비주얼 추론
|
1601
|
+
"qvq-max": {
|
1602
|
+
"description": "통의천문 QVQ 비주얼 추론 모델은 비주얼 입력과 사고 체인 출력을 지원하며, 수학, 프로그래밍, 비주얼 분석, 창작 및 일반 작업에서 더 강력한 능력을 보여줍니다."
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
1602
1605
|
"description": "통의 천문 코드 모델입니다."
|
package/locales/ko-KR/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "키워드",
|
20
20
|
"tooltip": "검색 결과를 다시 가져오고 새로운 요약 메시지를 생성합니다."
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "카테고리 검색",
|
24
|
+
"title": "카테고리 검색:",
|
25
|
+
"value": {
|
26
|
+
"files": "파일",
|
27
|
+
"general": "일반",
|
28
|
+
"images": "이미지",
|
29
|
+
"it": "정보 기술",
|
30
|
+
"map": "지도",
|
31
|
+
"music": "음악",
|
32
|
+
"news": "뉴스",
|
33
|
+
"science": "과학",
|
34
|
+
"social_media": "소셜 미디어",
|
35
|
+
"videos": "비디오"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "검색 엔진",
|
40
|
+
"title": "검색 엔진:"
|
41
|
+
},
|
23
42
|
"searchResult": "검색 수:",
|
43
|
+
"searchTimeRange": "시간 범위:",
|
24
44
|
"summary": "요약",
|
25
45
|
"summaryTooltip": "현재 내용 요약",
|
26
46
|
"viewMoreResults": "{{results}}개의 결과 더 보기"
|
@@ -91,6 +91,10 @@
|
|
91
91
|
"provider": "Provider"
|
92
92
|
},
|
93
93
|
"OllamaSetupGuide": {
|
94
|
+
"action": {
|
95
|
+
"close": "Sluit melding",
|
96
|
+
"start": "Geïnstalleerd en actief, begin gesprek"
|
97
|
+
},
|
94
98
|
"cors": {
|
95
99
|
"description": "Vanwege beveiligingsbeperkingen in de browser moet je cross-origin configuratie voor Ollama instellen om het correct te kunnen gebruiken.",
|
96
100
|
"linux": {
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama is het model aan het downloaden, sluit deze pagina alstublieft niet af. Bij een herstart zal het downloaden op de onderbroken plaats verdergaan.",
|
167
|
+
"failed": "Het downloaden van het model is mislukt. Controleer uw netwerk of de Ollama-instellingen en probeer het opnieuw.",
|
167
168
|
"remainingTime": "Overgebleven tijd",
|
168
169
|
"speed": "Downloadsnelheid",
|
169
170
|
"title": "Model {{model}} wordt gedownload"
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tiny is een ultra-presterend groot taalmodel dat de laagste implementatie- en afstemmingskosten heeft binnen de Wenxin modelreeks."
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "Het Wenxin grote model X1 heeft sterkere capaciteiten in begrip, planning, reflectie en evolutie. Als een dieper denkmodel met bredere mogelijkheden, combineert Wenxin X1 nauwkeurigheid, creativiteit en stijl, en excelleert in Chinese kennisvragen, literaire creatie, tekstschrijven, dagelijkse gesprekken, logische redenering, complexe berekeningen en het aanroepen van tools."
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001 (Tuning) biedt stabiele en afstelbare prestaties, ideaal voor oplossingen voor complexe taken."
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "Gemini 2.5 Pro Experimental is Google's meest geavanceerde denkmodel, dat in staat is om te redeneren over complexe problemen in code, wiskunde en STEM-gebieden, en bovendien gebruik maakt van lange contexten om grote datasets, codebases en documenten te analyseren."
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "Gemini 2.5 Pro Preview is Google's meest geavanceerde denkmodel, dat in staat is om te redeneren over complexe problemen in code, wiskunde en STEM-gebieden, en grote datasets, codebases en documenten te analyseren met behulp van lange context."
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7B is geschikt voor het verwerken van middelgrote taken, met een goede kosteneffectiviteit."
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1 is het nieuwe redeneermodel van OpenAI, geschikt voor complexe taken die uitgebreide algemene kennis vereisen. Dit model heeft een context van 128K en een kennisafkapdatum van oktober 2023."
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7B is een open-source taalmodelbibliotheek die is geoptimaliseerd met de 'C-RLFT (Conditionele Versterkingsleer Fijnstelling)' strategie."
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "Afhankelijk van de contextlengte, het onderwerp en de complexiteit, wordt uw verzoek verzonden naar Llama 3 70B Instruct, Claude 3.5 Sonnet (zelfregulerend) of GPT-4o."
|
1579
1582
|
},
|
@@ -1595,8 +1598,8 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "Het QVQ-model is een experimenteel onderzoeksmodel ontwikkeld door het Qwen-team, gericht op het verbeteren van visuele redeneervaardigheden, vooral in het domein van wiskundige redenering."
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1599
|
-
"description": "De QVQ visuele redeneermodel
|
1601
|
+
"qvq-max": {
|
1602
|
+
"description": "De QVQ visuele redeneermodel ondersteunt visuele invoer en denktaken, en toont sterkere capaciteiten in wiskunde, programmeren, visuele analyse, creatie en algemene taken."
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
1602
1605
|
"description": "Tongyi Qianwen code model."
|
package/locales/nl-NL/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "Zoekwoorden",
|
20
20
|
"tooltip": "De zoekresultaten worden opnieuw opgehaald en er wordt een nieuwe samenvattingsbericht aangemaakt"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "Zoekcategorie",
|
24
|
+
"title": "Zoekcategorie:",
|
25
|
+
"value": {
|
26
|
+
"files": "Bestanden",
|
27
|
+
"general": "Algemeen",
|
28
|
+
"images": "Afbeeldingen",
|
29
|
+
"it": "Informatietechnologie",
|
30
|
+
"map": "Kaart",
|
31
|
+
"music": "Muziek",
|
32
|
+
"news": "Nieuws",
|
33
|
+
"science": "Wetenschap",
|
34
|
+
"social_media": "Sociale media",
|
35
|
+
"videos": "Video's"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "Zoekmachine",
|
40
|
+
"title": "Zoekmachine:"
|
41
|
+
},
|
23
42
|
"searchResult": "Aantal zoekresultaten:",
|
43
|
+
"searchTimeRange": "Tijdspanne:",
|
24
44
|
"summary": "Samenvatting",
|
25
45
|
"summaryTooltip": "Samenvatting van de huidige inhoud",
|
26
46
|
"viewMoreResults": "Bekijk meer {{results}} resultaten"
|
@@ -91,6 +91,10 @@
|
|
91
91
|
"provider": "Dostawca"
|
92
92
|
},
|
93
93
|
"OllamaSetupGuide": {
|
94
|
+
"action": {
|
95
|
+
"close": "Zamknij powiadomienie",
|
96
|
+
"start": "Zainstalowano i uruchomiono, rozpocznij rozmowę"
|
97
|
+
},
|
94
98
|
"cors": {
|
95
99
|
"description": "Z powodu ograniczeń bezpieczeństwa przeglądarki, musisz skonfigurować CORS dla Ollama, aby móc go używać.",
|
96
100
|
"linux": {
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama pobiera ten model, proszę nie zamykać tej strony. Wznowienie pobierania nastąpi od miejsca przerwania",
|
167
|
+
"failed": "Pobieranie modelu nie powiodło się, sprawdź połączenie sieciowe lub ustawienia Ollama, a następnie spróbuj ponownie",
|
167
168
|
"remainingTime": "Pozostały czas",
|
168
169
|
"speed": "Prędkość pobierania",
|
169
170
|
"title": "Pobieranie modelu {{model}}"
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tiny to model językowy o ultra wysokiej wydajności opracowany przez Baidu, charakteryzujący się najniższymi kosztami wdrożenia i dostosowania w serii modeli Wenxin."
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "Model ERNIE X1 charakteryzuje się silniejszymi zdolnościami w zakresie rozumienia, planowania, refleksji i ewolucji. Jako model głębokiego myślenia o szerszych możliwościach, ERNIE X1 łączy w sobie precyzję, kreatywność i styl, osiągając szczególne wyniki w chińskich pytaniach i odpowiedziach, twórczości literackiej, pisaniu tekstów, codziennych rozmowach, wnioskowaniu logicznym, złożonych obliczeniach oraz wywoływaniu narzędzi."
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001 (Tuning) oferuje stabilną i dostosowywalną wydajność, co czyni go idealnym wyborem dla rozwiązań złożonych zadań."
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "Gemini 2.5 Pro Experimental to najnowocześniejszy model myślenia Google, zdolny do wnioskowania w zakresie kodu, matematyki i złożonych problemów w dziedzinie STEM, a także do analizy dużych zbiorów danych, repozytoriów kodu i dokumentów, wykorzystując długi kontekst."
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "Gemini 2.5 Pro Preview to najnowocześniejszy model myślenia Google, zdolny do wnioskowania w zakresie kodu, matematyki i złożonych problemów w dziedzinie STEM, a także do analizy dużych zbiorów danych, repozytoriów kodu i dokumentów przy użyciu długiego kontekstu."
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7B nadaje się do przetwarzania zadań średniej i małej skali, łącząc efektywność kosztową."
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1 to nowy model wnioskowania OpenAI, odpowiedni do złożonych zadań wymagających szerokiej wiedzy ogólnej. Model ten ma kontekst 128K i datę graniczną wiedzy z października 2023 roku."
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7B to otwarta biblioteka modeli językowych, dostrojona przy użyciu strategii „C-RLFT (warunkowe uczenie ze wzmocnieniem)”."
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "W zależności od długości kontekstu, tematu i złożoności, Twoje zapytanie zostanie wysłane do Llama 3 70B Instruct, Claude 3.5 Sonnet (samoregulacja) lub GPT-4o."
|
1579
1582
|
},
|
@@ -1595,8 +1598,8 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "Model QVQ jest eksperymentalnym modelem badawczym opracowanym przez zespół Qwen, skoncentrowanym na zwiększeniu zdolności w zakresie rozumowania wizualnego, szczególnie w dziedzinie rozumowania matematycznego."
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1599
|
-
"description": "Model wnioskowania
|
1601
|
+
"qvq-max": {
|
1602
|
+
"description": "Model wizualnego wnioskowania QVQ, wspierający wejścia wizualne oraz wyjścia w postaci łańcucha myślowego, wykazuje silniejsze zdolności w matematyce, programowaniu, analizie wizualnej, twórczości oraz w zadaniach ogólnych."
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
1602
1605
|
"description": "Model kodowania Qwen, oparty na ogólnym zrozumieniu."
|
package/locales/pl-PL/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "Słowa kluczowe",
|
20
20
|
"tooltip": "Ponownie pobierze wyniki wyszukiwania i utworzy nową wiadomość podsumowującą"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "Wyszukaj kategorię",
|
24
|
+
"title": "Kategoria wyszukiwania:",
|
25
|
+
"value": {
|
26
|
+
"files": "Pliki",
|
27
|
+
"general": "Ogólne",
|
28
|
+
"images": "Obrazy",
|
29
|
+
"it": "Technologia informacyjna",
|
30
|
+
"map": "Mapa",
|
31
|
+
"music": "Muzyka",
|
32
|
+
"news": "Aktualności",
|
33
|
+
"science": "Nauka",
|
34
|
+
"social_media": "Media społecznościowe",
|
35
|
+
"videos": "Filmy"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "Silnik wyszukiwania",
|
40
|
+
"title": "Silnik wyszukiwania:"
|
41
|
+
},
|
23
42
|
"searchResult": "Liczba wyników:",
|
43
|
+
"searchTimeRange": "Zakres czasu:",
|
24
44
|
"summary": "Podsumowanie",
|
25
45
|
"summaryTooltip": "Podsumuj bieżącą treść",
|
26
46
|
"viewMoreResults": "Zobacz więcej {{results}} wyników"
|
@@ -91,6 +91,10 @@
|
|
91
91
|
"provider": "Fornecedor"
|
92
92
|
},
|
93
93
|
"OllamaSetupGuide": {
|
94
|
+
"action": {
|
95
|
+
"close": "Fechar aviso",
|
96
|
+
"start": "Instalado e em execução, iniciar conversa"
|
97
|
+
},
|
94
98
|
"cors": {
|
95
99
|
"description": "Devido a restrições de segurança do navegador, você precisa configurar o CORS para o Ollama antes de usá-lo normalmente.",
|
96
100
|
"linux": {
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama está baixando este modelo, por favor, evite fechar esta página. O download será retomado do ponto em que parou.",
|
167
|
+
"failed": "Falha ao baixar o modelo, por favor verifique a rede ou as configurações do Ollama e tente novamente",
|
167
168
|
"remainingTime": "Tempo restante",
|
168
169
|
"speed": "Velocidade de download",
|
169
170
|
"title": "Baixando o modelo {{model}} "
|
@@ -842,6 +842,9 @@
|
|
842
842
|
"ernie-tiny-8k": {
|
843
843
|
"description": "ERNIE Tiny é um modelo de linguagem de grande escala de alto desempenho desenvolvido pela Baidu, com os menores custos de implantação e ajuste entre os modelos da série Wenxin."
|
844
844
|
},
|
845
|
+
"ernie-x1-32k-preview": {
|
846
|
+
"description": "O modelo grande Wenxin X1 possui habilidades aprimoradas de compreensão, planejamento, reflexão e evolução. Como um modelo de pensamento profundo mais abrangente, o Wenxin X1 combina precisão, criatividade e eloquência, destacando-se em perguntas e respostas de conhecimento em chinês, criação literária, redação de documentos, diálogos cotidianos, raciocínio lógico, cálculos complexos e chamadas de ferramentas."
|
847
|
+
},
|
845
848
|
"gemini-1.0-pro-001": {
|
846
849
|
"description": "Gemini 1.0 Pro 001 (Ajuste) oferece desempenho estável e ajustável, sendo a escolha ideal para soluções de tarefas complexas."
|
847
850
|
},
|
@@ -914,6 +917,9 @@
|
|
914
917
|
"gemini-2.5-pro-exp-03-25": {
|
915
918
|
"description": "O Gemini 2.5 Pro Experimental é o modelo de pensamento mais avançado do Google, capaz de raciocinar sobre problemas complexos em código, matemática e áreas STEM, além de analisar grandes conjuntos de dados, repositórios de código e documentos utilizando contextos longos."
|
916
919
|
},
|
920
|
+
"gemini-2.5-pro-preview-03-25": {
|
921
|
+
"description": "O Gemini 2.5 Pro Preview é o modelo de pensamento mais avançado do Google, capaz de raciocinar sobre problemas complexos em código, matemática e áreas STEM, além de analisar grandes conjuntos de dados, bibliotecas de código e documentos usando longos contextos."
|
922
|
+
},
|
917
923
|
"gemma-7b-it": {
|
918
924
|
"description": "Gemma 7B é adequado para o processamento de tarefas de pequeno a médio porte, combinando custo e eficiência."
|
919
925
|
},
|
@@ -1571,9 +1577,6 @@
|
|
1571
1577
|
"openai/o1-preview": {
|
1572
1578
|
"description": "o1 é o novo modelo de raciocínio da OpenAI, adequado para tarefas complexas que exigem amplo conhecimento geral. Este modelo possui um contexto de 128K e uma data limite de conhecimento em outubro de 2023."
|
1573
1579
|
},
|
1574
|
-
"openchat/openchat-7b": {
|
1575
|
-
"description": "OpenChat 7B é uma biblioteca de modelos de linguagem de código aberto ajustada com a estratégia de 'C-RLFT (refinamento de aprendizado por reforço condicional)'."
|
1576
|
-
},
|
1577
1580
|
"openrouter/auto": {
|
1578
1581
|
"description": "Com base no comprimento do contexto, tema e complexidade, sua solicitação será enviada para Llama 3 70B Instruct, Claude 3.5 Sonnet (autoajustável) ou GPT-4o."
|
1579
1582
|
},
|
@@ -1595,7 +1598,7 @@
|
|
1595
1598
|
"qvq-72b-preview": {
|
1596
1599
|
"description": "O modelo QVQ é um modelo de pesquisa experimental desenvolvido pela equipe Qwen, focado em melhorar a capacidade de raciocínio visual, especialmente na área de raciocínio matemático."
|
1597
1600
|
},
|
1598
|
-
"qvq-max
|
1601
|
+
"qvq-max": {
|
1599
1602
|
"description": "O modelo de raciocínio visual QVQ, da Tongyi Qianwen, suporta entrada visual e saída de cadeia de pensamento, demonstrando habilidades superiores em matemática, programação, análise visual, criação e tarefas gerais."
|
1600
1603
|
},
|
1601
1604
|
"qwen-coder-plus-latest": {
|
package/locales/pt-BR/tool.json
CHANGED
@@ -19,8 +19,28 @@
|
|
19
19
|
"placeholder": "Palavras-chave",
|
20
20
|
"tooltip": "Isso irá recuperar os resultados da pesquisa novamente e criar uma nova mensagem de resumo"
|
21
21
|
},
|
22
|
-
"
|
22
|
+
"searchCategory": {
|
23
|
+
"placeholder": "Pesquisar categoria",
|
24
|
+
"title": "Categoria de pesquisa:",
|
25
|
+
"value": {
|
26
|
+
"files": "Arquivos",
|
27
|
+
"general": "Geral",
|
28
|
+
"images": "Imagens",
|
29
|
+
"it": "Tecnologia da Informação",
|
30
|
+
"map": "Mapa",
|
31
|
+
"music": "Música",
|
32
|
+
"news": "Notícias",
|
33
|
+
"science": "Ciência",
|
34
|
+
"social_media": "Mídias sociais",
|
35
|
+
"videos": "Vídeos"
|
36
|
+
}
|
37
|
+
},
|
38
|
+
"searchEngine": {
|
39
|
+
"placeholder": "Motor de busca",
|
40
|
+
"title": "Motor de busca:"
|
41
|
+
},
|
23
42
|
"searchResult": "Número de pesquisas:",
|
43
|
+
"searchTimeRange": "Intervalo de tempo:",
|
24
44
|
"summary": "Resumo",
|
25
45
|
"summaryTooltip": "Resumir o conteúdo atual",
|
26
46
|
"viewMoreResults": "Ver mais {{results}} resultados"
|
@@ -91,6 +91,10 @@
|
|
91
91
|
"provider": "Поставщик"
|
92
92
|
},
|
93
93
|
"OllamaSetupGuide": {
|
94
|
+
"action": {
|
95
|
+
"close": "Закрыть уведомление",
|
96
|
+
"start": "Установлено и запущено, начать разговор"
|
97
|
+
},
|
94
98
|
"cors": {
|
95
99
|
"description": "Из-за ограничений безопасности браузера вам необходимо настроить кросс-доменные запросы для корректного использования Ollama.",
|
96
100
|
"linux": {
|
@@ -164,6 +164,7 @@
|
|
164
164
|
},
|
165
165
|
"download": {
|
166
166
|
"desc": "Ollama загружает эту модель, пожалуйста, старайтесь не закрывать эту страницу. При повторной загрузке процесс будет продолжен с места остановки",
|
167
|
+
"failed": "Не удалось загрузить модель, пожалуйста, проверьте сеть или настройки Ollama и попробуйте снова",
|
167
168
|
"remainingTime": "Оставшееся время",
|
168
169
|
"speed": "Скорость загрузки",
|
169
170
|
"title": "Загрузка модели {{model}} "
|