@lobehub/lobehub 2.0.0-next.331 → 2.0.0-next.333
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +60 -0
- package/apps/desktop/src/main/const/dir.ts +3 -0
- package/apps/desktop/src/main/controllers/SystemCtr.ts +19 -0
- package/apps/desktop/src/main/controllers/__tests__/SystemCtr.test.ts +1 -0
- package/apps/desktop/src/main/menus/impls/macOS.test.ts +1 -0
- package/changelog/v1.json +21 -0
- package/docs/changelog/2023-09-09-plugin-system.mdx +3 -2
- package/docs/changelog/2023-11-14-gpt4-vision.mdx +6 -4
- package/docs/changelog/2023-11-19-tts-stt.mdx +3 -2
- package/docs/changelog/2023-12-22-dalle-3.mdx +5 -2
- package/docs/changelog/2023-12-22-dalle-3.zh-CN.mdx +2 -2
- package/docs/changelog/2024-02-08-sso-oauth.mdx +2 -2
- package/docs/changelog/2024-06-19-lobe-chat-v1.mdx +3 -2
- package/docs/changelog/2024-06-19-lobe-chat-v1.zh-CN.mdx +2 -2
- package/docs/changelog/2024-07-19-gpt-4o-mini.mdx +3 -2
- package/docs/changelog/2024-07-19-gpt-4o-mini.zh-CN.mdx +2 -2
- package/docs/changelog/2024-08-02-lobe-chat-database-docker.mdx +3 -2
- package/docs/changelog/2024-08-21-file-upload-and-knowledge-base.mdx +5 -4
- package/docs/changelog/2024-09-13-openai-o1-models.mdx +2 -2
- package/docs/changelog/2024-09-20-artifacts.mdx +3 -2
- package/docs/changelog/2024-09-20-artifacts.zh-CN.mdx +2 -2
- package/docs/changelog/2024-10-27-pin-assistant.mdx +3 -2
- package/docs/changelog/2024-11-06-share-text-json.mdx +4 -2
- package/docs/changelog/2024-11-06-share-text-json.zh-CN.mdx +2 -2
- package/docs/changelog/2024-11-25-november-providers.mdx +2 -2
- package/docs/changelog/2024-11-27-forkable-chat.mdx +2 -2
- package/docs/changelog/2025-01-03-user-profile.mdx +2 -2
- package/docs/changelog/2025-01-22-new-ai-provider.mdx +2 -2
- package/docs/changelog/2025-02-02-deepseek-r1.mdx +4 -4
- package/docs/development/basic/add-new-authentication-providers.mdx +4 -0
- package/docs/development/basic/add-new-authentication-providers.zh-CN.mdx +4 -0
- package/docs/development/basic/add-new-image-model.mdx +4 -0
- package/docs/development/basic/add-new-image-model.zh-CN.mdx +4 -0
- package/docs/development/basic/architecture.mdx +4 -0
- package/docs/development/basic/architecture.zh-CN.mdx +4 -0
- package/docs/development/basic/chat-api.mdx +4 -0
- package/docs/development/basic/chat-api.zh-CN.mdx +4 -0
- package/docs/development/basic/comfyui-development.mdx +3 -1
- package/docs/development/basic/contributing-guidelines.mdx +4 -0
- package/docs/development/basic/contributing-guidelines.zh-CN.mdx +4 -0
- package/docs/development/basic/feature-development-frontend.mdx +11 -3
- package/docs/development/basic/feature-development-frontend.zh-CN.mdx +11 -3
- package/docs/development/basic/feature-development.mdx +14 -5
- package/docs/development/basic/feature-development.zh-CN.mdx +14 -5
- package/docs/development/basic/folder-structure.mdx +7 -0
- package/docs/development/basic/folder-structure.zh-CN.mdx +7 -0
- package/docs/development/basic/resources.mdx +4 -0
- package/docs/development/basic/resources.zh-CN.mdx +4 -0
- package/docs/development/basic/setup-development.mdx +4 -0
- package/docs/development/basic/setup-development.zh-CN.mdx +4 -0
- package/docs/development/basic/test.mdx +4 -0
- package/docs/development/basic/test.zh-CN.mdx +4 -0
- package/docs/development/basic/work-with-server-side-database.mdx +5 -5
- package/docs/development/basic/work-with-server-side-database.zh-CN.mdx +5 -5
- package/docs/development/internationalization/add-new-locale.mdx +4 -0
- package/docs/development/internationalization/add-new-locale.zh-CN.mdx +4 -0
- package/docs/development/internationalization/internationalization-implementation.mdx +4 -0
- package/docs/development/internationalization/internationalization-implementation.zh-CN.mdx +4 -0
- package/docs/development/others/lighthouse.mdx +4 -0
- package/docs/development/others/lighthouse.zh-CN.mdx +4 -0
- package/docs/development/start.mdx +4 -0
- package/docs/development/start.zh-CN.mdx +4 -0
- package/docs/development/state-management/state-management-intro.mdx +4 -2
- package/docs/development/state-management/state-management-intro.zh-CN.mdx +4 -2
- package/docs/development/state-management/state-management-selectors.mdx +6 -1
- package/docs/development/state-management/state-management-selectors.zh-CN.mdx +6 -1
- package/docs/development/tests/integration-testing.zh-CN.mdx +4 -0
- package/docs/self-hosting/advanced/analytics.mdx +2 -2
- package/docs/self-hosting/advanced/auth/better-auth/apple.mdx +132 -0
- package/docs/self-hosting/advanced/auth/better-auth/apple.zh-CN.mdx +127 -0
- package/docs/self-hosting/advanced/auth/better-auth/auth0.mdx +111 -0
- package/docs/self-hosting/advanced/auth/better-auth/auth0.zh-CN.mdx +107 -0
- package/docs/self-hosting/advanced/auth/better-auth/authelia.mdx +66 -0
- package/docs/self-hosting/advanced/auth/better-auth/authelia.zh-CN.mdx +62 -0
- package/docs/self-hosting/advanced/auth/better-auth/authentik.mdx +67 -0
- package/docs/self-hosting/advanced/auth/better-auth/authentik.zh-CN.mdx +63 -0
- package/docs/self-hosting/advanced/auth/better-auth/casdoor.mdx +62 -0
- package/docs/self-hosting/advanced/auth/better-auth/casdoor.zh-CN.mdx +58 -0
- package/docs/self-hosting/advanced/auth/better-auth/cloudflare-zero-trust.mdx +59 -0
- package/docs/self-hosting/advanced/auth/better-auth/cloudflare-zero-trust.zh-CN.mdx +55 -0
- package/docs/self-hosting/advanced/auth/better-auth/cognito.mdx +88 -0
- package/docs/self-hosting/advanced/auth/better-auth/cognito.zh-CN.mdx +85 -0
- package/docs/self-hosting/advanced/auth/better-auth/feishu.mdx +73 -0
- package/docs/self-hosting/advanced/auth/better-auth/feishu.zh-CN.mdx +69 -0
- package/docs/self-hosting/advanced/auth/better-auth/generic-oidc.mdx +86 -0
- package/docs/self-hosting/advanced/auth/better-auth/generic-oidc.zh-CN.mdx +83 -0
- package/docs/self-hosting/advanced/auth/better-auth/github.mdx +93 -0
- package/docs/self-hosting/advanced/auth/better-auth/github.zh-CN.mdx +90 -0
- package/docs/self-hosting/advanced/auth/better-auth/google.mdx +80 -0
- package/docs/self-hosting/advanced/auth/better-auth/google.zh-CN.mdx +77 -0
- package/docs/self-hosting/advanced/auth/better-auth/keycloak.mdx +77 -0
- package/docs/self-hosting/advanced/auth/better-auth/keycloak.zh-CN.mdx +74 -0
- package/docs/self-hosting/advanced/auth/better-auth/logto.mdx +64 -0
- package/docs/self-hosting/advanced/auth/better-auth/logto.zh-CN.mdx +60 -0
- package/docs/self-hosting/advanced/auth/better-auth/microsoft.mdx +113 -0
- package/docs/self-hosting/advanced/auth/better-auth/microsoft.zh-CN.mdx +109 -0
- package/docs/self-hosting/advanced/auth/better-auth/okta.mdx +67 -0
- package/docs/self-hosting/advanced/auth/better-auth/okta.zh-CN.mdx +63 -0
- package/docs/self-hosting/advanced/auth/better-auth/wechat.mdx +77 -0
- package/docs/self-hosting/advanced/auth/better-auth/wechat.zh-CN.mdx +72 -0
- package/docs/self-hosting/advanced/auth/better-auth/zitadel.mdx +73 -0
- package/docs/self-hosting/advanced/auth/better-auth/zitadel.zh-CN.mdx +69 -0
- package/docs/self-hosting/advanced/auth/clerk.mdx +2 -2
- package/docs/self-hosting/advanced/auth/legacy.mdx +106 -0
- package/docs/self-hosting/advanced/auth/legacy.zh-CN.mdx +101 -0
- package/docs/self-hosting/advanced/auth/next-auth/auth0.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/authelia.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/authentik.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/casdoor.mdx +5 -2
- package/docs/self-hosting/advanced/auth/next-auth/casdoor.zh-CN.mdx +2 -0
- package/docs/self-hosting/advanced/auth/next-auth/cloudflare-zero-trust.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/cloudflare-zero-trust.zh-CN.mdx +2 -2
- package/docs/self-hosting/advanced/auth/next-auth/github.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/google.mdx +10 -12
- package/docs/self-hosting/advanced/auth/next-auth/keycloak.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/logto.mdx +2 -2
- package/docs/self-hosting/advanced/auth/next-auth/microsoft-entra-id.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/okta.mdx +3 -2
- package/docs/self-hosting/advanced/auth/next-auth/okta.zh-CN.mdx +1 -3
- package/docs/self-hosting/advanced/auth/next-auth/wechat.mdx +2 -2
- package/docs/self-hosting/advanced/auth/next-auth/zitadel.mdx +3 -2
- package/docs/self-hosting/advanced/auth.mdx +86 -139
- package/docs/self-hosting/advanced/auth.zh-CN.mdx +84 -135
- package/docs/self-hosting/advanced/desktop.mdx +9 -3
- package/docs/self-hosting/advanced/desktop.zh-CN.mdx +9 -3
- package/docs/self-hosting/advanced/feature-flags.mdx +3 -2
- package/docs/self-hosting/advanced/knowledge-base.mdx +2 -2
- package/docs/self-hosting/advanced/model-list.mdx +2 -2
- package/docs/self-hosting/advanced/observability/grafana.mdx +4 -2
- package/docs/self-hosting/advanced/observability/grafana.zh-CN.mdx +2 -1
- package/docs/self-hosting/advanced/observability/langfuse.mdx +3 -2
- package/docs/self-hosting/advanced/online-search.mdx +4 -6
- package/docs/self-hosting/advanced/s3/tencent-cloud.mdx +2 -2
- package/docs/self-hosting/advanced/settings-url-share.mdx +3 -2
- package/docs/self-hosting/advanced/upstream-sync.mdx +3 -4
- package/docs/self-hosting/advanced/upstream-sync.zh-CN.mdx +0 -2
- package/docs/self-hosting/environment-variables/analytics.mdx +3 -2
- package/docs/self-hosting/environment-variables/auth.mdx +5 -12
- package/docs/self-hosting/environment-variables/auth.zh-CN.mdx +2 -9
- package/docs/self-hosting/environment-variables/basic.mdx +3 -10
- package/docs/self-hosting/environment-variables/basic.zh-CN.mdx +0 -7
- package/docs/self-hosting/environment-variables/model-provider.mdx +3 -4
- package/docs/self-hosting/environment-variables/s3.mdx +2 -2
- package/docs/self-hosting/environment-variables.mdx +2 -3
- package/docs/self-hosting/examples/azure-openai.mdx +2 -3
- package/docs/self-hosting/examples/azure-openai.zh-CN.mdx +0 -1
- package/docs/self-hosting/examples/ollama.mdx +3 -2
- package/docs/self-hosting/faq/no-v1-suffix.mdx +4 -4
- package/docs/self-hosting/faq/proxy-with-unable-to-verify-leaf-signature.mdx +3 -2
- package/docs/self-hosting/faq/vercel-ai-image-timeout.mdx +2 -2
- package/docs/self-hosting/migration/v2/breaking-changes.mdx +73 -0
- package/docs/self-hosting/migration/v2/breaking-changes.zh-CN.mdx +71 -0
- package/docs/self-hosting/platform/alibaba-cloud.mdx +2 -7
- package/docs/self-hosting/platform/alibaba-cloud.zh-CN.mdx +1 -6
- package/docs/self-hosting/platform/btpanel.mdx +4 -2
- package/docs/self-hosting/platform/btpanel.zh-CN.mdx +2 -2
- package/docs/self-hosting/platform/docker-compose.mdx +3 -3
- package/docs/self-hosting/platform/docker-compose.zh-CN.mdx +0 -1
- package/docs/self-hosting/platform/docker.mdx +2 -11
- package/docs/self-hosting/platform/docker.zh-CN.mdx +0 -8
- package/docs/self-hosting/platform/netlify.mdx +5 -17
- package/docs/self-hosting/platform/netlify.zh-CN.mdx +3 -17
- package/docs/self-hosting/platform/railway.mdx +3 -7
- package/docs/self-hosting/platform/railway.zh-CN.mdx +1 -7
- package/docs/self-hosting/platform/repocloud.mdx +3 -7
- package/docs/self-hosting/platform/repocloud.zh-CN.mdx +1 -6
- package/docs/self-hosting/platform/sealos.mdx +2 -7
- package/docs/self-hosting/platform/sealos.zh-CN.mdx +1 -6
- package/docs/self-hosting/platform/tencentcloud-lighthouse.mdx +2 -7
- package/docs/self-hosting/platform/tencentcloud-lighthouse.zh-CN.mdx +1 -6
- package/docs/self-hosting/platform/vercel.mdx +4 -9
- package/docs/self-hosting/platform/vercel.zh-CN.mdx +3 -8
- package/docs/self-hosting/platform/zeabur.mdx +2 -11
- package/docs/self-hosting/platform/zeabur.zh-CN.mdx +1 -10
- package/docs/self-hosting/server-database/docker-compose.mdx +11 -19
- package/docs/self-hosting/server-database/docker-compose.zh-CN.mdx +12 -21
- package/docs/self-hosting/server-database/docker.mdx +9 -24
- package/docs/self-hosting/server-database/docker.zh-CN.mdx +7 -24
- package/docs/self-hosting/server-database/dokploy.mdx +27 -25
- package/docs/self-hosting/server-database/dokploy.zh-CN.mdx +23 -21
- package/docs/self-hosting/server-database/netlify.mdx +2 -2
- package/docs/self-hosting/server-database/netlify.zh-CN.mdx +2 -2
- package/docs/self-hosting/server-database/railway.mdx +2 -2
- package/docs/self-hosting/server-database/repocloud.mdx +2 -2
- package/docs/self-hosting/server-database/sealos.mdx +2 -2
- package/docs/self-hosting/server-database/vercel.mdx +19 -72
- package/docs/self-hosting/server-database/vercel.zh-CN.mdx +17 -68
- package/docs/self-hosting/server-database/zeabur.mdx +2 -2
- package/docs/self-hosting/server-database.mdx +1 -19
- package/docs/self-hosting/server-database.zh-CN.mdx +0 -17
- package/docs/self-hosting/start.mdx +2 -2
- package/docs/self-hosting/start.zh-CN.mdx +2 -2
- package/e2e/src/support/webServer.ts +95 -43
- package/locales/ar/chat.json +5 -0
- package/locales/ar/desktop-onboarding.json +5 -0
- package/locales/ar/discover.json +15 -0
- package/locales/ar/models.json +35 -0
- package/locales/bg-BG/chat.json +5 -0
- package/locales/bg-BG/desktop-onboarding.json +5 -0
- package/locales/bg-BG/discover.json +15 -0
- package/locales/bg-BG/models.json +30 -0
- package/locales/de-DE/chat.json +5 -0
- package/locales/de-DE/desktop-onboarding.json +5 -0
- package/locales/de-DE/discover.json +15 -0
- package/locales/de-DE/models.json +38 -0
- package/locales/en-US/desktop-onboarding.json +6 -0
- package/locales/en-US/discover.json +14 -0
- package/locales/es-ES/chat.json +5 -0
- package/locales/es-ES/desktop-onboarding.json +5 -0
- package/locales/es-ES/discover.json +15 -0
- package/locales/es-ES/models.json +38 -0
- package/locales/fa-IR/chat.json +5 -0
- package/locales/fa-IR/desktop-onboarding.json +5 -0
- package/locales/fa-IR/discover.json +15 -0
- package/locales/fa-IR/models.json +11 -0
- package/locales/fr-FR/chat.json +5 -0
- package/locales/fr-FR/desktop-onboarding.json +5 -0
- package/locales/fr-FR/discover.json +15 -0
- package/locales/fr-FR/models.json +36 -0
- package/locales/it-IT/chat.json +5 -0
- package/locales/it-IT/desktop-onboarding.json +5 -0
- package/locales/it-IT/discover.json +15 -0
- package/locales/it-IT/models.json +32 -0
- package/locales/ja-JP/chat.json +5 -0
- package/locales/ja-JP/desktop-onboarding.json +5 -0
- package/locales/ja-JP/discover.json +15 -0
- package/locales/ja-JP/models.json +42 -0
- package/locales/ko-KR/chat.json +5 -0
- package/locales/ko-KR/desktop-onboarding.json +5 -0
- package/locales/ko-KR/discover.json +15 -0
- package/locales/ko-KR/models.json +55 -0
- package/locales/nl-NL/chat.json +5 -0
- package/locales/nl-NL/desktop-onboarding.json +5 -0
- package/locales/nl-NL/discover.json +15 -0
- package/locales/nl-NL/models.json +34 -0
- package/locales/pl-PL/chat.json +5 -0
- package/locales/pl-PL/desktop-onboarding.json +5 -0
- package/locales/pl-PL/discover.json +15 -0
- package/locales/pl-PL/models.json +31 -0
- package/locales/pt-BR/chat.json +5 -0
- package/locales/pt-BR/desktop-onboarding.json +5 -0
- package/locales/pt-BR/discover.json +15 -0
- package/locales/pt-BR/models.json +42 -0
- package/locales/ru-RU/chat.json +5 -0
- package/locales/ru-RU/desktop-onboarding.json +5 -0
- package/locales/ru-RU/discover.json +15 -0
- package/locales/ru-RU/models.json +32 -0
- package/locales/tr-TR/chat.json +5 -0
- package/locales/tr-TR/desktop-onboarding.json +5 -0
- package/locales/tr-TR/discover.json +15 -0
- package/locales/tr-TR/models.json +48 -0
- package/locales/vi-VN/chat.json +5 -0
- package/locales/vi-VN/desktop-onboarding.json +5 -0
- package/locales/vi-VN/discover.json +15 -0
- package/locales/vi-VN/models.json +36 -0
- package/locales/zh-CN/desktop-onboarding.json +4 -3
- package/locales/zh-CN/discover.json +14 -0
- package/locales/zh-CN/models.json +1 -0
- package/locales/zh-TW/chat.json +5 -0
- package/locales/zh-TW/desktop-onboarding.json +5 -0
- package/locales/zh-TW/discover.json +15 -0
- package/locales/zh-TW/models.json +42 -0
- package/package.json +2 -2
- package/packages/builtin-tool-cloud-sandbox/src/systemRole.ts +62 -2
- package/packages/conversation-flow/src/__tests__/fixtures/inputs/tasks/index.ts +2 -0
- package/packages/conversation-flow/src/__tests__/fixtures/inputs/tasks/multi-tasks-with-summary.json +234 -0
- package/packages/conversation-flow/src/__tests__/parse.test.ts +25 -0
- package/packages/conversation-flow/src/transformation/ContextTreeBuilder.ts +15 -0
- package/packages/conversation-flow/src/transformation/FlatListBuilder.ts +20 -0
- package/packages/types/src/serverConfig.ts +0 -1
- package/src/app/[variants]/(desktop)/desktop-onboarding/features/LoginStep.tsx +39 -1
- package/src/app/[variants]/(main)/settings/common/features/Common/Common.tsx +34 -14
- package/src/envs/app.ts +0 -13
- package/src/locales/default/desktop-onboarding.ts +1 -0
- package/src/server/globalConfig/index.ts +1 -2
- package/src/services/electron/system.ts +4 -0
- package/src/store/serverConfig/selectors.ts +0 -1
|
@@ -141,6 +141,16 @@
|
|
|
141
141
|
"filterBy.timePeriod.year": "최근 1년",
|
|
142
142
|
"footer.desc": "전 세계 AI 사용자와 함께 발전하세요. 크리에이터가 되어 LobeHub 커뮤니티에 에이전트와 스킬을 제출해보세요.",
|
|
143
143
|
"footer.title": "오늘 LobeHub 커뮤니티에 당신의 창작물을 공유하세요",
|
|
144
|
+
"fork.alreadyForked": "이미 이 에이전트를 포크했습니다. 포크한 항목으로 이동 중...",
|
|
145
|
+
"fork.failed": "포크에 실패했습니다. 다시 시도해 주세요.",
|
|
146
|
+
"fork.forkAndChat": "포크 및 채팅",
|
|
147
|
+
"fork.forkedFrom": "포크된 원본",
|
|
148
|
+
"fork.forks": "포크",
|
|
149
|
+
"fork.forksCount": "{{count}}개 포크",
|
|
150
|
+
"fork.forksCount_other": "{{count}}개 포크",
|
|
151
|
+
"fork.success": "성공적으로 포크되었습니다!",
|
|
152
|
+
"fork.viewAllForks": "모든 포크 보기",
|
|
153
|
+
"groupAgents.tag": "그룹",
|
|
144
154
|
"home.communityAgents": "커뮤니티 도우미",
|
|
145
155
|
"home.featuredAssistants": "추천 도우미",
|
|
146
156
|
"home.featuredModels": "추천 모델",
|
|
@@ -471,14 +481,19 @@
|
|
|
471
481
|
"user.follow": "팔로우",
|
|
472
482
|
"user.followers": "팔로워",
|
|
473
483
|
"user.following": "팔로잉",
|
|
484
|
+
"user.forkedAgentGroups": "포크한 에이전트 그룹",
|
|
485
|
+
"user.forkedAgents": "포크한 에이전트",
|
|
474
486
|
"user.login": "크리에이터 되기",
|
|
475
487
|
"user.logout": "로그아웃",
|
|
476
488
|
"user.myProfile": "내 프로필",
|
|
477
489
|
"user.noAgents": "이 사용자가 아직 도우미를 게시하지 않았습니다",
|
|
478
490
|
"user.noFavoriteAgents": "즐겨찾는 어시스턴트가 없습니다",
|
|
479
491
|
"user.noFavoritePlugins": "즐겨찾는 플러그인이 없습니다",
|
|
492
|
+
"user.noForkedAgentGroups": "포크한 에이전트 그룹이 없습니다",
|
|
493
|
+
"user.noForkedAgents": "포크한 에이전트가 없습니다",
|
|
480
494
|
"user.publishedAgents": "제작한 도우미",
|
|
481
495
|
"user.tabs.favorites": "즐겨찾기",
|
|
496
|
+
"user.tabs.forkedAgents": "포크됨",
|
|
482
497
|
"user.tabs.publishedAgents": "창작",
|
|
483
498
|
"user.unfavorite": "즐겨찾기 해제",
|
|
484
499
|
"user.unfavoriteFailed": "즐겨찾기 해제 실패",
|
|
@@ -573,6 +573,61 @@
|
|
|
573
573
|
"glm-z1-flash.description": "GLM-Z1 시리즈는 논리, 수학, 프로그래밍 등 복잡한 추론에서 뛰어난 성능을 발휘합니다.",
|
|
574
574
|
"glm-z1-flashx.description": "빠르고 저비용: 초고속 추론과 높은 동시성을 제공하는 Flash 향상 버전입니다.",
|
|
575
575
|
"glm-zero-preview.description": "GLM-Zero-Preview는 논리, 수학, 프로그래밍 등 복잡한 추론에서 뛰어난 성능을 발휘합니다.",
|
|
576
|
+
"global.anthropic.claude-opus-4-5-20251101-v1:0.description": "Claude Opus 4.5는 Anthropic의 대표 모델로, 탁월한 지능과 확장 가능한 성능을 결합하여 최고 수준의 응답 품질과 추론 능력이 요구되는 복잡한 작업을 처리합니다.",
|
|
577
|
+
"google/gemini-2.0-flash-001.description": "Gemini 2.0 Flash는 차세대 기능을 제공하며, 뛰어난 속도, 기본 도구 사용, 멀티모달 생성, 1M 토큰 컨텍스트 윈도우를 지원합니다.",
|
|
578
|
+
"google/gemini-2.0-flash-exp:free.description": "Gemini 2.0 Flash Experimental은 Google의 최신 실험용 멀티모달 AI 모델로, 이전 버전 대비 세계 지식, 코드, 장문 컨텍스트 처리에서 품질이 향상되었습니다.",
|
|
579
|
+
"google/gemini-2.0-flash-lite-001.description": "Gemini 2.0 Flash Lite는 지연 시간과 비용을 줄이기 위해 기본적으로 사고 기능이 비활성화된 경량 Gemini 변형 모델이며, 매개변수를 통해 활성화할 수 있습니다.",
|
|
580
|
+
"google/gemini-2.0-flash-lite.description": "Gemini 2.0 Flash Lite는 차세대 기능을 제공하며, 뛰어난 속도, 내장 도구 사용, 멀티모달 생성, 1M 토큰 컨텍스트 윈도우를 지원합니다.",
|
|
581
|
+
"google/gemini-2.0-flash.description": "Gemini 2.0 Flash는 Google의 고성능 추론 모델로, 확장된 멀티모달 작업에 적합합니다.",
|
|
582
|
+
"google/gemini-2.5-flash-image-free.description": "Gemini 2.5 Flash Image 무료 등급은 제한된 할당량의 멀티모달 생성을 지원합니다.",
|
|
583
|
+
"google/gemini-2.5-flash-image-preview.description": "Gemini 2.5 Flash는 이미지 생성을 지원하는 실험용 모델입니다.",
|
|
584
|
+
"google/gemini-2.5-flash-image.description": "Gemini 2.5 Flash Image (Nano Banana)는 Google의 이미지 생성 모델로, 멀티모달 대화를 지원합니다.",
|
|
585
|
+
"google/gemini-2.5-flash-lite.description": "Gemini 2.5 Flash Lite는 지연 시간과 비용을 최적화한 경량 Gemini 2.5 변형 모델로, 대량 처리 시나리오에 적합합니다.",
|
|
586
|
+
"google/gemini-2.5-flash-preview.description": "Gemini 2.5 Flash는 고급 추론, 코딩, 수학, 과학 작업을 위해 설계된 Google의 최첨단 대표 모델입니다. 내장된 '사고(thinking)' 기능을 통해 더 높은 정확도의 응답과 정교한 컨텍스트 처리를 제공합니다.\n\n참고: 이 모델은 사고(thinking)와 비사고(non-thinking) 두 가지 변형이 있으며, 사고 기능의 활성화 여부에 따라 출력 비용이 크게 달라집니다. 기본 변형(:thinking 접미사 없이)을 선택하면 사고 토큰 생성을 명시적으로 피합니다.\n\n사고 기능을 사용하고 사고 토큰을 받으려면 ':thinking' 변형을 선택해야 하며, 이 경우 사고 출력 비용이 더 높습니다.\n\nGemini 2.5 Flash는 문서에 명시된 'max reasoning tokens' 매개변수를 통해 구성할 수 있습니다 (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).",
|
|
587
|
+
"google/gemini-2.5-flash-preview:thinking.description": "Gemini 2.5 Flash는 고급 추론, 코딩, 수학, 과학 작업을 위해 설계된 Google의 최첨단 대표 모델입니다. 내장된 '사고(thinking)' 기능을 통해 더 높은 정확도의 응답과 정교한 컨텍스트 처리를 제공합니다.\n\n참고: 이 모델은 사고(thinking)와 비사고(non-thinking) 두 가지 변형이 있으며, 사고 기능의 활성화 여부에 따라 출력 비용이 크게 달라집니다. 기본 변형(:thinking 접미사 없이)을 선택하면 사고 토큰 생성을 명시적으로 피합니다.\n\n사고 기능을 사용하고 사고 토큰을 받으려면 ':thinking' 변형을 선택해야 하며, 이 경우 사고 출력 비용이 더 높습니다.\n\nGemini 2.5 Flash는 문서에 명시된 'max reasoning tokens' 매개변수를 통해 구성할 수 있습니다 (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).",
|
|
588
|
+
"google/gemini-2.5-flash.description": "Gemini 2.5 Flash (Lite/Pro/Flash)는 Google의 저지연부터 고성능 추론까지 아우르는 모델군입니다.",
|
|
589
|
+
"google/gemini-2.5-pro-free.description": "Gemini 2.5 Pro 무료 등급은 제한된 할당량의 멀티모달 장문 컨텍스트를 제공하며, 시험 사용 및 가벼운 워크플로우에 적합합니다.",
|
|
590
|
+
"google/gemini-2.5-pro-preview.description": "Gemini 2.5 Pro Preview는 복잡한 코드, 수학, STEM 문제에 대한 추론과 대규모 데이터셋, 코드베이스, 문서 분석을 위한 Google의 가장 진보된 사고 모델입니다.",
|
|
591
|
+
"google/gemini-2.5-pro.description": "Gemini 2.5 Pro는 복잡한 작업을 위한 장문 컨텍스트 지원을 갖춘 Google의 대표 추론 모델입니다.",
|
|
592
|
+
"google/gemini-3-pro-image-preview-free.description": "Gemini 3 Pro Image 무료 등급은 제한된 할당량의 멀티모달 생성을 지원합니다.",
|
|
593
|
+
"google/gemini-3-pro-image-preview.description": "Gemini 3 Pro Image (Nano Banana Pro)는 Google의 이미지 생성 모델로, 멀티모달 대화를 지원합니다.",
|
|
594
|
+
"google/gemini-3-pro-preview-free.description": "Gemini 3 Pro Preview Free는 표준 버전과 동일한 멀티모달 이해 및 추론 기능을 제공하지만, 할당량 및 속도 제한이 있어 시험 사용 및 저빈도 사용에 적합합니다.",
|
|
595
|
+
"google/gemini-3-pro-preview.description": "Gemini 3 Pro는 Gemini 시리즈의 차세대 멀티모달 추론 모델로, 텍스트, 오디오, 이미지, 비디오를 이해하며 복잡한 작업과 대규모 코드베이스를 처리할 수 있습니다.",
|
|
596
|
+
"google/gemini-embedding-001.description": "영어, 다국어, 코드 작업에서 뛰어난 성능을 보이는 최첨단 임베딩 모델입니다.",
|
|
597
|
+
"google/gemini-flash-1.5.description": "Gemini 1.5 Flash는 다양한 복잡한 작업을 위한 최적화된 멀티모달 처리를 제공합니다.",
|
|
598
|
+
"google/gemini-pro-1.5.description": "Gemini 1.5 Pro는 멀티모달 데이터 처리를 보다 효율적으로 수행하기 위한 최신 최적화를 결합한 모델입니다.",
|
|
599
|
+
"google/gemma-2-27b-it.description": "Gemma 2 27B는 다양한 시나리오에서 강력한 성능을 발휘하는 범용 대형 언어 모델입니다.",
|
|
600
|
+
"google/gemma-2-27b.description": "Gemma 2는 소형 앱부터 복잡한 데이터 처리까지 다양한 사용 사례에 적합한 Google의 효율적인 모델군입니다.",
|
|
601
|
+
"google/gemma-2-2b-it.description": "엣지 애플리케이션을 위해 설계된 고급 소형 언어 모델입니다.",
|
|
602
|
+
"google/gemma-2-9b-it.description": "Google이 개발한 Gemma 2 9B는 효율적인 명령 수행과 전반적인 우수한 성능을 제공합니다.",
|
|
603
|
+
"google/gemma-2-9b-it:free.description": "Gemma 2는 Google의 경량 오픈소스 텍스트 모델군입니다.",
|
|
604
|
+
"google/gemma-2-9b.description": "Gemma 2는 소형 앱부터 복잡한 데이터 처리까지 다양한 사용 사례에 적합한 Google의 효율적인 모델군입니다.",
|
|
605
|
+
"google/gemma-2b-it.description": "Gemma Instruct (2B)는 경량 애플리케이션을 위한 기본 명령 처리 기능을 제공합니다.",
|
|
606
|
+
"google/gemma-3-12b-it.description": "Gemma 3 12B는 효율성과 성능 면에서 새로운 기준을 제시하는 Google의 오픈소스 언어 모델입니다.",
|
|
607
|
+
"google/gemma-3-27b-it.description": "Gemma 3 27B는 효율성과 성능 면에서 새로운 기준을 제시하는 Google의 오픈소스 언어 모델입니다.",
|
|
608
|
+
"google/text-embedding-005.description": "코드 및 영어 작업에 최적화된 영어 중심 텍스트 임베딩 모델입니다.",
|
|
609
|
+
"google/text-multilingual-embedding-002.description": "다양한 언어 간 작업에 최적화된 다국어 텍스트 임베딩 모델입니다.",
|
|
610
|
+
"gpt-3.5-turbo-0125.description": "텍스트 생성 및 이해를 위한 GPT 3.5 Turbo; 현재 gpt-3.5-turbo-0125를 가리킵니다.",
|
|
611
|
+
"gpt-3.5-turbo-1106.description": "텍스트 생성 및 이해를 위한 GPT 3.5 Turbo; 현재 gpt-3.5-turbo-0125를 가리킵니다.",
|
|
612
|
+
"gpt-3.5-turbo-instruct.description": "명령 수행에 최적화된 텍스트 생성 및 이해용 GPT 3.5 Turbo.",
|
|
613
|
+
"gpt-3.5-turbo.description": "텍스트 생성 및 이해를 위한 GPT 3.5 Turbo; 현재 gpt-3.5-turbo-0125를 가리킵니다.",
|
|
614
|
+
"gpt-35-turbo-16k.description": "GPT-3.5 Turbo 16k는 복잡한 작업을 위한 고용량 텍스트 생성 모델입니다.",
|
|
615
|
+
"gpt-35-turbo.description": "GPT-3.5 Turbo는 OpenAI의 효율적인 채팅 및 텍스트 생성 모델로, 병렬 함수 호출을 지원합니다.",
|
|
616
|
+
"gpt-4-0125-preview.description": "최신 GPT-4 Turbo는 비전 기능을 추가했습니다. 시각적 요청은 JSON 모드 및 함수 호출을 지원합니다. 정확성과 효율성의 균형을 갖춘 실시간 애플리케이션용 비용 효율적인 멀티모달 모델입니다.",
|
|
617
|
+
"gpt-4-0613.description": "GPT-4는 더 긴 입력을 처리할 수 있는 확장된 컨텍스트 윈도우를 제공하여, 광범위한 정보 통합 및 데이터 분석에 적합합니다.",
|
|
618
|
+
"gpt-4-1106-preview.description": "최신 GPT-4 Turbo는 비전 기능을 추가했습니다. 시각적 요청은 JSON 모드 및 함수 호출을 지원합니다. 정확성과 효율성의 균형을 갖춘 실시간 애플리케이션용 비용 효율적인 멀티모달 모델입니다.",
|
|
619
|
+
"gpt-4-32k-0613.description": "GPT-4는 더 긴 입력을 처리할 수 있는 확장된 컨텍스트 윈도우를 제공하여, 광범위한 정보 통합 및 데이터 분석에 적합합니다.",
|
|
620
|
+
"gpt-4-32k.description": "GPT-4는 더 긴 입력을 처리할 수 있는 확장된 컨텍스트 윈도우를 제공하여, 광범위한 정보 통합 및 데이터 분석에 적합합니다.",
|
|
621
|
+
"gpt-4-turbo-2024-04-09.description": "최신 GPT-4 Turbo는 비전 기능을 추가했습니다. 시각적 요청은 JSON 모드 및 함수 호출을 지원합니다. 정확성과 효율성의 균형을 갖춘 실시간 애플리케이션용 비용 효율적인 멀티모달 모델입니다.",
|
|
622
|
+
"gpt-4-turbo-preview.description": "최신 GPT-4 Turbo는 비전 기능을 추가했습니다. 시각적 요청은 JSON 모드 및 함수 호출을 지원합니다. 정확성과 효율성의 균형을 갖춘 실시간 애플리케이션용 비용 효율적인 멀티모달 모델입니다.",
|
|
623
|
+
"gpt-4-turbo.description": "최신 GPT-4 Turbo는 비전 기능을 추가했습니다. 시각적 요청은 JSON 모드 및 함수 호출을 지원합니다. 정확성과 효율성의 균형을 갖춘 실시간 애플리케이션용 비용 효율적인 멀티모달 모델입니다.",
|
|
624
|
+
"gpt-4-vision-preview.description": "이미지 분석 및 처리 작업을 위한 GPT-4 Vision 프리뷰 모델입니다.",
|
|
625
|
+
"gpt-4.1-mini.description": "GPT-4.1 mini는 지능, 속도, 비용의 균형을 이루어 다양한 사용 사례에 적합합니다.",
|
|
626
|
+
"gpt-4.1-nano.description": "GPT-4.1 nano는 가장 빠르고 비용 효율적인 GPT-4.1 모델입니다.",
|
|
627
|
+
"gpt-4.1.description": "GPT-4.1은 복잡한 작업과 도메인 간 문제 해결을 위한 대표 모델입니다.",
|
|
628
|
+
"gpt-4.5-preview.description": "GPT-4.5-preview는 세계 지식과 의도 이해 능력이 향상된 최신 범용 모델로, 창의적 작업과 에이전트 계획에 강합니다. 지식 기준일: 2023년 10월.",
|
|
629
|
+
"gpt-4.description": "GPT-4는 더 긴 입력을 처리할 수 있는 확장된 컨텍스트 윈도우를 제공하여, 광범위한 정보 통합 및 데이터 분석에 적합합니다.",
|
|
630
|
+
"gpt-4o-2024-05-13.description": "ChatGPT-4o는 실시간으로 업데이트되는 동적 모델로, 고객 지원, 교육, 기술 지원 등 대규모 사용 사례에 적합한 강력한 이해 및 생성 능력을 결합합니다.",
|
|
576
631
|
"meta/Llama-3.2-90B-Vision-Instruct.description": "시각 이해 에이전트 애플리케이션을 위한 고급 이미지 추론 기능.",
|
|
577
632
|
"meta/Llama-3.3-70B-Instruct.description": "Llama 3.3은 가장 진보된 다국어 오픈소스 Llama 모델로, 매우 낮은 비용으로 405B에 근접한 성능을 제공합니다. Transformer 기반이며, 유용성과 안전성을 위해 SFT 및 RLHF로 개선되었습니다. 명령어 튜닝 버전은 다국어 채팅에 최적화되어 있으며, 산업 벤치마크에서 많은 오픈 및 클로즈드 채팅 모델을 능가합니다. 지식 기준일: 2023년 12월.",
|
|
578
633
|
"meta/Meta-Llama-3-70B-Instruct.description": "추론, 코딩, 다양한 언어 작업에 뛰어난 성능을 보이는 강력한 70B 파라미터 모델.",
|
package/locales/nl-NL/chat.json
CHANGED
|
@@ -204,6 +204,7 @@
|
|
|
204
204
|
"noSelectedAgents": "Nog geen leden geselecteerd",
|
|
205
205
|
"openInNewWindow": "Openen in nieuw venster",
|
|
206
206
|
"operation.execAgentRuntime": "Reactie voorbereiden",
|
|
207
|
+
"operation.execClientTask": "Taak wordt uitgevoerd",
|
|
207
208
|
"operation.sendMessage": "Bericht verzenden",
|
|
208
209
|
"owner": "Groepseigenaar",
|
|
209
210
|
"pageCopilot.title": "Pagina-agent",
|
|
@@ -322,11 +323,15 @@
|
|
|
322
323
|
"tab.profile": "Agentprofiel",
|
|
323
324
|
"tab.search": "Zoeken",
|
|
324
325
|
"task.activity.calling": "Skill wordt aangeroepen...",
|
|
326
|
+
"task.activity.clientExecuting": "Lokaal uitvoeren...",
|
|
325
327
|
"task.activity.generating": "Reactie wordt gegenereerd...",
|
|
326
328
|
"task.activity.gotResult": "Toolresultaat ontvangen",
|
|
327
329
|
"task.activity.toolCalling": "{{toolName}} wordt aangeroepen...",
|
|
328
330
|
"task.activity.toolResult": "{{toolName}} resultaat ontvangen",
|
|
329
331
|
"task.batchTasks": "{{count}} batch-subtaken",
|
|
332
|
+
"task.instruction": "Taakinstructie",
|
|
333
|
+
"task.intermediateSteps": "{{count}} tussenstappen",
|
|
334
|
+
"task.metrics.duration": "(duurde {{duration}})",
|
|
330
335
|
"task.metrics.stepsShort": "stappen",
|
|
331
336
|
"task.metrics.toolCallsShort": "toolgebruik",
|
|
332
337
|
"task.status.cancelled": "Taak geannuleerd",
|
|
@@ -58,6 +58,7 @@
|
|
|
58
58
|
"screen4.title": "Hoe wil je gegevens delen?",
|
|
59
59
|
"screen4.title2": "Jouw keuze helpt ons verbeteren",
|
|
60
60
|
"screen4.title3": "Je kunt dit op elk moment wijzigen in de instellingen",
|
|
61
|
+
"screen5.actions.cancel": "Annuleren",
|
|
61
62
|
"screen5.actions.connectToServer": "Verbinden met server",
|
|
62
63
|
"screen5.actions.connecting": "Bezig met verbinden...",
|
|
63
64
|
"screen5.actions.signInCloud": "Aanmelden bij LobeHub Cloud",
|
|
@@ -65,6 +66,10 @@
|
|
|
65
66
|
"screen5.actions.signingIn": "Bezig met aanmelden...",
|
|
66
67
|
"screen5.actions.signingOut": "Bezig met afmelden...",
|
|
67
68
|
"screen5.actions.tryAgain": "Opnieuw proberen",
|
|
69
|
+
"screen5.auth.phase.browserOpened": "Browser geopend, log in alstublieft...",
|
|
70
|
+
"screen5.auth.phase.verifying": "Inloggegevens worden geverifieerd...",
|
|
71
|
+
"screen5.auth.phase.waitingForAuth": "Wachten op autorisatie...",
|
|
72
|
+
"screen5.auth.remaining": "Resterend: {{time}}s",
|
|
68
73
|
"screen5.badge": "Aanmelden",
|
|
69
74
|
"screen5.description": "Meld je aan om Agenten, Groepen, instellingen en Context te synchroniseren op al je apparaten.",
|
|
70
75
|
"screen5.errors.desktopOnlyOidc": "OIDC-autorisatie is alleen beschikbaar in de desktop-app.",
|
|
@@ -141,6 +141,16 @@
|
|
|
141
141
|
"filterBy.timePeriod.year": "Laatste jaar",
|
|
142
142
|
"footer.desc": "Ontwikkel mee met AI-gebruikers wereldwijd. Word een maker en dien je agents en vaardigheden in bij de LobeHub-community.",
|
|
143
143
|
"footer.title": "Deel vandaag nog je creatie in de LobeHub-community",
|
|
144
|
+
"fork.alreadyForked": "Je hebt deze agent al geforkt. Navigeren naar je fork...",
|
|
145
|
+
"fork.failed": "Forken is mislukt. Probeer het opnieuw.",
|
|
146
|
+
"fork.forkAndChat": "Fork & Chat",
|
|
147
|
+
"fork.forkedFrom": "Geforkt van",
|
|
148
|
+
"fork.forks": "forks",
|
|
149
|
+
"fork.forksCount": "{{count}} fork",
|
|
150
|
+
"fork.forksCount_other": "{{count}} forks",
|
|
151
|
+
"fork.success": "Succesvol geforkt!",
|
|
152
|
+
"fork.viewAllForks": "Bekijk alle forks",
|
|
153
|
+
"groupAgents.tag": "Groep",
|
|
144
154
|
"home.communityAgents": "Community Agents",
|
|
145
155
|
"home.featuredAssistants": "Uitgelichte Agents",
|
|
146
156
|
"home.featuredModels": "Uitgelichte Modellen",
|
|
@@ -471,14 +481,19 @@
|
|
|
471
481
|
"user.follow": "Volgen",
|
|
472
482
|
"user.followers": "Volgers",
|
|
473
483
|
"user.following": "Volgend",
|
|
484
|
+
"user.forkedAgentGroups": "Geforkte Agentgroepen",
|
|
485
|
+
"user.forkedAgents": "Geforkte Agents",
|
|
474
486
|
"user.login": "Word een Maker",
|
|
475
487
|
"user.logout": "Uitloggen",
|
|
476
488
|
"user.myProfile": "Mijn Profiel",
|
|
477
489
|
"user.noAgents": "Deze gebruiker heeft nog geen Agents gepubliceerd",
|
|
478
490
|
"user.noFavoriteAgents": "Nog geen opgeslagen Agents",
|
|
479
491
|
"user.noFavoritePlugins": "Nog geen opgeslagen Vaardigheden",
|
|
492
|
+
"user.noForkedAgentGroups": "Nog geen geforkte agentgroepen",
|
|
493
|
+
"user.noForkedAgents": "Nog geen geforkte agents",
|
|
480
494
|
"user.publishedAgents": "Gemaakte Agents",
|
|
481
495
|
"user.tabs.favorites": "Favorieten",
|
|
496
|
+
"user.tabs.forkedAgents": "Geforkt",
|
|
482
497
|
"user.tabs.publishedAgents": "Gemaakt",
|
|
483
498
|
"user.unfavorite": "Verwijderen uit opgeslagen",
|
|
484
499
|
"user.unfavoriteFailed": "Verwijderen mislukt",
|
|
@@ -643,6 +643,40 @@
|
|
|
643
643
|
"gpt-4o-search-preview.description": "GPT-4o Search Preview is getraind om webzoekopdrachten te begrijpen en uit te voeren via de Chat Completions API. Webzoekopdrachten worden per toolaanroep gefactureerd, naast tokenkosten.",
|
|
644
644
|
"gpt-4o-transcribe.description": "GPT-4o Transcribe is een spraak-naar-tekstmodel dat audio transcribeert met GPT-4o, met verbeterde woordherkenning, taalidentificatie en nauwkeurigheid ten opzichte van het oorspronkelijke Whisper-model.",
|
|
645
645
|
"gpt-4o.description": "ChatGPT-4o is een dynamisch model dat realtime wordt bijgewerkt en sterke taalbegrip- en generatiecapaciteiten combineert voor grootschalige toepassingen zoals klantenservice, onderwijs en technische ondersteuning.",
|
|
646
|
+
"gpt-5-chat-latest.description": "Het GPT-5-model dat wordt gebruikt in ChatGPT, combineert sterk begrip en generatie voor conversatietoepassingen.",
|
|
647
|
+
"gpt-5-chat.description": "GPT-5 Chat is een previewmodel geoptimaliseerd voor conversatiescenario's. Het ondersteunt tekst- en afbeeldingsinvoer, geeft alleen tekstuitvoer en is geschikt voor chatbots en conversatie-AI-toepassingen.",
|
|
648
|
+
"gpt-5-codex.description": "GPT-5 Codex is een GPT-5-variant geoptimaliseerd voor agentmatige programmeertaken in Codex-achtige omgevingen.",
|
|
649
|
+
"gpt-5-mini.description": "Een snellere, kostenefficiëntere GPT-5-variant voor goed gedefinieerde taken, met snelle reacties zonder kwaliteitsverlies.",
|
|
650
|
+
"gpt-5-nano.description": "De snelste en meest kosteneffectieve GPT-5-variant, ideaal voor toepassingen met strikte eisen op het gebied van latentie en kosten.",
|
|
651
|
+
"gpt-5-pro.description": "GPT-5 Pro gebruikt meer rekenkracht om dieper na te denken en consequent betere antwoorden te leveren.",
|
|
652
|
+
"gpt-5.1-chat-latest.description": "GPT-5.1 Chat: de ChatGPT-variant van GPT-5.1, ontworpen voor chatscenario's.",
|
|
653
|
+
"gpt-5.1-codex-mini.description": "GPT-5.1 Codex mini: een kleinere, goedkopere Codex-variant geoptimaliseerd voor agentmatige programmeertaken.",
|
|
654
|
+
"gpt-5.1-codex.description": "GPT-5.1 Codex: een GPT-5.1-variant geoptimaliseerd voor agentmatige programmeertaken, geschikt voor complexe code-/agentworkflows in de Responses API.",
|
|
655
|
+
"gpt-5.1.description": "GPT-5.1 — een toonaangevend model geoptimaliseerd voor programmeren en agenttaken met configureerbare redeneercapaciteit en langere context.",
|
|
656
|
+
"gpt-5.2-chat-latest.description": "GPT-5.2 Chat is de ChatGPT-variant (chat-latest) om de nieuwste verbeteringen in conversatie te ervaren.",
|
|
657
|
+
"gpt-5.2-pro.description": "GPT-5.2 Pro: een slimmer, preciezer GPT-5.2-model (alleen via Responses API), geschikt voor moeilijkere problemen en langere redeneerketens.",
|
|
658
|
+
"gpt-5.2.description": "GPT-5.2 is een toonaangevend model voor programmeren en agentworkflows met sterkere redeneercapaciteit en prestaties bij lange contexten.",
|
|
659
|
+
"gpt-5.description": "Het beste model voor domeinoverstijgende programmeer- en agenttaken. GPT-5 biedt sprongen in nauwkeurigheid, snelheid, redeneervermogen, contextbewustzijn, gestructureerd denken en probleemoplossing.",
|
|
660
|
+
"gpt-audio.description": "GPT Audio is een algemeen chatmodel voor audio-invoer/-uitvoer, ondersteund in de Chat Completions API.",
|
|
661
|
+
"gpt-image-1-mini.description": "Een goedkopere GPT Image 1-variant met native tekst- en afbeeldingsinvoer en afbeeldingsuitvoer.",
|
|
662
|
+
"gpt-image-1.5.description": "Een verbeterd GPT Image 1-model met 4× snellere generatie, nauwkeurigere bewerkingen en verbeterde tekstrendering.",
|
|
663
|
+
"gpt-image-1.description": "ChatGPT's native multimodale afbeeldingsgeneratiemodel.",
|
|
664
|
+
"gpt-oss-120b.description": "Toegang vereist een aanvraag. GPT-OSS-120B is een open-source groot taalmodel van OpenAI met sterke tekstgeneratiecapaciteit.",
|
|
665
|
+
"gpt-oss-20b.description": "Toegang vereist een aanvraag. GPT-OSS-20B is een open-source middelgroot taalmodel van OpenAI met efficiënte tekstgeneratie.",
|
|
666
|
+
"gpt-oss:120b.description": "GPT-OSS 120B is OpenAI’s grote open-source LLM met MXFP4-kwantisatie en gepositioneerd als vlaggenschipmodel. Het vereist multi-GPU- of high-end werkstationomgevingen en levert uitstekende prestaties in complexe redenering, codegeneratie en meertalige verwerking, met geavanceerde functieaanroepen en toolintegratie.",
|
|
667
|
+
"gpt-oss:20b.description": "GPT-OSS 20B is een open-source LLM van OpenAI met MXFP4-kwantisatie, geschikt voor high-end consument-GPU’s of Apple Silicon Macs. Het presteert goed in dialooggeneratie, programmeren en redeneertaken, en ondersteunt functieaanroepen en toolgebruik.",
|
|
668
|
+
"gpt-realtime.description": "Een algemeen realtime model dat realtime tekst- en audio-invoer/-uitvoer ondersteunt, plus afbeeldingsinvoer.",
|
|
669
|
+
"grok-2-image-1212.description": "Ons nieuwste afbeeldingsgeneratiemodel creëert levendige, realistische beelden op basis van prompts en blinkt uit in marketing-, sociale media- en entertainmenttoepassingen.",
|
|
670
|
+
"grok-2-vision-1212.description": "Verbeterde nauwkeurigheid, instructieopvolging en meertalige mogelijkheden.",
|
|
671
|
+
"grok-3-mini.description": "Een lichtgewicht model dat eerst nadenkt voordat het antwoordt. Snel en slim voor logische taken die geen diepgaande domeinkennis vereisen, met toegang tot ruwe redeneertracering.",
|
|
672
|
+
"grok-3.description": "Een vlaggenschipmodel dat uitblinkt in zakelijke toepassingen zoals gegevensextractie, programmeren en samenvatten, met diepgaande domeinkennis in financiën, gezondheidszorg, recht en wetenschap.",
|
|
673
|
+
"grok-4-0709.description": "xAI’s Grok 4 met sterke redeneercapaciteit.",
|
|
674
|
+
"grok-4-1-fast-non-reasoning.description": "Een grensverleggend multimodaal model geoptimaliseerd voor krachtige agenttoolgebruik.",
|
|
675
|
+
"grok-4-1-fast-reasoning.description": "Een grensverleggend multimodaal model geoptimaliseerd voor krachtige agenttoolgebruik.",
|
|
676
|
+
"grok-4-fast-non-reasoning.description": "We zijn verheugd Grok 4 Fast uit te brengen, onze nieuwste vooruitgang in kosteneffectieve redeneermodellen.",
|
|
677
|
+
"grok-4-fast-reasoning.description": "We zijn verheugd Grok 4 Fast uit te brengen, onze nieuwste vooruitgang in kosteneffectieve redeneermodellen.",
|
|
678
|
+
"grok-4.description": "Ons nieuwste en krachtigste vlaggenschipmodel, uitblinkend in NLP, wiskunde en redenering—een ideale allrounder.",
|
|
679
|
+
"grok-code-fast-1.description": "We zijn verheugd om grok-code-fast-1 te lanceren, een snel en kosteneffectief redeneermodel dat uitblinkt in agentmatig programmeren.",
|
|
646
680
|
"meta.llama3-8b-instruct-v1:0.description": "Meta Llama 3 is een open LLM voor ontwikkelaars, onderzoekers en bedrijven, ontworpen om hen te helpen bij het bouwen, experimenteren en verantwoord opschalen van generatieve AI-ideeën. Als onderdeel van de basis voor wereldwijde gemeenschapsinnovatie is het goed geschikt voor beperkte rekenkracht en middelen, edge-apparaten en snellere trainingstijden.",
|
|
647
681
|
"meta/Llama-3.2-11B-Vision-Instruct.description": "Sterke beeldredenering op afbeeldingen met hoge resolutie, geschikt voor toepassingen voor visueel begrip.",
|
|
648
682
|
"meta/Llama-3.2-90B-Vision-Instruct.description": "Geavanceerde beeldredenering voor toepassingen met visueel begrip en agentfunctionaliteit.",
|
package/locales/pl-PL/chat.json
CHANGED
|
@@ -204,6 +204,7 @@
|
|
|
204
204
|
"noSelectedAgents": "Nie wybrano jeszcze członków",
|
|
205
205
|
"openInNewWindow": "Otwórz w nowym oknie",
|
|
206
206
|
"operation.execAgentRuntime": "Przygotowywanie odpowiedzi",
|
|
207
|
+
"operation.execClientTask": "Wykonywanie zadania",
|
|
207
208
|
"operation.sendMessage": "Wysyłanie wiadomości",
|
|
208
209
|
"owner": "Właściciel grupy",
|
|
209
210
|
"pageCopilot.title": "Agent strony",
|
|
@@ -322,11 +323,15 @@
|
|
|
322
323
|
"tab.profile": "Profil agenta",
|
|
323
324
|
"tab.search": "Szukaj",
|
|
324
325
|
"task.activity.calling": "Wywoływanie umiejętności...",
|
|
326
|
+
"task.activity.clientExecuting": "Wykonywanie lokalnie...",
|
|
325
327
|
"task.activity.generating": "Generowanie odpowiedzi...",
|
|
326
328
|
"task.activity.gotResult": "Otrzymano wynik narzędzia",
|
|
327
329
|
"task.activity.toolCalling": "Wywoływanie {{toolName}}...",
|
|
328
330
|
"task.activity.toolResult": "Otrzymano wynik {{toolName}}",
|
|
329
331
|
"task.batchTasks": "{{count}} zadań zbiorczych",
|
|
332
|
+
"task.instruction": "Instrukcja zadania",
|
|
333
|
+
"task.intermediateSteps": "{{count}} kroków pośrednich",
|
|
334
|
+
"task.metrics.duration": "(zajęło {{duration}})",
|
|
330
335
|
"task.metrics.stepsShort": "kroków",
|
|
331
336
|
"task.metrics.toolCallsShort": "użycia narzędzi",
|
|
332
337
|
"task.status.cancelled": "Zadanie anulowane",
|
|
@@ -58,6 +58,7 @@
|
|
|
58
58
|
"screen4.title": "Jak chcesz udostępniać dane?",
|
|
59
59
|
"screen4.title2": "Twój wybór pomaga nam się rozwijać",
|
|
60
60
|
"screen4.title3": "Możesz to zmienić w każdej chwili w ustawieniach",
|
|
61
|
+
"screen5.actions.cancel": "Anuluj",
|
|
61
62
|
"screen5.actions.connectToServer": "Połącz z serwerem",
|
|
62
63
|
"screen5.actions.connecting": "Łączenie...",
|
|
63
64
|
"screen5.actions.signInCloud": "Zaloguj się do LobeHub Cloud",
|
|
@@ -65,6 +66,10 @@
|
|
|
65
66
|
"screen5.actions.signingIn": "Logowanie...",
|
|
66
67
|
"screen5.actions.signingOut": "Wylogowywanie...",
|
|
67
68
|
"screen5.actions.tryAgain": "Spróbuj ponownie",
|
|
69
|
+
"screen5.auth.phase.browserOpened": "Przeglądarka została otwarta, zaloguj się...",
|
|
70
|
+
"screen5.auth.phase.verifying": "Weryfikowanie danych logowania...",
|
|
71
|
+
"screen5.auth.phase.waitingForAuth": "Oczekiwanie na autoryzację...",
|
|
72
|
+
"screen5.auth.remaining": "Pozostało: {{time}}s",
|
|
68
73
|
"screen5.badge": "Logowanie",
|
|
69
74
|
"screen5.description": "Zaloguj się, aby synchronizować Agentów, Grupy, ustawienia i Kontekst na wszystkich urządzeniach.",
|
|
70
75
|
"screen5.errors.desktopOnlyOidc": "Autoryzacja OIDC jest dostępna tylko w wersji desktopowej aplikacji.",
|
|
@@ -141,6 +141,16 @@
|
|
|
141
141
|
"filterBy.timePeriod.year": "Ostatni rok",
|
|
142
142
|
"footer.desc": "Rozwijaj się z użytkownikami AI na całym świecie. Zostań twórcą i przesyłaj swoich agentów i umiejętności do Społeczności LobeHub.",
|
|
143
143
|
"footer.title": "Podziel się swoją twórczością w Społeczności LobeHub już dziś",
|
|
144
|
+
"fork.alreadyForked": "Już rozwidliłeś tego agenta. Przechodzenie do Twojego forka...",
|
|
145
|
+
"fork.failed": "Nie udało się rozwidlić. Spróbuj ponownie.",
|
|
146
|
+
"fork.forkAndChat": "Rozwidlij i czatuj",
|
|
147
|
+
"fork.forkedFrom": "Rozwidlono z",
|
|
148
|
+
"fork.forks": "rozwidlenia",
|
|
149
|
+
"fork.forksCount": "{{count}} rozwidlenie",
|
|
150
|
+
"fork.forksCount_other": "{{count}} rozwidleń",
|
|
151
|
+
"fork.success": "Pomyślnie rozwidlono!",
|
|
152
|
+
"fork.viewAllForks": "Zobacz wszystkie rozwidlenia",
|
|
153
|
+
"groupAgents.tag": "Grupa",
|
|
144
154
|
"home.communityAgents": "Agenci Społeczności",
|
|
145
155
|
"home.featuredAssistants": "Polecani Agenci",
|
|
146
156
|
"home.featuredModels": "Polecane Modele",
|
|
@@ -471,14 +481,19 @@
|
|
|
471
481
|
"user.follow": "Obserwuj",
|
|
472
482
|
"user.followers": "Obserwujący",
|
|
473
483
|
"user.following": "Obserwowani",
|
|
484
|
+
"user.forkedAgentGroups": "Rozwidlone grupy agentów",
|
|
485
|
+
"user.forkedAgents": "Rozwidleni agenci",
|
|
474
486
|
"user.login": "Zostań Twórcą",
|
|
475
487
|
"user.logout": "Wyloguj",
|
|
476
488
|
"user.myProfile": "Mój Profil",
|
|
477
489
|
"user.noAgents": "Ten użytkownik nie opublikował jeszcze żadnych agentów",
|
|
478
490
|
"user.noFavoriteAgents": "Brak zapisanych agentów",
|
|
479
491
|
"user.noFavoritePlugins": "Brak zapisanych umiejętności",
|
|
492
|
+
"user.noForkedAgentGroups": "Brak rozwidlonych grup agentów",
|
|
493
|
+
"user.noForkedAgents": "Brak rozwidlonych agentów",
|
|
480
494
|
"user.publishedAgents": "Utworzeni Agenci",
|
|
481
495
|
"user.tabs.favorites": "Ulubione",
|
|
496
|
+
"user.tabs.forkedAgents": "Rozwidleni",
|
|
482
497
|
"user.tabs.publishedAgents": "Utworzone",
|
|
483
498
|
"user.unfavorite": "Usuń z ulubionych",
|
|
484
499
|
"user.unfavoriteFailed": "Nie udało się usunąć z ulubionych",
|
|
@@ -565,6 +565,37 @@
|
|
|
565
565
|
"glm-4.7.description": "GLM-4.7 to najnowszy flagowy model Zhipu, ulepszony pod kątem scenariuszy Agentic Coding z lepszymi możliwościami kodowania, planowaniem długoterminowym i współpracą z narzędziami. Osiąga czołowe wyniki wśród modeli open-source w wielu publicznych benchmarkach. Ogólne możliwości zostały ulepszone dzięki bardziej zwięzłym i naturalnym odpowiedziom oraz bardziej wciągającemu stylowi pisania. W przypadku złożonych zadań agenta, przestrzeganie instrukcji podczas wywołań narzędzi jest silniejsze, a estetyka interfejsu i efektywność realizacji zadań długoterminowych w Artifacts i Agentic Coding zostały dodatkowo ulepszone.",
|
|
566
566
|
"glm-4.description": "GLM-4 to starszy flagowy model wydany w styczniu 2024 r., obecnie zastąpiony przez silniejszy GLM-4-0520.",
|
|
567
567
|
"glm-4v-flash.description": "GLM-4V-Flash koncentruje się na efektywnym rozumieniu pojedynczych obrazów w scenariuszach szybkiej analizy, takich jak przetwarzanie obrazów w czasie rzeczywistym lub wsadowo.",
|
|
568
|
+
"glm-4v-plus-0111.description": "GLM-4V-Plus rozumie wideo i wiele obrazów, idealny do zadań multimodalnych.",
|
|
569
|
+
"glm-4v-plus.description": "GLM-4V-Plus rozumie wideo i wiele obrazów, idealny do zadań multimodalnych.",
|
|
570
|
+
"glm-4v.description": "GLM-4V zapewnia zaawansowane rozumienie obrazów i wnioskowanie w zadaniach wizualnych.",
|
|
571
|
+
"glm-z1-air.description": "Model wnioskowania o wysokiej zdolności do głębokiej analizy i dedukcji.",
|
|
572
|
+
"glm-z1-airx.description": "Ultraszybkie wnioskowanie przy zachowaniu wysokiej jakości rozumowania.",
|
|
573
|
+
"glm-z1-flash.description": "Seria GLM-Z1 oferuje zaawansowane wnioskowanie z naciskiem na logikę, matematykę i programowanie.",
|
|
574
|
+
"glm-z1-flashx.description": "Szybki i ekonomiczny: ulepszony Flash z ultraszybkim wnioskowaniem i wyższą równoległością.",
|
|
575
|
+
"glm-zero-preview.description": "GLM-Zero-Preview zapewnia zaawansowane wnioskowanie, wyróżniając się w logice, matematyce i programowaniu.",
|
|
576
|
+
"global.anthropic.claude-opus-4-5-20251101-v1:0.description": "Claude Opus 4.5 to flagowy model firmy Anthropic, łączący wyjątkową inteligencję i skalowalną wydajność w złożonych zadaniach wymagających najwyższej jakości odpowiedzi i rozumowania.",
|
|
577
|
+
"google/gemini-2.0-flash-001.description": "Gemini 2.0 Flash oferuje nowej generacji możliwości, w tym doskonałą szybkość, natywne użycie narzędzi, generację multimodalną i kontekst do 1 miliona tokenów.",
|
|
578
|
+
"google/gemini-2.0-flash-exp:free.description": "Gemini 2.0 Flash Experimental to najnowszy eksperymentalny model multimodalny Google z poprawioną jakością względem poprzednich wersji, szczególnie w zakresie wiedzy o świecie, kodu i długiego kontekstu.",
|
|
579
|
+
"google/gemini-2.0-flash-lite-001.description": "Gemini 2.0 Flash Lite to lekka wersja Gemini z domyślnie wyłączonym myśleniem, co poprawia opóźnienia i koszty; można je włączyć za pomocą parametrów.",
|
|
580
|
+
"google/gemini-2.0-flash-lite.description": "Gemini 2.0 Flash Lite oferuje funkcje nowej generacji, w tym wyjątkową szybkość, wbudowane narzędzia, generację multimodalną i kontekst do 1 miliona tokenów.",
|
|
581
|
+
"google/gemini-2.0-flash.description": "Gemini 2.0 Flash to wysokowydajny model wnioskowania Google do zaawansowanych zadań multimodalnych.",
|
|
582
|
+
"google/gemini-2.5-flash-image-free.description": "Gemini 2.5 Flash Image – darmowy poziom z ograniczonym limitem generacji multimodalnej.",
|
|
583
|
+
"google/gemini-2.5-flash-image-preview.description": "Eksperymentalny model Gemini 2.5 Flash z obsługą generowania obrazów.",
|
|
584
|
+
"google/gemini-2.5-flash-image.description": "Gemini 2.5 Flash Image (Nano Banana) to model generowania obrazów Google z obsługą konwersacji multimodalnych.",
|
|
585
|
+
"google/gemini-2.5-flash-lite.description": "Gemini 2.5 Flash Lite to lekka wersja Gemini 2.5 zoptymalizowana pod kątem opóźnień i kosztów, odpowiednia do scenariuszy o dużym przepływie danych.",
|
|
586
|
+
"google/gemini-2.5-flash-preview.description": "Gemini 2.5 Flash to najbardziej zaawansowany flagowy model Google, stworzony do zaawansowanego wnioskowania, kodowania, matematyki i nauk ścisłych. Zawiera wbudowane „myślenie” dla dokładniejszych odpowiedzi i lepszego przetwarzania kontekstu.\n\nUwaga: model ma dwie wersje — z myśleniem i bez. Ceny różnią się w zależności od wybranej wersji. Wybierając standardową wersję (bez sufiksu „:thinking”), model unika generowania tokenów myślenia.\n\nAby korzystać z myślenia i otrzymywać tokeny myślenia, należy wybrać wariant „:thinking”, który wiąże się z wyższymi kosztami.\n\nGemini 2.5 Flash można również skonfigurować za pomocą parametru „max reasoning tokens” zgodnie z dokumentacją (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).",
|
|
587
|
+
"google/gemini-2.5-flash-preview:thinking.description": "Gemini 2.5 Flash to najbardziej zaawansowany flagowy model Google, stworzony do zaawansowanego wnioskowania, kodowania, matematyki i nauk ścisłych. Zawiera wbudowane „myślenie” dla dokładniejszych odpowiedzi i lepszego przetwarzania kontekstu.\n\nUwaga: model ma dwie wersje — z myśleniem i bez. Ceny różnią się w zależności od wybranej wersji. Wybierając standardową wersję (bez sufiksu „:thinking”), model unika generowania tokenów myślenia.\n\nAby korzystać z myślenia i otrzymywać tokeny myślenia, należy wybrać wariant „:thinking”, który wiąże się z wyższymi kosztami.\n\nGemini 2.5 Flash można również skonfigurować za pomocą parametru „max reasoning tokens” zgodnie z dokumentacją (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).",
|
|
588
|
+
"google/gemini-2.5-flash.description": "Gemini 2.5 Flash (Lite/Pro/Flash) to rodzina modeli Google obejmująca od niskich opóźnień po wysokowydajne wnioskowanie.",
|
|
589
|
+
"google/gemini-2.5-pro-free.description": "Darmowa wersja Gemini 2.5 Pro z ograniczonym limitem multimodalnego długiego kontekstu, odpowiednia do testów i lekkich zadań.",
|
|
590
|
+
"google/gemini-2.5-pro-preview.description": "Gemini 2.5 Pro Preview to najbardziej zaawansowany model myślący Google do rozwiązywania złożonych problemów w kodzie, matematyce i naukach ścisłych oraz do analizy dużych zbiorów danych, baz kodu i dokumentów z długim kontekstem.",
|
|
591
|
+
"google/gemini-2.5-pro.description": "Gemini 2.5 Pro to flagowy model wnioskowania Google z obsługą długiego kontekstu do złożonych zadań.",
|
|
592
|
+
"google/gemini-3-pro-image-preview-free.description": "Gemini 3 Pro Image – darmowy poziom z ograniczonym limitem generacji multimodalnej.",
|
|
593
|
+
"google/gemini-3-pro-image-preview.description": "Gemini 3 Pro Image (Nano Banana Pro) to model generowania obrazów Google z obsługą konwersacji multimodalnych.",
|
|
594
|
+
"google/gemini-3-pro-preview-free.description": "Gemini 3 Pro Preview Free oferuje te same możliwości rozumienia i wnioskowania multimodalnego co wersja standardowa, ale z limitami, co czyni go odpowiednim do testów i rzadkich zastosowań.",
|
|
595
|
+
"google/gemini-3-pro-preview.description": "Gemini 3 Pro to model nowej generacji do wnioskowania multimodalnego w rodzinie Gemini, rozumiejący tekst, dźwięk, obrazy i wideo, obsługujący złożone zadania i duże bazy kodu.",
|
|
596
|
+
"google/gemini-embedding-001.description": "Nowoczesny model osadzania tekstu o wysokiej wydajności w języku angielskim, wielojęzycznym i kodzie.",
|
|
597
|
+
"google/gemini-flash-1.5.description": "Gemini 1.5 Flash zapewnia zoptymalizowane przetwarzanie multimodalne dla szerokiego zakresu złożonych zadań.",
|
|
598
|
+
"google/gemini-pro-1.5.description": "Gemini 1.5 Pro łączy najnowsze optymalizacje dla bardziej efektywnego przetwarzania danych multimodalnych.",
|
|
568
599
|
"meta.llama3-8b-instruct-v1:0.description": "Meta Llama 3 to otwarty model językowy (LLM) stworzony z myślą o programistach, naukowcach i przedsiębiorstwach, zaprojektowany, by wspierać ich w budowaniu, eksperymentowaniu i odpowiedzialnym skalowaniu pomysłów z zakresu generatywnej sztucznej inteligencji. Jako fundament globalnej innowacji społecznościowej, doskonale sprawdza się przy ograniczonych zasobach obliczeniowych, na urządzeniach brzegowych oraz przy szybszym czasie trenowania.",
|
|
569
600
|
"meta/Llama-3.2-11B-Vision-Instruct.description": "Zaawansowane rozumowanie obrazów w wysokiej rozdzielczości, idealne do aplikacji zrozumienia wizualnego.",
|
|
570
601
|
"meta/Llama-3.2-90B-Vision-Instruct.description": "Zaawansowane rozumowanie obrazów dla aplikacji agentów opartych na zrozumieniu wizualnym.",
|
package/locales/pt-BR/chat.json
CHANGED
|
@@ -204,6 +204,7 @@
|
|
|
204
204
|
"noSelectedAgents": "Nenhum membro selecionado ainda",
|
|
205
205
|
"openInNewWindow": "Abrir em Nova Janela",
|
|
206
206
|
"operation.execAgentRuntime": "Preparando resposta",
|
|
207
|
+
"operation.execClientTask": "Executando tarefa",
|
|
207
208
|
"operation.sendMessage": "Enviando mensagem",
|
|
208
209
|
"owner": "Proprietário do Grupo",
|
|
209
210
|
"pageCopilot.title": "Agente da Página",
|
|
@@ -322,11 +323,15 @@
|
|
|
322
323
|
"tab.profile": "Perfil do Agente",
|
|
323
324
|
"tab.search": "Buscar",
|
|
324
325
|
"task.activity.calling": "Chamando Habilidade...",
|
|
326
|
+
"task.activity.clientExecuting": "Executando localmente...",
|
|
325
327
|
"task.activity.generating": "Gerando resposta...",
|
|
326
328
|
"task.activity.gotResult": "Resultado da ferramenta recebido",
|
|
327
329
|
"task.activity.toolCalling": "Chamando {{toolName}}...",
|
|
328
330
|
"task.activity.toolResult": "Resultado de {{toolName}} recebido",
|
|
329
331
|
"task.batchTasks": "{{count}} Subtarefas em Lote",
|
|
332
|
+
"task.instruction": "Instruções da Tarefa",
|
|
333
|
+
"task.intermediateSteps": "{{count}} etapas intermediárias",
|
|
334
|
+
"task.metrics.duration": "(levou {{duration}})",
|
|
330
335
|
"task.metrics.stepsShort": "etapas",
|
|
331
336
|
"task.metrics.toolCallsShort": "usos de ferramenta",
|
|
332
337
|
"task.status.cancelled": "Tarefa Cancelada",
|
|
@@ -58,6 +58,7 @@
|
|
|
58
58
|
"screen4.title": "Como você gostaria de compartilhar os dados?",
|
|
59
59
|
"screen4.title2": "Sua escolha nos ajuda a melhorar",
|
|
60
60
|
"screen4.title3": "Você pode alterar isso a qualquer momento nas configurações",
|
|
61
|
+
"screen5.actions.cancel": "Cancelar",
|
|
61
62
|
"screen5.actions.connectToServer": "Conectar ao Servidor",
|
|
62
63
|
"screen5.actions.connecting": "Conectando...",
|
|
63
64
|
"screen5.actions.signInCloud": "Entrar no LobeHub Cloud",
|
|
@@ -65,6 +66,10 @@
|
|
|
65
66
|
"screen5.actions.signingIn": "Entrando...",
|
|
66
67
|
"screen5.actions.signingOut": "Saindo...",
|
|
67
68
|
"screen5.actions.tryAgain": "Tentar Novamente",
|
|
69
|
+
"screen5.auth.phase.browserOpened": "Navegador aberto, faça login...",
|
|
70
|
+
"screen5.auth.phase.verifying": "Verificando credenciais...",
|
|
71
|
+
"screen5.auth.phase.waitingForAuth": "Aguardando autorização...",
|
|
72
|
+
"screen5.auth.remaining": "Restante: {{time}}s",
|
|
68
73
|
"screen5.badge": "Entrar",
|
|
69
74
|
"screen5.description": "Entre para sincronizar Agentes, Grupos, configurações e Contexto em todos os dispositivos.",
|
|
70
75
|
"screen5.errors.desktopOnlyOidc": "A autorização OIDC está disponível apenas no aplicativo desktop.",
|
|
@@ -141,6 +141,16 @@
|
|
|
141
141
|
"filterBy.timePeriod.year": "Último Ano",
|
|
142
142
|
"footer.desc": "Evolua com usuários de IA do mundo todo. Torne-se um criador e envie seus agentes e habilidades para a Comunidade LobeHub.",
|
|
143
143
|
"footer.title": "Compartilhe sua criação na Comunidade LobeHub hoje",
|
|
144
|
+
"fork.alreadyForked": "Você já fez um fork deste agente. Redirecionando para o seu fork...",
|
|
145
|
+
"fork.failed": "Falha ao fazer o fork. Por favor, tente novamente.",
|
|
146
|
+
"fork.forkAndChat": "Fazer Fork e Conversar",
|
|
147
|
+
"fork.forkedFrom": "Fork feito de",
|
|
148
|
+
"fork.forks": "forks",
|
|
149
|
+
"fork.forksCount": "{{count}} fork",
|
|
150
|
+
"fork.forksCount_other": "{{count}} forks",
|
|
151
|
+
"fork.success": "Fork realizado com sucesso!",
|
|
152
|
+
"fork.viewAllForks": "Ver todos os forks",
|
|
153
|
+
"groupAgents.tag": "Grupo",
|
|
144
154
|
"home.communityAgents": "Agentes da Comunidade",
|
|
145
155
|
"home.featuredAssistants": "Agentes em Destaque",
|
|
146
156
|
"home.featuredModels": "Modelos em Destaque",
|
|
@@ -471,14 +481,19 @@
|
|
|
471
481
|
"user.follow": "Seguir",
|
|
472
482
|
"user.followers": "Seguidores",
|
|
473
483
|
"user.following": "Seguindo",
|
|
484
|
+
"user.forkedAgentGroups": "Grupos de Agentes Forkados",
|
|
485
|
+
"user.forkedAgents": "Agentes Forkados",
|
|
474
486
|
"user.login": "Torne-se um Criador",
|
|
475
487
|
"user.logout": "Sair",
|
|
476
488
|
"user.myProfile": "Meu Perfil",
|
|
477
489
|
"user.noAgents": "Este usuário ainda não publicou nenhum agente",
|
|
478
490
|
"user.noFavoriteAgents": "Nenhum agente salvo ainda",
|
|
479
491
|
"user.noFavoritePlugins": "Nenhuma habilidade salva ainda",
|
|
492
|
+
"user.noForkedAgentGroups": "Nenhum grupo de agentes forkado ainda",
|
|
493
|
+
"user.noForkedAgents": "Nenhum agente forkado ainda",
|
|
480
494
|
"user.publishedAgents": "Agentes Criados",
|
|
481
495
|
"user.tabs.favorites": "Favoritos",
|
|
496
|
+
"user.tabs.forkedAgents": "Forkados",
|
|
482
497
|
"user.tabs.publishedAgents": "Criados",
|
|
483
498
|
"user.unfavorite": "Remover dos Favoritos",
|
|
484
499
|
"user.unfavoriteFailed": "Falha ao remover dos favoritos",
|
|
@@ -680,6 +680,48 @@
|
|
|
680
680
|
"groq/compound-mini.description": "Compound-mini é um sistema de IA composto alimentado por modelos públicos disponíveis no GroqCloud, utilizando ferramentas de forma inteligente e seletiva para responder às perguntas dos usuários.",
|
|
681
681
|
"groq/compound.description": "Compound é um sistema de IA composto alimentado por múltiplos modelos públicos disponíveis no GroqCloud, utilizando ferramentas de forma inteligente e seletiva para responder às perguntas dos usuários.",
|
|
682
682
|
"gryphe/mythomax-l2-13b.description": "MythoMax L2 13B é um modelo de linguagem criativo e inteligente, fundido a partir de diversos modelos de ponta.",
|
|
683
|
+
"hunyuan-a13b.description": "O primeiro modelo de raciocínio híbrido da Hunyuan, atualizado a partir do hunyuan-standard-256K (80B no total, 13B ativos). Opera por padrão em modo de pensamento lento e permite alternância entre rápido/lento via parâmetros ou prefixo /no_think. A capacidade geral foi aprimorada em relação à geração anterior, especialmente em matemática, ciências, compreensão de textos longos e tarefas de agentes.",
|
|
684
|
+
"hunyuan-code.description": "Modelo mais recente de geração de código, treinado com 200B de código de alta qualidade e seis meses de SFT; contexto expandido para 8K. Classificado entre os melhores em benchmarks automatizados para cinco linguagens e em avaliações humanas em dez critérios.",
|
|
685
|
+
"hunyuan-functioncall.description": "Modelo FunctionCall MoE mais recente, treinado com dados de chamadas de função de alta qualidade, com janela de contexto de 32K e resultados líderes em benchmarks em diversas dimensões.",
|
|
686
|
+
"hunyuan-large-longcontext.description": "Excelente em tarefas com documentos longos, como resumo e perguntas e respostas, além de lidar com geração geral. Forte em análise e geração de textos longos e complexos.",
|
|
687
|
+
"hunyuan-large-vision.description": "Modelo de linguagem e visão treinado a partir do Hunyuan Large para compreensão de imagem e texto. Suporta entrada de múltiplas imagens + texto em qualquer resolução e melhora a compreensão visual multilíngue.",
|
|
688
|
+
"hunyuan-large.description": "O hunyuan-large possui ~389B de parâmetros totais e ~52B ativados, sendo o maior e mais poderoso modelo MoE aberto com arquitetura Transformer.",
|
|
689
|
+
"hunyuan-lite-vision.description": "Modelo multimodal mais recente de 7B com janela de contexto de 32K, compatível com chat multimodal em chinês/inglês, reconhecimento de objetos, compreensão de tabelas em documentos e matemática multimodal, superando outros modelos de 7B em vários benchmarks.",
|
|
690
|
+
"hunyuan-lite.description": "Atualizado para uma arquitetura MoE com janela de contexto de 256K, liderando muitos modelos abertos em benchmarks de NLP, código, matemática e indústria.",
|
|
691
|
+
"hunyuan-pro.description": "Modelo MoE com trilhões de parâmetros e contexto longo de 32K, líder em benchmarks, excelente em instruções complexas e raciocínio, matemática avançada, chamadas de função e otimizado para tradução multilíngue, finanças, direito e medicina.",
|
|
692
|
+
"hunyuan-role.description": "Modelo de interpretação de papéis mais recente, ajustado oficialmente com conjuntos de dados específicos, oferecendo desempenho base mais forte para cenários de roleplay.",
|
|
693
|
+
"hunyuan-standard-256K.description": "Utiliza roteamento aprimorado para mitigar desequilíbrio de carga e colapso de especialistas. Alcança 99,9% em 'agulha no palheiro' com contexto longo. O MOE-256K expande ainda mais o comprimento e a qualidade do contexto.",
|
|
694
|
+
"hunyuan-standard-vision.description": "Modelo multimodal mais recente com respostas multilíngues e equilíbrio entre habilidades em chinês e inglês.",
|
|
695
|
+
"hunyuan-standard.description": "Utiliza roteamento aprimorado para mitigar desequilíbrio de carga e colapso de especialistas. Alcança 99,9% em 'agulha no palheiro' com contexto longo. O MOE-32K oferece ótimo custo-benefício ao lidar com entradas extensas.",
|
|
696
|
+
"hunyuan-t1-20250321.description": "Desenvolve capacidades equilibradas em artes e STEM com forte captura de informações em textos longos. Suporta respostas com raciocínio para problemas de matemática, lógica, ciência e programação em diversos níveis de dificuldade.",
|
|
697
|
+
"hunyuan-t1-20250403.description": "Melhora a geração de código em nível de projeto e a qualidade da escrita, fortalece a compreensão de tópicos em múltiplas interações e o seguimento de instruções ToB, aprimora a compreensão lexical e reduz problemas de mistura entre chinês simplificado/tradicional e chinês/inglês.",
|
|
698
|
+
"hunyuan-t1-20250529.description": "Aprimora a escrita criativa e composição, fortalece a programação frontend, raciocínio matemático e lógico, e melhora o seguimento de instruções.",
|
|
699
|
+
"hunyuan-t1-20250711.description": "Melhora significativamente matemática avançada, lógica e programação, aumenta a estabilidade das respostas e aprimora a capacidade com textos longos.",
|
|
700
|
+
"hunyuan-t1-latest.description": "Melhora significativamente o modelo de pensamento lento em matemática avançada, raciocínio complexo, programação difícil, seguimento de instruções e qualidade da escrita criativa.",
|
|
701
|
+
"hunyuan-t1-vision-20250619.description": "Modelo multimodal de raciocínio profundo t1-vision mais recente com cadeia de pensamento nativa longa, significativamente melhorado em relação à versão padrão anterior.",
|
|
702
|
+
"hunyuan-t1-vision-20250916.description": "Modelo de raciocínio profundo t1-vision mais recente com grandes avanços em VQA, ancoragem visual, OCR, gráficos, resolução de problemas fotografados e criação baseada em imagem, além de melhor desempenho em inglês e idiomas de poucos recursos.",
|
|
703
|
+
"hunyuan-turbo-20241223.description": "Esta versão amplia a escalabilidade de instruções para melhor generalização, melhora significativamente o raciocínio em matemática/código/lógica, aprimora a compreensão lexical e eleva a qualidade da escrita.",
|
|
704
|
+
"hunyuan-turbo-latest.description": "Melhorias gerais na experiência em compreensão de NLP, escrita, chat, perguntas e respostas, tradução e domínios; respostas mais humanas, melhor esclarecimento de intenções ambíguas, análise lexical aprimorada, maior criatividade e interatividade, e conversas multi-turno mais robustas.",
|
|
705
|
+
"hunyuan-turbo-vision.description": "Modelo de linguagem e visão de próxima geração com nova arquitetura MoE, com amplas melhorias em reconhecimento, criação de conteúdo, perguntas e respostas baseadas em conhecimento e raciocínio analítico.",
|
|
706
|
+
"hunyuan-turbo.description": "Prévia do LLM de próxima geração da Hunyuan com nova arquitetura MoE, oferecendo raciocínio mais rápido e resultados superiores ao hunyuan-pro.",
|
|
707
|
+
"hunyuan-turbos-20250313.description": "Unifica o estilo de resolução de problemas matemáticos e fortalece perguntas e respostas matemáticas em múltiplas interações. O estilo de escrita foi refinado para reduzir o tom artificial e adicionar sofisticação.",
|
|
708
|
+
"hunyuan-turbos-20250416.description": "Base de pré-treinamento atualizada para melhorar a compreensão e seguimento de instruções; alinhamento aprimora matemática, código, lógica e ciência; melhora a qualidade da escrita, compreensão, precisão de tradução e perguntas e respostas baseadas em conhecimento; fortalece habilidades de agente, especialmente em múltiplas interações.",
|
|
709
|
+
"hunyuan-turbos-20250604.description": "Base de pré-treinamento atualizada com melhorias na escrita e compreensão de leitura, avanços significativos em código e STEM, e melhor seguimento de instruções complexas.",
|
|
710
|
+
"hunyuan-turbos-20250926.description": "Melhoria na qualidade dos dados de pré-treinamento e estratégia de pós-treinamento, aprimorando agentes, idiomas de poucos recursos/inglês, seguimento de instruções, código e capacidades STEM.",
|
|
711
|
+
"hunyuan-turbos-latest.description": "O mais recente modelo principal Hunyuan TurboS com raciocínio mais forte e melhor experiência geral.",
|
|
712
|
+
"hunyuan-turbos-longtext-128k-20250325.description": "Excelente em tarefas com documentos longos, como resumo e perguntas e respostas, além de lidar com geração geral. Forte em análise e geração de textos longos e complexos.",
|
|
713
|
+
"hunyuan-turbos-role-plus.description": "Modelo de interpretação de papéis mais recente, ajustado oficialmente com conjuntos de dados específicos, oferecendo desempenho base mais forte para cenários de roleplay.",
|
|
714
|
+
"hunyuan-turbos-vision-20250619.description": "Modelo principal TurboS de linguagem e visão mais recente com grandes avanços em tarefas imagem-texto como reconhecimento de entidades, perguntas e respostas baseadas em conhecimento, redação publicitária e resolução de problemas com fotos.",
|
|
715
|
+
"hunyuan-turbos-vision.description": "Modelo principal de linguagem e visão de próxima geração baseado no mais recente TurboS, focado em tarefas de compreensão imagem-texto como reconhecimento de entidades, perguntas e respostas baseadas em conhecimento, redação publicitária e resolução de problemas com fotos.",
|
|
716
|
+
"hunyuan-vision-1.5-instruct.description": "Modelo de pensamento rápido para geração de texto a partir de imagem baseado no TurboS textual. Em comparação com a versão anterior, apresenta melhorias significativas em reconhecimento básico de imagem e raciocínio analítico visual.",
|
|
717
|
+
"hunyuan-vision.description": "Modelo multimodal mais recente com suporte a entrada de imagem + texto para geração de texto.",
|
|
718
|
+
"image-01-live.description": "Modelo de geração de imagem com detalhes refinados, suportando geração de imagem a partir de texto e estilos controláveis.",
|
|
719
|
+
"image-01.description": "Novo modelo de geração de imagem com detalhes refinados, suportando geração de imagem a partir de texto e de imagem para imagem.",
|
|
720
|
+
"imagen-4.0-fast-generate-001.description": "Versão rápida da série de modelos de geração de imagem a partir de texto Imagen de quarta geração.",
|
|
721
|
+
"imagen-4.0-generate-001.description": "Série de modelos de geração de imagem a partir de texto Imagen de quarta geração.",
|
|
722
|
+
"imagen-4.0-generate-preview-06-06.description": "Família de modelos de geração de imagem a partir de texto da quarta geração Imagen.",
|
|
723
|
+
"imagen-4.0-ultra-generate-001.description": "Versão Ultra da série de modelos de geração de imagem a partir de texto Imagen de quarta geração.",
|
|
724
|
+
"imagen-4.0-ultra-generate-preview-06-06.description": "Variante Ultra da quarta geração de modelos de geração de imagem a partir de texto Imagen.",
|
|
683
725
|
"meta.llama3-8b-instruct-v1:0.description": "O Meta Llama 3 é um modelo de linguagem aberto para desenvolvedores, pesquisadores e empresas, projetado para ajudá-los a construir, experimentar e escalar ideias de IA generativa de forma responsável. Como parte da base para a inovação da comunidade global, é ideal para ambientes com recursos computacionais limitados, dispositivos de borda e tempos de treinamento mais rápidos.",
|
|
684
726
|
"mistral-large-latest.description": "Mistral Large é o modelo principal, com excelente desempenho em tarefas multilíngues, raciocínio complexo e geração de código — ideal para aplicações de alto nível.",
|
|
685
727
|
"mistral-large.description": "Mixtral Large é o modelo principal da Mistral, combinando geração de código, matemática e raciocínio com uma janela de contexto de 128K.",
|
package/locales/ru-RU/chat.json
CHANGED
|
@@ -204,6 +204,7 @@
|
|
|
204
204
|
"noSelectedAgents": "Участники ещё не выбраны",
|
|
205
205
|
"openInNewWindow": "Открыть в новом окне",
|
|
206
206
|
"operation.execAgentRuntime": "Подготовка ответа",
|
|
207
|
+
"operation.execClientTask": "Выполнение задачи",
|
|
207
208
|
"operation.sendMessage": "Отправка сообщения",
|
|
208
209
|
"owner": "Владелец группы",
|
|
209
210
|
"pageCopilot.title": "Агент страницы",
|
|
@@ -322,11 +323,15 @@
|
|
|
322
323
|
"tab.profile": "Профиль агента",
|
|
323
324
|
"tab.search": "Поиск",
|
|
324
325
|
"task.activity.calling": "Вызов навыка...",
|
|
326
|
+
"task.activity.clientExecuting": "Выполняется локально...",
|
|
325
327
|
"task.activity.generating": "Генерация ответа...",
|
|
326
328
|
"task.activity.gotResult": "Результат инструмента получен",
|
|
327
329
|
"task.activity.toolCalling": "Вызов {{toolName}}...",
|
|
328
330
|
"task.activity.toolResult": "Результат {{toolName}} получен",
|
|
329
331
|
"task.batchTasks": "{{count}} пакетных подзадач",
|
|
332
|
+
"task.instruction": "Инструкция к задаче",
|
|
333
|
+
"task.intermediateSteps": "{{count}} промежуточных этапов",
|
|
334
|
+
"task.metrics.duration": "(заняло {{duration}})",
|
|
330
335
|
"task.metrics.stepsShort": "шагов",
|
|
331
336
|
"task.metrics.toolCallsShort": "вызовов",
|
|
332
337
|
"task.status.cancelled": "Задача отменена",
|
|
@@ -58,6 +58,7 @@
|
|
|
58
58
|
"screen4.title": "Как вы хотите делиться данными?",
|
|
59
59
|
"screen4.title2": "Ваш выбор помогает нам становиться лучше",
|
|
60
60
|
"screen4.title3": "Вы можете изменить это в любое время в настройках",
|
|
61
|
+
"screen5.actions.cancel": "Отмена",
|
|
61
62
|
"screen5.actions.connectToServer": "Подключиться к серверу",
|
|
62
63
|
"screen5.actions.connecting": "Подключение...",
|
|
63
64
|
"screen5.actions.signInCloud": "Войти в LobeHub Cloud",
|
|
@@ -65,6 +66,10 @@
|
|
|
65
66
|
"screen5.actions.signingIn": "Вход...",
|
|
66
67
|
"screen5.actions.signingOut": "Выход...",
|
|
67
68
|
"screen5.actions.tryAgain": "Попробовать снова",
|
|
69
|
+
"screen5.auth.phase.browserOpened": "Браузер открыт, пожалуйста, выполните вход...",
|
|
70
|
+
"screen5.auth.phase.verifying": "Проверка учетных данных...",
|
|
71
|
+
"screen5.auth.phase.waitingForAuth": "Ожидание авторизации...",
|
|
72
|
+
"screen5.auth.remaining": "Осталось: {{time}} с",
|
|
68
73
|
"screen5.badge": "Вход",
|
|
69
74
|
"screen5.description": "Войдите, чтобы синхронизировать Агентов, Группы, настройки и Контекст на всех устройствах.",
|
|
70
75
|
"screen5.errors.desktopOnlyOidc": "Авторизация OIDC доступна только в настольной версии приложения.",
|