opc-agent 4.1.0 → 4.1.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.github/ISSUE_TEMPLATE/bug_report.md +20 -20
- package/.github/ISSUE_TEMPLATE/feature_request.md +14 -14
- package/.github/PULL_REQUEST_TEMPLATE.md +13 -13
- package/CHANGELOG.md +48 -48
- package/CONTRIBUTING.md +36 -36
- package/README.zh-CN.md +497 -497
- package/USABILITY-ISSUES.md +73 -0
- package/dist/channels/web.js +8 -2
- package/dist/channels/wechat.js +6 -6
- package/dist/cli.js +200 -85
- package/dist/core/runtime.js +37 -15
- package/dist/deploy/index.js +56 -56
- package/dist/doctor.d.ts +1 -0
- package/dist/doctor.js +105 -10
- package/dist/memory/deepbrain.d.ts +1 -1
- package/dist/memory/deepbrain.js +95 -4
- package/dist/scheduler/cron-engine.js +3 -36
- package/dist/studio/server.js +30 -1
- package/dist/studio-ui/index.html +230 -10
- package/dist/ui/components.js +105 -105
- package/examples/README.md +22 -22
- package/examples/basic-agent.ts +90 -90
- package/examples/brain-integration.ts +71 -71
- package/examples/multi-channel.ts +74 -74
- package/fix-sidebar.mjs +188 -188
- package/install.ps1 +154 -154
- package/install.sh +164 -164
- package/package.json +1 -1
- package/scripts/install.ps1 +31 -31
- package/scripts/install.sh +40 -40
- package/serve-studio.js +13 -13
- package/serve-test.js +25 -25
- package/src/channels/dingtalk.ts +46 -46
- package/src/channels/email.ts +351 -351
- package/src/channels/feishu.ts +349 -349
- package/src/channels/googlechat.ts +42 -42
- package/src/channels/imessage.ts +31 -31
- package/src/channels/irc.ts +82 -82
- package/src/channels/line.ts +32 -32
- package/src/channels/matrix.ts +33 -33
- package/src/channels/mattermost.ts +57 -57
- package/src/channels/msteams.ts +32 -32
- package/src/channels/nostr.ts +32 -32
- package/src/channels/qq.ts +33 -33
- package/src/channels/signal.ts +32 -32
- package/src/channels/sms.ts +33 -33
- package/src/channels/telegram.ts +616 -616
- package/src/channels/twitch.ts +65 -65
- package/src/channels/voice-call.ts +100 -100
- package/src/channels/web.ts +8 -2
- package/src/channels/websocket.ts +399 -399
- package/src/channels/wechat.ts +329 -329
- package/src/channels/whatsapp.ts +32 -32
- package/src/cli/chat.ts +99 -99
- package/src/cli/setup.ts +314 -314
- package/src/cli.ts +195 -92
- package/src/core/agent.ts +476 -476
- package/src/core/api-server.ts +277 -277
- package/src/core/audio.ts +98 -98
- package/src/core/collaboration.ts +275 -275
- package/src/core/context-discovery.ts +85 -85
- package/src/core/context-refs.ts +140 -140
- package/src/core/gateway.ts +106 -106
- package/src/core/heartbeat.ts +51 -51
- package/src/core/hooks.ts +105 -105
- package/src/core/ide-bridge.ts +133 -133
- package/src/core/node-network.ts +86 -86
- package/src/core/profiles.ts +122 -122
- package/src/core/runtime.ts +25 -0
- package/src/core/scheduler.ts +187 -187
- package/src/core/session-manager.ts +137 -137
- package/src/core/subagent.ts +98 -98
- package/src/core/vision.ts +180 -180
- package/src/core/workflow-graph.ts +365 -365
- package/src/daemon.ts +96 -96
- package/src/deploy/index.ts +255 -255
- package/src/doctor.ts +98 -11
- package/src/eval/index.ts +211 -211
- package/src/eval/suites/basic.json +16 -16
- package/src/eval/suites/memory.json +12 -12
- package/src/eval/suites/safety.json +14 -14
- package/src/hub/brain-seed.ts +54 -54
- package/src/hub/client.ts +60 -60
- package/src/mcp/servers/calculator-mcp.ts +65 -65
- package/src/mcp/servers/crypto-mcp.ts +73 -73
- package/src/mcp/servers/database-mcp.ts +72 -72
- package/src/mcp/servers/datetime-mcp.ts +69 -69
- package/src/mcp/servers/filesystem.ts +66 -66
- package/src/mcp/servers/github-mcp.ts +58 -58
- package/src/mcp/servers/index.ts +63 -63
- package/src/mcp/servers/json-mcp.ts +102 -102
- package/src/mcp/servers/memory-mcp.ts +56 -56
- package/src/mcp/servers/regex-mcp.ts +53 -53
- package/src/mcp/servers/web-mcp.ts +49 -49
- package/src/memory/context-compressor.ts +189 -189
- package/src/memory/deepbrain.ts +99 -5
- package/src/memory/seed-loader.ts +212 -212
- package/src/memory/user-profiler.ts +215 -215
- package/src/plugins/content-filter.ts +23 -23
- package/src/plugins/logger.ts +18 -18
- package/src/plugins/rate-limiter.ts +38 -38
- package/src/protocols/a2a/client.ts +132 -132
- package/src/protocols/a2a/index.ts +8 -8
- package/src/protocols/a2a/server.ts +333 -333
- package/src/protocols/a2a/types.ts +88 -88
- package/src/protocols/a2a/utils.ts +50 -50
- package/src/protocols/agui/client.ts +83 -83
- package/src/protocols/agui/index.ts +4 -4
- package/src/protocols/agui/server.ts +218 -218
- package/src/protocols/agui/types.ts +153 -153
- package/src/protocols/index.ts +2 -2
- package/src/protocols/mcp/agent-tools.ts +134 -134
- package/src/protocols/mcp/index.ts +8 -8
- package/src/protocols/mcp/server.ts +262 -262
- package/src/protocols/mcp/types.ts +69 -69
- package/src/providers/index.ts +632 -632
- package/src/publish/index.ts +376 -376
- package/src/scheduler/cron-engine.ts +191 -191
- package/src/scheduler/index.ts +2 -2
- package/src/schema/oad.ts +217 -217
- package/src/security/approval.ts +131 -131
- package/src/security/approvals.ts +143 -143
- package/src/security/elevated.ts +105 -105
- package/src/security/guardrails.ts +248 -248
- package/src/security/index.ts +9 -9
- package/src/security/keys.ts +87 -87
- package/src/security/secrets.ts +129 -129
- package/src/skills/builtin/index.ts +408 -408
- package/src/skills/marketplace.ts +113 -113
- package/src/skills/types.ts +42 -42
- package/src/studio/server.ts +31 -1
- package/src/studio/templates-data.ts +178 -178
- package/src/studio-ui/index.html +230 -10
- package/src/telemetry/index.ts +324 -324
- package/src/tools/builtin/browser.ts +299 -299
- package/src/tools/builtin/datetime.ts +41 -41
- package/src/tools/builtin/file.ts +107 -107
- package/src/tools/builtin/home-assistant.ts +116 -116
- package/src/tools/builtin/rl-tools.ts +243 -243
- package/src/tools/builtin/shell.ts +43 -43
- package/src/tools/builtin/vision.ts +64 -64
- package/src/tools/builtin/web-search.ts +126 -126
- package/src/tools/builtin/web.ts +35 -35
- package/src/tools/document-processor.ts +213 -213
- package/src/tools/image-generator.ts +150 -150
- package/src/tools/integrations/calendar.ts +73 -73
- package/src/tools/integrations/code-exec.ts +39 -39
- package/src/tools/integrations/csv-analyzer.ts +92 -92
- package/src/tools/integrations/database.ts +44 -44
- package/src/tools/integrations/email-send.ts +76 -76
- package/src/tools/integrations/git-tool.ts +42 -42
- package/src/tools/integrations/github-tool.ts +76 -76
- package/src/tools/integrations/image-gen.ts +56 -56
- package/src/tools/integrations/index.ts +92 -92
- package/src/tools/integrations/jira.ts +83 -83
- package/src/tools/integrations/notion.ts +71 -71
- package/src/tools/integrations/npm-tool.ts +48 -48
- package/src/tools/integrations/pdf-reader.ts +58 -58
- package/src/tools/integrations/slack.ts +65 -65
- package/src/tools/integrations/summarizer.ts +49 -49
- package/src/tools/integrations/translator.ts +48 -48
- package/src/tools/integrations/trello.ts +60 -60
- package/src/tools/integrations/vector-search.ts +42 -42
- package/src/tools/integrations/web-scraper.ts +47 -47
- package/src/tools/integrations/web-search.ts +58 -58
- package/src/tools/integrations/webhook.ts +38 -38
- package/src/tools/mcp-client.ts +131 -131
- package/src/tools/web-scraper.ts +179 -179
- package/src/tools/web-search.ts +180 -180
- package/src/ui/components.ts +127 -127
- package/srv-out.txt +1 -1
- package/templates/ecommerce-assistant/README.md +45 -45
- package/templates/ecommerce-assistant/oad.yaml +47 -47
- package/templates/tech-support/README.md +43 -43
- package/templates/tech-support/oad.yaml +45 -45
- package/test-agent/Dockerfile +9 -9
- package/test-agent/README.md +50 -50
- package/test-agent/agent.yaml +23 -23
- package/test-agent/docker-compose.yml +11 -11
- package/test-agent/oad.yaml +31 -31
- package/test-agent/package-lock.json +1492 -1492
- package/test-agent/package.json +17 -17
- package/test-agent/src/index.ts +24 -24
- package/test-agent/src/skills/echo.ts +15 -15
- package/test-agent/tsconfig.json +24 -24
- package/test-full.js +43 -43
- package/test-sidebar.js +22 -22
- package/test-studio3.js +75 -75
- package/test-studio4.js +41 -41
- package/tests/a2a-protocol.test.ts +285 -285
- package/tests/agui-protocol.test.ts +246 -246
- package/tests/api-server.test.ts +148 -148
- package/tests/approvals.test.ts +89 -89
- package/tests/audio.test.ts +40 -40
- package/tests/brain-seed-extended.test.ts +490 -490
- package/tests/brain-seed.test.ts +239 -239
- package/tests/browser.test.ts +179 -179
- package/tests/channels/discord.test.ts +79 -79
- package/tests/channels/email.test.ts +148 -148
- package/tests/channels/feishu.test.ts +123 -123
- package/tests/channels/telegram.test.ts +129 -129
- package/tests/channels/websocket.test.ts +53 -53
- package/tests/channels/wechat.test.ts +170 -170
- package/tests/channels-extra.test.ts +45 -45
- package/tests/chat-cli.test.ts +160 -160
- package/tests/cli.test.ts +46 -46
- package/tests/context-compressor.test.ts +172 -172
- package/tests/context-refs.test.ts +121 -121
- package/tests/cron-engine.test.ts +101 -101
- package/tests/daemon.test.ts +135 -135
- package/tests/deepbrain-wire.test.ts +234 -234
- package/tests/deploy-and-dag.test.ts +196 -196
- package/tests/doctor.test.ts +38 -38
- package/tests/document-processor.test.ts +69 -69
- package/tests/e2e-nocode.test.ts +442 -442
- package/tests/elevated.test.ts +69 -69
- package/tests/eval.test.ts +173 -173
- package/tests/gateway.test.ts +63 -63
- package/tests/guardrails.test.ts +177 -177
- package/tests/home-assistant.test.ts +40 -40
- package/tests/hooks.test.ts +79 -79
- package/tests/ide-bridge.test.ts +38 -38
- package/tests/image-generator.test.ts +84 -84
- package/tests/init-role.test.ts +124 -124
- package/tests/integrations.test.ts +249 -249
- package/tests/mcp-client.test.ts +92 -92
- package/tests/mcp-server.test.ts +178 -178
- package/tests/mcp-servers.test.ts +260 -260
- package/tests/node-network.test.ts +74 -74
- package/tests/plugin-a2a-enhanced.test.ts +230 -230
- package/tests/profiles.test.ts +61 -61
- package/tests/publish.test.ts +231 -231
- package/tests/rl-tools.test.ts +93 -93
- package/tests/sandbox-manager.test.ts +46 -46
- package/tests/scheduler.test.ts +200 -200
- package/tests/secrets.test.ts +107 -107
- package/tests/security-enhanced.test.ts +233 -233
- package/tests/settings-api.test.ts +148 -148
- package/tests/setup.test.ts +73 -73
- package/tests/subagent.test.ts +193 -193
- package/tests/telegram-discord.test.ts +60 -60
- package/tests/telemetry.test.ts +186 -186
- package/tests/user-profiler.test.ts +169 -169
- package/tests/v090-features.test.ts +254 -254
- package/tests/vision.test.ts +61 -61
- package/tests/voice-call.test.ts +47 -47
- package/tests/voice-enhanced.test.ts +169 -169
- package/tests/voice-interaction.test.ts +38 -38
- package/tests/web-search.test.ts +155 -155
- package/tests/workflow-graph.test.ts +279 -279
- package/tutorial/customer-service-agent/README.md +612 -612
- package/tutorial/customer-service-agent/SOUL.md +26 -26
- package/tutorial/customer-service-agent/agent.yaml +63 -63
- package/tutorial/customer-service-agent/package.json +19 -19
- package/tutorial/customer-service-agent/src/index.ts +69 -69
- package/tutorial/customer-service-agent/src/skills/faq.ts +27 -27
- package/tutorial/customer-service-agent/src/skills/ticket.ts +22 -22
- package/tutorial/customer-service-agent/tsconfig.json +14 -14
package/src/cli/setup.ts
CHANGED
|
@@ -1,314 +1,314 @@
|
|
|
1
|
-
import * as readline from 'readline';
|
|
2
|
-
import * as fs from 'fs';
|
|
3
|
-
import * as path from 'path';
|
|
4
|
-
import * as http from 'http';
|
|
5
|
-
import * as https from 'https';
|
|
6
|
-
import * as os from 'os';
|
|
7
|
-
|
|
8
|
-
// ── Colors ──────────────────────────────────────────────────────────────────
|
|
9
|
-
const c = {
|
|
10
|
-
green: (s: string) => `\x1b[32m${s}\x1b[0m`,
|
|
11
|
-
red: (s: string) => `\x1b[31m${s}\x1b[0m`,
|
|
12
|
-
yellow: (s: string) => `\x1b[33m${s}\x1b[0m`,
|
|
13
|
-
blue: (s: string) => `\x1b[34m${s}\x1b[0m`,
|
|
14
|
-
cyan: (s: string) => `\x1b[36m${s}\x1b[0m`,
|
|
15
|
-
bold: (s: string) => `\x1b[1m${s}\x1b[0m`,
|
|
16
|
-
dim: (s: string) => `\x1b[2m${s}\x1b[0m`,
|
|
17
|
-
magenta: (s: string) => `\x1b[35m${s}\x1b[0m`,
|
|
18
|
-
};
|
|
19
|
-
|
|
20
|
-
// ── Types ───────────────────────────────────────────────────────────────────
|
|
21
|
-
interface SetupConfig {
|
|
22
|
-
provider: 'ollama' | 'openai' | 'deepseek' | 'qwen' | 'anthropic';
|
|
23
|
-
model?: string;
|
|
24
|
-
embeddingModel?: string;
|
|
25
|
-
apiKey?: string;
|
|
26
|
-
baseUrl?: string;
|
|
27
|
-
}
|
|
28
|
-
|
|
29
|
-
interface AgentTemplate {
|
|
30
|
-
id: string;
|
|
31
|
-
icon: string;
|
|
32
|
-
name: string;
|
|
33
|
-
description: string;
|
|
34
|
-
defaultAgentName: string;
|
|
35
|
-
}
|
|
36
|
-
|
|
37
|
-
const TEMPLATES: AgentTemplate[] = [
|
|
38
|
-
{ id: 'customer-service', icon: '🎧', name: '客服助手', description: '回答客户问题,自动学习产品知识', defaultAgentName: '我的客服助手' },
|
|
39
|
-
{ id: 'content-writer', icon: '✍️', name: '写作助手', description: '帮你写文章、邮件、报告', defaultAgentName: '我的写作助手' },
|
|
40
|
-
{ id: 'data-analyst', icon: '📊', name: '数据分析师', description: '分析数据,生成洞察', defaultAgentName: '我的数据分析师' },
|
|
41
|
-
{ id: 'translator', icon: '🌐', name: '翻译助手', description: '多语言翻译,越用越准', defaultAgentName: '我的翻译助手' },
|
|
42
|
-
{ id: 'general', icon: '🤖', name: '通用助手', description: '什么都能聊', defaultAgentName: '我的AI助手' },
|
|
43
|
-
];
|
|
44
|
-
|
|
45
|
-
const OPC_HOME = path.join(os.homedir(), '.opc');
|
|
46
|
-
const CONFIG_PATH = path.join(OPC_HOME, 'config.json');
|
|
47
|
-
|
|
48
|
-
// ── Readline helpers ────────────────────────────────────────────────────────
|
|
49
|
-
export function createRL(input?: NodeJS.ReadableStream, output?: NodeJS.WritableStream): readline.Interface {
|
|
50
|
-
return readline.createInterface({
|
|
51
|
-
input: input ?? process.stdin,
|
|
52
|
-
output: output ?? process.stdout,
|
|
53
|
-
});
|
|
54
|
-
}
|
|
55
|
-
|
|
56
|
-
function ask(rl: readline.Interface, question: string): Promise<string> {
|
|
57
|
-
return new Promise((resolve) => rl.question(question, (a) => resolve(a.trim())));
|
|
58
|
-
}
|
|
59
|
-
|
|
60
|
-
// ── HTTP helpers ────────────────────────────────────────────────────────────
|
|
61
|
-
function httpGet(url: string, timeout = 5000): Promise<{ status: number; body: string }> {
|
|
62
|
-
return new Promise((resolve, reject) => {
|
|
63
|
-
const mod = url.startsWith('https') ? https : http;
|
|
64
|
-
const req = mod.get(url, { timeout }, (res) => {
|
|
65
|
-
let body = '';
|
|
66
|
-
res.on('data', (chunk: Buffer) => { body += chunk; });
|
|
67
|
-
res.on('end', () => resolve({ status: res.statusCode ?? 0, body }));
|
|
68
|
-
});
|
|
69
|
-
req.on('error', reject);
|
|
70
|
-
req.on('timeout', () => { req.destroy(); reject(new Error('timeout')); });
|
|
71
|
-
});
|
|
72
|
-
}
|
|
73
|
-
|
|
74
|
-
// ── Step 1: Welcome ─────────────────────────────────────────────────────────
|
|
75
|
-
function printWelcome(): void {
|
|
76
|
-
console.log('');
|
|
77
|
-
console.log(c.bold(' 🎉 欢迎使用 OPC Agent!'));
|
|
78
|
-
console.log(c.dim(' 让我们用 3 分钟配置你的第一个 AI Agent。'));
|
|
79
|
-
console.log('');
|
|
80
|
-
}
|
|
81
|
-
|
|
82
|
-
// ── Step 2: Model detection ─────────────────────────────────────────────────
|
|
83
|
-
async function detectOllama(): Promise<{ running: boolean; models: string[] }> {
|
|
84
|
-
try {
|
|
85
|
-
const { body } = await httpGet('http://localhost:11434/api/tags');
|
|
86
|
-
const data = JSON.parse(body);
|
|
87
|
-
const models = (data.models || []).map((m: any) => m.name || m.model);
|
|
88
|
-
return { running: true, models };
|
|
89
|
-
} catch {
|
|
90
|
-
return { running: false, models: [] };
|
|
91
|
-
}
|
|
92
|
-
}
|
|
93
|
-
|
|
94
|
-
async function stepModel(rl: readline.Interface): Promise<SetupConfig> {
|
|
95
|
-
console.log(c.bold('📡 Step 1/4: 配置 AI 模型'));
|
|
96
|
-
console.log(c.dim(' 正在检测本地 Ollama...'));
|
|
97
|
-
console.log('');
|
|
98
|
-
|
|
99
|
-
const ollama = await detectOllama();
|
|
100
|
-
|
|
101
|
-
if (ollama.running) {
|
|
102
|
-
console.log(c.green(' ✔ Ollama 运行中!已安装模型:'));
|
|
103
|
-
if (ollama.models.length === 0) {
|
|
104
|
-
console.log(c.dim(' (无模型)'));
|
|
105
|
-
} else {
|
|
106
|
-
for (const m of ollama.models) {
|
|
107
|
-
console.log(` • ${m}`);
|
|
108
|
-
}
|
|
109
|
-
}
|
|
110
|
-
console.log('');
|
|
111
|
-
|
|
112
|
-
const hasChat = ollama.models.some((m) => m.startsWith('qwen2.5'));
|
|
113
|
-
const hasEmbed = ollama.models.some((m) => m.startsWith('nomic-embed-text'));
|
|
114
|
-
|
|
115
|
-
if (!hasChat) {
|
|
116
|
-
const dl = await ask(rl, ` 推荐模型 ${c.cyan('qwen2.5:7b')} 未安装,是否下载? [Y/n] `);
|
|
117
|
-
if (dl.toLowerCase() !== 'n') {
|
|
118
|
-
console.log(c.dim(' → 请在终端运行: ollama pull qwen2.5:7b'));
|
|
119
|
-
}
|
|
120
|
-
}
|
|
121
|
-
|
|
122
|
-
if (!hasEmbed) {
|
|
123
|
-
const dl = await ask(rl, ` 推荐 Embedding 模型 ${c.cyan('nomic-embed-text')} 未安装,是否下载? [Y/n] `);
|
|
124
|
-
if (dl.toLowerCase() !== 'n') {
|
|
125
|
-
console.log(c.dim(' → 请在终端运行: ollama pull nomic-embed-text'));
|
|
126
|
-
}
|
|
127
|
-
}
|
|
128
|
-
|
|
129
|
-
console.log('');
|
|
130
|
-
return {
|
|
131
|
-
provider: 'ollama',
|
|
132
|
-
model: hasChat ? ollama.models.find((m) => m.startsWith('qwen2.5'))! : 'qwen2.5:7b',
|
|
133
|
-
embeddingModel: hasEmbed ? 'nomic-embed-text' : 'nomic-embed-text',
|
|
134
|
-
baseUrl: 'http://localhost:11434',
|
|
135
|
-
};
|
|
136
|
-
}
|
|
137
|
-
|
|
138
|
-
// Ollama not running
|
|
139
|
-
console.log(c.yellow(' ⚠ 未检测到 Ollama'));
|
|
140
|
-
console.log('');
|
|
141
|
-
console.log(' 请选择:');
|
|
142
|
-
console.log(` ${c.cyan('A')} ) 安装 Ollama(推荐,免费本地运行)`);
|
|
143
|
-
console.log(` ${c.cyan('B')} ) 使用云端 API`);
|
|
144
|
-
console.log('');
|
|
145
|
-
|
|
146
|
-
const choice = await ask(rl, ' 你的选择 [A/B]: ');
|
|
147
|
-
|
|
148
|
-
if (choice.toUpperCase() !== 'B') {
|
|
149
|
-
console.log('');
|
|
150
|
-
console.log(c.bold(' 📦 安装 Ollama:'));
|
|
151
|
-
console.log('');
|
|
152
|
-
if (process.platform === 'win32') {
|
|
153
|
-
console.log(' 1. 访问 https://ollama.com/download');
|
|
154
|
-
console.log(' 2. 下载 Windows 版并安装');
|
|
155
|
-
console.log(' 3. 安装后运行: ollama pull qwen2.5:7b');
|
|
156
|
-
} else if (process.platform === 'darwin') {
|
|
157
|
-
console.log(' brew install ollama');
|
|
158
|
-
console.log(' ollama serve &');
|
|
159
|
-
console.log(' ollama pull qwen2.5:7b');
|
|
160
|
-
} else {
|
|
161
|
-
console.log(' curl -fsSL https://ollama.com/install.sh | sh');
|
|
162
|
-
console.log(' ollama serve &');
|
|
163
|
-
console.log(' ollama pull qwen2.5:7b');
|
|
164
|
-
}
|
|
165
|
-
console.log('');
|
|
166
|
-
console.log(c.dim(' 安装完成后重新运行 opc setup'));
|
|
167
|
-
return { provider: 'ollama', model: 'qwen2.5:7b', embeddingModel: 'nomic-embed-text', baseUrl: 'http://localhost:11434' };
|
|
168
|
-
}
|
|
169
|
-
|
|
170
|
-
// Cloud API
|
|
171
|
-
console.log('');
|
|
172
|
-
console.log(' 选择云端 Provider:');
|
|
173
|
-
console.log(` ${c.cyan('1')} ) OpenAI`);
|
|
174
|
-
console.log(` ${c.cyan('2')} ) DeepSeek`);
|
|
175
|
-
console.log(` ${c.cyan('3')} ) 通义千问 (Qwen)`);
|
|
176
|
-
console.log(` ${c.cyan('4')} ) Anthropic`);
|
|
177
|
-
console.log('');
|
|
178
|
-
|
|
179
|
-
const providerChoice = await ask(rl, ' 选择 [1-4]: ');
|
|
180
|
-
const providers: Record<string, { provider: SetupConfig['provider']; baseUrl: string; model: string; testUrl: string }> = {
|
|
181
|
-
'1': { provider: 'openai', baseUrl: 'https://api.openai.com/v1', model: 'gpt-4o-mini', testUrl: 'https://api.openai.com/v1/models' },
|
|
182
|
-
'2': { provider: 'deepseek', baseUrl: 'https://api.deepseek.com/v1', model: 'deepseek-chat', testUrl: 'https://api.deepseek.com/v1/models' },
|
|
183
|
-
'3': { provider: 'qwen', baseUrl: 'https://dashscope.aliyuncs.com/compatible-mode/v1', model: 'qwen-plus', testUrl: 'https://dashscope.aliyuncs.com/compatible-mode/v1/models' },
|
|
184
|
-
'4': { provider: 'anthropic', baseUrl: 'https://api.anthropic.com/v1', model: 'claude-3-5-sonnet-20241022', testUrl: 'https://api.anthropic.com/v1/models' },
|
|
185
|
-
};
|
|
186
|
-
|
|
187
|
-
const p = providers[providerChoice] ?? providers['1'];
|
|
188
|
-
console.log('');
|
|
189
|
-
const apiKey = await ask(rl, ` 请输入 ${c.bold(p.provider)} API Key: `);
|
|
190
|
-
|
|
191
|
-
if (apiKey) {
|
|
192
|
-
console.log(c.dim(' 正在测试连接...'));
|
|
193
|
-
try {
|
|
194
|
-
await httpGet(p.testUrl);
|
|
195
|
-
console.log(c.green(' ✔ 连接成功!'));
|
|
196
|
-
} catch {
|
|
197
|
-
console.log(c.yellow(' ⚠ 无法验证连接,但配置已保存,稍后可测试'));
|
|
198
|
-
}
|
|
199
|
-
}
|
|
200
|
-
|
|
201
|
-
console.log('');
|
|
202
|
-
return { provider: p.provider, model: p.model, apiKey, baseUrl: p.baseUrl };
|
|
203
|
-
}
|
|
204
|
-
|
|
205
|
-
// ── Step 3: Choose template ─────────────────────────────────────────────────
|
|
206
|
-
async function stepTemplate(rl: readline.Interface): Promise<AgentTemplate> {
|
|
207
|
-
console.log(c.bold('📋 Step 2/4: 选择 Agent 模板'));
|
|
208
|
-
console.log('');
|
|
209
|
-
|
|
210
|
-
for (let i = 0; i < TEMPLATES.length; i++) {
|
|
211
|
-
const t = TEMPLATES[i];
|
|
212
|
-
console.log(` ${c.cyan(String(i + 1))} ) ${t.icon} ${c.bold(t.name)} — ${c.dim(t.description)}`);
|
|
213
|
-
}
|
|
214
|
-
console.log(` ${c.cyan('6')} ) 📋 更多模板...(打开 Studio 浏览)`);
|
|
215
|
-
console.log('');
|
|
216
|
-
|
|
217
|
-
const choice = await ask(rl, ' 选择模板 [1-6]: ');
|
|
218
|
-
const idx = parseInt(choice, 10) - 1;
|
|
219
|
-
|
|
220
|
-
if (choice === '6') {
|
|
221
|
-
console.log(c.dim(' → 运行 opc studio 浏览更多模板'));
|
|
222
|
-
return TEMPLATES[4]; // default to general
|
|
223
|
-
}
|
|
224
|
-
|
|
225
|
-
const template = TEMPLATES[idx] ?? TEMPLATES[4];
|
|
226
|
-
console.log(c.green(` ✔ 已选择: ${template.icon} ${template.name}`));
|
|
227
|
-
console.log('');
|
|
228
|
-
return template;
|
|
229
|
-
}
|
|
230
|
-
|
|
231
|
-
// ── Step 4: Create agent ────────────────────────────────────────────────────
|
|
232
|
-
async function stepCreateAgent(rl: readline.Interface, template: AgentTemplate, config: SetupConfig): Promise<{ name: string; id: string; dir: string }> {
|
|
233
|
-
console.log(c.bold('🤖 Step 3/4: 创建 Agent'));
|
|
234
|
-
console.log('');
|
|
235
|
-
|
|
236
|
-
const name = (await ask(rl, ` Agent 名称 [${template.defaultAgentName}]: `)) || template.defaultAgentName;
|
|
237
|
-
const description = await ask(rl, ` 简短描述(可选): `);
|
|
238
|
-
|
|
239
|
-
// Generate a slug id
|
|
240
|
-
const id = name
|
|
241
|
-
.replace(/[^\w\u4e00-\u9fff]/g, '-')
|
|
242
|
-
.replace(/-+/g, '-')
|
|
243
|
-
.replace(/^-|-$/g, '')
|
|
244
|
-
.toLowerCase() || `agent-${Date.now()}`;
|
|
245
|
-
|
|
246
|
-
const agentDir = path.join(OPC_HOME, 'agents', id);
|
|
247
|
-
fs.mkdirSync(agentDir, { recursive: true });
|
|
248
|
-
|
|
249
|
-
// Write agent config
|
|
250
|
-
const agentConfig = {
|
|
251
|
-
id,
|
|
252
|
-
name,
|
|
253
|
-
description: description || template.description,
|
|
254
|
-
template: template.id,
|
|
255
|
-
provider: config.provider,
|
|
256
|
-
model: config.model,
|
|
257
|
-
embeddingModel: config.embeddingModel,
|
|
258
|
-
apiKey: config.apiKey,
|
|
259
|
-
baseUrl: config.baseUrl,
|
|
260
|
-
createdAt: new Date().toISOString(),
|
|
261
|
-
};
|
|
262
|
-
|
|
263
|
-
fs.writeFileSync(path.join(agentDir, 'config.json'), JSON.stringify(agentConfig, null, 2));
|
|
264
|
-
|
|
265
|
-
// Initialize brain directory
|
|
266
|
-
const brainDir = path.join(agentDir, 'brain');
|
|
267
|
-
fs.mkdirSync(brainDir, { recursive: true });
|
|
268
|
-
fs.writeFileSync(path.join(brainDir, 'README.md'), `# ${name}\n\n将知识文档放在这里,Agent 会自动学习。\n`);
|
|
269
|
-
|
|
270
|
-
// Save global config
|
|
271
|
-
fs.mkdirSync(OPC_HOME, { recursive: true });
|
|
272
|
-
const globalConfig = fs.existsSync(CONFIG_PATH) ? JSON.parse(fs.readFileSync(CONFIG_PATH, 'utf-8')) : {};
|
|
273
|
-
globalConfig.defaultAgent = id;
|
|
274
|
-
globalConfig.provider = config.provider;
|
|
275
|
-
globalConfig.model = config.model;
|
|
276
|
-
globalConfig.baseUrl = config.baseUrl;
|
|
277
|
-
if (config.apiKey) globalConfig.apiKey = config.apiKey;
|
|
278
|
-
if (config.embeddingModel) globalConfig.embeddingModel = config.embeddingModel;
|
|
279
|
-
fs.writeFileSync(CONFIG_PATH, JSON.stringify(globalConfig, null, 2));
|
|
280
|
-
|
|
281
|
-
console.log('');
|
|
282
|
-
console.log(` ${c.green('✔')} Agent 已创建: ${c.bold(name)}`);
|
|
283
|
-
console.log(` ${c.dim(' 目录: ' + agentDir)}`);
|
|
284
|
-
console.log('');
|
|
285
|
-
|
|
286
|
-
return { name, id, dir: agentDir };
|
|
287
|
-
}
|
|
288
|
-
|
|
289
|
-
// ── Step 5: Completion ──────────────────────────────────────────────────────
|
|
290
|
-
function printCompletion(agentName: string): void {
|
|
291
|
-
console.log(c.bold(` ✅ 你的 AI Agent「${agentName}」已创建!`));
|
|
292
|
-
console.log('');
|
|
293
|
-
console.log(' 启动方式:');
|
|
294
|
-
console.log(` ${c.cyan('opc studio')} — 打开网页管理界面`);
|
|
295
|
-
console.log(` ${c.cyan('opc chat')} — 终端直接对话`);
|
|
296
|
-
console.log(` ${c.cyan('opc start')} — 后台运行`);
|
|
297
|
-
console.log('');
|
|
298
|
-
console.log(c.dim(' 你的 Agent 会自动学习和进化,越用越聪明!🧬'));
|
|
299
|
-
console.log('');
|
|
300
|
-
}
|
|
301
|
-
|
|
302
|
-
// ── Main entry ──────────────────────────────────────────────────────────────
|
|
303
|
-
export async function runSetup(input?: NodeJS.ReadableStream, output?: NodeJS.WritableStream): Promise<void> {
|
|
304
|
-
const rl = createRL(input, output);
|
|
305
|
-
try {
|
|
306
|
-
printWelcome();
|
|
307
|
-
const config = await stepModel(rl);
|
|
308
|
-
const template = await stepTemplate(rl);
|
|
309
|
-
const agent = await stepCreateAgent(rl, template, config);
|
|
310
|
-
printCompletion(agent.name);
|
|
311
|
-
} finally {
|
|
312
|
-
rl.close();
|
|
313
|
-
}
|
|
314
|
-
}
|
|
1
|
+
import * as readline from 'readline';
|
|
2
|
+
import * as fs from 'fs';
|
|
3
|
+
import * as path from 'path';
|
|
4
|
+
import * as http from 'http';
|
|
5
|
+
import * as https from 'https';
|
|
6
|
+
import * as os from 'os';
|
|
7
|
+
|
|
8
|
+
// ── Colors ──────────────────────────────────────────────────────────────────
|
|
9
|
+
const c = {
|
|
10
|
+
green: (s: string) => `\x1b[32m${s}\x1b[0m`,
|
|
11
|
+
red: (s: string) => `\x1b[31m${s}\x1b[0m`,
|
|
12
|
+
yellow: (s: string) => `\x1b[33m${s}\x1b[0m`,
|
|
13
|
+
blue: (s: string) => `\x1b[34m${s}\x1b[0m`,
|
|
14
|
+
cyan: (s: string) => `\x1b[36m${s}\x1b[0m`,
|
|
15
|
+
bold: (s: string) => `\x1b[1m${s}\x1b[0m`,
|
|
16
|
+
dim: (s: string) => `\x1b[2m${s}\x1b[0m`,
|
|
17
|
+
magenta: (s: string) => `\x1b[35m${s}\x1b[0m`,
|
|
18
|
+
};
|
|
19
|
+
|
|
20
|
+
// ── Types ───────────────────────────────────────────────────────────────────
|
|
21
|
+
interface SetupConfig {
|
|
22
|
+
provider: 'ollama' | 'openai' | 'deepseek' | 'qwen' | 'anthropic';
|
|
23
|
+
model?: string;
|
|
24
|
+
embeddingModel?: string;
|
|
25
|
+
apiKey?: string;
|
|
26
|
+
baseUrl?: string;
|
|
27
|
+
}
|
|
28
|
+
|
|
29
|
+
interface AgentTemplate {
|
|
30
|
+
id: string;
|
|
31
|
+
icon: string;
|
|
32
|
+
name: string;
|
|
33
|
+
description: string;
|
|
34
|
+
defaultAgentName: string;
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
const TEMPLATES: AgentTemplate[] = [
|
|
38
|
+
{ id: 'customer-service', icon: '🎧', name: '客服助手', description: '回答客户问题,自动学习产品知识', defaultAgentName: '我的客服助手' },
|
|
39
|
+
{ id: 'content-writer', icon: '✍️', name: '写作助手', description: '帮你写文章、邮件、报告', defaultAgentName: '我的写作助手' },
|
|
40
|
+
{ id: 'data-analyst', icon: '📊', name: '数据分析师', description: '分析数据,生成洞察', defaultAgentName: '我的数据分析师' },
|
|
41
|
+
{ id: 'translator', icon: '🌐', name: '翻译助手', description: '多语言翻译,越用越准', defaultAgentName: '我的翻译助手' },
|
|
42
|
+
{ id: 'general', icon: '🤖', name: '通用助手', description: '什么都能聊', defaultAgentName: '我的AI助手' },
|
|
43
|
+
];
|
|
44
|
+
|
|
45
|
+
const OPC_HOME = path.join(os.homedir(), '.opc');
|
|
46
|
+
const CONFIG_PATH = path.join(OPC_HOME, 'config.json');
|
|
47
|
+
|
|
48
|
+
// ── Readline helpers ────────────────────────────────────────────────────────
|
|
49
|
+
export function createRL(input?: NodeJS.ReadableStream, output?: NodeJS.WritableStream): readline.Interface {
|
|
50
|
+
return readline.createInterface({
|
|
51
|
+
input: input ?? process.stdin,
|
|
52
|
+
output: output ?? process.stdout,
|
|
53
|
+
});
|
|
54
|
+
}
|
|
55
|
+
|
|
56
|
+
function ask(rl: readline.Interface, question: string): Promise<string> {
|
|
57
|
+
return new Promise((resolve) => rl.question(question, (a) => resolve(a.trim())));
|
|
58
|
+
}
|
|
59
|
+
|
|
60
|
+
// ── HTTP helpers ────────────────────────────────────────────────────────────
|
|
61
|
+
function httpGet(url: string, timeout = 5000): Promise<{ status: number; body: string }> {
|
|
62
|
+
return new Promise((resolve, reject) => {
|
|
63
|
+
const mod = url.startsWith('https') ? https : http;
|
|
64
|
+
const req = mod.get(url, { timeout }, (res) => {
|
|
65
|
+
let body = '';
|
|
66
|
+
res.on('data', (chunk: Buffer) => { body += chunk; });
|
|
67
|
+
res.on('end', () => resolve({ status: res.statusCode ?? 0, body }));
|
|
68
|
+
});
|
|
69
|
+
req.on('error', reject);
|
|
70
|
+
req.on('timeout', () => { req.destroy(); reject(new Error('timeout')); });
|
|
71
|
+
});
|
|
72
|
+
}
|
|
73
|
+
|
|
74
|
+
// ── Step 1: Welcome ─────────────────────────────────────────────────────────
|
|
75
|
+
function printWelcome(): void {
|
|
76
|
+
console.log('');
|
|
77
|
+
console.log(c.bold(' 🎉 欢迎使用 OPC Agent!'));
|
|
78
|
+
console.log(c.dim(' 让我们用 3 分钟配置你的第一个 AI Agent。'));
|
|
79
|
+
console.log('');
|
|
80
|
+
}
|
|
81
|
+
|
|
82
|
+
// ── Step 2: Model detection ─────────────────────────────────────────────────
|
|
83
|
+
async function detectOllama(): Promise<{ running: boolean; models: string[] }> {
|
|
84
|
+
try {
|
|
85
|
+
const { body } = await httpGet('http://localhost:11434/api/tags');
|
|
86
|
+
const data = JSON.parse(body);
|
|
87
|
+
const models = (data.models || []).map((m: any) => m.name || m.model);
|
|
88
|
+
return { running: true, models };
|
|
89
|
+
} catch {
|
|
90
|
+
return { running: false, models: [] };
|
|
91
|
+
}
|
|
92
|
+
}
|
|
93
|
+
|
|
94
|
+
async function stepModel(rl: readline.Interface): Promise<SetupConfig> {
|
|
95
|
+
console.log(c.bold('📡 Step 1/4: 配置 AI 模型'));
|
|
96
|
+
console.log(c.dim(' 正在检测本地 Ollama...'));
|
|
97
|
+
console.log('');
|
|
98
|
+
|
|
99
|
+
const ollama = await detectOllama();
|
|
100
|
+
|
|
101
|
+
if (ollama.running) {
|
|
102
|
+
console.log(c.green(' ✔ Ollama 运行中!已安装模型:'));
|
|
103
|
+
if (ollama.models.length === 0) {
|
|
104
|
+
console.log(c.dim(' (无模型)'));
|
|
105
|
+
} else {
|
|
106
|
+
for (const m of ollama.models) {
|
|
107
|
+
console.log(` • ${m}`);
|
|
108
|
+
}
|
|
109
|
+
}
|
|
110
|
+
console.log('');
|
|
111
|
+
|
|
112
|
+
const hasChat = ollama.models.some((m) => m.startsWith('qwen2.5'));
|
|
113
|
+
const hasEmbed = ollama.models.some((m) => m.startsWith('nomic-embed-text'));
|
|
114
|
+
|
|
115
|
+
if (!hasChat) {
|
|
116
|
+
const dl = await ask(rl, ` 推荐模型 ${c.cyan('qwen2.5:7b')} 未安装,是否下载? [Y/n] `);
|
|
117
|
+
if (dl.toLowerCase() !== 'n') {
|
|
118
|
+
console.log(c.dim(' → 请在终端运行: ollama pull qwen2.5:7b'));
|
|
119
|
+
}
|
|
120
|
+
}
|
|
121
|
+
|
|
122
|
+
if (!hasEmbed) {
|
|
123
|
+
const dl = await ask(rl, ` 推荐 Embedding 模型 ${c.cyan('nomic-embed-text')} 未安装,是否下载? [Y/n] `);
|
|
124
|
+
if (dl.toLowerCase() !== 'n') {
|
|
125
|
+
console.log(c.dim(' → 请在终端运行: ollama pull nomic-embed-text'));
|
|
126
|
+
}
|
|
127
|
+
}
|
|
128
|
+
|
|
129
|
+
console.log('');
|
|
130
|
+
return {
|
|
131
|
+
provider: 'ollama',
|
|
132
|
+
model: hasChat ? ollama.models.find((m) => m.startsWith('qwen2.5'))! : 'qwen2.5:7b',
|
|
133
|
+
embeddingModel: hasEmbed ? 'nomic-embed-text' : 'nomic-embed-text',
|
|
134
|
+
baseUrl: 'http://localhost:11434',
|
|
135
|
+
};
|
|
136
|
+
}
|
|
137
|
+
|
|
138
|
+
// Ollama not running
|
|
139
|
+
console.log(c.yellow(' ⚠ 未检测到 Ollama'));
|
|
140
|
+
console.log('');
|
|
141
|
+
console.log(' 请选择:');
|
|
142
|
+
console.log(` ${c.cyan('A')} ) 安装 Ollama(推荐,免费本地运行)`);
|
|
143
|
+
console.log(` ${c.cyan('B')} ) 使用云端 API`);
|
|
144
|
+
console.log('');
|
|
145
|
+
|
|
146
|
+
const choice = await ask(rl, ' 你的选择 [A/B]: ');
|
|
147
|
+
|
|
148
|
+
if (choice.toUpperCase() !== 'B') {
|
|
149
|
+
console.log('');
|
|
150
|
+
console.log(c.bold(' 📦 安装 Ollama:'));
|
|
151
|
+
console.log('');
|
|
152
|
+
if (process.platform === 'win32') {
|
|
153
|
+
console.log(' 1. 访问 https://ollama.com/download');
|
|
154
|
+
console.log(' 2. 下载 Windows 版并安装');
|
|
155
|
+
console.log(' 3. 安装后运行: ollama pull qwen2.5:7b');
|
|
156
|
+
} else if (process.platform === 'darwin') {
|
|
157
|
+
console.log(' brew install ollama');
|
|
158
|
+
console.log(' ollama serve &');
|
|
159
|
+
console.log(' ollama pull qwen2.5:7b');
|
|
160
|
+
} else {
|
|
161
|
+
console.log(' curl -fsSL https://ollama.com/install.sh | sh');
|
|
162
|
+
console.log(' ollama serve &');
|
|
163
|
+
console.log(' ollama pull qwen2.5:7b');
|
|
164
|
+
}
|
|
165
|
+
console.log('');
|
|
166
|
+
console.log(c.dim(' 安装完成后重新运行 opc setup'));
|
|
167
|
+
return { provider: 'ollama', model: 'qwen2.5:7b', embeddingModel: 'nomic-embed-text', baseUrl: 'http://localhost:11434' };
|
|
168
|
+
}
|
|
169
|
+
|
|
170
|
+
// Cloud API
|
|
171
|
+
console.log('');
|
|
172
|
+
console.log(' 选择云端 Provider:');
|
|
173
|
+
console.log(` ${c.cyan('1')} ) OpenAI`);
|
|
174
|
+
console.log(` ${c.cyan('2')} ) DeepSeek`);
|
|
175
|
+
console.log(` ${c.cyan('3')} ) 通义千问 (Qwen)`);
|
|
176
|
+
console.log(` ${c.cyan('4')} ) Anthropic`);
|
|
177
|
+
console.log('');
|
|
178
|
+
|
|
179
|
+
const providerChoice = await ask(rl, ' 选择 [1-4]: ');
|
|
180
|
+
const providers: Record<string, { provider: SetupConfig['provider']; baseUrl: string; model: string; testUrl: string }> = {
|
|
181
|
+
'1': { provider: 'openai', baseUrl: 'https://api.openai.com/v1', model: 'gpt-4o-mini', testUrl: 'https://api.openai.com/v1/models' },
|
|
182
|
+
'2': { provider: 'deepseek', baseUrl: 'https://api.deepseek.com/v1', model: 'deepseek-chat', testUrl: 'https://api.deepseek.com/v1/models' },
|
|
183
|
+
'3': { provider: 'qwen', baseUrl: 'https://dashscope.aliyuncs.com/compatible-mode/v1', model: 'qwen-plus', testUrl: 'https://dashscope.aliyuncs.com/compatible-mode/v1/models' },
|
|
184
|
+
'4': { provider: 'anthropic', baseUrl: 'https://api.anthropic.com/v1', model: 'claude-3-5-sonnet-20241022', testUrl: 'https://api.anthropic.com/v1/models' },
|
|
185
|
+
};
|
|
186
|
+
|
|
187
|
+
const p = providers[providerChoice] ?? providers['1'];
|
|
188
|
+
console.log('');
|
|
189
|
+
const apiKey = await ask(rl, ` 请输入 ${c.bold(p.provider)} API Key: `);
|
|
190
|
+
|
|
191
|
+
if (apiKey) {
|
|
192
|
+
console.log(c.dim(' 正在测试连接...'));
|
|
193
|
+
try {
|
|
194
|
+
await httpGet(p.testUrl);
|
|
195
|
+
console.log(c.green(' ✔ 连接成功!'));
|
|
196
|
+
} catch {
|
|
197
|
+
console.log(c.yellow(' ⚠ 无法验证连接,但配置已保存,稍后可测试'));
|
|
198
|
+
}
|
|
199
|
+
}
|
|
200
|
+
|
|
201
|
+
console.log('');
|
|
202
|
+
return { provider: p.provider, model: p.model, apiKey, baseUrl: p.baseUrl };
|
|
203
|
+
}
|
|
204
|
+
|
|
205
|
+
// ── Step 3: Choose template ─────────────────────────────────────────────────
|
|
206
|
+
async function stepTemplate(rl: readline.Interface): Promise<AgentTemplate> {
|
|
207
|
+
console.log(c.bold('📋 Step 2/4: 选择 Agent 模板'));
|
|
208
|
+
console.log('');
|
|
209
|
+
|
|
210
|
+
for (let i = 0; i < TEMPLATES.length; i++) {
|
|
211
|
+
const t = TEMPLATES[i];
|
|
212
|
+
console.log(` ${c.cyan(String(i + 1))} ) ${t.icon} ${c.bold(t.name)} — ${c.dim(t.description)}`);
|
|
213
|
+
}
|
|
214
|
+
console.log(` ${c.cyan('6')} ) 📋 更多模板...(打开 Studio 浏览)`);
|
|
215
|
+
console.log('');
|
|
216
|
+
|
|
217
|
+
const choice = await ask(rl, ' 选择模板 [1-6]: ');
|
|
218
|
+
const idx = parseInt(choice, 10) - 1;
|
|
219
|
+
|
|
220
|
+
if (choice === '6') {
|
|
221
|
+
console.log(c.dim(' → 运行 opc studio 浏览更多模板'));
|
|
222
|
+
return TEMPLATES[4]; // default to general
|
|
223
|
+
}
|
|
224
|
+
|
|
225
|
+
const template = TEMPLATES[idx] ?? TEMPLATES[4];
|
|
226
|
+
console.log(c.green(` ✔ 已选择: ${template.icon} ${template.name}`));
|
|
227
|
+
console.log('');
|
|
228
|
+
return template;
|
|
229
|
+
}
|
|
230
|
+
|
|
231
|
+
// ── Step 4: Create agent ────────────────────────────────────────────────────
|
|
232
|
+
async function stepCreateAgent(rl: readline.Interface, template: AgentTemplate, config: SetupConfig): Promise<{ name: string; id: string; dir: string }> {
|
|
233
|
+
console.log(c.bold('🤖 Step 3/4: 创建 Agent'));
|
|
234
|
+
console.log('');
|
|
235
|
+
|
|
236
|
+
const name = (await ask(rl, ` Agent 名称 [${template.defaultAgentName}]: `)) || template.defaultAgentName;
|
|
237
|
+
const description = await ask(rl, ` 简短描述(可选): `);
|
|
238
|
+
|
|
239
|
+
// Generate a slug id
|
|
240
|
+
const id = name
|
|
241
|
+
.replace(/[^\w\u4e00-\u9fff]/g, '-')
|
|
242
|
+
.replace(/-+/g, '-')
|
|
243
|
+
.replace(/^-|-$/g, '')
|
|
244
|
+
.toLowerCase() || `agent-${Date.now()}`;
|
|
245
|
+
|
|
246
|
+
const agentDir = path.join(OPC_HOME, 'agents', id);
|
|
247
|
+
fs.mkdirSync(agentDir, { recursive: true });
|
|
248
|
+
|
|
249
|
+
// Write agent config
|
|
250
|
+
const agentConfig = {
|
|
251
|
+
id,
|
|
252
|
+
name,
|
|
253
|
+
description: description || template.description,
|
|
254
|
+
template: template.id,
|
|
255
|
+
provider: config.provider,
|
|
256
|
+
model: config.model,
|
|
257
|
+
embeddingModel: config.embeddingModel,
|
|
258
|
+
apiKey: config.apiKey,
|
|
259
|
+
baseUrl: config.baseUrl,
|
|
260
|
+
createdAt: new Date().toISOString(),
|
|
261
|
+
};
|
|
262
|
+
|
|
263
|
+
fs.writeFileSync(path.join(agentDir, 'config.json'), JSON.stringify(agentConfig, null, 2));
|
|
264
|
+
|
|
265
|
+
// Initialize brain directory
|
|
266
|
+
const brainDir = path.join(agentDir, 'brain');
|
|
267
|
+
fs.mkdirSync(brainDir, { recursive: true });
|
|
268
|
+
fs.writeFileSync(path.join(brainDir, 'README.md'), `# ${name}\n\n将知识文档放在这里,Agent 会自动学习。\n`);
|
|
269
|
+
|
|
270
|
+
// Save global config
|
|
271
|
+
fs.mkdirSync(OPC_HOME, { recursive: true });
|
|
272
|
+
const globalConfig = fs.existsSync(CONFIG_PATH) ? JSON.parse(fs.readFileSync(CONFIG_PATH, 'utf-8')) : {};
|
|
273
|
+
globalConfig.defaultAgent = id;
|
|
274
|
+
globalConfig.provider = config.provider;
|
|
275
|
+
globalConfig.model = config.model;
|
|
276
|
+
globalConfig.baseUrl = config.baseUrl;
|
|
277
|
+
if (config.apiKey) globalConfig.apiKey = config.apiKey;
|
|
278
|
+
if (config.embeddingModel) globalConfig.embeddingModel = config.embeddingModel;
|
|
279
|
+
fs.writeFileSync(CONFIG_PATH, JSON.stringify(globalConfig, null, 2));
|
|
280
|
+
|
|
281
|
+
console.log('');
|
|
282
|
+
console.log(` ${c.green('✔')} Agent 已创建: ${c.bold(name)}`);
|
|
283
|
+
console.log(` ${c.dim(' 目录: ' + agentDir)}`);
|
|
284
|
+
console.log('');
|
|
285
|
+
|
|
286
|
+
return { name, id, dir: agentDir };
|
|
287
|
+
}
|
|
288
|
+
|
|
289
|
+
// ── Step 5: Completion ──────────────────────────────────────────────────────
|
|
290
|
+
function printCompletion(agentName: string): void {
|
|
291
|
+
console.log(c.bold(` ✅ 你的 AI Agent「${agentName}」已创建!`));
|
|
292
|
+
console.log('');
|
|
293
|
+
console.log(' 启动方式:');
|
|
294
|
+
console.log(` ${c.cyan('opc studio')} — 打开网页管理界面`);
|
|
295
|
+
console.log(` ${c.cyan('opc chat')} — 终端直接对话`);
|
|
296
|
+
console.log(` ${c.cyan('opc start')} — 后台运行`);
|
|
297
|
+
console.log('');
|
|
298
|
+
console.log(c.dim(' 你的 Agent 会自动学习和进化,越用越聪明!🧬'));
|
|
299
|
+
console.log('');
|
|
300
|
+
}
|
|
301
|
+
|
|
302
|
+
// ── Main entry ──────────────────────────────────────────────────────────────
|
|
303
|
+
export async function runSetup(input?: NodeJS.ReadableStream, output?: NodeJS.WritableStream): Promise<void> {
|
|
304
|
+
const rl = createRL(input, output);
|
|
305
|
+
try {
|
|
306
|
+
printWelcome();
|
|
307
|
+
const config = await stepModel(rl);
|
|
308
|
+
const template = await stepTemplate(rl);
|
|
309
|
+
const agent = await stepCreateAgent(rl, template, config);
|
|
310
|
+
printCompletion(agent.name);
|
|
311
|
+
} finally {
|
|
312
|
+
rl.close();
|
|
313
|
+
}
|
|
314
|
+
}
|