@sleep2agi/agent-node 2.2.0-preview.0 → 2.2.0-preview.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cli.js +1 -1
- package/package.json +1 -1
package/dist/cli.js
CHANGED
|
@@ -36,7 +36,7 @@ Runtime:
|
|
|
36
36
|
收到来自 ${Z} 的任务:
|
|
37
37
|
|
|
38
38
|
${z}`:W,B=process.env.COMMHUB_URL||d,K=process.env.COMMHUB_TOKEN||L,j={};if(B)j.commhub={type:"http",url:`${B}/mcp`,headers:K?{Authorization:`Bearer ${K}`}:void 0};let F=(()=>{try{let{execSync:J}=k("child_process"),w=k("fs");try{let D=k.resolve("@anthropic-ai/claude-agent-sdk-linux-x64/claude");if(w.existsSync(D))return J(`${D} --version`,{stdio:"pipe"}),V(`[claude] using glibc binary: ${D}`),D}catch{}try{let D=J("which claude",{encoding:"utf-8"}).trim();if(D)return V(`[claude] using global binary: ${D}`),D}catch{}V("[claude] no binary resolved, falling back to SDK default");return}catch{return}})(),b={model:y||void 0,tools:I.length?I:void 0,maxTurns:Az,permissionMode:"bypassPermissions",allowDangerouslySkipPermissions:!0,settingSources:[],mcpServers:Object.keys(j).length?j:void 0,pathToClaudeCodeExecutable:F,env:process.env,cwd:process.cwd(),stderr:(J)=>{if(J.trim())V(`[stderr] ${J.trim().slice(0,300)}`)},hooks:{PreToolUse:[{hooks:[async(J)=>{return V(`[tool] ${J.tool_name}(${JSON.stringify(J.tool_input).slice(0,80)})`),{continue:!0}}]}]}};if($z>0)b.maxBudgetUsd=$z;if(E)b.systemPrompt=E;if(m)b.resume=m;let O="",U=Date.now();V(`[claude] claudePath=${F||"SDK default"}, mcpServers=${Object.keys(j).join(",")||"none"}`);for await(let J of Y({prompt:q,options:b})){let w=J;if(w.type==="system"&&w.subtype==="init")m=w.session_id,V(`[claude] session=${w.session_id?.slice(0,8)} model=${y||"default"}`),Fz(w.session_id);if(w.type==="result"){let D=Date.now()-U,e=w.usage||{};V(`[claude] ${w.subtype} | ${D}ms | $${w.total_cost_usd?.toFixed(4)||"?"} | in=${e.input_tokens||0} out=${e.output_tokens||0} | turns=${w.num_turns}`),O=w.subtype==="success"?w.result||"任务完成":`执行出错: ${w.error||w.result||"未知错误"}`}}return O}var x=null,mz=E||[`你是 ${H},一个 AI Agent 节点,工作目录:${process.cwd()}。`,"你通过通信网络(CommHub)接收任务并和其他 agent 协作。","","【可用通信工具】","- mcp_commhub__send_task(alias, task):派任务给指定 agent,等其 LLM 处理完返回 reply(同步语义)。","- mcp_commhub__send_message(alias, message):发聊天消息(不要求对方回复)。","- mcp_commhub__get_task(task_id):查询某任务的当前状态/reply。","- mcp_commhub__get_all_status():查看网络上所有在线 agent。","","【协作模式】","当你的任务需要其他 agent 的能力时:","1. 先 get_all_status 看哪些 agent 在线。","2. 用 send_task(alias, task, parent_task_id=<env CURRENT_TASK_ID>) 派给合适的 agent。"," ⚠ 必须把 parent_task_id 设成你当前任务的 ID(环境变量 CURRENT_TASK_ID 里),系统会自动把子任务最终结果串回给你的上游。",'3. 用 get_task 轮询子任务直到 status=replied/failed。允许中途汇报"还在等",但要继续轮询。',"4. 拿到 reply 后整合进你给上游的最终汇报。"," 即使你的 session 中途断开,只要 parent_task_id 设了,结果也会被系统自动 chain 回上游,不必焦虑。","","【禁止】",'- 不要回复"收到""好的""ok""在线""待命"等无内容确认。',"- 不要给自己发任务(会死循环)。","- 收到的若是 reply 类型,不要再 send_task 给原方(会乒乓回复)。","- 没有新任务时保持沉默,不主动发消息。","","你的最终回复会被系统自动 send_reply 给任务发起者。"].join(`
|
|
39
|
-
`),Xz={model_auto_compact_token_limit:200000,developer_instructions:mz};async function nz(z,Z,$){try{let{execSync:B}=await import("child_process"),K=B("which codex 2>/dev/null",{encoding:"utf-8"}).trim();if(K){let j=K.replace(/\/codex$/,"");if(!process.env.PATH?.includes(j))process.env.PATH=`${j}:${process.env.PATH}`}}catch{}let Q;try{({Codex:Q}=await import("@openai/codex-sdk"))}catch{throw Error("@openai/codex-sdk not installed. Run: npm install -g @openai/codex-sdk @openai/codex")}if(!x){let B=new Q({config:Xz}),j={skipGitRepoCheck:!0,approvalPolicy:"never",model:y||"gpt-5.4",sandboxMode:"danger-full-access",modelReasoningEffort:"low"};if(A)x=B.resumeThread(A,j),V(`codex resumed thread: ${A}`);else x=B.startThread(j)}V(`[codex] model=${y||"gpt-5.4"} thread=${x?.id||"new"}`);let X=z,W=$?.length?[{type:"text",text:X},...$.map((B)=>({type:"local_image",path:B}))]:X,q=Date.now();try{let{events:B}=await x.runStreamed(W),K="",j=null,F=0;for await(let U of B)if(U.type==="item.started"){let J=U.item;P(`[codex] ${J.type}${J.command?`: ${J.command.slice(0,60)}`:J.tool?`: ${J.server}/${J.tool}`:""}`)}else if(U.type==="item.completed"){F++;let J=U.item;if(J.type==="agent_message")K=J.text||"";if(J.type==="command_execution")P(`[codex] cmd exit=${J.exit_code} | ${J.aggregated_output?.slice(0,80)}`);if(J.type==="reasoning")P(`[codex] thinking: ${J.text?.slice(0,80)}`);if(J.type==="mcp_tool_call")P(`[codex] mcp: ${J.server}/${J.tool} → ${J.status}`)}else if(U.type==="turn.completed")j=U.usage;let b=Date.now()-q,O=j?.input_tokens||0;if(V(`[codex] done | ${b}ms | in=${O} out=${j?.output_tokens||0} | items=${F}`),x?.id)Fz(x.id);return K||"(无回复)"}catch(B){V(`codex thread error: ${B.message}, 重建`),x=new Q({config:Xz}).startThread({skipGitRepoCheck:!0,approvalPolicy:"never",model:y||"gpt-5.4",sandboxMode:"danger-full-access",modelReasoningEffort:"low"});let j=await x.run(W),F=Date.now()-q;return V(`[codex] retry done | ${F}ms`),j.finalResponse||"(无回复)"}}async function lz(z,Z){let $=process.env.ANTHROPIC_API_KEY||process.env.OPENAI_API_KEY||process.env.MINIMAX_CODING_API_KEY||G.apiKey||"",Q=process.env.ANTHROPIC_BASE_URL||G.anthropicBaseUrl||"",Y=process.env.OPENAI_BASE_URL||G.apiBaseUrl||"https://api.openai.com/v1",X=y||"gpt-4o-mini",W=!!Q,B=(Q||Y).replace(/\/v1\/?$/,"");if(!$)return"错误: 需要设置 ANTHROPIC_API_KEY, OPENAI_API_KEY, 或 MINIMAX_CODING_API_KEY";let K=E||`你是 ${H},一个 AI 助手。收到来自 ${Z} 的任务后简要执行并汇报。`,j=Date.now();V(`[http-api] model=${X} format=${W?"anthropic":"openai"} base=${B.replace(/\/v1$/,"")}`);let F="",b=null;if(W){let U=await fetch(`${B}/v1/messages`,{method:"POST",headers:{"Content-Type":"application/json","x-api-key":$,"anthropic-version":"2023-06-01"},body:JSON.stringify({model:X,system:K,messages:[{role:"user",content:z}],max_tokens:2000})});if(!U.ok){let D=await U.text();return`Anthropic API 错误 ${U.status}: ${D.slice(0,200)}`}let J=await U.json();F=(Array.isArray(J.content)?J.content:[]).filter((D)=>D.type==="text").map((D)=>D.text).join(`
|
|
39
|
+
`),Xz={model_auto_compact_token_limit:200000,developer_instructions:mz};async function nz(z,Z,$){try{let{execSync:B}=await import("child_process"),K=B("which codex 2>/dev/null",{encoding:"utf-8"}).trim();if(K){let j=K.replace(/\/codex$/,"");if(!process.env.PATH?.includes(j))process.env.PATH=`${j}:${process.env.PATH}`}}catch{}let Q;try{({Codex:Q}=await import("@openai/codex-sdk"))}catch{throw Error("@openai/codex-sdk not installed. Run: npm install -g @openai/codex-sdk @openai/codex")}if(!x){let B=new Q({config:Xz}),j={skipGitRepoCheck:!0,approvalPolicy:"never",model:y||"gpt-5.4",sandboxMode:"danger-full-access",modelReasoningEffort:"low"};if(A)x=B.resumeThread(A,j),V(`codex resumed thread: ${A}`);else x=B.startThread(j)}V(`[codex] model=${y||"gpt-5.4"} thread=${x?.id||"new"}`);let X=z,W=$?.length?[{type:"text",text:X},...$.map((B)=>({type:"local_image",path:B}))]:X,q=Date.now();try{let{events:B}=await x.runStreamed(W),K="",j=null,F=0;for await(let U of B)if(U.type==="item.started"){let J=U.item;P(`[codex] ${J.type}${J.command?`: ${J.command.slice(0,60)}`:J.tool?`: ${J.server}/${J.tool}`:""}`)}else if(U.type==="item.completed"){F++;let J=U.item;if(J.type==="agent_message")K=J.text||"";if(J.type==="command_execution")P(`[codex] cmd exit=${J.exit_code} | ${J.aggregated_output?.slice(0,80)}`);if(J.type==="reasoning")P(`[codex] thinking: ${J.text?.slice(0,80)}`);if(J.type==="mcp_tool_call")P(`[codex] mcp: ${J.server}/${J.tool} → ${J.status}`)}else if(U.type==="turn.completed")j=U.usage;let b=Date.now()-q,O=j?.input_tokens||0;if(V(`[codex] done | ${b}ms | in=${O} out=${j?.output_tokens||0} | items=${F}`),x?.id)Fz(x.id);return K||"(无回复)"}catch(B){V(`codex thread error: ${B.message}, 重建`),x=new Q({config:Xz}).startThread({skipGitRepoCheck:!0,approvalPolicy:"never",model:y||"gpt-5.4",sandboxMode:"danger-full-access",modelReasoningEffort:"low"});let j=await x.run(W),F=Date.now()-q;return V(`[codex] retry done | ${F}ms`),j.finalResponse||"(无回复)"}}async function lz(z,Z){let $=process.env.ANTHROPIC_API_KEY||process.env.ANTHROPIC_AUTH_TOKEN||process.env.OPENAI_API_KEY||process.env.MINIMAX_CODING_API_KEY||G.apiKey||"",Q=process.env.ANTHROPIC_BASE_URL||G.anthropicBaseUrl||"",Y=process.env.OPENAI_BASE_URL||G.apiBaseUrl||"https://api.openai.com/v1",X=y||"gpt-4o-mini",W=!!Q,B=(Q||Y).replace(/\/v1\/?$/,"");if(!$)return"错误: 需要设置 ANTHROPIC_API_KEY, OPENAI_API_KEY, 或 MINIMAX_CODING_API_KEY";let K=E||`你是 ${H},一个 AI 助手。收到来自 ${Z} 的任务后简要执行并汇报。`,j=Date.now();V(`[http-api] model=${X} format=${W?"anthropic":"openai"} base=${B.replace(/\/v1$/,"")}`);let F="",b=null;if(W){let U=await fetch(`${B}/v1/messages`,{method:"POST",headers:{"Content-Type":"application/json","x-api-key":$,"anthropic-version":"2023-06-01"},body:JSON.stringify({model:X,system:K,messages:[{role:"user",content:z}],max_tokens:2000})});if(!U.ok){let D=await U.text();return`Anthropic API 错误 ${U.status}: ${D.slice(0,200)}`}let J=await U.json();F=(Array.isArray(J.content)?J.content:[]).filter((D)=>D.type==="text").map((D)=>D.text).join(`
|
|
40
40
|
`)||"",b=J.usage}else{let U=await fetch(`${B}/chat/completions`,{method:"POST",headers:{"Content-Type":"application/json",Authorization:`Bearer ${$}`},body:JSON.stringify({model:X,messages:[{role:"system",content:K},{role:"user",content:z}],max_tokens:2000})});if(!U.ok){let w=await U.text();return`OpenAI API 错误 ${U.status}: ${w.slice(0,200)}`}let J=await U.json();F=J.choices?.[0]?.message?.content||"",b=J.usage}let O=Date.now()-j;return V(`[http-api] done | ${O}ms | in=${b?.input_tokens||b?.prompt_tokens||0} out=${b?.output_tokens||b?.completion_tokens||0}`),F||"(无回复)"}var Yz=Promise.resolve();function vz(z,Z,$,Q){let Y=async()=>{let W=process.env.CURRENT_TASK_ID;if($)process.env.CURRENT_TASK_ID=$;else delete process.env.CURRENT_TASK_ID;try{if(C==="codex")return await nz(z,Z,Q);if(C==="http")return await lz(z,Z);return await cz(z,Z)}finally{if(W!==void 0)process.env.CURRENT_TASK_ID=W;else delete process.env.CURRENT_TASK_ID}},X=Yz.then(Y,Y);return Yz=X.then(()=>{},()=>{}),X}async function rz(z,Z,$=null){V(`→ processing [${C}]: ${z.slice(0,80)}`),await r("working",z.slice(0,200)).catch(()=>{});let Q,Y=!1;try{Q=await vz(z,Z,$)}catch(X){Q=`${C} 错误: ${X.message}`,Y=!0,h(`✗ ${X.message}`)}finally{await r("idle").catch(()=>{})}if(!Y&&/(API 错误|API error|需要设置.*KEY|missing.*key|issue with the selected model|may not have access|may not exist|model.+not.+(found|available))/i.test(Q))Y=!0;return{text:Q,failed:Y}}var o={},gz=5000,iz=new Set(["收到","好的","ok","嗯","是的","了解","明白","确认","done","ack","roger","yes","no","在线","待命","正常","保持在线","通信正常","已收到","收到了","好","行","noted","copy","received","understood","等待任务","等待中","等待指令","无新任务","idle","waiting"]);function Uz(z,Z=!1){if(!z)return!0;let Q=z.trim().replace(/^[\[【].+?[\]】]\s*/,"").trim().toLowerCase().replace(/[\s。!?.!?✅❌👀⏳,,]+$/g,"").trim();if(iz.has(Q))return!0;if(/^[\p{Emoji}\s]+$/u.test(z.trim())&&!/[0-9a-zA-Z#*]/.test(z))return!0;return!1}function oz(z,Z,$){if(z===H)return"self";if(Z.startsWith(`[${H}]`))return"own-prefix";if(z!=="hub"&&z!=="api"){let Q=Date.now();if(o[z]&&Q-o[z]<gz)return"cooldown"}if($!=="task"&&$!=="broadcast"&&Uz(Z))return"low-value-inbound";return null}async function az(){let z=await pz();if(!z.length)return;for(let Z of z){let $=Z.from_session||"hub",Q=Z.content,Y=Z.type||"task";if(V(`← [${$}] (${Y}/${Z.priority||"normal"}) ${Q.slice(0,100)}`),await fz(Z.id),Y!=="task"&&Y!=="broadcast"){P(`skip non-task message: type=${Y}`);continue}let X=oz($,Q,Y);if(X){P(`skip message from ${$}: ${X}`);continue}let{text:W,failed:q}=await rz(Q,$,Z.id);if(V(`processTask returned: "${W.slice(0,80)}" (${W.length} chars, failed=${q})`),!q&&Uz(W,!0)){V(`skip reply: low-value (${W.slice(0,30)})`);continue}try{V(`sending reply to ${$} (task ${Z.id.slice(0,8)}, status=${q?"failed":"replied"})...`),await dz($,`[${H}] ${W.slice(0,2000)}`,Z.id,q),o[$]=Date.now(),V(`→ [${$}] ${W.slice(0,100)}`)}catch(B){_(`reply failed: ${B.message}`)}}}function Dz(z){return String(z.from?.id||z.chat?.id||"")}function Mz(z){return z.from?.username||z.from?.first_name||Dz(z)||"telegram"}function tz(z,Z){if(z.allowFrom.length===0)return!0;let $=Dz(Z),Q=Z.from?.username?String(Z.from.username):"";return z.allowFrom.includes($)||!!Q&&z.allowFrom.includes(Q)}async function g(z,Z,$){let Q=await fetch(`${z.apiBase}/${Z}`,{method:"POST",headers:{"Content-Type":"application/json"},body:JSON.stringify($)}),Y=await Q.json();if(!Y.ok)throw Error(`telegram ${Z} failed: ${Y.description||Q.statusText}`);return Y.result}async function Vz(z,Z,$,Q){let Y=$.match(/[\s\S]{1,4096}/g)||["(无回复)"];for(let X=0;X<Y.length;X++)await g(z,"sendMessage",{chat_id:Z,text:Y[X],...Q&&X===0?{reply_to_message_id:Q}:{}})}async function Jz(z,Z,$){let Q=await g(z,"getFile",{file_id:Z}),Y=String(Q.file_path||""),X=await fetch(`${z.fileBase}/${Y}`);if(!X.ok)throw Error(`telegram file download failed: ${X.status} ${X.statusText}`);let W=Y.split(".").pop(),q=($||Y.split("/").pop()||Z).replace(/[^a-zA-Z0-9._-]/g,"_"),B=q.includes(".")||!W?q:`${q}.${W}`,K=M(z.channel.inboxDir,`${Date.now()}_${B}`);return t(K,Buffer.from(await X.arrayBuffer())),K}async function sz(z,Z){let $=Z.text||Z.caption||"",Q=[];if(Array.isArray(Z.photo)&&Z.photo.length>0){let X=Z.photo[Z.photo.length-1],W=await Jz(z,X.file_id,`photo_${Z.message_id}.jpg`);Q.push(W)}let Y=String(Z.document?.mime_type||"");if(Z.document&&Y.startsWith("image/")){let X=await Jz(z,Z.document.file_id,Z.document.file_name||`image_${Z.message_id}`);Q.push(X)}if(Q.length)$+=`
|
|
41
41
|
|
|
42
42
|
[Telegram 附件已下载]
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@sleep2agi/agent-node",
|
|
3
|
-
"version": "2.2.0-preview.
|
|
3
|
+
"version": "2.2.0-preview.1",
|
|
4
4
|
"description": "AI Agent runtime for CommHub networks. Supports Claude Agent SDK, Codex SDK, and OpenAI/Anthropic-compatible HTTP API.",
|
|
5
5
|
"bin": {
|
|
6
6
|
"agent-node": "./dist/cli.js"
|