@hhsw2015/task-master-ai 0.43.11 → 0.43.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai-services-unified-BPC7geqj.js +1 -0
- package/dist/{ai-services-unified-CEISDqFA.js → ai-services-unified-DRuFwL_e.js} +1 -1
- package/dist/assets/config.json +20 -10
- package/dist/{commands-KgO10GUj.js → commands-Dz0lboMw.js} +5 -5
- package/dist/{config-manager-3pTgfD7M.js → config-manager-ClruewMP.js} +2 -2
- package/dist/{config-manager-CgVZwH7E.js → config-manager-DbOMSXgZ.js} +1 -1
- package/dist/{dependency-manager-C-vZzEg8.js → dependency-manager-CvhwFgTc.js} +5 -5
- package/dist/mcp-server.js +2 -2
- package/dist/{profiles-Bcr6LYd4.js → profiles-7bbMp9Tz.js} +4 -4
- package/dist/research-Dk2DpbPy.js +1 -0
- package/dist/response-language-DOZ-CzdZ.js +1 -0
- package/dist/{response-language-CoZ1-GFh.js → response-language-DpiGaJZK.js} +1 -1
- package/dist/{sentry-BU9KKY_b.js → sentry-DP3nr70b.js} +1 -1
- package/dist/tag-management-CFQqYS7R.js +1 -0
- package/dist/{task-manager-usj-gUpk.js → task-manager-H0H_vNIg.js} +1 -1
- package/dist/task-master.js +1 -1
- package/dist/update-subtask-by-id-r9mGhUE1.js +1 -0
- package/dist/update-task-by-id-B9n9vQcI.js +1 -0
- package/dist/{utils-Dr4DuQDO.js → utils-CEHCzNvD.js} +1 -1
- package/package.json +1 -1
- package/dist/ai-services-unified-BpfeMi5b.js +0 -1
- package/dist/research-LCg5MHnP.js +0 -1
- package/dist/response-language-Cqon4F_J.js +0 -1
- package/dist/tag-management-DvcGPdM_.js +0 -1
- package/dist/update-subtask-by-id-B5uMpDq6.js +0 -1
- package/dist/update-task-by-id-D_HOOs3f.js +0 -1
package/dist/mcp-server.js
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
import{o as e,s as t}from"./ai-services-unified-
|
|
2
|
+
import{o as e,s as t}from"./ai-services-unified-DRuFwL_e.js";import{$t as n,Dt as r,Ht as i,Kt as a,Mt as o,Nt as s,On as c,Qt as l,Sn as u,Tt as d,fn as f,ht as p,in as m,ln as h,m as g,ot as _,pn as v,rn as y,st as b,tn as x,tt as S,vn as C,vt as w,wt as ee,x as te,zt as ne}from"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import{r as re}from"./sentry-DP3nr70b.js";import{$ as T,A as ie,B as E,C as ae,D as oe,Dt as se,Et as ce,F as le,G as ue,H as de,I as fe,J as pe,M as me,N as he,O as ge,P as _e,Q as ve,R as D,S as ye,U as be,V as xe,W as Se,X as Ce,Y as we,Z as Te,at as Ee,ct as De,dt as Oe,gt as ke,ht as Ae,i as je,it as Me,j as Ne,lt as Pe,n as Fe,pt as Ie,r as Le,rt as Re,st as ze,t as Be,ut as Ve,w as He,z as Ue}from"./dependency-manager-CvhwFgTc.js";import{t as We}from"./response-language-DpiGaJZK.js";import{a as Ge,c as Ke,f as qe,g as Je,h as Ye,l as O,m as k,p as Xe,v as A}from"./profiles-7bbMp9Tz.js";import j from"node:path";import M from"chalk";import N from"fs";import P from"path";import Ze from"os";import F from"node:fs";import{z as I}from"zod";import{fileURLToPath as Qe}from"url";import L from"dotenv";import*as R from"@sentry/node";import{FastMCP as $e}from"fastmcp";const z={debug:0,info:1,warn:2,error:3,success:4},et=z[te().toLowerCase()]??z.info;function B(e,...t){if(w())return;let n={debug:M.gray(`[DEBUG]`),info:M.blue(`[INFO]`),warn:M.yellow(`[WARN]`),error:M.red(`[ERROR]`),success:M.green(`[SUCCESS]`)};if(z[e]!==void 0&&z[e]>=et){let r=n[e]||``,i=t;try{switch(e){case`error`:i=t.map(e=>typeof e==`string`?M.red(e):e);break;case`warn`:i=t.map(e=>typeof e==`string`?M.yellow(e):e);break;case`success`:i=t.map(e=>typeof e==`string`?M.green(e):e);break;case`info`:i=t.map(e=>typeof e==`string`?M.blue(e):e);break;case`debug`:i=t.map(e=>typeof e==`string`?M.gray(e):e);break}}catch(e){console.error(`Internal Logger Error applying chalk color:`,e),i=t}console.error(r,...i)}}function tt(){let e=e=>(...t)=>B(e,...t);return{debug:e(`debug`),info:e(`info`),warn:e(`warn`),error:e(`error`),success:e(`success`),log:B}}var V=tt(),H=class extends Error{constructor(e,t={}){super(e),this.name=`MCPError`,this.code=t.code,this.cause=t.cause,this.mcpResponse=t.mcpResponse}},U=class extends H{constructor(e,t={}){super(e,t),this.name=`MCPSessionError`}},nt=class extends H{constructor(e,t={}){super(e,t),this.name=`MCPSamplingError`}};function W(e){if(e instanceof H)return e;let t=e.message||`Unknown MCP error`,n=e;return t.includes(`session`)||t.includes(`connection`)?new U(t,{cause:n,code:`SESSION_ERROR`}):t.includes(`sampling`)||t.includes(`timeout`)?new nt(t,{cause:n,code:`SAMPLING_ERROR`}):t.includes(`capabilities`)||t.includes(`not supported`)?new U(t,{cause:n,code:`CAPABILITY_ERROR`}):new H(t,{cause:n,code:`UNKNOWN_ERROR`})}function rt(e){let t=e.trim();t=t.replace(/^```json\s*/gm,``),t=t.replace(/^```\s*/gm,``),t=t.replace(/```\s*$/gm,``),t=t.replace(/^const\s+\w+\s*=\s*/,``),t=t.replace(/^let\s+\w+\s*=\s*/,``),t=t.replace(/^var\s+\w+\s*=\s*/,``),t=t.replace(/;?\s*$/,``),t=t.replace(/^.*?(?=\{|\[)/s,``),t.split(`
|
|
3
3
|
`);let n=-1,r=0,i=!1,a=!1;for(let e=0;e<t.length;e++){let o=t[e];if(a){a=!1;continue}if(o===`\\`){a=!0;continue}if(o===`"`&&!a){i=!i;continue}if(!i){if(o===`{`||o===`[`)r++;else if((o===`}`||o===`]`)&&(r--,r===0)){n=e;break}}}if(n>-1&&(t=t.substring(0,n+1)),n===-1){let e=t.match(/{[\s\S]*}/),n=t.match(/\[[\s\S]*\]/);e?t=e[0]:n&&(t=n[0])}try{return JSON.parse(t),t}catch{try{let e=t.replace(/([{,]\s*)([a-zA-Z_$][a-zA-Z0-9_$]*)\s*:/g,`$1"$2":`).replace(/'/g,`"`).replace(/,\s*([}\]])/g,`$1`);return JSON.parse(e),e}catch{return e}}}function G(e){let t=[],n=``;for(let r of e)r.role===`system`?n=K(r.content):(r.role===`user`||r.role===`assistant`)&&t.push({role:r.role,content:{type:`text`,text:K(r.content)}});return{messages:t,systemPrompt:n}}function it(e){let t=``,n=null,r=`stop`,i=[];return typeof e==`string`?t=e:e.content?(t=K(e.content),n=e.usage,r=e.finishReason||`stop`):e.text?(t=e.text,n=e.usage,r=e.finishReason||`stop`):(t=JSON.stringify(e),i.push(`Unexpected MCP response format, used JSON fallback`)),{text:t,usage:n,finishReason:r,warnings:i}}function K(e){return typeof e==`string`?e:Array.isArray(e)?e.map(e=>typeof e==`string`?e:e.type===`text`&&e.text||e.text?e.text:``).filter(e=>e.length>0).join(` `):e&&typeof e==`object`&&(e.type===`text`&&e.text||e.text)?e.text:String(e||``)}function at(e,t=`result`){try{let n=q(e);return`
|
|
4
4
|
CRITICAL JSON GENERATION INSTRUCTIONS:
|
|
5
5
|
|
|
@@ -41,4 +41,4 @@ Current Tag: ${n.currentTag}`),{content:[{type:`text`,text:r}],isError:!0}}funct
|
|
|
41
41
|
|
|
42
42
|
### Results
|
|
43
43
|
|
|
44
|
-
${y.result}`;if(e){let{updateSubtaskById:e}=await import(`./update-subtask-by-id-B5uMpDq6.js`);await e(P.join(d,`.taskmaster`,`tasks`,`tasks.json`),l,n,!1,{session:p,mcpLog:m,commandName:`research-save`,outputType:`mcp`,projectRoot:d,tag:f},`json`),t.info(`Research saved to subtask ${l}`)}else{let e=(await import(`./update-task-by-id-D_HOOs3f.js`)).default,r=parseInt(l,10);await e(P.join(d,`.taskmaster`,`tasks`,`tasks.json`),r,n,!1,{session:p,mcpLog:m,commandName:`research-save`,outputType:`mcp`,projectRoot:d,tag:f},`json`,!0),t.info(`Research saved to task ${l}`)}}catch(e){t.warn(`Error saving research to task/subtask: ${e.message}`)}return _(),{success:!0,data:{query:y.query,result:y.result,contextSize:y.contextSize,contextTokens:y.contextTokens,tokenBreakdown:y.tokenBreakdown,systemPromptTokens:y.systemPromptTokens,userPromptTokens:y.userPromptTokens,totalInputTokens:y.totalInputTokens,detailLevel:y.detailLevel,telemetryData:y.telemetryData,tagInfo:y.tagInfo,savedFilePath:y.savedFilePath}}}catch(e){return _(),t.error(`Error in researchDirect: ${e.message}`),{success:!1,error:{code:e.code||`RESEARCH_ERROR`,message:e.message}}}}async function mn(e,t,n={}){let{tasksJsonPath:r,id:i,strength:a=`regular`,prompt:o,research:s=!1,projectRoot:c,tag:l}=e,{session:u}=n;b();let d=D(t);try{if(!r)return t.error(`scopeDownDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i)return t.error(`Missing required parameter: id`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`The id parameter is required for scoping down tasks`}};let e=i.split(`,`).map(e=>parseInt(e.trim(),10));t.info(`Scoping down tasks: ${e.join(`, `)}, strength: ${a}, research: ${s}`);let n=await oe(r,e,a,o,{session:u,mcpLog:d,projectRoot:c,commandName:`scope-down`,outputType:`mcp`,tag:l,research:s},`json`);return _(),{success:!0,data:{updatedTasks:n.updatedTasks,tasksUpdated:n.updatedTasks.length,message:`Successfully scoped down ${n.updatedTasks.length} task(s)`,telemetryData:n.telemetryData}}}catch(e){return _(),t.error(`Error in scopeDownDirect: ${e.message}`),{success:!1,error:{code:e.code||`SCOPE_DOWN_ERROR`,message:e.message}}}}async function hn(e,t,n={}){let{tasksJsonPath:r,id:i,strength:a=`regular`,prompt:o,research:s=!1,projectRoot:c,tag:l}=e,{session:u}=n;b();let d=D(t);try{if(!r)return t.error(`scopeUpDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i)return t.error(`Missing required parameter: id`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`The id parameter is required for scoping up tasks`}};let e=i.split(`,`).map(e=>parseInt(e.trim(),10));t.info(`Scoping up tasks: ${e.join(`, `)}, strength: ${a}, research: ${s}`);let n=await ge(r,e,a,o,{session:u,mcpLog:d,projectRoot:c,commandName:`scope-up`,outputType:`mcp`,tag:l,research:s},`json`);return _(),{success:!0,data:{updatedTasks:n.updatedTasks,tasksUpdated:n.updatedTasks.length,message:`Successfully scoped up ${n.updatedTasks.length} task(s)`,telemetryData:n.telemetryData}}}catch(e){return _(),t.error(`Error in scopeUpDirect: ${e.message}`),{success:!1,error:{code:e.code||`SCOPE_UP_ERROR`,message:e.message}}}}async function gn(e,t,n={}){let{session:r}=n,{tasksJsonPath:i,id:a,prompt:o,research:s,metadata:c,projectRoot:l,tag:u}=e,d=D(t);try{if(d.info(`Updating subtask by ID via direct function. ID: ${a}, ProjectRoot: ${l}`),!i){let e=`tasksJsonPath is required but was not provided.`;return d.error(e),{success:!1,error:{code:`MISSING_ARGUMENT`,message:e}}}if(!a||typeof a!=`string`||!a.trim()){let e=`Subtask ID cannot be empty.`;return d.error(e),{success:!1,error:{code:`INVALID_SUBTASK_ID`,message:e}}}if(!o&&!c){let e=`No prompt or metadata specified. Please provide information to append or metadata to update.`;return d.error(e),{success:!1,error:{code:`MISSING_PROMPT`,message:e}}}let e=String(a).trim(),n=i,f=s===!0;t.info(`Updating subtask with ID ${e} with prompt "${o||`(metadata-only)`}" and research: ${f}`);let p=w();p||b();try{let t=await He(n,e,o,f,{mcpLog:d,session:r,projectRoot:l,tag:u,commandName:`update-subtask`,outputType:`mcp`,metadata:c},`json`);if(!t||t.updatedSubtask===null){let e=`Subtask ${a} or its parent task not found.`;return d.error(e),{success:!1,error:{code:`SUBTASK_NOT_FOUND`,message:e}}}let i=e.split(`.`)[0],s=`Successfully updated subtask with ID ${e}`;return d.success(s),{success:!0,data:{message:`Successfully updated subtask with ID ${e}`,subtaskId:e,parentId:i,subtask:t.updatedSubtask,tasksPath:n,useResearch:f,telemetryData:t.telemetryData,tagInfo:t.tagInfo}}}catch(e){return d.error(`Error updating subtask by ID: ${e.message}`),{success:!1,error:{code:`UPDATE_SUBTASK_CORE_ERROR`,message:e.message||`Unknown error updating subtask`}}}finally{!p&&w()&&_()}}catch(e){return d.error(`Setup error in updateSubtaskByIdDirect: ${e.message}`),w()&&_(),{success:!1,error:{code:`DIRECT_FUNCTION_SETUP_ERROR`,message:e.message||`Unknown setup error`}}}}async function _n(e,t,n={}){let{session:r}=n,{tasksJsonPath:i,id:a,prompt:o,research:s,append:c,metadata:l,projectRoot:u,tag:d}=e,f=D(t);try{if(f.info(`Updating task by ID via direct function. ID: ${a}, ProjectRoot: ${u}`),!a){let e=`No task ID specified. Please provide a task ID to update.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}if(!o&&!l){let e=`No prompt or metadata specified. Please provide a prompt with new information or metadata for the task update.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let t;if(typeof a==`string`)t=a;else if(typeof a==`number`)t=String(a);else{let e=`Invalid task ID type: ${typeof a}. Task ID must be a string or number.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let n=i||E({projectRoot:u,file:e.file},f);if(!n){let e=`tasks.json path could not be resolved.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let p=s===!0;f.info(`Updating task with ID ${t} with prompt "${o||`(metadata-only)`}" and research: ${p}`);let m=w();m||b();try{let e=await ae(n,t,o,p,{mcpLog:f,session:r,projectRoot:u,tag:d,commandName:`update-task`,outputType:`mcp`,metadata:l},`json`,c||!1);if(!e||e.updatedTask===null){let n=`Task ${t} was not updated (likely already completed).`;return f.info(n),{success:!0,data:{message:n,taskId:t,updated:!1,telemetryData:e?.telemetryData,tagInfo:e?.tagInfo}}}let i=`Successfully updated task with ID ${t} based on the prompt`;return f.info(i),{success:!0,data:{message:i,taskId:t,tasksPath:n,useResearch:p,updated:!0,updatedTask:e.updatedTask,telemetryData:e.telemetryData,tagInfo:e.tagInfo}}}catch(e){return f.error(`Error updating task by ID: ${e.message}`),{success:!1,error:{code:`UPDATE_TASK_CORE_ERROR`,message:e.message||`Unknown error updating task`}}}finally{!m&&w()&&_()}}catch(e){return f.error(`Setup error in updateTaskByIdDirect: ${e.message}`),w()&&_(),{success:!1,error:{code:`DIRECT_FUNCTION_SETUP_ERROR`,message:e.message||`Unknown setup error`}}}}async function vn(e,t,n={}){let{session:r}=n,{from:i,prompt:a,research:o,tasksJsonPath:s,projectRoot:c,tag:l}=e,u=D(t);if(!c)return u.error(`updateTasksDirect requires a projectRoot argument.`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`projectRoot is required.`}};if(!i)return u.error(`updateTasksDirect called without from ID`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`Starting task ID (from) is required`}};if(!a)return u.error(`updateTasksDirect called without prompt`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`Update prompt is required`}};u.info(`Updating tasks via direct function. From: ${i}, Research: ${o}, File: ${s}, ProjectRoot: ${c}`),b();try{let e=await ye(s,i,a,o,{session:r,mcpLog:u,projectRoot:c,tag:l},`json`);return e&&e.success&&Array.isArray(e.updatedTasks)?(u.success(`Successfully updated ${e.updatedTasks.length} tasks.`),{success:!0,data:{message:`Successfully updated ${e.updatedTasks.length} tasks.`,tasksPath:s,updatedCount:e.updatedTasks.length,telemetryData:e.telemetryData,tagInfo:e.tagInfo}}):(u.error(`Core updateTasks function did not return a successful structure.`),{success:!1,error:{code:`CORE_FUNCTION_ERROR`,message:e?.message||`Core function failed to update tasks or returned unexpected result.`}})}catch(e){return u.error(`Error executing core updateTasks: ${e.message}`),{success:!1,error:{code:`UPDATE_TASKS_CORE_ERROR`,message:e.message||`Unknown error updating tasks`}}}finally{_()}}async function yn(e,t,n={}){let{tasksJsonPath:r,name:i,projectRoot:a}=e,{session:o}=n;b();let s=D(t);try{if(!r)return t.error(`useTagDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i||typeof i!=`string`)return t.error(`Missing required parameter: name`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`Tag name is required and must be a string`}};t.info(`Switching to tag: ${i}`);let e=await Ae(r,i,{},{session:o,mcpLog:s,projectRoot:a},`json`);return _(),{success:!0,data:{tagName:e.currentTag,switched:e.switched,previousTag:e.previousTag,taskCount:e.taskCount,message:`Successfully switched to tag "${e.currentTag}"`}}}catch(e){return _(),t.error(`Error in useTagDirect: ${e.message}`),{success:!1,error:{code:e.code||`USE_TAG_ERROR`,message:e.message}}}}async function bn(e,t){let{tasksJsonPath:n,projectRoot:r,tag:i}=e;if(!n)return t.error(`validateDependenciesDirect called without tasksJsonPath`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};try{t.info(`Validating dependencies in tasks: ${n}`);let e=n;return N.existsSync(e)?(b(),await je(e,{projectRoot:r,tag:i}),_(),{success:!0,data:{message:`Dependencies validated successfully`,tasksPath:e}}):{success:!1,error:{code:`FILE_NOT_FOUND`,message:`Tasks file not found at ${e}`}}}catch(e){return _(),t.error(`Error validating dependencies: ${e.message}`),{success:!1,error:{code:`VALIDATION_ERROR`,message:e.message}}}}function xn(e){e.addTool({name:`add_dependency`,description:`Add a dependency relationship between two tasks`,parameters:I.object({id:I.string().describe(`ID of task that will depend on another task`),dependsOn:I.string().describe(`ID of task that will become a dependency`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Add Dependency`,destructiveHint:!0},execute:Q(`add-dependency`,async(e,{log:t,session:n})=>{try{t.info(`Adding dependency for task ${e.id} to depend on ${e.dependsOn}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await zt({tasksJsonPath:i,id:e.id,dependsOn:e.dependsOn,projectRoot:e.projectRoot,tag:n},t);return a.success?t.info(`Successfully added dependency: ${a.data.message}`):t.error(`Failed to add dependency: ${a.error.message}`),Y({result:a,log:t,errorPrefix:`Error adding dependency`,projectRoot:e.projectRoot,tag:n})}catch(e){return t.error(`Error in addDependency tool: ${e.message}`),J(e.message)}})})}function Sn(e){e.addTool({name:`add_subtask`,description:`Add a subtask to an existing task`,parameters:I.object({id:I.string().describe(`Parent task ID (required)`),taskId:I.string().optional().describe(`Existing task ID to convert to subtask`),title:I.string().optional().describe(`Title for the new subtask (when creating a new subtask)`),description:I.string().optional().describe(`Description for the new subtask`),details:I.string().optional().describe(`Implementation details for the new subtask`),status:I.string().optional().describe(`Status for the new subtask (default: 'pending')`),dependencies:I.string().optional().describe(`Comma-separated list of dependency IDs for the new subtask`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),skipGenerate:I.boolean().optional().describe(`Skip regenerating task files`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Add Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Adding subtask with args: ${JSON.stringify(e)}`);let a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=await Bt({tasksJsonPath:a,id:e.id,taskId:e.taskId,title:e.title,description:e.description,details:e.details,status:e.status,dependencies:e.dependencies,skipGenerate:e.skipGenerate,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Subtask added successfully: ${o.data.message}`):t.error(`Failed to add subtask: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error adding subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in addSubtask tool: ${e.message}`),J(e.message)}})})}function Cn(e){e.addTool({name:`add_tag`,description:`Create a new tag for organizing tasks in different contexts`,parameters:I.object({name:I.string().describe(`Name of the new tag to create`),copyFromCurrent:I.boolean().optional().describe(`Whether to copy tasks from the current tag (default: false)`),copyFromTag:I.string().optional().describe(`Specific tag to copy tasks from`),fromBranch:I.boolean().optional().describe(`Create tag name from current git branch (ignores name parameter)`),description:I.string().optional().describe(`Optional description for the tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Add Tag`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting add-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Vt({tasksJsonPath:r,name:e.name,copyFromCurrent:e.copyFromCurrent,copyFromTag:e.copyFromTag,fromBranch:e.fromBranch,description:e.description,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error creating tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in add-tag tool: ${e.message}`),J(e.message)}})})}function wn(e){e.addTool({name:`add_task`,description:`Add a new task using AI`,parameters:I.object({prompt:I.string().optional().describe(`Description of the task to add (required if not using manual fields)`),title:I.string().optional().describe(`Task title (for manual task creation)`),description:I.string().optional().describe(`Task description (for manual task creation)`),details:I.string().optional().describe(`Implementation details (for manual task creation)`),testStrategy:I.string().optional().describe(`Test strategy (for manual task creation)`),dependencies:I.string().optional().describe(`Comma-separated list of task IDs this task depends on`),priority:I.string().optional().describe(`Task priority (high, medium, low)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for task creation`)}),annotations:{title:`Add Task`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting add-task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Ht({tasksJsonPath:a,prompt:e.prompt,title:e.title,description:e.description,details:e.details,testStrategy:e.testStrategy,dependencies:e.dependencies,priority:e.priority,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error adding task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in add-task tool: ${e.message}`),J(e.message)}})})}function Tn(e){e.addTool({name:`analyze_project_complexity`,description:`Analyze task complexity and generate expansion recommendations.`,parameters:I.object({threshold:I.coerce.number().int().min(1).max(10).optional().default(5).describe(`Complexity score threshold (1-10) to recommend expansion.`),research:I.boolean().optional().default(!1).describe(`Use Perplexity AI for research-backed analysis.`),output:I.string().optional().describe(`Output file path relative to project root (default: ${s}).`),file:I.string().optional().describe(`Path to the tasks file relative to project root (default: tasks/tasks.json).`),ids:I.string().optional().describe(`Comma-separated list of task IDs to analyze specifically (e.g., "1,3,5").`),from:I.coerce.number().int().positive().optional().describe(`Starting task ID in a range to analyze.`),to:I.coerce.number().int().positive().optional().describe(`Ending task ID in a range to analyze.`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Analyze Project Complexity`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`analyze_project_complexity`;try{t.info(`Executing ${i} tool with args: ${JSON.stringify(e)}`);let a=r({projectRoot:e.projectRoot,tag:e.tag}),o;try{o=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`${i}: Resolved tasks path: ${o}`)}catch(n){return t.error(`${i}: Error finding tasks.json: ${n.message}`),J(`Failed to find tasks.json within project root '${e.projectRoot}': ${n.message}`)}let s=S(e.output,{projectRoot:e.projectRoot,tag:a},t);t.info(`${i}: Report output path: ${s}`);let c=P.dirname(s);try{N.existsSync(c)||(N.mkdirSync(c,{recursive:!0}),t.info(`${i}: Created output directory: ${c}`))}catch(e){return t.error(`${i}: Failed to create output directory ${c}: ${e.message}`),J(`Failed to create output directory: ${e.message}`)}let l=await Ut({tasksJsonPath:o,outputPath:s,threshold:e.threshold,research:e.research,projectRoot:e.projectRoot,tag:a,ids:e.ids,from:e.from,to:e.to},t,{session:n});return t.info(`${i}: Direct function result: success=${l.success}`),Y({result:l,log:t,errorPrefix:`Error analyzing task complexity`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function En(e){e.addTool({name:`clear_subtasks`,description:`Clear subtasks from specified tasks`,parameters:I.object({id:I.string().optional().describe(`Task IDs (comma-separated) to clear subtasks from`),all:I.boolean().optional().describe(`Clear subtasks from all tasks`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}).refine(e=>e.id||e.all,{message:`Either 'id' or 'all' parameter must be provided`,path:[`id`,`all`]}),annotations:{title:`Clear Subtasks`,destructiveHint:!0},execute:Q(`clear-subtasks`,async(e,t)=>{try{t.log.info(`Clearing subtasks with args: ${JSON.stringify(e)}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await Wt({tasksJsonPath:i,id:e.id,all:e.all,projectRoot:e.projectRoot,tag:n},t.log,{session:t.session});return a.success?t.log.info(`Subtasks cleared successfully: ${a.data.message}`):t.log.error(`Failed to clear subtasks: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error clearing subtasks`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in clearSubtasks tool: ${e.message}`),J(e.message)}})})}function Dn(e){e.addTool({name:`complexity_report`,description:`Display the complexity analysis report in a readable format`,parameters:I.object({file:I.string().optional().describe(`Path to the report file (default: ${s})`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Complexity Report`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Getting complexity report with args: ${JSON.stringify(e)}`);let n=p(e.projectRoot),r=Ue({projectRoot:e.projectRoot,complexityReport:e.file,tag:n},t);if(t.info(`Reading complexity report from path: `,r),!r)return J(`No complexity report found. Run task-master analyze-complexity first.`);let i=await Gt({reportPath:r},t);return i.success?t.info(`Successfully retrieved complexity report`):t.error(`Failed to retrieve complexity report: ${i.error.message}`),Y({result:i,log:t,errorPrefix:`Error retrieving complexity report`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in complexity-report tool: ${e.message}`),J(`Failed to retrieve complexity report: ${e.message}`)}})})}function On(e){e.addTool({name:`copy_tag`,description:`Copy an existing tag to create a new tag with all tasks and metadata`,parameters:I.object({sourceName:I.string().describe(`Name of the source tag to copy from`),targetName:I.string().describe(`Name of the new tag to create`),description:I.string().optional().describe(`Optional description for the new tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Copy Tag`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting copy-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Kt({tasksJsonPath:r,sourceName:e.sourceName,targetName:e.targetName,description:e.description,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error copying tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in copy-tag tool: ${e.message}`),J(e.message)}})})}function kn(e){e.addTool({name:`delete_tag`,description:`Delete an existing tag and all its tasks`,parameters:I.object({name:I.string().describe(`Name of the tag to delete`),yes:I.boolean().optional().describe(`Skip confirmation prompts (default: true for MCP)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Delete Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting delete-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await qt({tasksJsonPath:r,name:e.name,yes:e.yes===void 0?!0:e.yes,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error deleting tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in delete-tag tool: ${e.message}`),J(e.message)}})})}function An(e){e.addTool({name:`expand_all`,description:`Expand all pending tasks into subtasks based on complexity or defaults`,parameters:I.object({num:I.string().optional().describe(`Target number of subtasks per task (uses complexity/defaults otherwise)`),research:I.boolean().optional().describe(`Enable research-backed subtask generation (e.g., using Perplexity)`),prompt:I.string().optional().describe(`Additional context to guide subtask generation for all tasks`),force:I.boolean().optional().describe(`Force regeneration of subtasks for tasks that already have them`),file:I.string().optional().describe(`Absolute path to the tasks file in the /tasks folder inside the project root (default: tasks/tasks.json)`),projectRoot:I.string().optional().describe(`Absolute path to the project root directory (derived from session if possible)`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Expand All Tasks`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Tool expand_all execution started with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`Resolved tasks.json path: ${a}`)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=xe(null,{projectRoot:e.projectRoot,tag:i},t);return t.info(`Using complexity report path: ${o}`),Y({result:await Jt({tasksJsonPath:a,num:e.num,research:e.research,prompt:e.prompt,force:e.force,projectRoot:e.projectRoot,tag:i,complexityReportPath:o},t,{session:n}),log:t,errorPrefix:`Error expanding all tasks`,projectRoot:e.projectRoot})}catch(e){return t.error(`Unexpected error in expand_all tool execute: ${e.message}`),e.stack&&t.error(e.stack),J(`An unexpected error occurred: ${e.message}`)}})})}function jn(e){e.addTool({name:`expand_task`,description:`Expand a task into subtasks for detailed implementation`,parameters:I.object({id:I.string().describe(`ID of task to expand`),num:I.string().optional().describe(`Number of subtasks to generate`),research:I.boolean().optional().default(!1).describe(`Use research role for generation`),prompt:I.string().optional().describe(`Additional context for subtask generation`),file:I.string().optional().describe(`Path to the tasks file relative to project root (e.g., tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),force:I.boolean().optional().default(!1).describe(`Force expansion even if subtasks exist`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Expand Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting expand-task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=Ue({...e,tag:i},t);return Y({result:await Yt({tasksJsonPath:a,id:e.id,num:e.num,research:e.research,prompt:e.prompt,force:e.force,complexityReportPath:o,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error expanding task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in expand-task tool: ${e.message}`),J(e.message)}})})}function Mn(e){e.addTool({name:`fix_dependencies`,description:`Fix invalid dependencies in tasks automatically`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Fix Dependencies`,destructiveHint:!0},execute:Q(`fix-dependencies`,async(e,t)=>{try{t.log.info(`Fixing dependencies with args: ${JSON.stringify(e)}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await Xt({tasksJsonPath:i,projectRoot:e.projectRoot,tag:n},t.log);return a.success?t.log.info(`Successfully fixed dependencies: ${a.data.message}`):t.log.error(`Failed to fix dependencies: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error fixing dependencies`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in fixDependencies tool: ${e.message}`),J(e.message)}})})}function Nn(e){e.addTool({name:`initialize_project`,description:`Initializes a new Task Master project structure by calling the core initialization logic. Creates necessary folders and configuration files for Task Master in the current directory.`,parameters:I.object({skipInstall:I.boolean().optional().default(!1).describe(`Skip installing dependencies automatically. Never do this unless you are sure the project is already installed.`),addAliases:I.boolean().optional().default(!0).describe(`Add shell aliases (tm, taskmaster, hamster, ham) to shell config file.`),initGit:I.boolean().optional().default(!0).describe(`Initialize Git repository in project root.`),storeTasksInGit:I.boolean().optional().default(!0).describe(`Store tasks in Git (tasks.json and tasks/ directory).`),yes:I.boolean().optional().default(!0).describe(`Skip prompts and use default values. Always set to true for MCP tools.`),projectRoot:I.string().describe(`The root directory for the project. ALWAYS SET THIS TO THE PROJECT ROOT DIRECTORY. IF NOT SET, THE TOOL WILL NOT WORK.`),rules:I.array(I.enum(A)).optional().describe(`List of rule profiles to include at initialization. If omitted, defaults to Cursor profile only. Available options: ${A.join(`, `)}`)}),annotations:{title:`Initialize Project`,destructiveHint:!0},execute:Z(async(e,t)=>{let{log:n}=t,r=t.session;try{return n.info(`Executing initialize_project tool with args: ${JSON.stringify(e)}`),Y({result:await Zt(e,n,{session:r}),log:n,errorPrefix:`Initialization failed`,projectRoot:e.projectRoot})}catch(e){let t=`Project initialization tool failed: ${e.message||`Unknown error`}`;return n.error(t,e),J(t,{details:e.stack})}})})}function Pn(e){e.addTool({name:`list_tags`,description:`List all available tags with task counts and metadata`,parameters:I.object({showMetadata:I.boolean().optional().describe(`Whether to include metadata in the output (default: false)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`List Tags`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting list-tags with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Qt({tasksJsonPath:r,showMetadata:e.showMetadata,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error listing tags`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in list-tags tool: ${e.message}`),J(e.message)}})})}function Fn(e){e.addTool({name:`models`,description:`Get information about available AI models or set model configurations. Run without arguments to get the current model configuration and API key status for the selected model providers.`,parameters:I.object({setMain:I.string().optional().describe(`Set the primary model for task generation/updates. Model provider API key is required in the MCP config ENV.`),setResearch:I.string().optional().describe(`Set the model for research-backed operations. Model provider API key is required in the MCP config ENV.`),setFallback:I.string().optional().describe(`Set the model to use if the primary fails. Model provider API key is required in the MCP config ENV.`),listAvailableModels:I.boolean().optional().describe(`List all available models not currently in use. Input/output costs values are in dollars (3 is $3.00).`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),openrouter:I.boolean().optional().describe(`Indicates the set model ID is a custom OpenRouter model.`),ollama:I.boolean().optional().describe(`Indicates the set model ID is a custom Ollama model.`),bedrock:I.boolean().optional().describe(`Indicates the set model ID is a custom AWS Bedrock model.`),azure:I.boolean().optional().describe(`Indicates the set model ID is a custom Azure OpenAI model.`),vertex:I.boolean().optional().describe(`Indicates the set model ID is a custom Google Vertex AI model.`),"openai-compatible":I.boolean().optional().describe(`Indicates the set model ID is a custom OpenAI-compatible model. Requires baseURL parameter.`),baseURL:I.string().optional().describe(`Custom base URL for providers that support it (e.g., https://api.example.com/v1).`)}),annotations:{title:`Models`,destructiveHint:!0},execute:Q(`models`,async(e,t)=>{try{return t.log.info(`Starting models tool with args: ${JSON.stringify(e)}`),Y({result:await nn({...e,projectRoot:e.projectRoot},t.log,{session:t.session}),log:t.log,errorPrefix:`Error managing models`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in models tool: ${e.message}`),J(e.message)}})})}function In(e){e.addTool({name:`move_task`,description:`Move a task or subtask to a new position`,parameters:I.object({from:I.string().describe(`ID of the task/subtask to move (e.g., "5" or "5.2"). Can be comma-separated to move multiple tasks (e.g., "5,6,7")`),to:I.string().optional().describe(`ID of the destination (e.g., "7" or "7.3"). Required for within-tag moves. For cross-tag moves, if omitted, task will be moved to the target tag maintaining its ID`),file:I.string().optional().describe(`Custom path to tasks.json file`),projectRoot:I.string().describe(`Root directory of the project (typically derived from session)`),tag:I.string().optional().describe(`Tag context to operate on`),fromTag:I.string().optional().describe(`Source tag for cross-tag moves`),toTag:I.string().optional().describe(`Target tag for cross-tag moves`),withDependencies:I.boolean().optional().describe(`Move dependent tasks along with main task`),ignoreDependencies:I.boolean().optional().describe(`Break cross-tag dependencies during move`)}),annotations:{title:`Move Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{if(e.fromTag&&e.toTag&&e.fromTag!==e.toTag){if(!e.from)return J(`Source IDs are required for cross-tag moves`,`MISSING_SOURCE_IDS`);e.to&&t.warn(`The "to" parameter is not used for cross-tag moves and will be ignored. Tasks retain their original IDs in the target tag.`);let r=e.file;return r||=E(e,t),Y({result:await rn({sourceIds:e.from,sourceTag:e.fromTag,targetTag:e.toTag,withDependencies:e.withDependencies||!1,ignoreDependencies:e.ignoreDependencies||!1,tasksJsonPath:r,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error moving tasks between tags`,projectRoot:e.projectRoot})}else{if(!e.to)return J(`Destination ID is required for within-tag moves`,`MISSING_DESTINATION_ID`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a=e.file;a||=E(e,t);let o=e.from.split(`,`).map(e=>e.trim()),s=e.to.split(`,`).map(e=>e.trim());if(o.length!==s.length){if(o.length>1){let r=[],c=[];for(let l=0;l<o.length;l++){let u=o[l],d=s[l];if(u===d){t.info(`Skipping ${u} -> ${d} (same ID)`),c.push({fromId:u,toId:d,reason:`same ID`});continue}let f=l===o.length-1,p=await an({sourceId:u,destinationId:d,tasksJsonPath:a,projectRoot:e.projectRoot,tag:i,generateFiles:f},t,{session:n});p.success?r.push(p.data):t.error(`Failed to move ${u} to ${d}: ${p.error.message}`)}return Y({result:{success:!0,data:{moves:r,skipped:c.length>0?c:void 0,message:`Successfully moved ${r.length} tasks${c.length>0?`, skipped ${c.length}`:``}`}},log:t,errorPrefix:`Error moving multiple tasks`,projectRoot:e.projectRoot})}return Y({result:{success:!0,data:{moves:results,skippedMoves,message:`Successfully moved ${results.length} tasks${skippedMoves.length>0?`, skipped ${skippedMoves.length} moves`:``}`}},log:t,errorPrefix:`Error moving multiple tasks`,projectRoot:e.projectRoot})}else return Y({result:await an({sourceId:e.from,destinationId:e.to,tasksJsonPath:a,projectRoot:e.projectRoot,tag:i,generateFiles:!0},t,{session:n}),log:t,errorPrefix:`Error moving task`,projectRoot:e.projectRoot})}}catch(e){return J(`Failed to move task: ${e.message}`,`MOVE_TASK_ERROR`)}})})}function Ln(e){e.addTool({name:`next_task`,description:`Find the next task to work on based on dependencies and status`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),complexityReport:I.string().optional().describe(`Path to the complexity report file (relative to project root or absolute)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Next Task`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Finding next task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=ue(e,n)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o;try{o=de({...e,tag:i},n)}catch(e){t.error(`Error finding complexity report: ${e.message}`),o=null}let s=await on({tasksJsonPath:a,reportPath:o,projectRoot:e.projectRoot,tag:i},t,{session:n});return t.info(`Next task result: ${s.success?`found`:`none`}`),Y({result:s,log:t,errorPrefix:`Error finding next task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error finding next task: ${e.message}`),J(e.message)}})})}function Rn(e){e.addTool({name:`parse_prd`,description:`Parse a Product Requirements Document (PRD) text file to automatically generate initial tasks. Reinitializing the project is not necessary to run this tool. It is recommended to run parse-prd after initializing the project and creating/importing a prd.txt file in the project root's ${i} directory.`,parameters:I.object({input:I.string().optional().default(ne).describe(`Absolute path to the PRD document file (.txt, .md, etc.)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),destination:I.enum([`local`,`hamster`]).optional().default(`local`).describe(`Where to parse the PRD: "local" writes tasks.json, "hamster" creates a cloud brief and tasks`),output:I.string().optional().describe(`Output path for tasks.json file (default: ${a})`),numTasks:I.string().optional().describe(`Approximate number of top-level tasks to generate (default: 10). As the agent, if you have enough information, ensure to enter a number of tasks that would logically scale with project complexity. Setting to 0 will allow Taskmaster to determine the appropriate number of tasks based on the complexity of the PRD. Avoid entering numbers above 50 due to context window limitations.`),force:I.boolean().optional().default(!1).describe(`Overwrite existing output file without prompting.`),research:I.boolean().optional().describe(`Enable Taskmaster to use the research role for potentially more informed task generation. Requires appropriate API key.`),append:I.boolean().optional().describe(`Append generated tasks to existing file.`)}),annotations:{title:`Parse PRD`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n,reportProgress:i})=>{try{let a=r({projectRoot:e.projectRoot,tag:e.tag}),o=dt(i,t);return Y({result:await sn({...e,tag:a},t,{session:n,reportProgress:o}),log:t,errorPrefix:`Error parsing PRD`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in parse_prd: ${e.message}`),J(`Failed to parse PRD: ${e.message}`)}})})}function zn(e){e.addTool({name:`remove_dependency`,description:`Remove a dependency from a task`,parameters:I.object({id:I.string().describe(`Task ID to remove dependency from`),dependsOn:I.string().describe(`Task ID to remove as a dependency`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Remove Dependency`,destructiveHint:!0},execute:Q(`remove-dependency`,async(e,t)=>{try{let n=r({projectRoot:e.projectRoot,tag:e.tag});t.log.info(`Removing dependency for task ${e.id} from ${e.dependsOn} with args: ${JSON.stringify(e)}`);let i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await ln({tasksJsonPath:i,id:e.id,dependsOn:e.dependsOn,projectRoot:e.projectRoot,tag:n},t.log);return a.success?t.log.info(`Successfully removed dependency: ${a.data.message}`):t.log.error(`Failed to remove dependency: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error removing dependency`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in removeDependency tool: ${e.message}`),J(e.message)}})})}function Bn(e){e.addTool({name:`remove_subtask`,description:`Remove a subtask from its parent task`,parameters:I.object({id:I.string().describe(`Subtask ID to remove in format 'parentId.subtaskId' (required)`),convert:I.boolean().optional().describe(`Convert the subtask to a standalone task instead of deleting it`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),skipGenerate:I.boolean().optional().describe(`Skip regenerating task files`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Remove Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Removing subtask with args: ${JSON.stringify(e)}`);let a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=await un({tasksJsonPath:a,id:e.id,convert:e.convert,skipGenerate:e.skipGenerate,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Subtask removed successfully: ${o.data.message}`):t.error(`Failed to remove subtask: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error removing subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in removeSubtask tool: ${e.message}`),J(e.message)}})})}function Vn(e){e.addTool({name:`remove_task`,description:`Remove a task or subtask permanently from the tasks list`,parameters:I.object({id:I.string().describe(`ID of the task or subtask to remove (e.g., '5' or '5.2'). Can be comma-separated to update multiple tasks/subtasks at once.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),confirm:I.boolean().optional().describe(`Whether to skip confirmation prompt (default: false)`),tag:I.string().optional().describe(`Specify which tag context to operate on. Defaults to the current active tag.`)}),annotations:{title:`Remove Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Removing task(s) with ID(s): ${e.id}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}t.info(`Using tasks file path: ${a}`);let o=await dn({tasksJsonPath:a,id:e.id,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Successfully removed task: ${e.id}`):t.error(`Failed to remove task: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error removing task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in remove-task tool: ${e.message}`),J(`Failed to remove task: ${e.message}`)}})})}function Hn(e){e.addTool({name:`rename_tag`,description:`Rename an existing tag`,parameters:I.object({oldName:I.string().describe(`Current name of the tag to rename`),newName:I.string().describe(`New name for the tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Rename Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting rename-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await fn({tasksJsonPath:r,oldName:e.oldName,newName:e.newName,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error renaming tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in rename-tag tool: ${e.message}`),J(e.message)}})})}function Un(e){e.addTool({name:`research`,description:`Perform AI-powered research queries with project context`,parameters:I.object({query:I.string().describe(`Research query/prompt (required)`),taskIds:I.string().optional().describe(`Comma-separated list of task/subtask IDs for context (e.g., "15,16.2,17")`),filePaths:I.string().optional().describe(`Comma-separated list of file paths for context (e.g., "src/api.js,docs/readme.md")`),customContext:I.string().optional().describe(`Additional custom context text to include in the research`),includeProjectTree:I.boolean().optional().describe(`Include project file tree structure in context (default: false)`),detailLevel:I.enum([`low`,`medium`,`high`]).optional().describe(`Detail level for the research response (default: medium)`),saveTo:I.string().optional().describe(`Automatically save research results to specified task/subtask ID (e.g., "15" or "15.2")`),saveToFile:I.boolean().optional().describe(`Save research results to .taskmaster/docs/research/ directory (default: false)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Research`,destructiveHint:!0,openWorldHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});return t.info(`Starting research with query: "${e.query.substring(0,100)}${e.query.length>100?`...`:``}"`),Y({result:await pn({query:e.query,taskIds:e.taskIds,filePaths:e.filePaths,customContext:e.customContext,includeProjectTree:e.includeProjectTree||!1,detailLevel:e.detailLevel||`medium`,saveTo:e.saveTo,saveToFile:e.saveToFile||!1,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error performing research`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in research tool: ${e.message}`),J(e.message)}})})}async function Wn(e,t,n={}){let{projectRoot:r,language:i}=e,a=D(t);t.info(`Executing response-language_direct with args: ${JSON.stringify(e)}`),t.info(`Using project root: ${r}`);try{return b(),We(i,{mcpLog:a,projectRoot:r})}catch(e){return{success:!1,error:{code:`DIRECT_FUNCTION_ERROR`,message:e.message,details:e.stack}}}finally{_()}}function Gn(e){e.addTool({name:`response-language`,description:`Get or set the response language for the project`,parameters:I.object({projectRoot:I.string().describe(`The root directory for the project. ALWAYS SET THIS TO THE PROJECT ROOT DIRECTORY. IF NOT SET, THE TOOL WILL NOT WORK.`),language:I.string().describe(`The new response language to set. like "中文" "English" or "español".`)}),annotations:{title:`Response Language`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{return t.info(`Executing response-language tool with args: ${JSON.stringify(e)}`),Y({result:await Wn({...e,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error setting response language`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in response-language tool: ${e.message}`),J(e.message)}})})}async function Kn(e,t,n={}){b();try{let{action:t,profiles:n,projectRoot:r,yes:i,force:a}=e;if(!t||!Array.isArray(n)||n.length===0||!r)return{success:!1,error:{code:`MISSING_ARGUMENT`,message:`action, profiles, and projectRoot are required.`}};let o=[],s=[];if(t===O.REMOVE){if(!a&&Ke(r,n)){let e=Ge(r);return e.filter(e=>!n.includes(e)),{success:!1,error:{code:`CRITICAL_REMOVAL_BLOCKED`,message:`CRITICAL: This operation would remove ALL remaining rule profiles (${n.join(`, `)}), leaving your project with no rules configurations. This could significantly impact functionality. Currently installed profiles: ${e.join(`, `)}. If you're certain you want to proceed, set force: true or use the CLI with --force flag.`}}}for(let e of n){if(!Ye(e)){o.push({profileName:e,success:!1,error:`The requested rule profile for '${e}' is unavailable. Supported profiles are: ${A.join(`, `)}.`});continue}let t=Je(r,k(e));o.push(t)}let e=o.filter(e=>e.success).map(e=>e.profileName),t=o.filter(e=>e.skipped).map(e=>e.profileName),i=o.filter(e=>e.error&&!e.success&&!e.skipped),s=o.filter(e=>e.notice),c=``;return e.length>0&&(c+=`Successfully removed Task Master rules: ${e.join(`, `)}.`),t.length>0&&(c+=`Skipped (default or protected): ${t.join(`, `)}.`),i.length>0&&(c+=i.map(e=>`Error removing ${e.profileName}: ${e.error}`).join(` `)),s.length>0&&(c+=` Notices: ${s.map(e=>`${e.profileName} - ${e.notice}`).join(`; `)}.`),_(),{success:i.length===0,data:{summary:c,results:o}}}else if(t===O.ADD){for(let e of n){if(!Ye(e)){s.push({profileName:e,success:!1,error:`Profile not found: static import missing for '${e}'. Valid profiles: ${A.join(`, `)}`});continue}let t=k(e),{success:n,failed:i}=Xe(r,t),a=t.rulesDir,o=P.join(r,a),c=t.profileDir,l=t.mcpConfig!==!1,u=l&&t.mcpConfigPath?P.join(r,t.mcpConfigPath):null,d=l&&u?N.existsSync(u):void 0,f=N.existsSync(o),p=N.existsSync(P.join(r,c)),m=i>0?`${i} rule files failed to convert.`:null,h={profileName:e,mcpConfigCreated:d,rulesDirCreated:f,profileFolderCreated:p,skipped:!1,error:m,success:(l?d:!0)&&f&&n>0&&!m};s.push(h)}let e=s.filter(e=>e.success).map(e=>e.profileName),t=s.filter(e=>e.error&&!e.success),i=``;return e.length>0&&(i+=`Successfully added rules: ${e.join(`, `)}.`),t.length>0&&(i+=t.map(e=>` Error adding ${e.profileName}: ${e.error}`).join(` `)),_(),{success:t.length===0,data:{summary:i,results:s}}}else return _(),{success:!1,error:{code:`INVALID_ACTION`,message:`Unknown action. Use "${O.ADD}" or "${O.REMOVE}".`}}}catch(e){return _(),t.error(`[rulesDirect] Error: ${e.message}`),{success:!1,error:{code:e.code||`RULES_ERROR`,message:e.message}}}}function qn(e){e.addTool({name:`rules`,description:`Add or remove rule profiles from the project.`,parameters:I.object({action:I.enum([`add`,`remove`]).describe(`Whether to add or remove rule profiles.`),profiles:I.array(I.enum(A)).min(1).describe(`List of rule profiles to add or remove (e.g., [\"cursor\", \"roo\"]). Available options: ${A.join(`, `)}`),projectRoot:I.string().describe(`The root directory of the project. Must be an absolute path.`),force:I.boolean().optional().default(!1).describe(`DANGEROUS: Force removal even if it would leave no rule profiles. Only use if you are absolutely certain.`)}),annotations:{title:`Rules`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{return t.info(`[rules tool] Executing action: ${e.action} for profiles: ${e.profiles.join(`, `)} in ${e.projectRoot}`),Y({result:await Kn(e,t,{session:n}),log:t,projectRoot:e.projectRoot})}catch(e){return t.error(`[rules tool] Error: ${e.message}`),J(e.message,{details:e.stack})}})})}function Jn(e){e.addTool({name:`scope_down_task`,description:`Decrease the complexity of one or more tasks using AI`,parameters:I.object({id:I.string().describe(`Comma-separated list of task IDs to scope down (e.g., "1,3,5")`),strength:I.string().optional().describe(`Strength level: light, regular, or heavy (default: regular)`),prompt:I.string().optional().describe(`Custom prompt for specific scoping adjustments`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for scoping`)}),annotations:{title:`Scope Down Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting scope-down with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await mn({tasksJsonPath:a,id:e.id,strength:e.strength,prompt:e.prompt,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error scoping down task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in scope-down tool: ${e.message}`),J(e.message)}})})}function Yn(e){e.addTool({name:`scope_up_task`,description:`Increase the complexity of one or more tasks using AI`,parameters:I.object({id:I.string().describe(`Comma-separated list of task IDs to scope up (e.g., "1,3,5")`),strength:I.string().optional().describe(`Strength level: light, regular, or heavy (default: regular)`),prompt:I.string().optional().describe(`Custom prompt for specific scoping adjustments`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for scoping`)}),annotations:{title:`Scope Up Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting scope-up with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await hn({tasksJsonPath:a,id:e.id,strength:e.strength,prompt:e.prompt,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error scoping up task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in scope-up tool: ${e.message}`),J(e.message)}})})}function Xn(e){e.addTool({name:`update_subtask`,description:`Appends timestamped information to a specific subtask without replacing existing content. If you just want to update the subtask status, use set_task_status instead.`,parameters:I.object({id:y.describe(`ID of the subtask to update in format "parentId.subtaskId" (e.g., "5.2"). Parent ID is the ID of the task that contains the subtask.`),prompt:I.string().optional().describe(`Information to add to the subtask. Required unless only updating metadata.`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),metadata:I.string().optional().describe(`JSON string of metadata to merge into subtask metadata. Example: '{"ticketId": "JIRA-456", "reviewed": true}'. Requires TASK_MASTER_ALLOW_METADATA_UPDATES=true in MCP environment.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update_subtask`;try{let a=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Updating subtask with args: ${JSON.stringify(e)}`);let o;try{o=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let s=mt(e.metadata,J);if(s.error)return s.error;let c=s.parsedMetadata;if(!e.prompt&&!c)return J(`Either prompt or metadata must be provided for update-subtask`);let l=await gn({tasksJsonPath:o,id:e.id,prompt:e.prompt,research:e.research,metadata:c,projectRoot:e.projectRoot,tag:a},t,{session:n});return l.success?t.info(`Successfully updated subtask with ID ${e.id}`):t.error(`Failed to update subtask: ${l.error?.message||`Unknown error`}`),Y({result:l,log:t,errorPrefix:`Error updating subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function Zn(e){e.addTool({name:`update_task`,description:`Updates a single task by ID with new information or context provided in the prompt.`,parameters:I.object({id:I.string().describe(`ID of the task (e.g., '15') to update. Subtasks are supported using the update-subtask tool.`),prompt:I.string().optional().describe(`New information or context to incorporate into the task. Required unless only updating metadata.`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),append:I.boolean().optional().describe(`Append timestamped information to task details instead of full update`),metadata:I.string().optional().describe(`JSON string of metadata to merge into task metadata. Example: '{"githubIssue": 42, "sprint": "Q1-S3"}'. Requires TASK_MASTER_ALLOW_METADATA_UPDATES=true in MCP environment.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update_task`;try{let a=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Executing ${i} tool with args: ${JSON.stringify(e)}`);let o;try{o=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`${i}: Resolved tasks path: ${o}`)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let s=mt(e.metadata,J);if(s.error)return s.error;let c=s.parsedMetadata;if(!e.prompt&&!c)return J(`Either prompt or metadata must be provided for update-task`);let l=await _n({tasksJsonPath:o,id:e.id,prompt:e.prompt,research:e.research,append:e.append,metadata:c,projectRoot:e.projectRoot,tag:a},t,{session:n});return t.info(`${i}: Direct function result: success=${l.success}`),Y({result:l,log:t,errorPrefix:`Error updating task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function Qn(e){e.addTool({name:`update`,description:`Update multiple upcoming tasks (with ID >= 'from' ID) based on new context or changes provided in the prompt. Use 'update_task' instead for a single specific task or 'update_subtask' for subtasks.`,parameters:I.object({from:I.string().describe(`Task ID from which to start updating (inclusive). IMPORTANT: This tool uses 'from', not 'id'`),prompt:I.string().describe(`Explanation of changes or new context to apply`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),file:I.string().optional().describe(`Path to the tasks file relative to project root`),projectRoot:I.string().optional().describe(`The directory of the project. (Optional, usually from session)`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Tasks`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update`,{from:a,prompt:o,research:s,file:c,projectRoot:l,tag:u}=e,d=r({projectRoot:e.projectRoot,tag:e.tag});try{t.info(`Executing ${i} tool with normalized root: ${l}`);let r;try{r=E({projectRoot:l,file:c},t),t.info(`${i}: Resolved tasks path: ${r}`)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json within project root '${l}': ${e.message}`)}let u=await vn({tasksJsonPath:r,from:a,prompt:o,research:s,projectRoot:l,tag:d},t,{session:n});return t.info(`${i}: Direct function result: success=${u.success}`),Y({result:u,log:t,errorPrefix:`Error updating tasks`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function $n(e){e.addTool({name:`use_tag`,description:`Switch to a different tag context for task operations`,parameters:I.object({name:I.string().describe(`Name of the tag to switch to`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Use Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting use-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await yn({tasksJsonPath:r,name:e.name,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error switching tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in use-tag tool: ${e.message}`),J(e.message)}})})}function er(e){e.addTool({name:`validate_dependencies`,description:`Check tasks for dependency issues (like circular references or links to non-existent tasks) without making changes.`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Validate Dependencies`,readOnlyHint:!0},execute:Q(`validate-dependencies`,async(e,{log:t,session:n})=>{try{let n=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Validating dependencies with args: ${JSON.stringify(e)}`);let i;try{i=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await bn({tasksJsonPath:i,projectRoot:e.projectRoot,tag:n},t);return a.success?t.info(`Successfully validated dependencies: ${a.data.message}`):t.error(`Failed to validate dependencies: ${a.error.message}`),Y({result:a,log:t,errorPrefix:`Error validating dependencies`,projectRoot:e.projectRoot,tag:n})}catch(e){return t.error(`Error in validateDependencies tool: ${e.message}`),J(e.message)}})})}const $={initialize_project:Nn,models:Fn,rules:qn,parse_prd:Rn,"response-language":Gn,analyze_project_complexity:Tn,expand_task:jn,expand_all:An,scope_up_task:Yn,scope_down_task:Jn,get_tasks:Mt,get_task:Pt,next_task:Ln,complexity_report:Dn,set_task_status:Rt,add_task:wn,add_subtask:Sn,update:Qn,update_task:Zn,update_subtask:Xn,remove_task:Vn,remove_subtask:Bn,clear_subtasks:En,move_task:In,add_dependency:xn,remove_dependency:zn,validate_dependencies:er,fix_dependencies:Mn,list_tags:Pn,add_tag:Cn,delete_tag:kn,use_tag:$n,rename_tag:Hn,copy_tag:On,research:Un,autopilot_start:gt,autopilot_resume:vt,autopilot_next:bt,autopilot_status:St,autopilot_complete:wt,autopilot_commit:Et,autopilot_finalize:Ot,autopilot_abort:At,generate:It},tr=[`get_tasks`,`next_task`,`get_task`,`set_task_status`,`update_subtask`,`parse_prd`,`expand_task`],nr=[...tr,`initialize_project`,`analyze_project_complexity`,`expand_all`,`add_subtask`,`remove_task`,`add_task`,`complexity_report`];function rr(e){return $[e]||null}function ir(){let e=process.env.TASK_MASTER_TOOLS;if(!e||e.trim()===``)return V.debug(`No TASK_MASTER_TOOLS env var found, defaulting to "core"`),`core`;let t=e.trim();return V.debug(`TASK_MASTER_TOOLS env var: "${t}"`),t}function ar(e,t=`core`){let n=[],r=[];try{let i=t.trim(),a=[],o=i.toLowerCase();switch(o){case`all`:a=Object.keys($),V.info(`Loading all available tools`);break;case`core`:case`lean`:a=tr,V.info(`Loading core tools only`);break;case`standard`:a=nr,V.info(`Loading standard tools`);break;default:let e=i.split(`,`).map(e=>e.trim()).filter(e=>e.length>0),t=new Set,n=[],r={response_language:`response-language`};for(let i of e){let e=null,a=i.toLowerCase();if(r[a]){let t=r[a];for(let n of Object.keys($))if(n.toLowerCase()===t.toLowerCase()){e=n;break}}if(!e){for(let t of Object.keys($))if(t.toLowerCase()===a){e=t;break}}if(!e){let t=a.replace(/_/g,`-`);for(let n of Object.keys($))if(n.toLowerCase()===t){e=n;break}}if(!e){let t=a.replace(/-/g,`_`);for(let n of Object.keys($))if(n.toLowerCase()===t){e=n;break}}e?(t.add(e),V.debug(`Resolved tool "${i}" to "${e}"`)):(n.push(i),V.warn(`Unknown tool specified: "${i}"`))}a=Array.from(t),n.length>0&&V.warn(`Unknown tools: ${n.join(`, `)}`),a.length===0?(V.warn(`No valid tools found in custom list. Loading all tools as fallback.`),a=Object.keys($)):V.info(`Loading ${a.length} custom tools from list (${t.size} unique after normalization)`);break}return V.info(`Registering ${a.length} MCP tools (mode: ${i})`),a.forEach(t=>{try{let i=rr(t);i?(i(e),V.debug(`Registered tool: ${t}`),n.push(t)):(V.warn(`Tool ${t} not found in registry`),r.push(t))}catch(e){e.message&&e.message.includes(`already registered`)?(V.debug(`Tool ${t} already registered, skipping`),n.push(t)):(V.error(`Failed to register tool ${t}: ${e.message}`),r.push(t))}}),V.info(`Successfully registered ${n.length}/${a.length} tools`),r.length>0&&V.warn(`Failed tools: ${r.join(`, `)}`),{registeredTools:n,failedTools:r,normalizedMode:o}}catch(t){V.error(`Error parsing TASK_MASTER_TOOLS environment variable: ${t.message}`),V.info(`Falling back to loading all tools`);let i=Object.keys($);for(let t of i){let i=rr(t);if(i)try{i(e),n.push(t)}catch(e){e.message&&e.message.includes(`already registered`)?(V.debug(`Fallback tool ${t} already registered, skipping`),n.push(t)):(V.warn(`Failed to register fallback tool '${t}': ${e.message}`),r.push(t))}else V.warn(`Tool '${t}' not found in registry`),r.push(t)}return V.info(`Successfully registered ${n.length} fallback tools`),{registeredTools:n,failedTools:r,normalizedMode:`all`}}}L.config(),re();const or=Qe(import.meta.url);P.dirname(or);var sr=class{constructor(){this.options={name:`Task Master MCP Server`,version:se};let e=new $e(this.options);if(e._mcpServer&&R.wrapMcpServerWithSentry)try{e._mcpServer=R.wrapMcpServerWithSentry(e._mcpServer)}catch(e){V.warn(`Failed to wrap MCP server with Sentry: ${e.message}`)}this.server=e,this.initialized=!1,this.init=this.init.bind(this),this.start=this.start.bind(this),this.stop=this.stop.bind(this),this.logger=V}async init(){if(this.initialized)return;let e=ir();this.logger.info(`Task Master MCP Server starting...`),this.logger.info(`Tool mode configuration: ${e}`);let t=ar(this.server,e);return this.logger.info(`Normalized tool mode: ${t.normalizedMode}`),this.logger.info(`Registered ${t.registeredTools.length} tools successfully`),t.registeredTools.length>0&&this.logger.debug(`Registered tools: ${t.registeredTools.join(`, `)}`),t.failedTools.length>0&&this.logger.warn(`Failed to register ${t.failedTools.length} tools: ${t.failedTools.join(`, `)}`),this.initialized=!0,this}async start(){return this.initialized||await this.init(),this.server.on(`connect`,e=>{e.session.server.sendLoggingMessage({data:{context:e.session.context,message:`MCP Server connected: ${e.session.name}`},level:`info`}),this.registerRemoteProvider(e.session)}),await this.server.start({transportType:`stdio`,timeout:12e4}),this}registerRemoteProvider(e){if(e){if(!e.clientCapabilities||!e.clientCapabilities.sampling){e.server.sendLoggingMessage({data:{context:e.context,message:`MCP session missing required sampling capabilities, providers not registered`},level:`info`});return}let n=new lt;n.setSession(e),t.getInstance().registerProvider(`mcp`,n),e.server.sendLoggingMessage({data:{context:e.context,message:`MCP Server connected`},level:`info`})}else e.server.sendLoggingMessage({data:{context:e.context,message:`No MCP sessions available, providers not registered`},level:`warn`})}async stop(){this.server&&await this.server.stop()}};L.config(),process.env.TASK_MASTER_MCP=`true`;async function cr(){let e=new sr;process.on(`SIGINT`,async()=>{await e.stop(),process.exit(0)}),process.on(`SIGTERM`,async()=>{await e.stop(),process.exit(0)});try{await e.start()}catch(e){V.error(`Failed to start MCP server: ${e.message}`),process.exit(1)}}cr();export{};
|
|
44
|
+
${y.result}`;if(e){let{updateSubtaskById:e}=await import(`./update-subtask-by-id-r9mGhUE1.js`);await e(P.join(d,`.taskmaster`,`tasks`,`tasks.json`),l,n,!1,{session:p,mcpLog:m,commandName:`research-save`,outputType:`mcp`,projectRoot:d,tag:f},`json`),t.info(`Research saved to subtask ${l}`)}else{let e=(await import(`./update-task-by-id-B9n9vQcI.js`)).default,r=parseInt(l,10);await e(P.join(d,`.taskmaster`,`tasks`,`tasks.json`),r,n,!1,{session:p,mcpLog:m,commandName:`research-save`,outputType:`mcp`,projectRoot:d,tag:f},`json`,!0),t.info(`Research saved to task ${l}`)}}catch(e){t.warn(`Error saving research to task/subtask: ${e.message}`)}return _(),{success:!0,data:{query:y.query,result:y.result,contextSize:y.contextSize,contextTokens:y.contextTokens,tokenBreakdown:y.tokenBreakdown,systemPromptTokens:y.systemPromptTokens,userPromptTokens:y.userPromptTokens,totalInputTokens:y.totalInputTokens,detailLevel:y.detailLevel,telemetryData:y.telemetryData,tagInfo:y.tagInfo,savedFilePath:y.savedFilePath}}}catch(e){return _(),t.error(`Error in researchDirect: ${e.message}`),{success:!1,error:{code:e.code||`RESEARCH_ERROR`,message:e.message}}}}async function mn(e,t,n={}){let{tasksJsonPath:r,id:i,strength:a=`regular`,prompt:o,research:s=!1,projectRoot:c,tag:l}=e,{session:u}=n;b();let d=D(t);try{if(!r)return t.error(`scopeDownDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i)return t.error(`Missing required parameter: id`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`The id parameter is required for scoping down tasks`}};let e=i.split(`,`).map(e=>parseInt(e.trim(),10));t.info(`Scoping down tasks: ${e.join(`, `)}, strength: ${a}, research: ${s}`);let n=await oe(r,e,a,o,{session:u,mcpLog:d,projectRoot:c,commandName:`scope-down`,outputType:`mcp`,tag:l,research:s},`json`);return _(),{success:!0,data:{updatedTasks:n.updatedTasks,tasksUpdated:n.updatedTasks.length,message:`Successfully scoped down ${n.updatedTasks.length} task(s)`,telemetryData:n.telemetryData}}}catch(e){return _(),t.error(`Error in scopeDownDirect: ${e.message}`),{success:!1,error:{code:e.code||`SCOPE_DOWN_ERROR`,message:e.message}}}}async function hn(e,t,n={}){let{tasksJsonPath:r,id:i,strength:a=`regular`,prompt:o,research:s=!1,projectRoot:c,tag:l}=e,{session:u}=n;b();let d=D(t);try{if(!r)return t.error(`scopeUpDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i)return t.error(`Missing required parameter: id`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`The id parameter is required for scoping up tasks`}};let e=i.split(`,`).map(e=>parseInt(e.trim(),10));t.info(`Scoping up tasks: ${e.join(`, `)}, strength: ${a}, research: ${s}`);let n=await ge(r,e,a,o,{session:u,mcpLog:d,projectRoot:c,commandName:`scope-up`,outputType:`mcp`,tag:l,research:s},`json`);return _(),{success:!0,data:{updatedTasks:n.updatedTasks,tasksUpdated:n.updatedTasks.length,message:`Successfully scoped up ${n.updatedTasks.length} task(s)`,telemetryData:n.telemetryData}}}catch(e){return _(),t.error(`Error in scopeUpDirect: ${e.message}`),{success:!1,error:{code:e.code||`SCOPE_UP_ERROR`,message:e.message}}}}async function gn(e,t,n={}){let{session:r}=n,{tasksJsonPath:i,id:a,prompt:o,research:s,metadata:c,projectRoot:l,tag:u}=e,d=D(t);try{if(d.info(`Updating subtask by ID via direct function. ID: ${a}, ProjectRoot: ${l}`),!i){let e=`tasksJsonPath is required but was not provided.`;return d.error(e),{success:!1,error:{code:`MISSING_ARGUMENT`,message:e}}}if(!a||typeof a!=`string`||!a.trim()){let e=`Subtask ID cannot be empty.`;return d.error(e),{success:!1,error:{code:`INVALID_SUBTASK_ID`,message:e}}}if(!o&&!c){let e=`No prompt or metadata specified. Please provide information to append or metadata to update.`;return d.error(e),{success:!1,error:{code:`MISSING_PROMPT`,message:e}}}let e=String(a).trim(),n=i,f=s===!0;t.info(`Updating subtask with ID ${e} with prompt "${o||`(metadata-only)`}" and research: ${f}`);let p=w();p||b();try{let t=await He(n,e,o,f,{mcpLog:d,session:r,projectRoot:l,tag:u,commandName:`update-subtask`,outputType:`mcp`,metadata:c},`json`);if(!t||t.updatedSubtask===null){let e=`Subtask ${a} or its parent task not found.`;return d.error(e),{success:!1,error:{code:`SUBTASK_NOT_FOUND`,message:e}}}let i=e.split(`.`)[0],s=`Successfully updated subtask with ID ${e}`;return d.success(s),{success:!0,data:{message:`Successfully updated subtask with ID ${e}`,subtaskId:e,parentId:i,subtask:t.updatedSubtask,tasksPath:n,useResearch:f,telemetryData:t.telemetryData,tagInfo:t.tagInfo}}}catch(e){return d.error(`Error updating subtask by ID: ${e.message}`),{success:!1,error:{code:`UPDATE_SUBTASK_CORE_ERROR`,message:e.message||`Unknown error updating subtask`}}}finally{!p&&w()&&_()}}catch(e){return d.error(`Setup error in updateSubtaskByIdDirect: ${e.message}`),w()&&_(),{success:!1,error:{code:`DIRECT_FUNCTION_SETUP_ERROR`,message:e.message||`Unknown setup error`}}}}async function _n(e,t,n={}){let{session:r}=n,{tasksJsonPath:i,id:a,prompt:o,research:s,append:c,metadata:l,projectRoot:u,tag:d}=e,f=D(t);try{if(f.info(`Updating task by ID via direct function. ID: ${a}, ProjectRoot: ${u}`),!a){let e=`No task ID specified. Please provide a task ID to update.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}if(!o&&!l){let e=`No prompt or metadata specified. Please provide a prompt with new information or metadata for the task update.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let t;if(typeof a==`string`)t=a;else if(typeof a==`number`)t=String(a);else{let e=`Invalid task ID type: ${typeof a}. Task ID must be a string or number.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let n=i||E({projectRoot:u,file:e.file},f);if(!n){let e=`tasks.json path could not be resolved.`;return f.error(e),{success:!1,error:{code:`INPUT_VALIDATION_ERROR`,message:e}}}let p=s===!0;f.info(`Updating task with ID ${t} with prompt "${o||`(metadata-only)`}" and research: ${p}`);let m=w();m||b();try{let e=await ae(n,t,o,p,{mcpLog:f,session:r,projectRoot:u,tag:d,commandName:`update-task`,outputType:`mcp`,metadata:l},`json`,c||!1);if(!e||e.updatedTask===null){let n=`Task ${t} was not updated (likely already completed).`;return f.info(n),{success:!0,data:{message:n,taskId:t,updated:!1,telemetryData:e?.telemetryData,tagInfo:e?.tagInfo}}}let i=`Successfully updated task with ID ${t} based on the prompt`;return f.info(i),{success:!0,data:{message:i,taskId:t,tasksPath:n,useResearch:p,updated:!0,updatedTask:e.updatedTask,telemetryData:e.telemetryData,tagInfo:e.tagInfo}}}catch(e){return f.error(`Error updating task by ID: ${e.message}`),{success:!1,error:{code:`UPDATE_TASK_CORE_ERROR`,message:e.message||`Unknown error updating task`}}}finally{!m&&w()&&_()}}catch(e){return f.error(`Setup error in updateTaskByIdDirect: ${e.message}`),w()&&_(),{success:!1,error:{code:`DIRECT_FUNCTION_SETUP_ERROR`,message:e.message||`Unknown setup error`}}}}async function vn(e,t,n={}){let{session:r}=n,{from:i,prompt:a,research:o,tasksJsonPath:s,projectRoot:c,tag:l}=e,u=D(t);if(!c)return u.error(`updateTasksDirect requires a projectRoot argument.`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`projectRoot is required.`}};if(!i)return u.error(`updateTasksDirect called without from ID`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`Starting task ID (from) is required`}};if(!a)return u.error(`updateTasksDirect called without prompt`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`Update prompt is required`}};u.info(`Updating tasks via direct function. From: ${i}, Research: ${o}, File: ${s}, ProjectRoot: ${c}`),b();try{let e=await ye(s,i,a,o,{session:r,mcpLog:u,projectRoot:c,tag:l},`json`);return e&&e.success&&Array.isArray(e.updatedTasks)?(u.success(`Successfully updated ${e.updatedTasks.length} tasks.`),{success:!0,data:{message:`Successfully updated ${e.updatedTasks.length} tasks.`,tasksPath:s,updatedCount:e.updatedTasks.length,telemetryData:e.telemetryData,tagInfo:e.tagInfo}}):(u.error(`Core updateTasks function did not return a successful structure.`),{success:!1,error:{code:`CORE_FUNCTION_ERROR`,message:e?.message||`Core function failed to update tasks or returned unexpected result.`}})}catch(e){return u.error(`Error executing core updateTasks: ${e.message}`),{success:!1,error:{code:`UPDATE_TASKS_CORE_ERROR`,message:e.message||`Unknown error updating tasks`}}}finally{_()}}async function yn(e,t,n={}){let{tasksJsonPath:r,name:i,projectRoot:a}=e,{session:o}=n;b();let s=D(t);try{if(!r)return t.error(`useTagDirect called without tasksJsonPath`),_(),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};if(!i||typeof i!=`string`)return t.error(`Missing required parameter: name`),_(),{success:!1,error:{code:`MISSING_PARAMETER`,message:`Tag name is required and must be a string`}};t.info(`Switching to tag: ${i}`);let e=await Ae(r,i,{},{session:o,mcpLog:s,projectRoot:a},`json`);return _(),{success:!0,data:{tagName:e.currentTag,switched:e.switched,previousTag:e.previousTag,taskCount:e.taskCount,message:`Successfully switched to tag "${e.currentTag}"`}}}catch(e){return _(),t.error(`Error in useTagDirect: ${e.message}`),{success:!1,error:{code:e.code||`USE_TAG_ERROR`,message:e.message}}}}async function bn(e,t){let{tasksJsonPath:n,projectRoot:r,tag:i}=e;if(!n)return t.error(`validateDependenciesDirect called without tasksJsonPath`),{success:!1,error:{code:`MISSING_ARGUMENT`,message:`tasksJsonPath is required`}};try{t.info(`Validating dependencies in tasks: ${n}`);let e=n;return N.existsSync(e)?(b(),await je(e,{projectRoot:r,tag:i}),_(),{success:!0,data:{message:`Dependencies validated successfully`,tasksPath:e}}):{success:!1,error:{code:`FILE_NOT_FOUND`,message:`Tasks file not found at ${e}`}}}catch(e){return _(),t.error(`Error validating dependencies: ${e.message}`),{success:!1,error:{code:`VALIDATION_ERROR`,message:e.message}}}}function xn(e){e.addTool({name:`add_dependency`,description:`Add a dependency relationship between two tasks`,parameters:I.object({id:I.string().describe(`ID of task that will depend on another task`),dependsOn:I.string().describe(`ID of task that will become a dependency`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Add Dependency`,destructiveHint:!0},execute:Q(`add-dependency`,async(e,{log:t,session:n})=>{try{t.info(`Adding dependency for task ${e.id} to depend on ${e.dependsOn}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await zt({tasksJsonPath:i,id:e.id,dependsOn:e.dependsOn,projectRoot:e.projectRoot,tag:n},t);return a.success?t.info(`Successfully added dependency: ${a.data.message}`):t.error(`Failed to add dependency: ${a.error.message}`),Y({result:a,log:t,errorPrefix:`Error adding dependency`,projectRoot:e.projectRoot,tag:n})}catch(e){return t.error(`Error in addDependency tool: ${e.message}`),J(e.message)}})})}function Sn(e){e.addTool({name:`add_subtask`,description:`Add a subtask to an existing task`,parameters:I.object({id:I.string().describe(`Parent task ID (required)`),taskId:I.string().optional().describe(`Existing task ID to convert to subtask`),title:I.string().optional().describe(`Title for the new subtask (when creating a new subtask)`),description:I.string().optional().describe(`Description for the new subtask`),details:I.string().optional().describe(`Implementation details for the new subtask`),status:I.string().optional().describe(`Status for the new subtask (default: 'pending')`),dependencies:I.string().optional().describe(`Comma-separated list of dependency IDs for the new subtask`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),skipGenerate:I.boolean().optional().describe(`Skip regenerating task files`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Add Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Adding subtask with args: ${JSON.stringify(e)}`);let a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=await Bt({tasksJsonPath:a,id:e.id,taskId:e.taskId,title:e.title,description:e.description,details:e.details,status:e.status,dependencies:e.dependencies,skipGenerate:e.skipGenerate,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Subtask added successfully: ${o.data.message}`):t.error(`Failed to add subtask: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error adding subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in addSubtask tool: ${e.message}`),J(e.message)}})})}function Cn(e){e.addTool({name:`add_tag`,description:`Create a new tag for organizing tasks in different contexts`,parameters:I.object({name:I.string().describe(`Name of the new tag to create`),copyFromCurrent:I.boolean().optional().describe(`Whether to copy tasks from the current tag (default: false)`),copyFromTag:I.string().optional().describe(`Specific tag to copy tasks from`),fromBranch:I.boolean().optional().describe(`Create tag name from current git branch (ignores name parameter)`),description:I.string().optional().describe(`Optional description for the tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Add Tag`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting add-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Vt({tasksJsonPath:r,name:e.name,copyFromCurrent:e.copyFromCurrent,copyFromTag:e.copyFromTag,fromBranch:e.fromBranch,description:e.description,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error creating tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in add-tag tool: ${e.message}`),J(e.message)}})})}function wn(e){e.addTool({name:`add_task`,description:`Add a new task using AI`,parameters:I.object({prompt:I.string().optional().describe(`Description of the task to add (required if not using manual fields)`),title:I.string().optional().describe(`Task title (for manual task creation)`),description:I.string().optional().describe(`Task description (for manual task creation)`),details:I.string().optional().describe(`Implementation details (for manual task creation)`),testStrategy:I.string().optional().describe(`Test strategy (for manual task creation)`),dependencies:I.string().optional().describe(`Comma-separated list of task IDs this task depends on`),priority:I.string().optional().describe(`Task priority (high, medium, low)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for task creation`)}),annotations:{title:`Add Task`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting add-task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Ht({tasksJsonPath:a,prompt:e.prompt,title:e.title,description:e.description,details:e.details,testStrategy:e.testStrategy,dependencies:e.dependencies,priority:e.priority,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error adding task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in add-task tool: ${e.message}`),J(e.message)}})})}function Tn(e){e.addTool({name:`analyze_project_complexity`,description:`Analyze task complexity and generate expansion recommendations.`,parameters:I.object({threshold:I.coerce.number().int().min(1).max(10).optional().default(5).describe(`Complexity score threshold (1-10) to recommend expansion.`),research:I.boolean().optional().default(!1).describe(`Use Perplexity AI for research-backed analysis.`),output:I.string().optional().describe(`Output file path relative to project root (default: ${s}).`),file:I.string().optional().describe(`Path to the tasks file relative to project root (default: tasks/tasks.json).`),ids:I.string().optional().describe(`Comma-separated list of task IDs to analyze specifically (e.g., "1,3,5").`),from:I.coerce.number().int().positive().optional().describe(`Starting task ID in a range to analyze.`),to:I.coerce.number().int().positive().optional().describe(`Ending task ID in a range to analyze.`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Analyze Project Complexity`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`analyze_project_complexity`;try{t.info(`Executing ${i} tool with args: ${JSON.stringify(e)}`);let a=r({projectRoot:e.projectRoot,tag:e.tag}),o;try{o=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`${i}: Resolved tasks path: ${o}`)}catch(n){return t.error(`${i}: Error finding tasks.json: ${n.message}`),J(`Failed to find tasks.json within project root '${e.projectRoot}': ${n.message}`)}let s=S(e.output,{projectRoot:e.projectRoot,tag:a},t);t.info(`${i}: Report output path: ${s}`);let c=P.dirname(s);try{N.existsSync(c)||(N.mkdirSync(c,{recursive:!0}),t.info(`${i}: Created output directory: ${c}`))}catch(e){return t.error(`${i}: Failed to create output directory ${c}: ${e.message}`),J(`Failed to create output directory: ${e.message}`)}let l=await Ut({tasksJsonPath:o,outputPath:s,threshold:e.threshold,research:e.research,projectRoot:e.projectRoot,tag:a,ids:e.ids,from:e.from,to:e.to},t,{session:n});return t.info(`${i}: Direct function result: success=${l.success}`),Y({result:l,log:t,errorPrefix:`Error analyzing task complexity`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function En(e){e.addTool({name:`clear_subtasks`,description:`Clear subtasks from specified tasks`,parameters:I.object({id:I.string().optional().describe(`Task IDs (comma-separated) to clear subtasks from`),all:I.boolean().optional().describe(`Clear subtasks from all tasks`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}).refine(e=>e.id||e.all,{message:`Either 'id' or 'all' parameter must be provided`,path:[`id`,`all`]}),annotations:{title:`Clear Subtasks`,destructiveHint:!0},execute:Q(`clear-subtasks`,async(e,t)=>{try{t.log.info(`Clearing subtasks with args: ${JSON.stringify(e)}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await Wt({tasksJsonPath:i,id:e.id,all:e.all,projectRoot:e.projectRoot,tag:n},t.log,{session:t.session});return a.success?t.log.info(`Subtasks cleared successfully: ${a.data.message}`):t.log.error(`Failed to clear subtasks: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error clearing subtasks`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in clearSubtasks tool: ${e.message}`),J(e.message)}})})}function Dn(e){e.addTool({name:`complexity_report`,description:`Display the complexity analysis report in a readable format`,parameters:I.object({file:I.string().optional().describe(`Path to the report file (default: ${s})`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Complexity Report`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Getting complexity report with args: ${JSON.stringify(e)}`);let n=p(e.projectRoot),r=Ue({projectRoot:e.projectRoot,complexityReport:e.file,tag:n},t);if(t.info(`Reading complexity report from path: `,r),!r)return J(`No complexity report found. Run task-master analyze-complexity first.`);let i=await Gt({reportPath:r},t);return i.success?t.info(`Successfully retrieved complexity report`):t.error(`Failed to retrieve complexity report: ${i.error.message}`),Y({result:i,log:t,errorPrefix:`Error retrieving complexity report`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in complexity-report tool: ${e.message}`),J(`Failed to retrieve complexity report: ${e.message}`)}})})}function On(e){e.addTool({name:`copy_tag`,description:`Copy an existing tag to create a new tag with all tasks and metadata`,parameters:I.object({sourceName:I.string().describe(`Name of the source tag to copy from`),targetName:I.string().describe(`Name of the new tag to create`),description:I.string().optional().describe(`Optional description for the new tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Copy Tag`,destructiveHint:!1},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting copy-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Kt({tasksJsonPath:r,sourceName:e.sourceName,targetName:e.targetName,description:e.description,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error copying tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in copy-tag tool: ${e.message}`),J(e.message)}})})}function kn(e){e.addTool({name:`delete_tag`,description:`Delete an existing tag and all its tasks`,parameters:I.object({name:I.string().describe(`Name of the tag to delete`),yes:I.boolean().optional().describe(`Skip confirmation prompts (default: true for MCP)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Delete Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting delete-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await qt({tasksJsonPath:r,name:e.name,yes:e.yes===void 0?!0:e.yes,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error deleting tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in delete-tag tool: ${e.message}`),J(e.message)}})})}function An(e){e.addTool({name:`expand_all`,description:`Expand all pending tasks into subtasks based on complexity or defaults`,parameters:I.object({num:I.string().optional().describe(`Target number of subtasks per task (uses complexity/defaults otherwise)`),research:I.boolean().optional().describe(`Enable research-backed subtask generation (e.g., using Perplexity)`),prompt:I.string().optional().describe(`Additional context to guide subtask generation for all tasks`),force:I.boolean().optional().describe(`Force regeneration of subtasks for tasks that already have them`),file:I.string().optional().describe(`Absolute path to the tasks file in the /tasks folder inside the project root (default: tasks/tasks.json)`),projectRoot:I.string().optional().describe(`Absolute path to the project root directory (derived from session if possible)`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Expand All Tasks`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Tool expand_all execution started with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`Resolved tasks.json path: ${a}`)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=xe(null,{projectRoot:e.projectRoot,tag:i},t);return t.info(`Using complexity report path: ${o}`),Y({result:await Jt({tasksJsonPath:a,num:e.num,research:e.research,prompt:e.prompt,force:e.force,projectRoot:e.projectRoot,tag:i,complexityReportPath:o},t,{session:n}),log:t,errorPrefix:`Error expanding all tasks`,projectRoot:e.projectRoot})}catch(e){return t.error(`Unexpected error in expand_all tool execute: ${e.message}`),e.stack&&t.error(e.stack),J(`An unexpected error occurred: ${e.message}`)}})})}function jn(e){e.addTool({name:`expand_task`,description:`Expand a task into subtasks for detailed implementation`,parameters:I.object({id:I.string().describe(`ID of task to expand`),num:I.string().optional().describe(`Number of subtasks to generate`),research:I.boolean().optional().default(!1).describe(`Use research role for generation`),prompt:I.string().optional().describe(`Additional context for subtask generation`),file:I.string().optional().describe(`Path to the tasks file relative to project root (e.g., tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),force:I.boolean().optional().default(!1).describe(`Force expansion even if subtasks exist`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Expand Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting expand-task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=Ue({...e,tag:i},t);return Y({result:await Yt({tasksJsonPath:a,id:e.id,num:e.num,research:e.research,prompt:e.prompt,force:e.force,complexityReportPath:o,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error expanding task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in expand-task tool: ${e.message}`),J(e.message)}})})}function Mn(e){e.addTool({name:`fix_dependencies`,description:`Fix invalid dependencies in tasks automatically`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Fix Dependencies`,destructiveHint:!0},execute:Q(`fix-dependencies`,async(e,t)=>{try{t.log.info(`Fixing dependencies with args: ${JSON.stringify(e)}`);let n=r({projectRoot:e.projectRoot,tag:e.tag}),i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await Xt({tasksJsonPath:i,projectRoot:e.projectRoot,tag:n},t.log);return a.success?t.log.info(`Successfully fixed dependencies: ${a.data.message}`):t.log.error(`Failed to fix dependencies: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error fixing dependencies`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in fixDependencies tool: ${e.message}`),J(e.message)}})})}function Nn(e){e.addTool({name:`initialize_project`,description:`Initializes a new Task Master project structure by calling the core initialization logic. Creates necessary folders and configuration files for Task Master in the current directory.`,parameters:I.object({skipInstall:I.boolean().optional().default(!1).describe(`Skip installing dependencies automatically. Never do this unless you are sure the project is already installed.`),addAliases:I.boolean().optional().default(!0).describe(`Add shell aliases (tm, taskmaster, hamster, ham) to shell config file.`),initGit:I.boolean().optional().default(!0).describe(`Initialize Git repository in project root.`),storeTasksInGit:I.boolean().optional().default(!0).describe(`Store tasks in Git (tasks.json and tasks/ directory).`),yes:I.boolean().optional().default(!0).describe(`Skip prompts and use default values. Always set to true for MCP tools.`),projectRoot:I.string().describe(`The root directory for the project. ALWAYS SET THIS TO THE PROJECT ROOT DIRECTORY. IF NOT SET, THE TOOL WILL NOT WORK.`),rules:I.array(I.enum(A)).optional().describe(`List of rule profiles to include at initialization. If omitted, defaults to Cursor profile only. Available options: ${A.join(`, `)}`)}),annotations:{title:`Initialize Project`,destructiveHint:!0},execute:Z(async(e,t)=>{let{log:n}=t,r=t.session;try{return n.info(`Executing initialize_project tool with args: ${JSON.stringify(e)}`),Y({result:await Zt(e,n,{session:r}),log:n,errorPrefix:`Initialization failed`,projectRoot:e.projectRoot})}catch(e){let t=`Project initialization tool failed: ${e.message||`Unknown error`}`;return n.error(t,e),J(t,{details:e.stack})}})})}function Pn(e){e.addTool({name:`list_tags`,description:`List all available tags with task counts and metadata`,parameters:I.object({showMetadata:I.boolean().optional().describe(`Whether to include metadata in the output (default: false)`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`List Tags`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting list-tags with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await Qt({tasksJsonPath:r,showMetadata:e.showMetadata,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error listing tags`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in list-tags tool: ${e.message}`),J(e.message)}})})}function Fn(e){e.addTool({name:`models`,description:`Get information about available AI models or set model configurations. Run without arguments to get the current model configuration and API key status for the selected model providers.`,parameters:I.object({setMain:I.string().optional().describe(`Set the primary model for task generation/updates. Model provider API key is required in the MCP config ENV.`),setResearch:I.string().optional().describe(`Set the model for research-backed operations. Model provider API key is required in the MCP config ENV.`),setFallback:I.string().optional().describe(`Set the model to use if the primary fails. Model provider API key is required in the MCP config ENV.`),listAvailableModels:I.boolean().optional().describe(`List all available models not currently in use. Input/output costs values are in dollars (3 is $3.00).`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),openrouter:I.boolean().optional().describe(`Indicates the set model ID is a custom OpenRouter model.`),ollama:I.boolean().optional().describe(`Indicates the set model ID is a custom Ollama model.`),bedrock:I.boolean().optional().describe(`Indicates the set model ID is a custom AWS Bedrock model.`),azure:I.boolean().optional().describe(`Indicates the set model ID is a custom Azure OpenAI model.`),vertex:I.boolean().optional().describe(`Indicates the set model ID is a custom Google Vertex AI model.`),"openai-compatible":I.boolean().optional().describe(`Indicates the set model ID is a custom OpenAI-compatible model. Requires baseURL parameter.`),baseURL:I.string().optional().describe(`Custom base URL for providers that support it (e.g., https://api.example.com/v1).`)}),annotations:{title:`Models`,destructiveHint:!0},execute:Q(`models`,async(e,t)=>{try{return t.log.info(`Starting models tool with args: ${JSON.stringify(e)}`),Y({result:await nn({...e,projectRoot:e.projectRoot},t.log,{session:t.session}),log:t.log,errorPrefix:`Error managing models`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in models tool: ${e.message}`),J(e.message)}})})}function In(e){e.addTool({name:`move_task`,description:`Move a task or subtask to a new position`,parameters:I.object({from:I.string().describe(`ID of the task/subtask to move (e.g., "5" or "5.2"). Can be comma-separated to move multiple tasks (e.g., "5,6,7")`),to:I.string().optional().describe(`ID of the destination (e.g., "7" or "7.3"). Required for within-tag moves. For cross-tag moves, if omitted, task will be moved to the target tag maintaining its ID`),file:I.string().optional().describe(`Custom path to tasks.json file`),projectRoot:I.string().describe(`Root directory of the project (typically derived from session)`),tag:I.string().optional().describe(`Tag context to operate on`),fromTag:I.string().optional().describe(`Source tag for cross-tag moves`),toTag:I.string().optional().describe(`Target tag for cross-tag moves`),withDependencies:I.boolean().optional().describe(`Move dependent tasks along with main task`),ignoreDependencies:I.boolean().optional().describe(`Break cross-tag dependencies during move`)}),annotations:{title:`Move Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{if(e.fromTag&&e.toTag&&e.fromTag!==e.toTag){if(!e.from)return J(`Source IDs are required for cross-tag moves`,`MISSING_SOURCE_IDS`);e.to&&t.warn(`The "to" parameter is not used for cross-tag moves and will be ignored. Tasks retain their original IDs in the target tag.`);let r=e.file;return r||=E(e,t),Y({result:await rn({sourceIds:e.from,sourceTag:e.fromTag,targetTag:e.toTag,withDependencies:e.withDependencies||!1,ignoreDependencies:e.ignoreDependencies||!1,tasksJsonPath:r,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error moving tasks between tags`,projectRoot:e.projectRoot})}else{if(!e.to)return J(`Destination ID is required for within-tag moves`,`MISSING_DESTINATION_ID`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a=e.file;a||=E(e,t);let o=e.from.split(`,`).map(e=>e.trim()),s=e.to.split(`,`).map(e=>e.trim());if(o.length!==s.length){if(o.length>1){let r=[],c=[];for(let l=0;l<o.length;l++){let u=o[l],d=s[l];if(u===d){t.info(`Skipping ${u} -> ${d} (same ID)`),c.push({fromId:u,toId:d,reason:`same ID`});continue}let f=l===o.length-1,p=await an({sourceId:u,destinationId:d,tasksJsonPath:a,projectRoot:e.projectRoot,tag:i,generateFiles:f},t,{session:n});p.success?r.push(p.data):t.error(`Failed to move ${u} to ${d}: ${p.error.message}`)}return Y({result:{success:!0,data:{moves:r,skipped:c.length>0?c:void 0,message:`Successfully moved ${r.length} tasks${c.length>0?`, skipped ${c.length}`:``}`}},log:t,errorPrefix:`Error moving multiple tasks`,projectRoot:e.projectRoot})}return Y({result:{success:!0,data:{moves:results,skippedMoves,message:`Successfully moved ${results.length} tasks${skippedMoves.length>0?`, skipped ${skippedMoves.length} moves`:``}`}},log:t,errorPrefix:`Error moving multiple tasks`,projectRoot:e.projectRoot})}else return Y({result:await an({sourceId:e.from,destinationId:e.to,tasksJsonPath:a,projectRoot:e.projectRoot,tag:i,generateFiles:!0},t,{session:n}),log:t,errorPrefix:`Error moving task`,projectRoot:e.projectRoot})}}catch(e){return J(`Failed to move task: ${e.message}`,`MOVE_TASK_ERROR`)}})})}function Ln(e){e.addTool({name:`next_task`,description:`Find the next task to work on based on dependencies and status`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),complexityReport:I.string().optional().describe(`Path to the complexity report file (relative to project root or absolute)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Next Task`,readOnlyHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Finding next task with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=ue(e,n)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o;try{o=de({...e,tag:i},n)}catch(e){t.error(`Error finding complexity report: ${e.message}`),o=null}let s=await on({tasksJsonPath:a,reportPath:o,projectRoot:e.projectRoot,tag:i},t,{session:n});return t.info(`Next task result: ${s.success?`found`:`none`}`),Y({result:s,log:t,errorPrefix:`Error finding next task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error finding next task: ${e.message}`),J(e.message)}})})}function Rn(e){e.addTool({name:`parse_prd`,description:`Parse a Product Requirements Document (PRD) text file to automatically generate initial tasks. Reinitializing the project is not necessary to run this tool. It is recommended to run parse-prd after initializing the project and creating/importing a prd.txt file in the project root's ${i} directory.`,parameters:I.object({input:I.string().optional().default(ne).describe(`Absolute path to the PRD document file (.txt, .md, etc.)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),destination:I.enum([`local`,`hamster`]).optional().default(`local`).describe(`Where to parse the PRD: "local" writes tasks.json, "hamster" creates a cloud brief and tasks`),output:I.string().optional().describe(`Output path for tasks.json file (default: ${a})`),numTasks:I.string().optional().describe(`Approximate number of top-level tasks to generate (default: 10). As the agent, if you have enough information, ensure to enter a number of tasks that would logically scale with project complexity. Setting to 0 will allow Taskmaster to determine the appropriate number of tasks based on the complexity of the PRD. Avoid entering numbers above 50 due to context window limitations.`),force:I.boolean().optional().default(!1).describe(`Overwrite existing output file without prompting.`),research:I.boolean().optional().describe(`Enable Taskmaster to use the research role for potentially more informed task generation. Requires appropriate API key.`),append:I.boolean().optional().describe(`Append generated tasks to existing file.`)}),annotations:{title:`Parse PRD`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n,reportProgress:i})=>{try{let a=r({projectRoot:e.projectRoot,tag:e.tag}),o=dt(i,t);return Y({result:await sn({...e,tag:a},t,{session:n,reportProgress:o}),log:t,errorPrefix:`Error parsing PRD`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in parse_prd: ${e.message}`),J(`Failed to parse PRD: ${e.message}`)}})})}function zn(e){e.addTool({name:`remove_dependency`,description:`Remove a dependency from a task`,parameters:I.object({id:I.string().describe(`Task ID to remove dependency from`),dependsOn:I.string().describe(`Task ID to remove as a dependency`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Remove Dependency`,destructiveHint:!0},execute:Q(`remove-dependency`,async(e,t)=>{try{let n=r({projectRoot:e.projectRoot,tag:e.tag});t.log.info(`Removing dependency for task ${e.id} from ${e.dependsOn} with args: ${JSON.stringify(e)}`);let i;try{i=E({projectRoot:e.projectRoot,file:e.file},t.log)}catch(e){return t.log.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await ln({tasksJsonPath:i,id:e.id,dependsOn:e.dependsOn,projectRoot:e.projectRoot,tag:n},t.log);return a.success?t.log.info(`Successfully removed dependency: ${a.data.message}`):t.log.error(`Failed to remove dependency: ${a.error.message}`),Y({result:a,log:t.log,errorPrefix:`Error removing dependency`,projectRoot:e.projectRoot})}catch(e){return t.log.error(`Error in removeDependency tool: ${e.message}`),J(e.message)}})})}function Bn(e){e.addTool({name:`remove_subtask`,description:`Remove a subtask from its parent task`,parameters:I.object({id:I.string().describe(`Subtask ID to remove in format 'parentId.subtaskId' (required)`),convert:I.boolean().optional().describe(`Convert the subtask to a standalone task instead of deleting it`),file:I.string().optional().describe(`Absolute path to the tasks file (default: tasks/tasks.json)`),skipGenerate:I.boolean().optional().describe(`Skip regenerating task files`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Remove Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Removing subtask with args: ${JSON.stringify(e)}`);let a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let o=await un({tasksJsonPath:a,id:e.id,convert:e.convert,skipGenerate:e.skipGenerate,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Subtask removed successfully: ${o.data.message}`):t.error(`Failed to remove subtask: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error removing subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in removeSubtask tool: ${e.message}`),J(e.message)}})})}function Vn(e){e.addTool({name:`remove_task`,description:`Remove a task or subtask permanently from the tasks list`,parameters:I.object({id:I.string().describe(`ID of the task or subtask to remove (e.g., '5' or '5.2'). Can be comma-separated to update multiple tasks/subtasks at once.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),confirm:I.boolean().optional().describe(`Whether to skip confirmation prompt (default: false)`),tag:I.string().optional().describe(`Specify which tag context to operate on. Defaults to the current active tag.`)}),annotations:{title:`Remove Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Removing task(s) with ID(s): ${e.id}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}t.info(`Using tasks file path: ${a}`);let o=await dn({tasksJsonPath:a,id:e.id,projectRoot:e.projectRoot,tag:i},t,{session:n});return o.success?t.info(`Successfully removed task: ${e.id}`):t.error(`Failed to remove task: ${o.error.message}`),Y({result:o,log:t,errorPrefix:`Error removing task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in remove-task tool: ${e.message}`),J(`Failed to remove task: ${e.message}`)}})})}function Hn(e){e.addTool({name:`rename_tag`,description:`Rename an existing tag`,parameters:I.object({oldName:I.string().describe(`Current name of the tag to rename`),newName:I.string().describe(`New name for the tag`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Rename Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting rename-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await fn({tasksJsonPath:r,oldName:e.oldName,newName:e.newName,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error renaming tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in rename-tag tool: ${e.message}`),J(e.message)}})})}function Un(e){e.addTool({name:`research`,description:`Perform AI-powered research queries with project context`,parameters:I.object({query:I.string().describe(`Research query/prompt (required)`),taskIds:I.string().optional().describe(`Comma-separated list of task/subtask IDs for context (e.g., "15,16.2,17")`),filePaths:I.string().optional().describe(`Comma-separated list of file paths for context (e.g., "src/api.js,docs/readme.md")`),customContext:I.string().optional().describe(`Additional custom context text to include in the research`),includeProjectTree:I.boolean().optional().describe(`Include project file tree structure in context (default: false)`),detailLevel:I.enum([`low`,`medium`,`high`]).optional().describe(`Detail level for the research response (default: medium)`),saveTo:I.string().optional().describe(`Automatically save research results to specified task/subtask ID (e.g., "15" or "15.2")`),saveToFile:I.boolean().optional().describe(`Save research results to .taskmaster/docs/research/ directory (default: false)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Research`,destructiveHint:!0,openWorldHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{let i=r({projectRoot:e.projectRoot,tag:e.tag});return t.info(`Starting research with query: "${e.query.substring(0,100)}${e.query.length>100?`...`:``}"`),Y({result:await pn({query:e.query,taskIds:e.taskIds,filePaths:e.filePaths,customContext:e.customContext,includeProjectTree:e.includeProjectTree||!1,detailLevel:e.detailLevel||`medium`,saveTo:e.saveTo,saveToFile:e.saveToFile||!1,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error performing research`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in research tool: ${e.message}`),J(e.message)}})})}async function Wn(e,t,n={}){let{projectRoot:r,language:i}=e,a=D(t);t.info(`Executing response-language_direct with args: ${JSON.stringify(e)}`),t.info(`Using project root: ${r}`);try{return b(),We(i,{mcpLog:a,projectRoot:r})}catch(e){return{success:!1,error:{code:`DIRECT_FUNCTION_ERROR`,message:e.message,details:e.stack}}}finally{_()}}function Gn(e){e.addTool({name:`response-language`,description:`Get or set the response language for the project`,parameters:I.object({projectRoot:I.string().describe(`The root directory for the project. ALWAYS SET THIS TO THE PROJECT ROOT DIRECTORY. IF NOT SET, THE TOOL WILL NOT WORK.`),language:I.string().describe(`The new response language to set. like "中文" "English" or "español".`)}),annotations:{title:`Response Language`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{return t.info(`Executing response-language tool with args: ${JSON.stringify(e)}`),Y({result:await Wn({...e,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error setting response language`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in response-language tool: ${e.message}`),J(e.message)}})})}async function Kn(e,t,n={}){b();try{let{action:t,profiles:n,projectRoot:r,yes:i,force:a}=e;if(!t||!Array.isArray(n)||n.length===0||!r)return{success:!1,error:{code:`MISSING_ARGUMENT`,message:`action, profiles, and projectRoot are required.`}};let o=[],s=[];if(t===O.REMOVE){if(!a&&Ke(r,n)){let e=Ge(r);return e.filter(e=>!n.includes(e)),{success:!1,error:{code:`CRITICAL_REMOVAL_BLOCKED`,message:`CRITICAL: This operation would remove ALL remaining rule profiles (${n.join(`, `)}), leaving your project with no rules configurations. This could significantly impact functionality. Currently installed profiles: ${e.join(`, `)}. If you're certain you want to proceed, set force: true or use the CLI with --force flag.`}}}for(let e of n){if(!Ye(e)){o.push({profileName:e,success:!1,error:`The requested rule profile for '${e}' is unavailable. Supported profiles are: ${A.join(`, `)}.`});continue}let t=Je(r,k(e));o.push(t)}let e=o.filter(e=>e.success).map(e=>e.profileName),t=o.filter(e=>e.skipped).map(e=>e.profileName),i=o.filter(e=>e.error&&!e.success&&!e.skipped),s=o.filter(e=>e.notice),c=``;return e.length>0&&(c+=`Successfully removed Task Master rules: ${e.join(`, `)}.`),t.length>0&&(c+=`Skipped (default or protected): ${t.join(`, `)}.`),i.length>0&&(c+=i.map(e=>`Error removing ${e.profileName}: ${e.error}`).join(` `)),s.length>0&&(c+=` Notices: ${s.map(e=>`${e.profileName} - ${e.notice}`).join(`; `)}.`),_(),{success:i.length===0,data:{summary:c,results:o}}}else if(t===O.ADD){for(let e of n){if(!Ye(e)){s.push({profileName:e,success:!1,error:`Profile not found: static import missing for '${e}'. Valid profiles: ${A.join(`, `)}`});continue}let t=k(e),{success:n,failed:i}=Xe(r,t),a=t.rulesDir,o=P.join(r,a),c=t.profileDir,l=t.mcpConfig!==!1,u=l&&t.mcpConfigPath?P.join(r,t.mcpConfigPath):null,d=l&&u?N.existsSync(u):void 0,f=N.existsSync(o),p=N.existsSync(P.join(r,c)),m=i>0?`${i} rule files failed to convert.`:null,h={profileName:e,mcpConfigCreated:d,rulesDirCreated:f,profileFolderCreated:p,skipped:!1,error:m,success:(l?d:!0)&&f&&n>0&&!m};s.push(h)}let e=s.filter(e=>e.success).map(e=>e.profileName),t=s.filter(e=>e.error&&!e.success),i=``;return e.length>0&&(i+=`Successfully added rules: ${e.join(`, `)}.`),t.length>0&&(i+=t.map(e=>` Error adding ${e.profileName}: ${e.error}`).join(` `)),_(),{success:t.length===0,data:{summary:i,results:s}}}else return _(),{success:!1,error:{code:`INVALID_ACTION`,message:`Unknown action. Use "${O.ADD}" or "${O.REMOVE}".`}}}catch(e){return _(),t.error(`[rulesDirect] Error: ${e.message}`),{success:!1,error:{code:e.code||`RULES_ERROR`,message:e.message}}}}function qn(e){e.addTool({name:`rules`,description:`Add or remove rule profiles from the project.`,parameters:I.object({action:I.enum([`add`,`remove`]).describe(`Whether to add or remove rule profiles.`),profiles:I.array(I.enum(A)).min(1).describe(`List of rule profiles to add or remove (e.g., [\"cursor\", \"roo\"]). Available options: ${A.join(`, `)}`),projectRoot:I.string().describe(`The root directory of the project. Must be an absolute path.`),force:I.boolean().optional().default(!1).describe(`DANGEROUS: Force removal even if it would leave no rule profiles. Only use if you are absolutely certain.`)}),annotations:{title:`Rules`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{return t.info(`[rules tool] Executing action: ${e.action} for profiles: ${e.profiles.join(`, `)} in ${e.projectRoot}`),Y({result:await Kn(e,t,{session:n}),log:t,projectRoot:e.projectRoot})}catch(e){return t.error(`[rules tool] Error: ${e.message}`),J(e.message,{details:e.stack})}})})}function Jn(e){e.addTool({name:`scope_down_task`,description:`Decrease the complexity of one or more tasks using AI`,parameters:I.object({id:I.string().describe(`Comma-separated list of task IDs to scope down (e.g., "1,3,5")`),strength:I.string().optional().describe(`Strength level: light, regular, or heavy (default: regular)`),prompt:I.string().optional().describe(`Custom prompt for specific scoping adjustments`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for scoping`)}),annotations:{title:`Scope Down Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting scope-down with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await mn({tasksJsonPath:a,id:e.id,strength:e.strength,prompt:e.prompt,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error scoping down task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in scope-down tool: ${e.message}`),J(e.message)}})})}function Yn(e){e.addTool({name:`scope_up_task`,description:`Increase the complexity of one or more tasks using AI`,parameters:I.object({id:I.string().describe(`Comma-separated list of task IDs to scope up (e.g., "1,3,5")`),strength:I.string().optional().describe(`Strength level: light, regular, or heavy (default: regular)`),prompt:I.string().optional().describe(`Custom prompt for specific scoping adjustments`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`),research:I.boolean().optional().describe(`Whether to use research capabilities for scoping`)}),annotations:{title:`Scope Up Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting scope-up with args: ${JSON.stringify(e)}`);let i=r({projectRoot:e.projectRoot,tag:e.tag}),a;try{a=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await hn({tasksJsonPath:a,id:e.id,strength:e.strength,prompt:e.prompt,research:e.research,projectRoot:e.projectRoot,tag:i},t,{session:n}),log:t,errorPrefix:`Error scoping up task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in scope-up tool: ${e.message}`),J(e.message)}})})}function Xn(e){e.addTool({name:`update_subtask`,description:`Appends timestamped information to a specific subtask without replacing existing content. If you just want to update the subtask status, use set_task_status instead.`,parameters:I.object({id:y.describe(`ID of the subtask to update in format "parentId.subtaskId" (e.g., "5.2"). Parent ID is the ID of the task that contains the subtask.`),prompt:I.string().optional().describe(`Information to add to the subtask. Required unless only updating metadata.`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),metadata:I.string().optional().describe(`JSON string of metadata to merge into subtask metadata. Example: '{"ticketId": "JIRA-456", "reviewed": true}'. Requires TASK_MASTER_ALLOW_METADATA_UPDATES=true in MCP environment.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Subtask`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update_subtask`;try{let a=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Updating subtask with args: ${JSON.stringify(e)}`);let o;try{o=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let s=mt(e.metadata,J);if(s.error)return s.error;let c=s.parsedMetadata;if(!e.prompt&&!c)return J(`Either prompt or metadata must be provided for update-subtask`);let l=await gn({tasksJsonPath:o,id:e.id,prompt:e.prompt,research:e.research,metadata:c,projectRoot:e.projectRoot,tag:a},t,{session:n});return l.success?t.info(`Successfully updated subtask with ID ${e.id}`):t.error(`Failed to update subtask: ${l.error?.message||`Unknown error`}`),Y({result:l,log:t,errorPrefix:`Error updating subtask`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function Zn(e){e.addTool({name:`update_task`,description:`Updates a single task by ID with new information or context provided in the prompt.`,parameters:I.object({id:I.string().describe(`ID of the task (e.g., '15') to update. Subtasks are supported using the update-subtask tool.`),prompt:I.string().optional().describe(`New information or context to incorporate into the task. Required unless only updating metadata.`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),append:I.boolean().optional().describe(`Append timestamped information to task details instead of full update`),metadata:I.string().optional().describe(`JSON string of metadata to merge into task metadata. Example: '{"githubIssue": 42, "sprint": "Q1-S3"}'. Requires TASK_MASTER_ALLOW_METADATA_UPDATES=true in MCP environment.`),file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Task`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update_task`;try{let a=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Executing ${i} tool with args: ${JSON.stringify(e)}`);let o;try{o=E({projectRoot:e.projectRoot,file:e.file},t),t.info(`${i}: Resolved tasks path: ${o}`)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let s=mt(e.metadata,J);if(s.error)return s.error;let c=s.parsedMetadata;if(!e.prompt&&!c)return J(`Either prompt or metadata must be provided for update-task`);let l=await _n({tasksJsonPath:o,id:e.id,prompt:e.prompt,research:e.research,append:e.append,metadata:c,projectRoot:e.projectRoot,tag:a},t,{session:n});return t.info(`${i}: Direct function result: success=${l.success}`),Y({result:l,log:t,errorPrefix:`Error updating task`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function Qn(e){e.addTool({name:`update`,description:`Update multiple upcoming tasks (with ID >= 'from' ID) based on new context or changes provided in the prompt. Use 'update_task' instead for a single specific task or 'update_subtask' for subtasks.`,parameters:I.object({from:I.string().describe(`Task ID from which to start updating (inclusive). IMPORTANT: This tool uses 'from', not 'id'`),prompt:I.string().describe(`Explanation of changes or new context to apply`),research:I.boolean().optional().describe(`Use Perplexity AI for research-backed updates`),file:I.string().optional().describe(`Path to the tasks file relative to project root`),projectRoot:I.string().optional().describe(`The directory of the project. (Optional, usually from session)`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Update Tasks`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{let i=`update`,{from:a,prompt:o,research:s,file:c,projectRoot:l,tag:u}=e,d=r({projectRoot:e.projectRoot,tag:e.tag});try{t.info(`Executing ${i} tool with normalized root: ${l}`);let r;try{r=E({projectRoot:l,file:c},t),t.info(`${i}: Resolved tasks path: ${r}`)}catch(e){return t.error(`${i}: Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json within project root '${l}': ${e.message}`)}let u=await vn({tasksJsonPath:r,from:a,prompt:o,research:s,projectRoot:l,tag:d},t,{session:n});return t.info(`${i}: Direct function result: success=${u.success}`),Y({result:u,log:t,errorPrefix:`Error updating tasks`,projectRoot:e.projectRoot})}catch(e){return t.error(`Critical error in ${i} tool execute: ${e.message}`),J(`Internal tool error (${i}): ${e.message}`)}})})}function $n(e){e.addTool({name:`use_tag`,description:`Switch to a different tag context for task operations`,parameters:I.object({name:I.string().describe(`Name of the tag to switch to`),file:I.string().optional().describe(`Path to the tasks file (default: tasks/tasks.json)`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`)}),annotations:{title:`Use Tag`,destructiveHint:!0},execute:Z(async(e,{log:t,session:n})=>{try{t.info(`Starting use-tag with args: ${JSON.stringify(e)}`);let r;try{r=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}return Y({result:await yn({tasksJsonPath:r,name:e.name,projectRoot:e.projectRoot},t,{session:n}),log:t,errorPrefix:`Error switching tag`,projectRoot:e.projectRoot})}catch(e){return t.error(`Error in use-tag tool: ${e.message}`),J(e.message)}})})}function er(e){e.addTool({name:`validate_dependencies`,description:`Check tasks for dependency issues (like circular references or links to non-existent tasks) without making changes.`,parameters:I.object({file:I.string().optional().describe(`Absolute path to the tasks file`),projectRoot:I.string().describe(`The directory of the project. Must be an absolute path.`),tag:I.string().optional().describe(`Tag context to operate on`)}),annotations:{title:`Validate Dependencies`,readOnlyHint:!0},execute:Q(`validate-dependencies`,async(e,{log:t,session:n})=>{try{let n=r({projectRoot:e.projectRoot,tag:e.tag});t.info(`Validating dependencies with args: ${JSON.stringify(e)}`);let i;try{i=E({projectRoot:e.projectRoot,file:e.file},t)}catch(e){return t.error(`Error finding tasks.json: ${e.message}`),J(`Failed to find tasks.json: ${e.message}`)}let a=await bn({tasksJsonPath:i,projectRoot:e.projectRoot,tag:n},t);return a.success?t.info(`Successfully validated dependencies: ${a.data.message}`):t.error(`Failed to validate dependencies: ${a.error.message}`),Y({result:a,log:t,errorPrefix:`Error validating dependencies`,projectRoot:e.projectRoot,tag:n})}catch(e){return t.error(`Error in validateDependencies tool: ${e.message}`),J(e.message)}})})}const $={initialize_project:Nn,models:Fn,rules:qn,parse_prd:Rn,"response-language":Gn,analyze_project_complexity:Tn,expand_task:jn,expand_all:An,scope_up_task:Yn,scope_down_task:Jn,get_tasks:Mt,get_task:Pt,next_task:Ln,complexity_report:Dn,set_task_status:Rt,add_task:wn,add_subtask:Sn,update:Qn,update_task:Zn,update_subtask:Xn,remove_task:Vn,remove_subtask:Bn,clear_subtasks:En,move_task:In,add_dependency:xn,remove_dependency:zn,validate_dependencies:er,fix_dependencies:Mn,list_tags:Pn,add_tag:Cn,delete_tag:kn,use_tag:$n,rename_tag:Hn,copy_tag:On,research:Un,autopilot_start:gt,autopilot_resume:vt,autopilot_next:bt,autopilot_status:St,autopilot_complete:wt,autopilot_commit:Et,autopilot_finalize:Ot,autopilot_abort:At,generate:It},tr=[`get_tasks`,`next_task`,`get_task`,`set_task_status`,`update_subtask`,`parse_prd`,`expand_task`],nr=[...tr,`initialize_project`,`analyze_project_complexity`,`expand_all`,`add_subtask`,`remove_task`,`add_task`,`complexity_report`];function rr(e){return $[e]||null}function ir(){let e=process.env.TASK_MASTER_TOOLS;if(!e||e.trim()===``)return V.debug(`No TASK_MASTER_TOOLS env var found, defaulting to "core"`),`core`;let t=e.trim();return V.debug(`TASK_MASTER_TOOLS env var: "${t}"`),t}function ar(e,t=`core`){let n=[],r=[];try{let i=t.trim(),a=[],o=i.toLowerCase();switch(o){case`all`:a=Object.keys($),V.info(`Loading all available tools`);break;case`core`:case`lean`:a=tr,V.info(`Loading core tools only`);break;case`standard`:a=nr,V.info(`Loading standard tools`);break;default:let e=i.split(`,`).map(e=>e.trim()).filter(e=>e.length>0),t=new Set,n=[],r={response_language:`response-language`};for(let i of e){let e=null,a=i.toLowerCase();if(r[a]){let t=r[a];for(let n of Object.keys($))if(n.toLowerCase()===t.toLowerCase()){e=n;break}}if(!e){for(let t of Object.keys($))if(t.toLowerCase()===a){e=t;break}}if(!e){let t=a.replace(/_/g,`-`);for(let n of Object.keys($))if(n.toLowerCase()===t){e=n;break}}if(!e){let t=a.replace(/-/g,`_`);for(let n of Object.keys($))if(n.toLowerCase()===t){e=n;break}}e?(t.add(e),V.debug(`Resolved tool "${i}" to "${e}"`)):(n.push(i),V.warn(`Unknown tool specified: "${i}"`))}a=Array.from(t),n.length>0&&V.warn(`Unknown tools: ${n.join(`, `)}`),a.length===0?(V.warn(`No valid tools found in custom list. Loading all tools as fallback.`),a=Object.keys($)):V.info(`Loading ${a.length} custom tools from list (${t.size} unique after normalization)`);break}return V.info(`Registering ${a.length} MCP tools (mode: ${i})`),a.forEach(t=>{try{let i=rr(t);i?(i(e),V.debug(`Registered tool: ${t}`),n.push(t)):(V.warn(`Tool ${t} not found in registry`),r.push(t))}catch(e){e.message&&e.message.includes(`already registered`)?(V.debug(`Tool ${t} already registered, skipping`),n.push(t)):(V.error(`Failed to register tool ${t}: ${e.message}`),r.push(t))}}),V.info(`Successfully registered ${n.length}/${a.length} tools`),r.length>0&&V.warn(`Failed tools: ${r.join(`, `)}`),{registeredTools:n,failedTools:r,normalizedMode:o}}catch(t){V.error(`Error parsing TASK_MASTER_TOOLS environment variable: ${t.message}`),V.info(`Falling back to loading all tools`);let i=Object.keys($);for(let t of i){let i=rr(t);if(i)try{i(e),n.push(t)}catch(e){e.message&&e.message.includes(`already registered`)?(V.debug(`Fallback tool ${t} already registered, skipping`),n.push(t)):(V.warn(`Failed to register fallback tool '${t}': ${e.message}`),r.push(t))}else V.warn(`Tool '${t}' not found in registry`),r.push(t)}return V.info(`Successfully registered ${n.length} fallback tools`),{registeredTools:n,failedTools:r,normalizedMode:`all`}}}L.config(),re();const or=Qe(import.meta.url);P.dirname(or);var sr=class{constructor(){this.options={name:`Task Master MCP Server`,version:se};let e=new $e(this.options);if(e._mcpServer&&R.wrapMcpServerWithSentry)try{e._mcpServer=R.wrapMcpServerWithSentry(e._mcpServer)}catch(e){V.warn(`Failed to wrap MCP server with Sentry: ${e.message}`)}this.server=e,this.initialized=!1,this.init=this.init.bind(this),this.start=this.start.bind(this),this.stop=this.stop.bind(this),this.logger=V}async init(){if(this.initialized)return;let e=ir();this.logger.info(`Task Master MCP Server starting...`),this.logger.info(`Tool mode configuration: ${e}`);let t=ar(this.server,e);return this.logger.info(`Normalized tool mode: ${t.normalizedMode}`),this.logger.info(`Registered ${t.registeredTools.length} tools successfully`),t.registeredTools.length>0&&this.logger.debug(`Registered tools: ${t.registeredTools.join(`, `)}`),t.failedTools.length>0&&this.logger.warn(`Failed to register ${t.failedTools.length} tools: ${t.failedTools.join(`, `)}`),this.initialized=!0,this}async start(){return this.initialized||await this.init(),this.server.on(`connect`,e=>{e.session.server.sendLoggingMessage({data:{context:e.session.context,message:`MCP Server connected: ${e.session.name}`},level:`info`}),this.registerRemoteProvider(e.session)}),await this.server.start({transportType:`stdio`,timeout:12e4}),this}registerRemoteProvider(e){if(e){if(!e.clientCapabilities||!e.clientCapabilities.sampling){e.server.sendLoggingMessage({data:{context:e.context,message:`MCP session missing required sampling capabilities, providers not registered`},level:`info`});return}let n=new lt;n.setSession(e),t.getInstance().registerProvider(`mcp`,n),e.server.sendLoggingMessage({data:{context:e.context,message:`MCP Server connected`},level:`info`})}else e.server.sendLoggingMessage({data:{context:e.context,message:`No MCP sessions available, providers not registered`},level:`warn`})}async stop(){this.server&&await this.server.stop()}};L.config(),process.env.TASK_MASTER_MCP=`true`;async function cr(){let e=new sr;process.on(`SIGINT`,async()=>{await e.stop(),process.exit(0)}),process.on(`SIGTERM`,async()=>{await e.stop(),process.exit(0)});try{await e.start()}catch(e){V.error(`Failed to start MCP server: ${e.message}`),process.exit(1)}}cr();export{};
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import{c as e}from"./ai-services-unified-
|
|
1
|
+
import{c as e}from"./ai-services-unified-DRuFwL_e.js";import{Bt as t,E as n,Ft as r,Gt as i,Ht as a,It as o,J as s,Pt as c,Ut as l,Vt as u,Wt as d,hn as f,qt as p,vt as m,yt as h}from"./config-manager-ClruewMP.js";import{r as g}from"./git-utils-DllbRE35.js";import{Ct as _,St as v,Tt as y,x as ee}from"./dependency-manager-CvhwFgTc.js";import*as b from"node:path";import x from"chalk";import S from"fs";import C from"path";import{randomUUID as w}from"crypto";import T from"os";import*as E from"node:fs";import*as te from"node:os";import{execSync as D}from"child_process";import{fileURLToPath as ne}from"url";import O from"boxen";import k from"readline";import re from"figlet";import ie from"gradient-string";import ae from"inquirer";import oe from"ora";import se from"open";const A=[`amp`,`claude`,`cline`,`codex`,`cursor`,`gemini`,`kiro`,`opencode`,`kilo`,`roo`,`trae`,`vscode`,`windsurf`,`zed`],ce=[`architect`,`ask`,`orchestrator`,`code`,`debug`,`test`],j=`# Task files`,M=`tasks.json`,N=`tasks/`;function le(e){return e.trim().replace(/^#/,``).trim()}function P(e){let t=le(e);return t===M||t===N}function ue(e,t){return e.map(e=>{if(P(e)){let n=le(e),r=e.match(/\s*$/)[0];return t?`# ${n}${r}`:`${n}${r}`}return e})}function de(e){let t=[],n=!1;for(let r of e){if(r.trim()===j){n=!0;continue}P(r)||n&&!r.trim()||(n&&r.trim()&&!P(r)&&(n=!1),n||t.push(r))}return t}function fe(e,t){return e.filter(e=>{let n=e.trim();return!n||P(e)||n===j?!1:!t.has(n)})}function pe(e){let t=[j];return e?t.push(`# ${M}`,`# ${N} `):t.push(M,`${N} `),t}function me(e){if(e.some(e=>e.trim())){let t=e[e.length-1];t&&t.trim()&&e.push(``)}}function he(e,t,n){if(!e||typeof e!=`string`)throw Error(`targetPath must be a non-empty string`);if(!e.endsWith(`.gitignore`))throw Error(`targetPath must end with .gitignore`);if(!t||typeof t!=`string`)throw Error(`content must be a non-empty string`);if(typeof n!=`boolean`)throw Error(`storeTasksInGit must be a boolean`)}function ge(e,t,n){try{S.writeFileSync(e,t.join(`
|
|
2
2
|
`)+`
|
|
3
3
|
`),typeof n==`function`&&n(`success`,`Created ${e} with full template`)}catch(t){throw typeof n==`function`&&n(`error`,`Failed to create ${e}: ${t.message}`),t}}function _e(e,t,n,r){try{let i=de(S.readFileSync(e,`utf8`).split(`
|
|
4
4
|
`)),a=fe(t,new Set(i.map(e=>e.trim()).filter(e=>e))),o=[...i];a.length>0&&(me(o),o.push(...a)),me(o),o.push(...pe(n)),S.writeFileSync(e,o.join(`
|
|
@@ -3500,17 +3500,17 @@ ${r}
|
|
|
3500
3500
|
`;return S.appendFileSync(t,i),X(`debug`,`Added ${n.length} alias(es): ${n.map(e=>e.name).join(`, `)}`),!0}catch(e){return X(`debug`,`Failed to add aliases: ${e.message}`),!1}}function hn(e){let t=C.join(e,d);if(S.existsSync(t)){X(`debug`,`State file already exists, preserving current configuration`);return}let n={currentTag:`master`,lastSwitched:new Date().toISOString(),branchTagMapping:{},migrationNoticeShown:!1};try{S.writeFileSync(t,JSON.stringify(n,null,2)),X(`success`,`Created initial state file: ${t}`),X(`info`,`Default tag set to "master" for task organization`)}catch(e){X(`error`,`Failed to create state file: ${e.message}`)}}function Q(e,t,n={}){if(!xe(e)){X(`error`,`Source file not found: ${e}`);return}let r=L(e,`utf8`);if(Object.entries(n).forEach(([e,t])=>{let n=RegExp(`\\{\\{${e}\\}\\}`,`g`);r=r.replace(n,t)}),S.existsSync(t)){let e=C.basename(t);if(e===`.gitignore`){X(`info`,`${t} already exists, merging content...`);let e=S.readFileSync(t,`utf8`),n=new Set(e.split(`
|
|
3501
3501
|
`).map(e=>e.trim())),i=r.split(`
|
|
3502
3502
|
`).filter(e=>!n.has(e.trim()));if(i.length>0){let n=`${e.trim()}\n\n# Added by Taskmaster\n${i.join(`
|
|
3503
|
-
`)}`;S.writeFileSync(t,n),X(`success`,`Updated ${t} with additional entries`)}else X(`info`,`No new content to add to ${t}`);return}if(e===`README-task-master.md`){X(`info`,`${t} already exists`);let e=C.join(C.dirname(t),`README-task-master.md`);S.writeFileSync(e,r),X(`success`,`Created ${e} (preserved original README-task-master.md)`);return}X(`debug`,`${t} already exists, skipping.`);return}S.writeFileSync(t,r),X(`info`,`Created file: ${t}`)}async function gn(e={}){m()||pn(),e.git===!0?e.initGit=!0:e.git===!1&&(e.initGit=!1),e.gitTasks===!0?e.storeTasksInGit=!0:e.gitTasks===!1&&(e.storeTasksInGit=!1);let t=e.yes||e.name&&e.description,n;if(n=e.rulesExplicitlyProvided?e.rules:[],t){m()||console.log(`SKIPPING PROMPTS - Using defaults or provided values`),e.name,e.description,e.version,e.author;let t=e.dryRun||!1,r=e.initGit===void 0?!0:e.initGit,i=e.storeTasksInGit===void 0?!0:e.storeTasksInGit;if(t)return X(`info`,`DRY RUN MODE: No files will be modified`),X(`info`,`Would initialize Task Master project`),X(`info`,`Would create/update necessary project files`),X(`info`,`${r?`Would initialize Git repository`:`Would skip Git initialization`}`),X(`info`,`${i?`Would store tasks in Git`:`Would exclude tasks from Git`}`),{dryRun:!0};let a=e.storage||`local`;await bn(!0,r,i,t,{...e,preferredLanguage:`
|
|
3503
|
+
`)}`;S.writeFileSync(t,n),X(`success`,`Updated ${t} with additional entries`)}else X(`info`,`No new content to add to ${t}`);return}if(e===`README-task-master.md`){X(`info`,`${t} already exists`);let e=C.join(C.dirname(t),`README-task-master.md`);S.writeFileSync(e,r),X(`success`,`Created ${e} (preserved original README-task-master.md)`);return}X(`debug`,`${t} already exists, skipping.`);return}S.writeFileSync(t,r),X(`info`,`Created file: ${t}`)}async function gn(e={}){m()||pn(),e.git===!0?e.initGit=!0:e.git===!1&&(e.initGit=!1),e.gitTasks===!0?e.storeTasksInGit=!0:e.gitTasks===!1&&(e.storeTasksInGit=!1);let t=e.yes||e.name&&e.description,n;if(n=e.rulesExplicitlyProvided?e.rules:[],t){m()||console.log(`SKIPPING PROMPTS - Using defaults or provided values`),e.name,e.description,e.version,e.author;let t=e.dryRun||!1,r=e.initGit===void 0?!0:e.initGit,i=e.storeTasksInGit===void 0?!0:e.storeTasksInGit;if(t)return X(`info`,`DRY RUN MODE: No files will be modified`),X(`info`,`Would initialize Task Master project`),X(`info`,`Would create/update necessary project files`),X(`info`,`${r?`Would initialize Git repository`:`Would skip Git initialization`}`),X(`info`,`${i?`Would store tasks in Git`:`Would exclude tasks from Git`}`),{dryRun:!0};let a=e.storage||`local`;await bn(!0,r,i,t,{...e,preferredLanguage:`Chinese`},n,a,null)}else{X(`debug`,`Required options not provided, proceeding with prompts.`);let t;try{let r=_n();X(`debug`,`Init started - taskmaster_id: ${r}`);let i=await yn();X(`debug`,`Storage selected: ${i} - taskmaster_id: ${r}`);let a=null;if(i===`cloud`)try{let e=f.getInstance(),t=await e.getAuthCredentials();t?(X(`success`,`Already authenticated with Hamster`),a=t):(X(`info`,`Starting authentication flow...`),console.log(x.blue(`
|
|
3504
3504
|
🔐 Authentication Required
|
|
3505
3505
|
`)),console.log(x.white(` Selecting cloud storage will open your browser for authentication.`)),console.log(x.gray(` This enables sync across devices with Hamster.
|
|
3506
|
-
`)),a=await _(e),X(`debug`,`Auth completed - taskmaster_id: ${r}`));let n=await v(e,{promptMessage:`Select an organization to continue:`});n.success||X(`warn`,n.message||`Organization selection required`)}catch(e){X(`error`,`Failed to authenticate: ${e.message}. Falling back to local storage.`),i=`local`}t=k.createInterface({input:process.stdin,output:process.stdout});let o=!0,s=!0;i===`local`?(o=e.initGit===void 0?(await $(t,x.cyan(`Initialize a Git repository in project root? (Y/n): `),e=>{let t=e.trim().toLowerCase()!==`n`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Initialize a Git repository in project root?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()!==`n`:e.initGit,s=e.storeTasksInGit===void 0?(await $(t,x.cyan(`Store tasks in Git (tasks.json and tasks/ directory)? (Y/n): `),e=>{let t=e.trim().toLowerCase()!==`n`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Store tasks in Git (tasks.json and tasks/ directory)?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()!==`n`:e.storeTasksInGit):(o=!0,s=!1);let c=!1;e.rulesExplicitlyProvided?X(`info`,`Using rule profiles provided via command line: ${n.join(`, `)}`):c=(await $(t,x.cyan(`Set up AI IDE rules for better integration? (Cursor, Windsurf, etc.) (y/N): `),e=>{let t=e.trim().toLowerCase()===`y`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Set up AI IDE rules for better integration?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()===`y`;let l=(await $(t,x.cyan(`Preferred response language (
|
|
3506
|
+
`)),a=await _(e),X(`debug`,`Auth completed - taskmaster_id: ${r}`));let n=await v(e,{promptMessage:`Select an organization to continue:`});n.success||X(`warn`,n.message||`Organization selection required`)}catch(e){X(`error`,`Failed to authenticate: ${e.message}. Falling back to local storage.`),i=`local`}t=k.createInterface({input:process.stdin,output:process.stdout});let o=!0,s=!0;i===`local`?(o=e.initGit===void 0?(await $(t,x.cyan(`Initialize a Git repository in project root? (Y/n): `),e=>{let t=e.trim().toLowerCase()!==`n`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Initialize a Git repository in project root?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()!==`n`:e.initGit,s=e.storeTasksInGit===void 0?(await $(t,x.cyan(`Store tasks in Git (tasks.json and tasks/ directory)? (Y/n): `),e=>{let t=e.trim().toLowerCase()!==`n`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Store tasks in Git (tasks.json and tasks/ directory)?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()!==`n`:e.storeTasksInGit):(o=!0,s=!1);let c=!1;e.rulesExplicitlyProvided?X(`info`,`Using rule profiles provided via command line: ${n.join(`, `)}`):c=(await $(t,x.cyan(`Set up AI IDE rules for better integration? (Cursor, Windsurf, etc.) (y/N): `),e=>{let t=e.trim().toLowerCase()===`y`,n=t?x.green(`✓`):x.red(`✗`);return x.cyan(`Set up AI IDE rules for better integration?`)+` `+n+` `+x.dim(t?`Yes`:`No`)})).trim().toLowerCase()===`y`;let l=(await $(t,x.cyan(`Preferred response language (Chinese): `))).trim()||`Chinese`;console.log(`
|
|
3507
3507
|
`+x.bold(`Taskmaster Project Settings:`)),console.log(x.dim(`─`.repeat(50))),console.log(` `+x.dim(`Storage:`.padEnd(32)),x.white(i===`cloud`?`Hamster Studio`:`Local File Storage`));let u=c?x.green(`✓`):x.dim(`✗`);if(console.log(` `+x.dim(`AI IDE rules:`.padEnd(32)),u+` `+x.dim(c?`Yes`:`No`)),console.log(` `+x.dim(`Response language:`.padEnd(32)),x.white(l)),i===`local`){let e=o?x.green(`✓`):x.dim(`✗`);console.log(` `+x.dim(`Initialize Git repository:`.padEnd(32)),e+` `+x.dim(o?`Yes`:`No`));let t=s?x.green(`✓`):x.dim(`✗`);console.log(` `+x.dim(`Store tasks in Git:`.padEnd(32)),t+` `+x.dim(s?`Yes`:`No`))}if(console.log(x.dim(`─`.repeat(50))),(await $(t,x.yellow(`
|
|
3508
3508
|
Do you want to continue with these settings? (Y/n): `))).trim().toLowerCase()===`n`){t.close(),X(`info`,`Project initialization cancelled by user`),process.exit(0);return}let d=e.dryRun||!1;if(d)return X(`info`,`DRY RUN MODE: No files will be modified`),X(`info`,`Would initialize Task Master project`),X(`info`,`Would create/update necessary project files`),X(`info`,`${o?`Would initialize Git repository`:`Would skip Git initialization`}`),X(`info`,`${s?`Would store tasks in Git`:`Would exclude tasks from Git`}`),{dryRun:!0};await bn(!0,o,s,d,{...e,shouldSetupRules:c,preferredLanguage:l},n,i,a),t.close()}catch(e){t&&t.close(),X(`error`,`Error during initialization process: ${e.message}`),process.exit(1)}}}function $(e,t,n){return new Promise(r=>{e.question(t,e=>{k.moveCursor(process.stdout,0,-1),k.cursorTo(process.stdout,0),k.clearLine(process.stdout,0),n&&process.stdout.write(n(e)+`
|
|
3509
3509
|
`),r(e)})})}function _n(){return w()}function vn(e,t,n){try{if(!S.existsSync(e)){X(`warn`,`Config file does not exist, skipping storage configuration`);return}let n=S.readFileSync(e,`utf8`),r=JSON.parse(n);r.storage||={},t===`cloud`?(r.storage.type=`api`,r.storage.apiEndpoint=process.env.TM_BASE_DOMAIN||process.env.TM_PUBLIC_BASE_DOMAIN||`https://tryhamster.com/api`,r.storage.operatingMode=`team`,X(`debug`,`Connected to Hamster Studio`)):(r.storage.type=`file`,r.storage.operatingMode=`solo`,X(`debug`,`Configured storage for local file storage`)),S.writeFileSync(e,JSON.stringify(r,null,2)),X(`debug`,`Storage configuration updated in config.json`)}catch(e){X(`error`,`Failed to update storage configuration: ${e.message}`)}}async function yn(){if(m())return`local`;try{console.log(x.bold.cyan(`You need a plan before you execute.
|
|
3510
3510
|
`));let{storageType:e}=await ae.prompt([{type:`list`,name:`storageType`,message:x.white(`How do you want to build it?
|
|
3511
3511
|
`),choices:[{name:[x.bold(`Solo (Taskmaster)`),``,x.white(` • Parse your own PRDs into structured task lists and build with any IDE or background agents`),x.white(` • Agents execute tasks with precision, no scope creep, no going off-track`),x.white(` • Tasks live in a local JSON file, everything stays in your repo`),x.white(` • Upgrade to Hamster to bring the Taskmaster experience to your team`),``].join(`
|
|
3512
3512
|
`),value:`local`,short:`Solo (Taskmaster)`},{name:[x.bold(`Together (Hamster)`),``,x.white(` • Write a brief with your team. Hamster refines it into a plan.`),x.white(` • Your team drafts, refines, and aligns on the same page before executing`),x.white(` • One brief, one plan, one source of truth for execution`),x.white(` • Access tasks on Taskmaster and execute with any AI agent`),``].join(`
|
|
3513
|
-
`),value:`cloud`,short:`Together (Hamster)`}],default:`local`,pageSize:20}]);return e}catch(e){if(e.isTtyError||e.name===`ExitPromptError`)return X(`warn`,`Storage selection cancelled, defaulting to local storage`),`local`;throw e}}async function bn(e,n,s,d,f,h=A,_=`local`,v=null){let y=process.cwd();X(`debug`,`Initializing project in ${y}`),Z(C.join(y,u)),Z(C.join(y,i)),Z(C.join(y,a)),Z(C.join(y,l)),Z(C.join(y,p)),hn(y);let b={year:new Date().getFullYear()},S=_===`cloud`?`team`:`solo`;function w(e){let t=q(e);t?J(y,t,{mode:S}):X(`warn`,`Unknown rule profile: ${e}`)}Q(`env.example`,C.join(y,c),b),Q(`config.json`,C.join(y,t),{...b});let T=C.join(y,t);dn(T)?X(`debug`,`Updated config with correct maxTokens values`):X(`debug`,`Could not update maxTokens in config`),vn(T,_,v);try{let e=L(`gitignore`,`utf8`);ve(C.join(y,o),e,s,X)}catch(e){X(`error`,`Failed to create .gitignore: ${e.message}`)}Q(`example_prd.txt`,C.join(y,r)),Q(`example_prd_rpg.txt`,C.join(y,p,`example_prd_rpg.txt`));try{n===!1?X(`info`,`Git initialization skipped due to --no-git flag.`):n===!0?g()?X(`debug`,`Existing Git repository detected – skipping git init despite --git flag.`):(X(`info`,`Initializing Git repository due to --git flag...`),D(`git init`,{cwd:y,stdio:`ignore`}),X(`success`,`Git repository initialized`)):g()?X(`debug`,`Existing Git repository detected – skipping git init.`):(X(`info`,`No Git repository detected. Initializing one in project root...`),D(`git init`,{cwd:y,stdio:`ignore`}),X(`success`,`Git repository initialized`))}catch{X(`warn`,`Git not available, skipping repository initialization`)}if(f.rulesExplicitlyProvided||f.yes){X(`info`,`Generating profile rules from command-line flags...`);for(let e of h)w(e)}e&&mn();let E={cwd:y,stdio:`inherit`};if(m()&&(E.stdio=`ignore`,X(`info`,`Running npm install silently...`)),f.shouldSetupRules&&!m()&&!d&&!f?.yes&&!f.rulesExplicitlyProvided){console.log(O(x.cyan(`Configuring Rule Profiles...`),{padding:.5,margin:{top:1,bottom:.5},borderStyle:`round`,borderColor:`cyan`,width:60})),X(`info`,`Running interactive rules setup. Please select which rule profiles to include.`);try{D(`npx task-master rules --setup`,{stdio:`inherit`,cwd:y}),X(`success`,`Rule profiles configured.`)}catch(e){X(`error`,`Failed to configure rule profiles:`,e.message),X(`warn`,`You may need to run "task-master rules --setup" manually.`)}}else m()||d||f?.yes?f.rulesExplicitlyProvided?X(`debug`,`Skipping interactive rules setup because --rules flag was used.`):X(`debug`,`Skipping interactive rules setup in non-interactive mode.`):f.shouldSetupRules||X(`debug`,`Skipping rules setup - user declined.`);if(f.preferredLanguage&&!d)try{let e=(await import(`./response-language-
|
|
3513
|
+
`),value:`cloud`,short:`Together (Hamster)`}],default:`local`,pageSize:20}]);return e}catch(e){if(e.isTtyError||e.name===`ExitPromptError`)return X(`warn`,`Storage selection cancelled, defaulting to local storage`),`local`;throw e}}async function bn(e,n,s,d,f,h=A,_=`local`,v=null){let y=process.cwd();X(`debug`,`Initializing project in ${y}`),Z(C.join(y,u)),Z(C.join(y,i)),Z(C.join(y,a)),Z(C.join(y,l)),Z(C.join(y,p)),hn(y);let b={year:new Date().getFullYear()},S=_===`cloud`?`team`:`solo`;function w(e){let t=q(e);t?J(y,t,{mode:S}):X(`warn`,`Unknown rule profile: ${e}`)}Q(`env.example`,C.join(y,c),b),Q(`config.json`,C.join(y,t),{...b});let T=C.join(y,t);dn(T)?X(`debug`,`Updated config with correct maxTokens values`):X(`debug`,`Could not update maxTokens in config`),vn(T,_,v);try{let e=L(`gitignore`,`utf8`);ve(C.join(y,o),e,s,X)}catch(e){X(`error`,`Failed to create .gitignore: ${e.message}`)}Q(`example_prd.txt`,C.join(y,r)),Q(`example_prd_rpg.txt`,C.join(y,p,`example_prd_rpg.txt`));try{n===!1?X(`info`,`Git initialization skipped due to --no-git flag.`):n===!0?g()?X(`debug`,`Existing Git repository detected – skipping git init despite --git flag.`):(X(`info`,`Initializing Git repository due to --git flag...`),D(`git init`,{cwd:y,stdio:`ignore`}),X(`success`,`Git repository initialized`)):g()?X(`debug`,`Existing Git repository detected – skipping git init.`):(X(`info`,`No Git repository detected. Initializing one in project root...`),D(`git init`,{cwd:y,stdio:`ignore`}),X(`success`,`Git repository initialized`))}catch{X(`warn`,`Git not available, skipping repository initialization`)}if(f.rulesExplicitlyProvided||f.yes){X(`info`,`Generating profile rules from command-line flags...`);for(let e of h)w(e)}e&&mn();let E={cwd:y,stdio:`inherit`};if(m()&&(E.stdio=`ignore`,X(`info`,`Running npm install silently...`)),f.shouldSetupRules&&!m()&&!d&&!f?.yes&&!f.rulesExplicitlyProvided){console.log(O(x.cyan(`Configuring Rule Profiles...`),{padding:.5,margin:{top:1,bottom:.5},borderStyle:`round`,borderColor:`cyan`,width:60})),X(`info`,`Running interactive rules setup. Please select which rule profiles to include.`);try{D(`npx task-master rules --setup`,{stdio:`inherit`,cwd:y}),X(`success`,`Rule profiles configured.`)}catch(e){X(`error`,`Failed to configure rule profiles:`,e.message),X(`warn`,`You may need to run "task-master rules --setup" manually.`)}}else m()||d||f?.yes?f.rulesExplicitlyProvided?X(`debug`,`Skipping interactive rules setup because --rules flag was used.`):X(`debug`,`Skipping interactive rules setup in non-interactive mode.`):f.shouldSetupRules||X(`debug`,`Skipping rules setup - user declined.`);if(f.preferredLanguage&&!d)try{let e=(await import(`./response-language-DOZ-CzdZ.js`)).default;e(f.preferredLanguage,{projectRoot:y,silent:!0}),X(`debug`,`Response language set to: ${f.preferredLanguage}`)}catch(e){X(`warn`,`Failed to set response language: ${e.message}`)}else m()&&!d?X(`debug`,`Skipping response language setup in silent (MCP) mode.`):d&&X(`debug`,`DRY RUN: Skipping response language setup.`);if(!m()&&!d&&!f?.yes&&_===`local`){console.log(O(x.cyan(`Configuring AI Models...`),{padding:.5,margin:{top:1,bottom:.5},borderStyle:`round`,borderColor:`cyan`,width:60})),X(`info`,`Running interactive model setup. Please select your preferred AI models.`);try{D(`npx task-master models --setup`,{stdio:`inherit`,cwd:y}),X(`success`,`AI Models configured.`)}catch(e){X(`error`,`Failed to configure AI models:`,e.message),X(`warn`,`You may need to run "task-master models --setup" manually.`)}}else _===`cloud`&&!d?console.log(O(x.green.bold(`✓ AI Models Managed by Hamster - go ham!
|
|
3514
3514
|
|
|
3515
3515
|
`)+x.white(`Hamster handles all AI model configuration for you.
|
|
3516
3516
|
`)+x.dim(`• Optimized model selection for your tasks
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import"./ai-services-unified-DRuFwL_e.js";import"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import"./sentry-DP3nr70b.js";import{A as e}from"./dependency-manager-CvhwFgTc.js";import"./response-language-DpiGaJZK.js";export{e as performResearch};
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import{t as e}from"./response-language-DpiGaJZK.js";export{e as default};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
import{B as e,X as t,f as n,q as r,yt as i}from"./config-manager-
|
|
1
|
+
import{B as e,X as t,f as n,q as r,yt as i}from"./config-manager-ClruewMP.js";function a(a,o={}){let{mcpLog:s,projectRoot:c}=o,l=(e,...t)=>{s&&typeof s[e]==`function`&&s[e](...t)},u=t(null,{projectRoot:c}),d=e(c);if(i(`debug`,`Checking for config file using findConfigPath, found: ${u}`),i(`debug`,`Checking config file using isConfigFilePresent(), exists: ${d}`),!d)return{success:!1,error:{code:`CONFIG_MISSING`,message:`The configuration file is missing. Run "task-master init" to create it.`}};if(typeof a!=`string`||a.trim()===``)return{success:!1,error:{code:`INVALID_RESPONSE_LANGUAGE`,message:`Invalid response language: ${a}. Must be a non-empty string.`}};try{let e=n(c),t=a.trim();return e.global||={},e.global.responseLanguage=t,r(e,c)?{success:!0,data:{responseLanguage:t,message:`Response language set to ${t}`}}:{success:!1,error:{code:`WRITE_ERROR`,message:`Error writing updated configuration to configuration file`}}}catch(e){return l(`error`,`Error setting response language: ${e.message}`),{success:!1,error:{code:`SET_RESPONSE_LANGUAGE_ERROR`,message:e.message}}}}var o=a;export{o as t};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
import{U as e,r as t}from"./config-manager-
|
|
1
|
+
import{U as e,r as t}from"./config-manager-ClruewMP.js";import{createHash as n}from"crypto";import*as r from"@sentry/node";let i=!1;function a(e){if(e)return n(`sha256`).update(e).digest(`hex`).substring(0,8)}function o(n={}){if(i)return;e(!0);try{if(!t(n.projectRoot)){console.log(`✓ Anonymous telemetry disabled per user preference. Set anonymousTelemetry: true in .taskmaster/config.json to re-enable.`);return}}catch{}finally{e(!1)}let a=n.dsn||`https://ce8c03ca1dd0da5b9837c6ba1b3a0f9d@o4510099843776512.ingest.us.sentry.io/4510381945585664`;if(a)try{r.init({dsn:a,environment:n.environment||process.env.NODE_ENV||`production`,integrations:[r.vercelAIIntegration({recordInputs:!0,recordOutputs:!0}),r.zodErrorsIntegration()],tracesSampleRate:n.tracesSampleRate??1,sendDefaultPii:n.sendDefaultPii??!0,debug:process.env.SENTRY_DEBUG===`true`}),i=!0,process.env.SENTRY_DEBUG===`true`&&(console.log(` DSN: ${a.substring(0,40)}...`),console.log(` Environment: ${n.environment||process.env.NODE_ENV||`production`}`),console.log(` Traces Sample Rate: ${n.tracesSampleRate??1}`))}catch(e){console.error(`Failed to initialize telemetry: ${e.message}`)}}function s(e,t={}){if(!i)return process.env.SENTRY_DEBUG===`true`&&console.log(`⚠️ Sentry not initialized, telemetry config not available`),null;let n={isEnabled:!0,recordInputs:!0,recordOutputs:!0};return e&&(n.functionId=e),Object.keys(t).length>0&&(n.metadata={},t.command&&(n.metadata.command=t.command),t.outputType&&(n.metadata.outputType=t.outputType),t.tag&&(n.metadata.tag=t.tag),t.taskId&&(n.metadata.taskId=t.taskId),t.userId&&(n.metadata.userId=t.userId),t.briefId&&(n.metadata.briefId=t.briefId),t.projectHash&&(n.metadata.projectHash=t.projectHash)),process.env.SENTRY_DEBUG===`true`&&console.log(`📊 Sentry telemetry config created:`,JSON.stringify(n,null,2)),n}export{a as n,o as r,s as t};
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import"./ai-services-unified-DRuFwL_e.js";import"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import"./sentry-DP3nr70b.js";import{ct as e,dt as t,ft as n,ht as r,lt as i,mt as a,pt as o,st as s,ut as c}from"./dependency-manager-CvhwFgTc.js";import"./response-language-DpiGaJZK.js";export{s as copyTag,e as createTag,i as createTagFromBranch,c as deleteTag,t as renameTag,n as switchCurrentTag,o as tags,a as updateBranchTagMapping,r as useTag};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
import"./ai-services-unified-
|
|
1
|
+
import"./ai-services-unified-DRuFwL_e.js";import{dt as e,wt as t}from"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import"./sentry-DP3nr70b.js";import{$ as n,A as r,C as i,D as a,E as o,I as s,J as c,K as l,L as u,M as d,N as f,O as p,P as m,Q as h,S as g,T as _,X as v,Y as y,Z as b,gt as x,j as S,k as C,q as w,w as T}from"./dependency-manager-CvhwFgTc.js";import{t as E}from"./response-language-DpiGaJZK.js";export{n as addSubtask,h as addTask,b as analyzeTaskComplexity,v as clearSubtasks,c as expandAllTasks,y as expandTask,x as findNextTask,e as findTaskById,w as isTaskDependentOn,l as listTasks,u as migrateProject,s as moveTask,m as parsePRD,r as performResearch,t as readComplexityReport,f as removeSubtask,S as removeTask,a as scopeDownTask,p as scopeUpTask,E as setResponseLanguage,_ as setTaskStatus,d as taskExists,o as updateSingleTaskStatus,T as updateSubtaskById,i as updateTaskById,g as updateTasks,C as validateStrength};
|
package/dist/task-master.js
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
import{U as e,cn as t,hn as n}from"./config-manager-
|
|
2
|
+
import{U as e,cn as t,hn as n}from"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import{r}from"./sentry-DP3nr70b.js";import{join as i}from"node:path";import a from"dotenv";const o=process.cwd(),s=t();a.config({path:i(s,`.env`)}),r({projectRoot:s}),process.env.TASKMASTER_ORIGINAL_CWD=o,process.env.DEBUG===`1`&&console.error(`DEBUG - dev.js received args:`,process.argv.slice(2));try{await n.getInstance().hasValidSession()&&e(!0)}catch{e(!1)}const{runCLI:c}=await import(`./commands-Dz0lboMw.js`);c(process.argv);export{};
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import"./ai-services-unified-DRuFwL_e.js";import"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import"./sentry-DP3nr70b.js";import{w as e}from"./dependency-manager-CvhwFgTc.js";import"./response-language-DpiGaJZK.js";export{e as default};
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import"./ai-services-unified-DRuFwL_e.js";import"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";import"./sentry-DP3nr70b.js";import{C as e}from"./dependency-manager-CvhwFgTc.js";import"./response-language-DpiGaJZK.js";export{e as default};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
import{At as e,Ct as t,Dt as n,Et as r,Mt as i,Ot as a,St as o,Tt as s,_t as c,at as l,bt as u,ct as d,dt as f,ft as p,gt as m,ht as h,it as g,jt as _,kt as v,lt as y,mt as b,nt as x,ot as S,pt as C,rt as w,st as T,ut as E,vt as D,wt as O,xt as k,yt as A}from"./config-manager-
|
|
1
|
+
import{At as e,Ct as t,Dt as n,Et as r,Mt as i,Ot as a,St as o,Tt as s,_t as c,at as l,bt as u,ct as d,dt as f,ft as p,gt as m,ht as h,it as g,jt as _,kt as v,lt as y,mt as b,nt as x,ot as S,pt as C,rt as w,st as T,ut as E,vt as D,wt as O,xt as k,yt as A}from"./config-manager-ClruewMP.js";import"./git-utils-DllbRE35.js";export{x as LOG_LEVELS,w as addComplexityToTask,g as aggregateTelemetry,l as createStateJson,S as disableSilentMode,T as enableSilentMode,d as ensureTagMetadata,y as findCycles,E as findProjectRoot,f as findTaskById,p as findTaskInComplexityReport,C as flattenTasksWithSubtasks,b as formatTaskId,h as getCurrentTag,m as getTasksForTag,c as isEmpty,D as isSilentMode,A as log,u as markMigrationForNotice,k as migrateConfigJson,o as normalizeTaskIds,t as performCompleteTagMigration,O as readComplexityReport,s as readJSON,r as resolveEnvVariable,n as resolveTag,a as taskExists,v as traverseDependencies,e as truncate,_ as withFileLockSync,i as writeJSON};
|
package/package.json
CHANGED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import{a as e,i as t,n,r,t as i}from"./ai-services-unified-CEISDqFA.js";import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import"./sentry-BU9KKY_b.js";export{i as generateObjectService,n as generateTextService,r as logAiUsage,t as streamObjectService,e as streamTextService};
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import"./ai-services-unified-CEISDqFA.js";import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import"./sentry-BU9KKY_b.js";import{A as e}from"./dependency-manager-C-vZzEg8.js";import"./response-language-CoZ1-GFh.js";export{e as performResearch};
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import{t as e}from"./response-language-CoZ1-GFh.js";export{e as default};
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import"./ai-services-unified-CEISDqFA.js";import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import"./sentry-BU9KKY_b.js";import{ct as e,dt as t,ft as n,ht as r,lt as i,mt as a,pt as o,st as s,ut as c}from"./dependency-manager-C-vZzEg8.js";import"./response-language-CoZ1-GFh.js";export{s as copyTag,e as createTag,i as createTagFromBranch,c as deleteTag,t as renameTag,n as switchCurrentTag,o as tags,a as updateBranchTagMapping,r as useTag};
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import"./ai-services-unified-CEISDqFA.js";import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import"./sentry-BU9KKY_b.js";import{w as e}from"./dependency-manager-C-vZzEg8.js";import"./response-language-CoZ1-GFh.js";export{e as default};
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
import"./ai-services-unified-CEISDqFA.js";import"./config-manager-3pTgfD7M.js";import"./git-utils-DllbRE35.js";import"./sentry-BU9KKY_b.js";import{C as e}from"./dependency-manager-C-vZzEg8.js";import"./response-language-CoZ1-GFh.js";export{e as default};
|