@hhsw2015/task-master-ai 0.43.23 → 0.43.25
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/{ai-services-unified-DtbZ97I9.js → ai-services-unified-BowF9Vy-.js} +1 -1
- package/dist/ai-services-unified-BrOxyGnK.js +1 -0
- package/dist/{commands-DQQiVMX6.js → commands-NfRQi5z9.js} +5 -5
- package/dist/{config-manager-CWpKNiYl.js → config-manager-DyVbAaPA.js} +1 -1
- package/dist/{config-manager-BuZz09-s.js → config-manager-sjIWkXoH.js} +3 -3
- package/dist/{dependency-manager-n41vxXSd.js → dependency-manager-DQ1KOp7k.js} +5 -5
- package/dist/mcp-server.js +3 -3
- package/dist/{profiles-Cboh56FD.js → profiles-DyfO-NpI.js} +82 -82
- package/dist/research-CRMEGWua.js +1 -0
- package/dist/response-language-B-J-_qEU.js +1 -0
- package/dist/{response-language-C7cKb2Ko.js → response-language-BecQLUIF.js} +1 -1
- package/dist/{sentry-BO485gI2.js → sentry-Be7jGuko.js} +1 -1
- package/dist/tag-management-DyNjp_pG.js +1 -0
- package/dist/{task-manager-DWiOU-Qz.js → task-manager-DAaLRrbY.js} +1 -1
- package/dist/task-master.js +1 -1
- package/dist/update-subtask-by-id-bImgwSPt.js +1 -0
- package/dist/update-task-by-id-DrnkDCmy.js +1 -0
- package/dist/{utils-C0PQr5j9.js → utils-5vnyX__r.js} +1 -1
- package/package.json +1 -1
- package/dist/ai-services-unified-DYnwtn3B.js +0 -1
- package/dist/research-BKRDZpbK.js +0 -1
- package/dist/response-language-Ck21x8Tz.js +0 -1
- package/dist/tag-management-D95A287N.js +0 -1
- package/dist/update-subtask-by-id-DXTuaE1V.js +0 -1
- package/dist/update-task-by-id-Dshn8T-y.js +0 -1
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import{C as e,E as t,Et as n,F as r,H as i,I as a,J as o,L as s,M as c,N as l,O as u,P as d,a as f,f as p,ht as m,j as h,l as g,m as _,o as ee,s as te,ut as v,v as y,w as b,x,y as S,yt as C}from"./config-manager-BuZz09-s.js";import{n as w,t as T}from"./sentry-BO485gI2.js";import{createRequire as E}from"node:module";import D,{promises as O}from"fs";import k,{join as A}from"path";import{homedir as j}from"os";import{execSync as M,spawn as N}from"child_process";import*as P from"ai";import{jsonrepair as F}from"jsonrepair";import{EnvHttpProxyAgent as I}from"undici";import{createAnthropic as ne}from"@ai-sdk/anthropic";import{createPerplexity as re}from"@ai-sdk/perplexity";import{createGoogleGenerativeAI as ie}from"@ai-sdk/google";import{createOpenAI as ae}from"@ai-sdk/openai";import{createXai as oe}from"@ai-sdk/xai";import{createGroq as se}from"@ai-sdk/groq";import{createOpenRouter as ce}from"@openrouter/ai-sdk-provider";import{createOllama as le}from"ollama-ai-provider-v2";import{createAmazonBedrock as ue}from"@ai-sdk/amazon-bedrock";import{fromNodeProviderChain as de}from"@aws-sdk/credential-providers";import{createAzure as fe}from"@ai-sdk/azure";import{createVertex as pe}from"@ai-sdk/google-vertex";import{createClaudeCode as me}from"ai-sdk-provider-claude-code";import{createGeminiProvider as he}from"ai-sdk-provider-gemini-cli";import{APICallError as L,LoadAPIKeyError as ge,NoSuchModelError as R}from"@ai-sdk/provider";import{generateId as z}from"@ai-sdk/provider-utils";import{parse as _e}from"jsonc-parser";import{createCodexCli as ve}from"ai-sdk-provider-codex-cli";import{createOpenAICompatible as ye}from"@ai-sdk/openai-compatible";var be=Object.defineProperty,xe=e=>{let t={};for(var n in e)be(t,n,{get:e[n],enumerable:!0});return t},B=E(import.meta.url);let V=null;var H=class e{constructor(){this._providers=new Map,this._initialized=!1}static getInstance(){return V||=new e,V}initialize(){return this._initialized||=!0,this}registerProvider(e,t,n={}){if(!e||typeof e!=`string`)throw Error(`Provider name must be a non-empty string`);if(!t)throw Error(`Provider instance is required`);if(typeof t.generateText!=`function`||typeof t.streamText!=`function`||typeof t.generateObject!=`function`)throw Error(`Provider must implement BaseAIProvider interface`);return this._providers.set(e,{instance:t,options:n,registeredAt:new Date}),this}hasProvider(e){return this._providers.has(e)}getProvider(e){let t=this._providers.get(e);return t?t.instance:null}getAllProviders(){return new Map(this._providers)}unregisterProvider(e){return this._providers.has(e)?(this._providers.delete(e),!0):!1}reset(){this._providers.clear(),this._initialized=!1}};H.getInstance().initialize();var U=H;const{JSONParseError:Se,NoObjectGeneratedError:Ce,generateObject:we,generateText:Te,streamObject:Ee,streamText:De,zodSchema:Oe}=P,W=P.jsonSchema,ke=new Set([`minimum`,`maximum`,`exclusiveMinimum`,`exclusiveMaximum`]),Ae=[`additionalProperties`,`contains`,`if`,`then`,`else`,`not`,`propertyNames`],je=[`allOf`,`anyOf`,`oneOf`,`prefixItems`],Me=[`definitions`,`$defs`,`dependentSchemas`,`patternProperties`,`properties`],Ne=e=>e?Array.isArray(e)?e.includes(`integer`):e===`integer`:!1,G=e=>{if(!e||typeof e!=`object`)return e;if(Array.isArray(e))return e.map(G);let t={...e};if(Ne(t.type))for(let e of ke)e in t&&delete t[e];for(let e of Ae)t[e]&&(t[e]=G(t[e]));for(let e of je)Array.isArray(t[e])&&(t[e]=t[e].map(G));for(let e of Me)if(t[e]&&typeof t[e]==`object`){let n={};for(let[r,i]of Object.entries(t[e]))n[r]=G(i);t[e]=n}return t.items&&=G(t.items),t},K=e=>{if(!e||typeof e!=`object`)return e;if(Array.isArray(e))return e.map(K);let t={};for(let[n,r]of Object.entries(e))t[n]=K(r);let n=t.type===`object`,r=t.properties&&typeof t.properties==`object`&&!Array.isArray(t.properties),i=Object.prototype.hasOwnProperty.call(t,`additionalProperties`),a=r?Object.keys(t.properties):[],o=Array.isArray(t.required),s=o?new Set(t.required):new Set,c=o&&a.every(e=>s.has(e))&&t.required.length===a.length;return n&&r&&!i&&(t.additionalProperties=!1),n&&r&&!c&&(t.required=a),t},q=e=>{let t=Oe(e);if(!t||typeof t!=`object`||!t.jsonSchema)return t;let n=K(G(t.jsonSchema));return typeof W==`function`?W(n,{validate:t.validate}):{...t,jsonSchema:n}};var J=class e{constructor(){if(this.constructor===e)throw Error(`BaseAIProvider cannot be instantiated directly`);this.name=this.constructor.name,this._proxyAgent=null,this.needsExplicitJsonSchema=!1,this.supportsTemperature=!0}validateAuth(e){if(!e.apiKey)throw Error(`${this.name} API key is required`)}createProxyFetch(){this._projectRoot||=v();let e=this._projectRoot;if(i(null,e))return this._proxyAgent||=new I,(e,t={})=>fetch(e,{...t,dispatcher:this._proxyAgent})}validateParams(e){if(this.validateAuth(e),!e.modelId)throw Error(`${this.name} Model ID is required`);this.validateOptionalParams(e)}validateOptionalParams(e){if(e.temperature!==void 0&&(e.temperature<0||e.temperature>1))throw Error(`Temperature must be between 0 and 1`);if(e.maxTokens!==void 0){let t=Number(e.maxTokens);if(!Number.isFinite(t)||t<=0)throw Error(`maxTokens must be a finite number greater than 0`)}}validateMessages(e){if(!e||!Array.isArray(e)||e.length===0)throw Error(`Invalid or empty messages array provided`);for(let t of e)if(!t.role||!t.content)throw Error(`Invalid message format. Each message must have role and content`)}handleError(e,t){let n=t.message||`Unknown error occurred`;throw C(`error`,`${this.name} ${e} failed: ${n}`,{error:t}),Error(`${this.name} API error during ${e}: ${n}`)}getClient(e){throw Error(`getClient must be implemented by provider`)}isRequiredApiKey(){return!0}getRequiredApiKeyName(){throw Error(`getRequiredApiKeyName must be implemented by provider`)}prepareTokenParam(e,t){return t===void 0?{}:{maxOutputTokens:Math.floor(Number(t))}}async generateText(e){try{this.validateParams(e),this.validateMessages(e.messages),C(`debug`,`Generating ${this.name} text with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.generateText`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=await Te({model:t(e.modelId),messages:e.messages,...this.prepareTokenParam(e.modelId,e.maxTokens),...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});C(`debug`,`${this.name} generateText completed successfully for model: ${e.modelId}`);let a=i.usage?.inputTokens??i.usage?.promptTokens??0,o=i.usage?.outputTokens??i.usage?.completionTokens??0,s=i.usage?.totalTokens??a+o;return{text:i.text,usage:{inputTokens:a,outputTokens:o,totalTokens:s}}}catch(e){this.handleError(`text generation`,e)}}async streamText(e){try{this.validateParams(e),this.validateMessages(e.messages),C(`debug`,`Streaming ${this.name} text with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.streamText`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=await De({model:t(e.modelId),messages:e.messages,...this.prepareTokenParam(e.modelId,e.maxTokens),...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r},...e.experimental_transform&&{experimental_transform:e.experimental_transform}});return C(`debug`,`${this.name} streamText initiated successfully for model: ${e.modelId}`),i}catch(e){this.handleError(`text streaming`,e)}}async streamObject(e){try{if(this.validateParams(e),this.validateMessages(e.messages),!e.schema)throw Error(`Schema is required for object streaming`);C(`debug`,`Streaming ${this.name} object with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.streamObject`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=q(e.schema),a=await Ee({model:t(e.modelId),messages:e.messages,schema:i,mode:e.mode||`auto`,maxOutputTokens:e.maxTokens,...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});return C(`debug`,`${this.name} streamObject initiated successfully for model: ${e.modelId}`),a}catch(e){this.handleError(`object streaming`,e)}}async generateObject(e){try{if(this.validateParams(e),this.validateMessages(e.messages),!e.schema)throw Error(`Schema is required for object generation`);if(!e.objectName)throw Error(`Object name is required for object generation`);C(`debug`,`Generating ${this.name} object ('${e.objectName}') with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.generateObject.${e.objectName}`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=q(e.schema),a=await we({model:t(e.modelId),messages:e.messages,schema:i,mode:this.needsExplicitJsonSchema?`json`:`auto`,schemaName:e.objectName,schemaDescription:`Generate a valid JSON object for ${e.objectName}`,maxTokens:e.maxTokens,...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});C(`debug`,`${this.name} generateObject completed successfully for model: ${e.modelId}`);let o=a.usage?.inputTokens??a.usage?.promptTokens??0,s=a.usage?.outputTokens??a.usage?.completionTokens??0,c=a.usage?.totalTokens??o+s;return{object:a.object,usage:{inputTokens:o,outputTokens:s,totalTokens:c}}}catch(e){if(Ce.isInstance(e)&&e.cause instanceof Se&&e.cause.text){C(`warn`,`${this.name} generated malformed JSON, attempting to repair...`);try{let t=F(e.cause.text),n=JSON.parse(t);return C(`info`,`Successfully repaired ${this.name} JSON output`),{object:n,usage:{inputTokens:e.usage?.promptTokens||e.usage?.inputTokens||0,outputTokens:e.usage?.completionTokens||e.usage?.outputTokens||0,totalTokens:e.usage?.totalTokens||0}}}catch(e){C(`error`,`Failed to repair ${this.name} JSON: ${e.message}`)}}this.handleError(`object generation`,e)}}},Pe=class extends J{constructor(){super(),this.name=`Anthropic`}getRequiredApiKeyName(){return`ANTHROPIC_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ne({apiKey:t,...n&&{baseURL:n},headers:{"anthropic-beta":`output-128k-2025-02-19`},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Fe=class extends J{constructor(){super(),this.name=`Perplexity`}getRequiredApiKeyName(){return`PERPLEXITY_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return re({apiKey:t,baseURL:n||`https://api.perplexity.ai`,...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}async generateObject(e){return super.generateObject({...e,mode:`json`})}},Ie=class extends J{constructor(){super(),this.name=`Google`}getRequiredApiKeyName(){return`GOOGLE_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ie({apiKey:t,...n&&{baseURL:n},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Le=class extends J{constructor(){super(),this.name=`OpenAI`}getRequiredApiKeyName(){return`OPENAI_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ae({apiKey:t,...n&&{baseURL:n},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Re=class extends J{constructor(){super(),this.name=`xAI`}getRequiredApiKeyName(){return`XAI_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return oe({apiKey:t,baseURL:n||`https://api.x.ai/v1`})}catch(e){this.handleError(`client initialization`,e)}}},ze=class extends J{constructor(){super(),this.name=`Groq`}getRequiredApiKeyName(){return`GROQ_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return se({apiKey:t,...n&&{baseURL:n}})}catch(e){this.handleError(`client initialization`,e)}}},Be=class extends J{constructor(){super(),this.name=`OpenRouter`}getRequiredApiKeyName(){return`OPENROUTER_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return ce({apiKey:t,...n&&{baseURL:n}})}catch(e){this.handleError(`client initialization`,e)}}},Ve=class extends J{constructor(){super(),this.name=`Ollama`}validateAuth(e){}getClient(e){try{let{baseURL:t}=e;return le({...t&&{baseURL:t}})}catch(e){this.handleError(`client initialization`,e)}}isRequiredApiKey(){return!1}getRequiredApiKeyName(){return`OLLAMA_API_KEY`}},He=class extends J{constructor(){super(),this.name=`Bedrock`}isRequiredApiKey(){return!1}getRequiredApiKeyName(){return`AWS_ACCESS_KEY_ID`}validateAuth(e){}getClient(e){try{let e=de(),t=this.createProxyFetch();return ue({credentialProvider:e,...t&&{fetch:t}})}catch(e){this.handleError(`client initialization`,e)}}},Ue=class extends J{constructor(){super(),this.name=`Azure OpenAI`}getRequiredApiKeyName(){return`AZURE_OPENAI_API_KEY`}validateAuth(e){if(!e.apiKey)throw Error(`Azure API key is required`);if(!e.baseURL)throw Error(`Azure endpoint URL is required. Set it in .taskmasterconfig global.azureBaseURL or models.[role].baseURL`)}normalizeBaseURL(e){if(!e)return e;try{let t=new URL(e),n=t.pathname.replace(/\/+$/,``);return n.endsWith(`/openai`)||(n=`${n}/openai`),t.pathname=n,t.toString()}catch{let t=e.replace(/\/+$/,``);return t.endsWith(`/openai`)?t:`${t}/openai`}}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.normalizeBaseURL(n),i=this.createProxyFetch();return fe({apiKey:t,baseURL:r,...i&&{fetch:i}})}catch(e){this.handleError(`client initialization`,e)}}},Y=class extends Error{constructor(e){super(e),this.name=`VertexAuthError`,this.code=`vertex_auth_error`}},X=class extends Error{constructor(e){super(e),this.name=`VertexConfigError`,this.code=`vertex_config_error`}},We=class extends Error{constructor(e,t){super(e),this.name=`VertexApiError`,this.code=`vertex_api_error`,this.statusCode=t}},Ge=class extends J{constructor(){super(),this.name=`Google Vertex AI`}getRequiredApiKeyName(){return`GOOGLE_API_KEY`}isRequiredApiKey(){return!1}isAuthenticationRequired(){return!0}isValidCredential(e){return e?typeof e==`string`?e.trim().length>0:typeof e==`object`:!1}validateAuth(e){let{apiKey:t,projectId:n,location:r,credentials:i}=e,a=this.isValidCredential(t),o=this.isValidCredential(i);if(!a&&!o)throw new Y(`Vertex AI requires authentication. Provide one of the following:
|
|
1
|
+
import{C as e,E as t,Et as n,F as r,H as i,I as a,J as o,L as s,M as c,N as l,O as u,P as d,a as f,f as p,ht as m,j as h,l as g,m as _,o as ee,s as te,ut as v,v as y,w as b,x,y as S,yt as C}from"./config-manager-sjIWkXoH.js";import{n as w,t as T}from"./sentry-Be7jGuko.js";import{createRequire as E}from"node:module";import D,{promises as O}from"fs";import k,{join as A}from"path";import{homedir as j}from"os";import{execSync as M,spawn as N}from"child_process";import*as P from"ai";import{jsonrepair as F}from"jsonrepair";import{EnvHttpProxyAgent as I}from"undici";import{createAnthropic as ne}from"@ai-sdk/anthropic";import{createPerplexity as re}from"@ai-sdk/perplexity";import{createGoogleGenerativeAI as ie}from"@ai-sdk/google";import{createOpenAI as ae}from"@ai-sdk/openai";import{createXai as oe}from"@ai-sdk/xai";import{createGroq as se}from"@ai-sdk/groq";import{createOpenRouter as ce}from"@openrouter/ai-sdk-provider";import{createOllama as le}from"ollama-ai-provider-v2";import{createAmazonBedrock as ue}from"@ai-sdk/amazon-bedrock";import{fromNodeProviderChain as de}from"@aws-sdk/credential-providers";import{createAzure as fe}from"@ai-sdk/azure";import{createVertex as pe}from"@ai-sdk/google-vertex";import{createClaudeCode as me}from"ai-sdk-provider-claude-code";import{createGeminiProvider as he}from"ai-sdk-provider-gemini-cli";import{APICallError as L,LoadAPIKeyError as ge,NoSuchModelError as R}from"@ai-sdk/provider";import{generateId as z}from"@ai-sdk/provider-utils";import{parse as _e}from"jsonc-parser";import{createCodexCli as ve}from"ai-sdk-provider-codex-cli";import{createOpenAICompatible as ye}from"@ai-sdk/openai-compatible";var be=Object.defineProperty,xe=e=>{let t={};for(var n in e)be(t,n,{get:e[n],enumerable:!0});return t},B=E(import.meta.url);let V=null;var H=class e{constructor(){this._providers=new Map,this._initialized=!1}static getInstance(){return V||=new e,V}initialize(){return this._initialized||=!0,this}registerProvider(e,t,n={}){if(!e||typeof e!=`string`)throw Error(`Provider name must be a non-empty string`);if(!t)throw Error(`Provider instance is required`);if(typeof t.generateText!=`function`||typeof t.streamText!=`function`||typeof t.generateObject!=`function`)throw Error(`Provider must implement BaseAIProvider interface`);return this._providers.set(e,{instance:t,options:n,registeredAt:new Date}),this}hasProvider(e){return this._providers.has(e)}getProvider(e){let t=this._providers.get(e);return t?t.instance:null}getAllProviders(){return new Map(this._providers)}unregisterProvider(e){return this._providers.has(e)?(this._providers.delete(e),!0):!1}reset(){this._providers.clear(),this._initialized=!1}};H.getInstance().initialize();var U=H;const{JSONParseError:Se,NoObjectGeneratedError:Ce,generateObject:we,generateText:Te,streamObject:Ee,streamText:De,zodSchema:Oe}=P,W=P.jsonSchema,ke=new Set([`minimum`,`maximum`,`exclusiveMinimum`,`exclusiveMaximum`]),Ae=[`additionalProperties`,`contains`,`if`,`then`,`else`,`not`,`propertyNames`],je=[`allOf`,`anyOf`,`oneOf`,`prefixItems`],Me=[`definitions`,`$defs`,`dependentSchemas`,`patternProperties`,`properties`],Ne=e=>e?Array.isArray(e)?e.includes(`integer`):e===`integer`:!1,G=e=>{if(!e||typeof e!=`object`)return e;if(Array.isArray(e))return e.map(G);let t={...e};if(Ne(t.type))for(let e of ke)e in t&&delete t[e];for(let e of Ae)t[e]&&(t[e]=G(t[e]));for(let e of je)Array.isArray(t[e])&&(t[e]=t[e].map(G));for(let e of Me)if(t[e]&&typeof t[e]==`object`){let n={};for(let[r,i]of Object.entries(t[e]))n[r]=G(i);t[e]=n}return t.items&&=G(t.items),t},K=e=>{if(!e||typeof e!=`object`)return e;if(Array.isArray(e))return e.map(K);let t={};for(let[n,r]of Object.entries(e))t[n]=K(r);let n=t.type===`object`,r=t.properties&&typeof t.properties==`object`&&!Array.isArray(t.properties),i=Object.prototype.hasOwnProperty.call(t,`additionalProperties`),a=r?Object.keys(t.properties):[],o=Array.isArray(t.required),s=o?new Set(t.required):new Set,c=o&&a.every(e=>s.has(e))&&t.required.length===a.length;return n&&r&&!i&&(t.additionalProperties=!1),n&&r&&!c&&(t.required=a),t},q=e=>{let t=Oe(e);if(!t||typeof t!=`object`||!t.jsonSchema)return t;let n=K(G(t.jsonSchema));return typeof W==`function`?W(n,{validate:t.validate}):{...t,jsonSchema:n}};var J=class e{constructor(){if(this.constructor===e)throw Error(`BaseAIProvider cannot be instantiated directly`);this.name=this.constructor.name,this._proxyAgent=null,this.needsExplicitJsonSchema=!1,this.supportsTemperature=!0}validateAuth(e){if(!e.apiKey)throw Error(`${this.name} API key is required`)}createProxyFetch(){this._projectRoot||=v();let e=this._projectRoot;if(i(null,e))return this._proxyAgent||=new I,(e,t={})=>fetch(e,{...t,dispatcher:this._proxyAgent})}validateParams(e){if(this.validateAuth(e),!e.modelId)throw Error(`${this.name} Model ID is required`);this.validateOptionalParams(e)}validateOptionalParams(e){if(e.temperature!==void 0&&(e.temperature<0||e.temperature>1))throw Error(`Temperature must be between 0 and 1`);if(e.maxTokens!==void 0){let t=Number(e.maxTokens);if(!Number.isFinite(t)||t<=0)throw Error(`maxTokens must be a finite number greater than 0`)}}validateMessages(e){if(!e||!Array.isArray(e)||e.length===0)throw Error(`Invalid or empty messages array provided`);for(let t of e)if(!t.role||!t.content)throw Error(`Invalid message format. Each message must have role and content`)}handleError(e,t){let n=t.message||`Unknown error occurred`;throw C(`error`,`${this.name} ${e} failed: ${n}`,{error:t}),Error(`${this.name} API error during ${e}: ${n}`)}getClient(e){throw Error(`getClient must be implemented by provider`)}isRequiredApiKey(){return!0}getRequiredApiKeyName(){throw Error(`getRequiredApiKeyName must be implemented by provider`)}prepareTokenParam(e,t){return t===void 0?{}:{maxOutputTokens:Math.floor(Number(t))}}async generateText(e){try{this.validateParams(e),this.validateMessages(e.messages),C(`debug`,`Generating ${this.name} text with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.generateText`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=await Te({model:t(e.modelId),messages:e.messages,...this.prepareTokenParam(e.modelId,e.maxTokens),...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});C(`debug`,`${this.name} generateText completed successfully for model: ${e.modelId}`);let a=i.usage?.inputTokens??i.usage?.promptTokens??0,o=i.usage?.outputTokens??i.usage?.completionTokens??0,s=i.usage?.totalTokens??a+o;return{text:i.text,usage:{inputTokens:a,outputTokens:o,totalTokens:s}}}catch(e){this.handleError(`text generation`,e)}}async streamText(e){try{this.validateParams(e),this.validateMessages(e.messages),C(`debug`,`Streaming ${this.name} text with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.streamText`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=await De({model:t(e.modelId),messages:e.messages,...this.prepareTokenParam(e.modelId,e.maxTokens),...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r},...e.experimental_transform&&{experimental_transform:e.experimental_transform}});return C(`debug`,`${this.name} streamText initiated successfully for model: ${e.modelId}`),i}catch(e){this.handleError(`text streaming`,e)}}async streamObject(e){try{if(this.validateParams(e),this.validateMessages(e.messages),!e.schema)throw Error(`Schema is required for object streaming`);C(`debug`,`Streaming ${this.name} object with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.streamObject`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=q(e.schema),a=await Ee({model:t(e.modelId),messages:e.messages,schema:i,mode:e.mode||`auto`,maxOutputTokens:e.maxTokens,...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});return C(`debug`,`${this.name} streamObject initiated successfully for model: ${e.modelId}`),a}catch(e){this.handleError(`object streaming`,e)}}async generateObject(e){try{if(this.validateParams(e),this.validateMessages(e.messages),!e.schema)throw Error(`Schema is required for object generation`);if(!e.objectName)throw Error(`Object name is required for object generation`);C(`debug`,`Generating ${this.name} object ('${e.objectName}') with model: ${e.modelId}`);let t=await this.getClient(e),n=e.commandName||`unknown`,r=T(`${this.name}.${e.modelId}.${n}.generateObject.${e.objectName}`,{command:n,outputType:e.outputType,tag:e.tag,projectHash:w(e.projectRoot),userId:e.userId,briefId:e.briefId}),i=q(e.schema),a=await we({model:t(e.modelId),messages:e.messages,schema:i,mode:this.needsExplicitJsonSchema?`json`:`auto`,schemaName:e.objectName,schemaDescription:`Generate a valid JSON object for ${e.objectName}`,maxTokens:e.maxTokens,...this.supportsTemperature&&e.temperature!==void 0?{temperature:e.temperature}:{},...r&&{experimental_telemetry:r}});C(`debug`,`${this.name} generateObject completed successfully for model: ${e.modelId}`);let o=a.usage?.inputTokens??a.usage?.promptTokens??0,s=a.usage?.outputTokens??a.usage?.completionTokens??0,c=a.usage?.totalTokens??o+s;return{object:a.object,usage:{inputTokens:o,outputTokens:s,totalTokens:c}}}catch(e){if(Ce.isInstance(e)&&e.cause instanceof Se&&e.cause.text){C(`warn`,`${this.name} generated malformed JSON, attempting to repair...`);try{let t=F(e.cause.text),n=JSON.parse(t);return C(`info`,`Successfully repaired ${this.name} JSON output`),{object:n,usage:{inputTokens:e.usage?.promptTokens||e.usage?.inputTokens||0,outputTokens:e.usage?.completionTokens||e.usage?.outputTokens||0,totalTokens:e.usage?.totalTokens||0}}}catch(e){C(`error`,`Failed to repair ${this.name} JSON: ${e.message}`)}}this.handleError(`object generation`,e)}}},Pe=class extends J{constructor(){super(),this.name=`Anthropic`}getRequiredApiKeyName(){return`ANTHROPIC_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ne({apiKey:t,...n&&{baseURL:n},headers:{"anthropic-beta":`output-128k-2025-02-19`},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Fe=class extends J{constructor(){super(),this.name=`Perplexity`}getRequiredApiKeyName(){return`PERPLEXITY_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return re({apiKey:t,baseURL:n||`https://api.perplexity.ai`,...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}async generateObject(e){return super.generateObject({...e,mode:`json`})}},Ie=class extends J{constructor(){super(),this.name=`Google`}getRequiredApiKeyName(){return`GOOGLE_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ie({apiKey:t,...n&&{baseURL:n},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Le=class extends J{constructor(){super(),this.name=`OpenAI`}getRequiredApiKeyName(){return`OPENAI_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.createProxyFetch();return ae({apiKey:t,...n&&{baseURL:n},...r&&{fetch:r}})}catch(e){this.handleError(`client initialization`,e)}}},Re=class extends J{constructor(){super(),this.name=`xAI`}getRequiredApiKeyName(){return`XAI_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return oe({apiKey:t,baseURL:n||`https://api.x.ai/v1`})}catch(e){this.handleError(`client initialization`,e)}}},ze=class extends J{constructor(){super(),this.name=`Groq`}getRequiredApiKeyName(){return`GROQ_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return se({apiKey:t,...n&&{baseURL:n}})}catch(e){this.handleError(`client initialization`,e)}}},Be=class extends J{constructor(){super(),this.name=`OpenRouter`}getRequiredApiKeyName(){return`OPENROUTER_API_KEY`}getClient(e){try{let{apiKey:t,baseURL:n}=e;return ce({apiKey:t,...n&&{baseURL:n}})}catch(e){this.handleError(`client initialization`,e)}}},Ve=class extends J{constructor(){super(),this.name=`Ollama`}validateAuth(e){}getClient(e){try{let{baseURL:t}=e;return le({...t&&{baseURL:t}})}catch(e){this.handleError(`client initialization`,e)}}isRequiredApiKey(){return!1}getRequiredApiKeyName(){return`OLLAMA_API_KEY`}},He=class extends J{constructor(){super(),this.name=`Bedrock`}isRequiredApiKey(){return!1}getRequiredApiKeyName(){return`AWS_ACCESS_KEY_ID`}validateAuth(e){}getClient(e){try{let e=de(),t=this.createProxyFetch();return ue({credentialProvider:e,...t&&{fetch:t}})}catch(e){this.handleError(`client initialization`,e)}}},Ue=class extends J{constructor(){super(),this.name=`Azure OpenAI`}getRequiredApiKeyName(){return`AZURE_OPENAI_API_KEY`}validateAuth(e){if(!e.apiKey)throw Error(`Azure API key is required`);if(!e.baseURL)throw Error(`Azure endpoint URL is required. Set it in .taskmasterconfig global.azureBaseURL or models.[role].baseURL`)}normalizeBaseURL(e){if(!e)return e;try{let t=new URL(e),n=t.pathname.replace(/\/+$/,``);return n.endsWith(`/openai`)||(n=`${n}/openai`),t.pathname=n,t.toString()}catch{let t=e.replace(/\/+$/,``);return t.endsWith(`/openai`)?t:`${t}/openai`}}getClient(e){try{let{apiKey:t,baseURL:n}=e,r=this.normalizeBaseURL(n),i=this.createProxyFetch();return fe({apiKey:t,baseURL:r,...i&&{fetch:i}})}catch(e){this.handleError(`client initialization`,e)}}},Y=class extends Error{constructor(e){super(e),this.name=`VertexAuthError`,this.code=`vertex_auth_error`}},X=class extends Error{constructor(e){super(e),this.name=`VertexConfigError`,this.code=`vertex_config_error`}},We=class extends Error{constructor(e,t){super(e),this.name=`VertexApiError`,this.code=`vertex_api_error`,this.statusCode=t}},Ge=class extends J{constructor(){super(),this.name=`Google Vertex AI`}getRequiredApiKeyName(){return`GOOGLE_API_KEY`}isRequiredApiKey(){return!1}isAuthenticationRequired(){return!0}isValidCredential(e){return e?typeof e==`string`?e.trim().length>0:typeof e==`object`:!1}validateAuth(e){let{apiKey:t,projectId:n,location:r,credentials:i}=e,a=this.isValidCredential(t),o=this.isValidCredential(i);if(!a&&!o)throw new Y(`Vertex AI requires authentication. Provide one of the following:
|
|
2
2
|
• GOOGLE_API_KEY environment variable (typical for API-based auth), OR
|
|
3
3
|
• GOOGLE_APPLICATION_CREDENTIALS pointing to a service account JSON file (recommended for production)`);if(!n||typeof n==`string`&&n.trim().length===0)throw new X(`Google Cloud project ID is required for Vertex AI. Set VERTEX_PROJECT_ID environment variable.`);if(!r||typeof r==`string`&&r.trim().length===0)throw new X(`Google Cloud location is required for Vertex AI. Set VERTEX_LOCATION environment variable (e.g., "us-central1").`)}getClient(e){try{let{apiKey:t,projectId:n,location:r,credentials:i,baseURL:a}=e,o=this.createProxyFetch(),s={};return t?s.googleAuthOptions={...i,apiKey:t}:i&&(s.googleAuthOptions=i),pe({...s,project:n,location:r,...a&&{baseURL:a},...o&&{fetch:o}})}catch(e){this.handleError(`client initialization`,e)}}handleError(e,t){if(C(`error`,`Vertex AI ${e} error:`,t),t.name===`VertexAuthError`||t.name===`VertexConfigError`||t.name===`VertexApiError`)throw t;if(t.response){let e=t.response.status,n=t.response.data?.error?.message||t.message;throw e===401||e===403?new Y(`Authentication failed: ${n}`):e===400?new X(`Invalid request: ${n}`):new We(`API error (${e}): ${n}`,e)}throw Error(`Vertex AI ${e} failed: ${t.message}`)}};let Ke=!1;var qe=class extends J{constructor(){super(),this.name=`Claude Code`,this.supportedModels=d(`claude-code`),this.supportedModels.length===0&&C(`warn`,`No supported models found for claude-code provider. Check supported-models.json configuration.`),this.needsExplicitJsonSchema=!0,this.supportsTemperature=!1}getRequiredApiKeyName(){return`CLAUDE_CODE_API_KEY`}isRequiredApiKey(){return!1}validateAuth(e){if(process.env.NODE_ENV!==`test`&&!Ke&&!process.env.CLAUDE_CODE_OAUTH_TOKEN)try{M(`claude --version`,{stdio:`pipe`,timeout:1e3})}catch{C(`warn`,`Claude Code CLI not detected. Install it with: npm install -g @anthropic-ai/claude-code`)}finally{Ke=!0}}getClient(e={}){try{let t=g(e.commandName)||{},n=process.env.ANTHROPIC_API_KEY,r=process.env.CLAUDE_CODE_API_KEY;try{return r?process.env.ANTHROPIC_API_KEY=r:n&&delete process.env.ANTHROPIC_API_KEY,me({defaultSettings:{systemPrompt:{type:`preset`,preset:`claude_code`},settingSources:[`user`,`project`,`local`],...t}})}finally{n?process.env.ANTHROPIC_API_KEY=n:delete process.env.ANTHROPIC_API_KEY}}catch(e){let t=String(e?.message||``);if(e?.code===`ENOENT`||/claude/i.test(t)){let t=Error(`Claude Code CLI not available. Please install Claude Code CLI first. Original error: ${e.message}`);t.cause=e,this.handleError(`Claude Code CLI initialization`,t)}else this.handleError(`client initialization`,e)}}getSupportedModels(){return this.supportedModels}isModelSupported(e){return e?this.supportedModels.includes(String(e).toLowerCase()):!1}},Je=class extends J{constructor(){super(),this.name=`Gemini CLI`,this.supportsTemperature=!1}validateAuth(e){}async getClient(e){try{let t={};return t=e.apiKey&&e.apiKey!==`gemini-cli-no-key-required`?{authType:`api-key`,apiKey:e.apiKey}:{authType:`oauth-personal`},e.baseURL&&(t.baseURL=e.baseURL),he(t)}catch(e){this.handleError(`client initialization`,e)}}getRequiredApiKeyName(){return`GEMINI_API_KEY`}isRequiredApiKey(){return!1}};function Z({message:e,code:t,exitCode:n,stderr:r,stdout:i,promptExcerpt:a,isRetryable:o=!1}){return new L({message:e,isRetryable:o,url:`grok-cli://command`,requestBodyValues:a?{prompt:a}:void 0,data:{code:t,exitCode:n,stderr:r,stdout:i,promptExcerpt:a}})}function Ye({message:e}){return new ge({message:e||`Authentication failed. Please ensure Grok CLI is properly configured with API key.`})}function Xe({message:e,promptExcerpt:t,timeoutMs:n}){return new L({message:e,isRetryable:!0,url:`grok-cli://command`,requestBodyValues:t?{prompt:t}:void 0,data:{code:`TIMEOUT`,promptExcerpt:t,timeoutMs:n}})}function Ze({message:e}){return new L({message:e||`Grok CLI is not installed or not found in PATH. Please install with: npm install -g @vibe-kit/grok-cli`,isRetryable:!1,url:`grok-cli://installation`,requestBodyValues:void 0})}function Qe(e){let t=e.trim(),n=/```(?:json)?\s*([\s\S]*?)\s*```/i.exec(t);n&&(t=n[1]);let r=/^\s*(?:const|let|var)\s+\w+\s*=\s*([\s\S]*)/i.exec(t);r&&(t=r[1],t.trim().endsWith(`;`)&&(t=t.trim().slice(0,-1)));let i=t.indexOf(`{`),a=t.indexOf(`[`);if(i===-1&&a===-1)return e;let o=a===-1?i:i===-1?a:Math.min(i,a);t=t.slice(o);let s=e=>{let t=[];try{let n=_e(e,t,{allowTrailingComma:!0});if(t.length===0)return JSON.stringify(n,null,2)}catch{}},c=s(t);if(c!==void 0)return c;let l=t[0],u=l===`{`?`}`:`]`,d=[],f=0,p=!1,m=!1;for(let e=0;e<t.length;e++){let n=t[e];if(m){m=!1;continue}if(n===`\\`){m=!0;continue}if(n===`"`&&!p){p=!0;continue}if(n===`"`&&p){p=!1;continue}p||(n===l?f++:n===u&&(f--,f===0&&d.push(e+1)))}for(let e=d.length-1;e>=0;e--){let n=s(t.slice(0,d[e]));if(n!==void 0)return n}let h=Math.max(0,t.length-1e3);for(let e=t.length-1;e>h;e--){let n=s(t.slice(0,e));if(n!==void 0)return n}return e}function $e(e){return e.map(e=>{let t=``;return typeof e.content==`string`?t=e.content:Array.isArray(e.content)?t=e.content.filter(e=>e.type===`text`).map(e=>e.text||``).join(`
|
|
4
4
|
`):e.content&&typeof e.content==`object`&&(t=e.content.text||JSON.stringify(e.content)),{role:e.role,content:t.trim()}})}function et(e){try{let t=e.trim().split(`
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
import{a as e,i as t,n,r,t as i}from"./ai-services-unified-BowF9Vy-.js";import"./config-manager-sjIWkXoH.js";import"./git-utils-DllbRE35.js";import"./sentry-Be7jGuko.js";export{i as generateObjectService,n as generateTextService,r as logAiUsage,t as streamObjectService,e as streamTextService};
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import"./ai-services-unified-
|
|
1
|
+
import"./ai-services-unified-BowF9Vy-.js";import{B as e,Bt as t,D as n,Ht as r,Kt as i,Lt as a,Nt as o,On as s,Q as c,Rt as l,Tt as u,U as d,Ut as f,Vt as p,dt as m,h,hn as g,ln as ee,m as _,mn as te,p as v,t as y,ut as ne,yt as b,z as x}from"./config-manager-sjIWkXoH.js";import"./git-utils-DllbRE35.js";import"./sentry-Be7jGuko.js";import{$ as S,C,D as w,F as T,I as E,J as re,L as ie,M as ae,N as oe,O as se,P as ce,Q as le,S as ue,St as de,X as D,Y as fe,Z as pe,_ as me,_t as he,a as ge,at as O,b as _e,bt as ve,c as ye,ct as be,d as k,dt as xe,et as A,f as Se,g as Ce,h as we,ht as Te,i as Ee,it as De,j as Oe,k as j,l as ke,m as Ae,n as je,nt as Me,o as Ne,ot as Pe,p as M,r as Fe,rt as Ie,s as Le,st as Re,t as ze,tt as Be,u as Ve,ut as He,v as N,vt as Ue,w as We,wt as Ge,xt as Ke,y as qe,yt as Je}from"./dependency-manager-DQ1KOp7k.js";import{t as Ye}from"./response-language-BecQLUIF.js";import{_ as Xe,a as Ze,c as Qe,d as $e,f as et,g as tt,h as nt,i as rt,l as P,m as it,n as at,o as F,p as ot,r as st,s as I,t as L,u as R,v as z}from"./profiles-DyfO-NpI.js";import B from"chalk";import V from"fs";import H from"path";import U from"boxen";import{Command as ct}from"commander";import W from"inquirer";const G={AUTHENTICATION:`authentication`,VALIDATION:`validation`,NETWORK:`network`,API:`api`,FILE_SYSTEM:`file_system`,TASK:`task`,PERMISSION:`permission`,TIMEOUT:`timeout`,GENERIC:`generic`},lt=[/\b[A-Za-z0-9_-]{20,}\b/g,/sk-[A-Za-z0-9]{32,}/g,/api[_-]?key[:\s=]+[^\s]+/gi,/bearer\s+[^\s]+/gi,/token[:\s=]+[^\s]+/gi,/\/Users\/[^/]+/g,/C:\\Users\\[^\\]+/g,/\/home\/[^/]+/g,/[a-zA-Z0-9._%+-]+@[a-zA-Z0-9.-]+\.[a-zA-Z]{2,}/g,/https?:\/\/[^:]+:[^@]+@/g];function K(e){if(!e||typeof e!=`string`)return e;let t=e;for(let e of lt)t=t.replace(e,`***REDACTED***`);return t}function ut(e){if(!e)return G.GENERIC;let t=(e.message||``).toLowerCase(),n=(e.code||``).toLowerCase();return t.includes(`auth`)||t.includes(`unauthorized`)||t.includes(`forbidden`)||t.includes(`api key`)||t.includes(`token`)||n.includes(`auth`)?G.AUTHENTICATION:t.includes(`invalid`)||t.includes(`validation`)||t.includes(`required`)||t.includes(`must be`)||n.includes(`validation`)?G.VALIDATION:t.includes(`network`)||t.includes(`connection`)||t.includes(`econnrefused`)||t.includes(`enotfound`)||n.includes(`network`)||n.includes(`econnrefused`)||n.includes(`enotfound`)?G.NETWORK:t.includes(`timeout`)||t.includes(`timed out`)||n.includes(`timeout`)?G.TIMEOUT:t.includes(`api`)||t.includes(`rate limit`)||t.includes(`quota`)||n.includes(`api`)?G.API:t.includes(`enoent`)||t.includes(`eacces`)||t.includes(`file`)||t.includes(`directory`)||n.includes(`enoent`)||n.includes(`eacces`)?G.FILE_SYSTEM:t.includes(`permission`)||t.includes(`access denied`)||n.includes(`eperm`)?G.PERMISSION:t.includes(`task`)||t.includes(`subtask`)?G.TASK:G.GENERIC}function dt(e,t,n){let r=[],i=(t.message||``).toLowerCase();switch(e){case G.AUTHENTICATION:i.includes(`api key`)?(r.push(`Check that your API key is correctly set in the .env file`),r.push(`Verify the API key has not expired or been revoked`)):i.includes(`token`)?(r.push(`Your authentication token may have expired`),r.push(`Try running: tm auth refresh`)):(r.push(`Verify your credentials are correctly configured`),r.push(`Check the authentication status with: tm auth status`));break;case G.VALIDATION:i.includes(`brief id`)?(r.push(`Brief IDs are case-insensitive (e.g., "ham32" = "HAM-32")`),r.push(`Check the brief ID format: usually LETTERS-NUMBERS`)):i.includes(`task id`)||i.includes(`invalid id`)?(r.push(`Task IDs should be numbers (e.g., 1, 2, 3)`),r.push(`Subtask IDs use dot notation (e.g., 1.1, 2.3)`)):(r.push(`Check that all required parameters are provided`),r.push(`Verify parameter values match expected formats`));break;case G.NETWORK:i.includes(`econnrefused`)?(r.push(`Could not connect to the server`),r.push(`Check your internet connection`),r.push(`Verify the API endpoint URL is correct`)):i.includes(`enotfound`)?(r.push(`Could not resolve the server hostname`),r.push(`Check your internet connection`)):(r.push(`Check your network connection`),r.push(`Verify firewall settings are not blocking the request`));break;case G.TIMEOUT:r.push(`The operation took too long to complete`),r.push(`Try again with a simpler request`),r.push(`Check your network speed and stability`);break;case G.API:i.includes(`rate limit`)?(r.push(`You have exceeded the API rate limit`),r.push(`Wait a few minutes before trying again`)):i.includes(`quota`)?(r.push(`You have reached your API quota`),r.push(`Check your account usage and limits`)):(r.push(`The API returned an error`),r.push(`Try again in a few moments`));break;case G.FILE_SYSTEM:i.includes(`enoent`)?(r.push(`The specified file or directory does not exist`),r.push(`Check the file path and ensure it is correct`),n.includes(`tasks.json`)&&r.push(`Initialize the project with: tm init`)):i.includes(`eacces`)?(r.push(`Permission denied to access the file`),r.push(`Check file permissions or run with appropriate privileges`)):r.push(`Check that the file or directory exists and is accessible`);break;case G.PERMISSION:r.push(`You do not have permission to perform this operation`),r.push(`Check file/directory permissions`),r.push(`You may need elevated privileges (sudo)`);break;case G.TASK:i.includes(`not found`)?(r.push(`The specified task does not exist`),r.push(`Use: tm list to see all available tasks`)):i.includes(`dependency`)||i.includes(`circular`)?(r.push(`Task dependencies form a circular reference`),r.push(`Use: tm validate-dependencies to identify issues`)):(r.push(`Check that the task ID is correct`),r.push(`Use: tm show <id> to view task details`));break;default:r.push(`Check the error message for specific details`),n&&r.push(`Operation failed while: ${n}`)}return r.slice(0,2)}function ft(e,t={}){let{context:n=``,debug:r=!1,command:i=``}=t;typeof e==`string`&&(e=Error(e)),(!e||typeof e!=`object`)&&(e=Error(`An unknown error occurred`));let a=K(e.message||`Unknown error`),o=ut(e),s=dt(o,e,n);return{type:o,message:a,context:n||`Unknown operation`,hints:s,command:i||null,code:e.code||null,stack:r?K(e.stack):null}}function q(e,t={}){let n=ft(e,t),r=B.red.bold(`✗ Error
|
|
2
2
|
|
|
3
3
|
`);r+=B.white(n.message)+`
|
|
4
4
|
|
|
@@ -106,7 +106,7 @@ Subtask update was not completed. Review the messages above for details.`))}catc
|
|
|
106
106
|
To fix this issue:`)),console.log(` 1. Run task-master list --with-subtasks to see all available subtask IDs`),console.log(` 2. Use a valid subtask ID with the --id parameter in format "parentId.subtaskId"`)):e.message.includes(`API key`)&&console.log(B.yellow(`
|
|
107
107
|
This error is related to API keys. Check your environment variables.`)),_()&&console.error(e),process.exit(1)}}),t.command(`scope-up`).description(`Increase task complexity with AI assistance`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-i, --id <ids>`,`Comma-separated task/subtask IDs to scope up (required)`).option(`-s, --strength <level>`,`Complexity increase strength: light, regular, heavy`,`regular`).option(`-p, --prompt <text>`,`Custom instructions for targeted scope adjustments`).option(`-r, --research`,`Use research AI for more informed adjustments`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{try{let t=J({tasksPath:e.file||!0,tag:e.tag}),n=t.getTasksPath(),r=t.getCurrentTag();await k(r),e.id||(console.error(B.red(`Error: --id parameter is required`)),console.log(B.yellow(`Usage example: task-master scope-up --id=1,2,3 --strength=regular`)),process.exit(1));let i=e.id.split(`,`).map(e=>{let t=parseInt(e.trim(),10);return(Number.isNaN(t)||t<=0)&&(console.error(B.red(`Error: Invalid task ID: ${e.trim()}`)),process.exit(1)),t});j(e.strength)||(console.error(B.red(`Error: Invalid strength level: ${e.strength}. Must be one of: light, regular, heavy`)),process.exit(1)),V.existsSync(n)||(console.error(B.red(`Error: Tasks file not found at path: ${n}`)),process.exit(1)),console.log(B.blue(`Scoping up ${i.length} task(s): ${i.join(`, `)}`)),console.log(B.blue(`Strength level: ${e.strength}`)),e.prompt&&console.log(B.blue(`Custom instructions: ${e.prompt}`));let a={projectRoot:t.getProjectRoot(),tag:r,commandName:`scope-up`,outputType:`cli`,research:e.research||!1},o=await se(n,i,e.strength,e.prompt||null,a,`text`);console.log(B.green(`✅ Successfully scoped up ${o.updatedTasks.length} task(s)`))}catch(e){console.error(B.red(`Error: ${e.message}`)),e.message.includes(`not found`)&&(console.log(B.yellow(`
|
|
108
108
|
To fix this issue:`)),console.log(` 1. Run task-master list to see all available task IDs`),console.log(` 2. Use valid task IDs with the --id parameter`)),_()&&console.error(e),process.exit(1)}}),t.command(`scope-down`).description(`Decrease task complexity with AI assistance`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-i, --id <ids>`,`Comma-separated task/subtask IDs to scope down (required)`).option(`-s, --strength <level>`,`Complexity decrease strength: light, regular, heavy`,`regular`).option(`-p, --prompt <text>`,`Custom instructions for targeted scope adjustments`).option(`-r, --research`,`Use research AI for more informed adjustments`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{try{let t=J({tasksPath:e.file||!0,tag:e.tag}),n=t.getTasksPath(),r=t.getCurrentTag();await k(r),e.id||(console.error(B.red(`Error: --id parameter is required`)),console.log(B.yellow(`Usage example: task-master scope-down --id=1,2,3 --strength=regular`)),process.exit(1));let i=e.id.split(`,`).map(e=>{let t=parseInt(e.trim(),10);return(Number.isNaN(t)||t<=0)&&(console.error(B.red(`Error: Invalid task ID: ${e.trim()}`)),process.exit(1)),t});j(e.strength)||(console.error(B.red(`Error: Invalid strength level: ${e.strength}. Must be one of: light, regular, heavy`)),process.exit(1)),V.existsSync(n)||(console.error(B.red(`Error: Tasks file not found at path: ${n}`)),process.exit(1)),console.log(B.blue(`Scoping down ${i.length} task(s): ${i.join(`, `)}`)),console.log(B.blue(`Strength level: ${e.strength}`)),e.prompt&&console.log(B.blue(`Custom instructions: ${e.prompt}`));let a={projectRoot:t.getProjectRoot(),tag:r,commandName:`scope-down`,outputType:`cli`,research:e.research||!1},o=await w(n,i,e.strength,e.prompt||null,a,`text`);console.log(B.green(`✅ Successfully scoped down ${o.updatedTasks.length} task(s)`))}catch(e){console.error(B.red(`Error: ${e.message}`)),e.message.includes(`not found`)&&(console.log(B.yellow(`
|
|
109
|
-
To fix this issue:`)),console.log(` 1. Run task-master list to see all available task IDs`),console.log(` 2. Use valid task IDs with the --id parameter`)),_()&&console.error(e),process.exit(1)}}),Pe(t),t.command(`expand`).description(`Expand a task into subtasks using AI`).option(`-i, --id <id>`,`ID of the task to expand`).option(`-a, --all`,`Expand all pending tasks based on complexity analysis`).option(`-n, --num <number>`,`Number of subtasks to generate (uses complexity analysis by default if available)`).option(`-r, --research`,`Enable research-backed generation (e.g., using Perplexity)`,!1).option(`-p, --prompt <text>`,`Additional context for subtask generation`).option(`-f, --force`,`Force expansion even if subtasks exist`,!1).option(`--file <file>`,`Path to the tasks file (relative to project root)`,i).option(`-cr, --complexity-report <file>`,`Path to the complexity report file (use this to specify the complexity report, not --file)`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let n={tasksPath:e.file||!0,tag:e.tag};e.complexityReport&&(n.complexityReportPath=e.complexityReport);let r=J(n),i=r.getCurrentTag();if(await k(i),e.all){console.log(B.blue(`Expanding all pending tasks...`));try{await re(r.getTasksPath(),e.num,e.research,e.prompt,e.force,{projectRoot:r.getProjectRoot(),tag:i,complexityReportPath:r.getComplexityReportPath()})}catch(e){console.error(B.red(`Error expanding all tasks: ${e.message}`)),process.exit(1)}}else if(e.id){e.id||(console.error(B.red(`Error: Task ID is required unless using --all.`)),process.exit(1)),console.log(B.blue(`Expanding task ${e.id}...`));try{await fe(r.getTasksPath(),e.id,e.num,e.research,e.prompt,{projectRoot:r.getProjectRoot(),tag:i,complexityReportPath:r.getComplexityReportPath()},e.force)}catch(t){console.error(B.red(`Error expanding task ${e.id}: ${t.message}`)),process.exit(1)}}else console.error(B.red(`Error: You must specify either a task ID (--id) or --all.`)),t.help()}),t.command(`analyze-complexity`).description(`Analyze tasks and generate expansion recommendations${B.reset(``)}`).option(`-o, --output <file>`,`Output file path for the report`).option(`-m, --model <model>`,`LLM model to use for analysis (defaults to configured model)`).option(`-t, --threshold <number>`,`Minimum complexity score to recommend expansion (1-10)`,`5`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-r, --research`,`Use configured research model for research-backed complexity analysis`).option(`-i, --id <ids>`,`Comma-separated list of specific task IDs to analyze (e.g., "1,3,5")`).option(`--from <id>`,`Starting task ID in a range to analyze`).option(`--to <id>`,`Ending task ID in a range to analyze`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let t={tasksPath:e.file||!0,tag:e.tag};e.output&&(t.complexityReportPath=e.output);let n=J(t);e.model,parseFloat(e.threshold);let r=e.research||!1,i=n.getCurrentTag();await k(i);let a=n.getComplexityReportPath();if(console.log(B.blue(`Analyzing task complexity from: ${n.getTasksPath()}`)),console.log(B.blue(`Output report will be saved to: ${a}`)),e.id)console.log(B.blue(`Analyzing specific task IDs: ${e.id}`));else if(e.from||e.to){let t=e.from?e.from:`first`,n=e.to?e.to:`last`;console.log(B.blue(`Analyzing tasks in range: ${t} to ${n}`))}r&&console.log(B.blue(`Using Perplexity AI for research-backed complexity analysis`)),await pe({...e,output:a,tag:i,projectRoot:n.getProjectRoot(),file:n.getTasksPath()})}),t.command(`research`).description(`Perform AI-powered research queries with project context`).argument(`[prompt]`,`Research prompt to investigate`).option(`--file <file>`,`Path to the tasks file`).option(`-i, --id <ids>`,`Comma-separated task/subtask IDs to include as context (e.g., "15,16.2")`).option(`-f, --files <paths>`,`Comma-separated file paths to include as context`).option(`-c, --context <text>`,`Additional custom context to include in the research prompt`).option(`-t, --tree`,`Include project file tree structure in the research context`).option(`-s, --save <file>`,`Save research results to the specified task/subtask(s)`).option(`-d, --detail <level>`,`Output detail level: low, medium, high`,`medium`).option(`--save-to <id>`,`Automatically save research results to specified task/subtask ID (e.g., "15" or "15.2")`).option(`--save-file`,`Save research results to .taskmaster/docs/research/ directory`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async(e,t)=>{let n=J({tasksPath:t.file||!0,tag:t.tag});(!e||typeof e!=`string`||e.trim().length===0)&&(console.error(B.red(`Error: Research prompt is required and cannot be empty`)),p(),process.exit(1));let r=[`low`,`medium`,`high`];t.detail&&!r.includes(t.detail.toLowerCase())&&(console.error(B.red(`Error: Detail level must be one of: ${r.join(`, `)}`)),process.exit(1));let i=[];if(t.id)try{i=t.id.split(`,`).map(e=>{let t=e.trim();if(!/^\d+(\.\d+)?$/.test(t))throw Error(`Invalid task ID format: "${t}". Expected format: "15" or "15.2"`);return t})}catch(e){console.error(B.red(`Error parsing task IDs: ${e.message}`)),process.exit(1)}let a=[];if(t.files)try{a=t.files.split(`,`).map(e=>{let t=e.trim();if(t.length===0)throw Error(`Empty file path provided`);return t})}catch(e){console.error(B.red(`Error parsing file paths: ${e.message}`)),process.exit(1)}if(t.saveTo){let e=t.saveTo.trim();e.length===0&&(console.error(B.red(`Error: Save-to ID cannot be empty`)),process.exit(1)),/^\d+(\.\d+)?$/.test(e)||(console.error(B.red(`Error: Save-to ID must be in format "15" for task or "15.2" for subtask`)),process.exit(1))}if(t.save){let e=t.save.trim();e.length===0&&(console.error(B.red(`Error: Save target cannot be empty`)),process.exit(1)),(e.includes(`..`)||e.startsWith(`/`))&&(console.error(B.red(`Error: Save path must be relative and cannot contain ".."`)),process.exit(1))}let o=n.getCurrentTag();if(await k(o),i.length>0)try{let e=u(n.getTasksPath(),n.getProjectRoot(),o);(!e||!e.tasks)&&(console.error(B.red(`Error: No valid tasks found in ${n.getTasksPath()} for tag '${o}'`)),process.exit(1))}catch(e){console.error(B.red(`Error reading tasks file: ${e.message}`)),process.exit(1)}if(a.length>0)for(let e of a){let t=H.isAbsolute(e)?e:H.join(n.getProjectRoot(),e);V.existsSync(t)||(console.error(B.red(`Error: File not found: ${e}`)),process.exit(1))}let s={prompt:e.trim(),taskIds:i,filePaths:a,customContext:t.context?t.context.trim():null,includeProjectTree:!!t.tree,saveTarget:t.save?t.save.trim():null,saveToId:t.saveTo?t.saveTo.trim():null,allowFollowUp:!0,detailLevel:t.detail?t.detail.toLowerCase():`medium`,tasksPath:n.getTasksPath(),projectRoot:n.getProjectRoot()};console.log(B.blue(`Researching: "${s.prompt}"`)),s.taskIds.length>0&&console.log(B.gray(`Task context: ${s.taskIds.join(`, `)}`)),s.filePaths.length>0&&console.log(B.gray(`File context: ${s.filePaths.join(`, `)}`)),s.customContext&&console.log(B.gray(`Custom context: ${s.customContext.substring(0,50)}${s.customContext.length>50?`...`:``}`)),s.includeProjectTree&&console.log(B.gray(`Including project file tree`)),console.log(B.gray(`Detail level: ${s.detailLevel}`));try{let{performResearch:e}=await import(`./research-
|
|
109
|
+
To fix this issue:`)),console.log(` 1. Run task-master list to see all available task IDs`),console.log(` 2. Use valid task IDs with the --id parameter`)),_()&&console.error(e),process.exit(1)}}),Pe(t),t.command(`expand`).description(`Expand a task into subtasks using AI`).option(`-i, --id <id>`,`ID of the task to expand`).option(`-a, --all`,`Expand all pending tasks based on complexity analysis`).option(`-n, --num <number>`,`Number of subtasks to generate (uses complexity analysis by default if available)`).option(`-r, --research`,`Enable research-backed generation (e.g., using Perplexity)`,!1).option(`-p, --prompt <text>`,`Additional context for subtask generation`).option(`-f, --force`,`Force expansion even if subtasks exist`,!1).option(`--file <file>`,`Path to the tasks file (relative to project root)`,i).option(`-cr, --complexity-report <file>`,`Path to the complexity report file (use this to specify the complexity report, not --file)`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let n={tasksPath:e.file||!0,tag:e.tag};e.complexityReport&&(n.complexityReportPath=e.complexityReport);let r=J(n),i=r.getCurrentTag();if(await k(i),e.all){console.log(B.blue(`Expanding all pending tasks...`));try{await re(r.getTasksPath(),e.num,e.research,e.prompt,e.force,{projectRoot:r.getProjectRoot(),tag:i,complexityReportPath:r.getComplexityReportPath()})}catch(e){console.error(B.red(`Error expanding all tasks: ${e.message}`)),process.exit(1)}}else if(e.id){e.id||(console.error(B.red(`Error: Task ID is required unless using --all.`)),process.exit(1)),console.log(B.blue(`Expanding task ${e.id}...`));try{await fe(r.getTasksPath(),e.id,e.num,e.research,e.prompt,{projectRoot:r.getProjectRoot(),tag:i,complexityReportPath:r.getComplexityReportPath()},e.force)}catch(t){console.error(B.red(`Error expanding task ${e.id}: ${t.message}`)),process.exit(1)}}else console.error(B.red(`Error: You must specify either a task ID (--id) or --all.`)),t.help()}),t.command(`analyze-complexity`).description(`Analyze tasks and generate expansion recommendations${B.reset(``)}`).option(`-o, --output <file>`,`Output file path for the report`).option(`-m, --model <model>`,`LLM model to use for analysis (defaults to configured model)`).option(`-t, --threshold <number>`,`Minimum complexity score to recommend expansion (1-10)`,`5`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-r, --research`,`Use configured research model for research-backed complexity analysis`).option(`-i, --id <ids>`,`Comma-separated list of specific task IDs to analyze (e.g., "1,3,5")`).option(`--from <id>`,`Starting task ID in a range to analyze`).option(`--to <id>`,`Ending task ID in a range to analyze`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let t={tasksPath:e.file||!0,tag:e.tag};e.output&&(t.complexityReportPath=e.output);let n=J(t);e.model,parseFloat(e.threshold);let r=e.research||!1,i=n.getCurrentTag();await k(i);let a=n.getComplexityReportPath();if(console.log(B.blue(`Analyzing task complexity from: ${n.getTasksPath()}`)),console.log(B.blue(`Output report will be saved to: ${a}`)),e.id)console.log(B.blue(`Analyzing specific task IDs: ${e.id}`));else if(e.from||e.to){let t=e.from?e.from:`first`,n=e.to?e.to:`last`;console.log(B.blue(`Analyzing tasks in range: ${t} to ${n}`))}r&&console.log(B.blue(`Using Perplexity AI for research-backed complexity analysis`)),await pe({...e,output:a,tag:i,projectRoot:n.getProjectRoot(),file:n.getTasksPath()})}),t.command(`research`).description(`Perform AI-powered research queries with project context`).argument(`[prompt]`,`Research prompt to investigate`).option(`--file <file>`,`Path to the tasks file`).option(`-i, --id <ids>`,`Comma-separated task/subtask IDs to include as context (e.g., "15,16.2")`).option(`-f, --files <paths>`,`Comma-separated file paths to include as context`).option(`-c, --context <text>`,`Additional custom context to include in the research prompt`).option(`-t, --tree`,`Include project file tree structure in the research context`).option(`-s, --save <file>`,`Save research results to the specified task/subtask(s)`).option(`-d, --detail <level>`,`Output detail level: low, medium, high`,`medium`).option(`--save-to <id>`,`Automatically save research results to specified task/subtask ID (e.g., "15" or "15.2")`).option(`--save-file`,`Save research results to .taskmaster/docs/research/ directory`).option(`--tag <tag>`,`Specify tag context for task operations`).action(async(e,t)=>{let n=J({tasksPath:t.file||!0,tag:t.tag});(!e||typeof e!=`string`||e.trim().length===0)&&(console.error(B.red(`Error: Research prompt is required and cannot be empty`)),p(),process.exit(1));let r=[`low`,`medium`,`high`];t.detail&&!r.includes(t.detail.toLowerCase())&&(console.error(B.red(`Error: Detail level must be one of: ${r.join(`, `)}`)),process.exit(1));let i=[];if(t.id)try{i=t.id.split(`,`).map(e=>{let t=e.trim();if(!/^\d+(\.\d+)?$/.test(t))throw Error(`Invalid task ID format: "${t}". Expected format: "15" or "15.2"`);return t})}catch(e){console.error(B.red(`Error parsing task IDs: ${e.message}`)),process.exit(1)}let a=[];if(t.files)try{a=t.files.split(`,`).map(e=>{let t=e.trim();if(t.length===0)throw Error(`Empty file path provided`);return t})}catch(e){console.error(B.red(`Error parsing file paths: ${e.message}`)),process.exit(1)}if(t.saveTo){let e=t.saveTo.trim();e.length===0&&(console.error(B.red(`Error: Save-to ID cannot be empty`)),process.exit(1)),/^\d+(\.\d+)?$/.test(e)||(console.error(B.red(`Error: Save-to ID must be in format "15" for task or "15.2" for subtask`)),process.exit(1))}if(t.save){let e=t.save.trim();e.length===0&&(console.error(B.red(`Error: Save target cannot be empty`)),process.exit(1)),(e.includes(`..`)||e.startsWith(`/`))&&(console.error(B.red(`Error: Save path must be relative and cannot contain ".."`)),process.exit(1))}let o=n.getCurrentTag();if(await k(o),i.length>0)try{let e=u(n.getTasksPath(),n.getProjectRoot(),o);(!e||!e.tasks)&&(console.error(B.red(`Error: No valid tasks found in ${n.getTasksPath()} for tag '${o}'`)),process.exit(1))}catch(e){console.error(B.red(`Error reading tasks file: ${e.message}`)),process.exit(1)}if(a.length>0)for(let e of a){let t=H.isAbsolute(e)?e:H.join(n.getProjectRoot(),e);V.existsSync(t)||(console.error(B.red(`Error: File not found: ${e}`)),process.exit(1))}let s={prompt:e.trim(),taskIds:i,filePaths:a,customContext:t.context?t.context.trim():null,includeProjectTree:!!t.tree,saveTarget:t.save?t.save.trim():null,saveToId:t.saveTo?t.saveTo.trim():null,allowFollowUp:!0,detailLevel:t.detail?t.detail.toLowerCase():`medium`,tasksPath:n.getTasksPath(),projectRoot:n.getProjectRoot()};console.log(B.blue(`Researching: "${s.prompt}"`)),s.taskIds.length>0&&console.log(B.gray(`Task context: ${s.taskIds.join(`, `)}`)),s.filePaths.length>0&&console.log(B.gray(`File context: ${s.filePaths.join(`, `)}`)),s.customContext&&console.log(B.gray(`Custom context: ${s.customContext.substring(0,50)}${s.customContext.length>50?`...`:``}`)),s.includeProjectTree&&console.log(B.gray(`Including project file tree`)),console.log(B.gray(`Detail level: ${s.detailLevel}`));try{let{performResearch:e}=await import(`./research-CRMEGWua.js`),n={taskIds:s.taskIds,filePaths:s.filePaths,customContext:s.customContext||``,includeProjectTree:s.includeProjectTree,detailLevel:s.detailLevel,projectRoot:s.projectRoot,saveToFile:!!t.saveFile,tag:o},r=await e(s.prompt,n,{commandName:`research`,outputType:`cli`,tag:o},`text`,s.allowFollowUp);if(s.saveToId&&!r.interactiveSaveOccurred)try{let e=s.saveToId.includes(`.`),t=`## Research Query: ${s.prompt}
|
|
110
110
|
|
|
111
111
|
**Detail Level:** ${r.detailLevel}
|
|
112
112
|
**Context Size:** ${r.contextSize} characters
|
|
@@ -114,7 +114,7 @@ To fix this issue:`)),console.log(` 1. Run task-master list to see all availabl
|
|
|
114
114
|
|
|
115
115
|
### Results
|
|
116
116
|
|
|
117
|
-
${r.result}`;if(e){let{updateSubtaskById:e}=await import(`./update-subtask-by-id-
|
|
117
|
+
${r.result}`;if(e){let{updateSubtaskById:e}=await import(`./update-subtask-by-id-bImgwSPt.js`);await e(s.tasksPath,s.saveToId,t,!1,{commandName:`research-save`,outputType:`cli`,projectRoot:s.projectRoot,tag:o},`text`),console.log(B.green(`✅ Research saved to subtask ${s.saveToId}`))}else{let e=(await import(`./update-task-by-id-DrnkDCmy.js`)).default,n=parseInt(s.saveToId,10);await e(s.tasksPath,n,t,!1,{commandName:`research-save`,outputType:`cli`,projectRoot:s.projectRoot,tag:o},`text`,!0),console.log(B.green(`✅ Research saved to task ${s.saveToId}`))}}catch(e){console.log(B.red(`❌ Error saving to task/subtask: ${e.message}`))}if(s.saveTarget){let e=`# Research Query: ${s.prompt}
|
|
118
118
|
|
|
119
119
|
**Detail Level:** ${r.detailLevel}
|
|
120
120
|
**Context Size:** ${r.contextSize} characters
|
|
@@ -229,7 +229,7 @@ ${r.result}
|
|
|
229
229
|
task-master research "How should I implement user authentication?"
|
|
230
230
|
task-master research "What's the best approach?" --id=15,23.2
|
|
231
231
|
task-master research "How does auth work?" --files=src/auth.js --tree
|
|
232
|
-
task-master research "Implementation steps?" --save-to=15.2 --detail=high`,{padding:1,borderColor:`blue`,borderStyle:`round`}))}return t.command(`remove-task`).description(`Remove one or more tasks or subtasks permanently`).option(`-i, --id <ids>`,`ID(s) of the task(s) or subtask(s) to remove (e.g., "5", "5.2", or "5,6.1,7")`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-y, --yes`,`Skip confirmation prompt`,!1).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let t=J({tasksPath:e.file||!0,tag:e.tag}),n=e.id,r=t.getCurrentTag();await k(r),n||(console.error(B.red(`Error: Task ID(s) are required`)),console.error(B.yellow(`Usage: task-master remove-task --id=<taskId1,taskId2...>`)),process.exit(1));let i=n.split(`,`).map(e=>e.trim()).filter(Boolean);i.length===0&&(console.error(B.red(`Error: No valid task IDs provided.`)),process.exit(1));try{let n=t.getTasksPath(),a=u(n,t.getProjectRoot(),r);(!a||!a.tasks)&&(console.error(B.red(`Error: No valid tasks found in ${n}`)),process.exit(1));let o=[],s=[],c=0,l=[];for(let e of i)if(!ae(a.tasks,e))s.push(e);else{let t=m(a.tasks,e).task;if(t){if(o.push({id:e,task:t}),!t.isSubtask){t.subtasks&&t.subtasks.length>0&&(c+=t.subtasks.length);let n=a.tasks.filter(t=>t.dependencies&&t.dependencies.includes(parseInt(e,10)));n.length>0&&l.push(` - Task ${e}: ${n.length} dependent tasks (${n.map(e=>e.id).join(`, `)})`)}}else s.push(`${e} (error finding details)`)}if(s.length>0&&console.warn(B.yellow(`Warning: The following task IDs were not found: ${s.join(`, `)}`)),o.length===0&&(console.log(B.blue(`No existing tasks found to remove.`)),process.exit(0)),!e.yes){console.log(),console.log(B.red.bold(`⚠️ WARNING: This will permanently delete the following ${o.length} item(s):`)),console.log(),o.forEach(({id:e,task:t})=>{t&&(t.isSubtask?(console.log(B.white(` Subtask ${e}: ${t.title||`(no title)`}`)),t.parentTask&&console.log(B.gray(` (Parent: ${t.parentTask.id} - ${t.parentTask.title||`(no title)`})`))):console.log(B.white.bold(` Task ${e}: ${t.title||`(no title)`}`)))}),c>0&&console.log(B.yellow(`⚠️ This will also delete ${c} subtasks associated with the selected main tasks!`)),l.length>0&&(console.log(B.yellow(`⚠️ Warning: Dependencies on the following tasks will be removed:`)),l.forEach(e=>console.log(B.yellow(e)))),console.log();let{confirm:e}=await W.prompt([{type:`confirm`,name:`confirm`,message:B.red.bold(`Are you sure you want to permanently delete these ${o.length} item(s)?`),default:!1}]);e||(console.log(B.blue(`Task deletion cancelled.`)),process.exit(0))}let d=qe(`Removing ${o.length} task(s)/subtask(s)...`),f=o.map(({id:e})=>e).join(`,`),p=await Oe(t.getTasksPath(),f,{projectRoot:t.getProjectRoot(),tag:r});_e(d),p.success?console.log(U(B.green(`Successfully removed ${p.removedTasks.length} task(s)/subtask(s).`)+(p.message?`\n\nDetails:\n${p.message}`:``)+(p.error?`\n\nWarnings:\n${B.yellow(p.error)}`:``),{padding:1,borderColor:`green`,borderStyle:`round`})):(console.error(U(B.red(`Operation completed with errors. Removed ${p.removedTasks.length} task(s)/subtask(s).`)+(p.message?`\n\nDetails:\n${p.message}`:``)+(p.error?`\n\nErrors:\n${B.red(p.error)}`:``),{padding:1,borderColor:`red`,borderStyle:`round`})),process.exit(1)),s.length>0&&(console.warn(B.yellow(`Note: The following IDs were not found initially and were skipped: ${s.join(`, `)}`)),p.removedTasks.length===0&&process.exit(1))}catch(e){console.error(B.red(`Error: ${e.message||`An unknown error occurred`}`)),process.exit(1)}}),t.command(`init`).description(`Initialize a new project with Task Master structure`).option(`-y, --yes`,`Skip prompts and use default values`).option(`-n, --name <name>`,`Project name`).option(`-d, --description <description>`,`Project description`).option(`-v, --version <version>`,`Project version`,`0.1.0`).option(`-a, --author <author>`,`Author name`).option(`-r, --rules <rules...>`,`List of rules to add (roo, windsurf, cursor, ...). Accepts comma or space separated values.`).option(`--skip-install`,`Skip installing dependencies`).option(`--dry-run`,`Show what would be done without making changes`).option(`--aliases`,`Add shell aliases (tm, taskmaster, hamster, ham)`).option(`--no-aliases`,`Skip shell aliases (tm, taskmaster, hamster, ham)`).option(`--git`,`Initialize Git repository`).option(`--no-git`,`Skip Git repository initialization`).option(`--git-tasks`,`Store tasks in Git`).option(`--no-git-tasks`,`No Git storage of tasks`).action(async e=>{let t=z,n=!1;if(e.rules&&Array.isArray(e.rules)){let r=e.rules.flatMap(e=>e.split(`,`)).map(e=>e.trim()).filter(Boolean);r.length>0&&(t=r,n=!0)}e.rules=t,e.rulesExplicitlyProvided=n;try{await et(e)}catch(e){console.error(B.red(`Error during initialization: ${e.message}`)),process.exit(1)}}),t.command(`models`).description(`Manage AI model configurations`).option(`--set-main <model_id>`,`Set the primary model for task generation/updates`).option(`--set-research <model_id>`,`Set the model for research-backed operations`).option(`--set-fallback <model_id>`,`Set the model to use if the primary fails`).option(`--setup`,`Run interactive setup to configure models`).option(`--openrouter`,`Allow setting a custom OpenRouter model ID (use with --set-*) `).option(`--ollama`,`Allow setting a custom Ollama model ID (use with --set-*) `).option(`--bedrock`,`Allow setting a custom Bedrock model ID (use with --set-*) `).option(`--claude-code`,`Allow setting a Claude Code model ID (use with --set-*)`).option(`--azure`,`Allow setting a custom Azure OpenAI model ID (use with --set-*) `).option(`--vertex`,`Allow setting a custom Vertex AI model ID (use with --set-*) `).option(`--gemini-cli`,`Allow setting a Gemini CLI model ID (use with --set-*)`).option(`--codex-cli`,`Allow setting a Codex CLI model ID (use with --set-*)`).option(`--lmstudio`,`Allow setting a custom LM Studio model ID (use with --set-*)`).option(`--openai-compatible`,`Allow setting a custom OpenAI-compatible model ID (use with --set-*)`).option(`--baseURL <url>`,`Custom base URL for openai-compatible, lmstudio, or ollama providers (e.g., http://localhost:8000/v1)`).addHelpText(`after`,`
|
|
232
|
+
task-master research "Implementation steps?" --save-to=15.2 --detail=high`,{padding:1,borderColor:`blue`,borderStyle:`round`}))}return t.command(`remove-task`).description(`Remove one or more tasks or subtasks permanently`).option(`-i, --id <ids>`,`ID(s) of the task(s) or subtask(s) to remove (e.g., "5", "5.2", or "5,6.1,7")`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-y, --yes`,`Skip confirmation prompt`,!1).option(`--tag <tag>`,`Specify tag context for task operations`).action(async e=>{let t=J({tasksPath:e.file||!0,tag:e.tag}),n=e.id,r=t.getCurrentTag();await k(r),n||(console.error(B.red(`Error: Task ID(s) are required`)),console.error(B.yellow(`Usage: task-master remove-task --id=<taskId1,taskId2...>`)),process.exit(1));let i=n.split(`,`).map(e=>e.trim()).filter(Boolean);i.length===0&&(console.error(B.red(`Error: No valid task IDs provided.`)),process.exit(1));try{let n=t.getTasksPath(),a=u(n,t.getProjectRoot(),r);(!a||!a.tasks)&&(console.error(B.red(`Error: No valid tasks found in ${n}`)),process.exit(1));let o=[],s=[],c=0,l=[];for(let e of i)if(!ae(a.tasks,e))s.push(e);else{let t=m(a.tasks,e).task;if(t){if(o.push({id:e,task:t}),!t.isSubtask){t.subtasks&&t.subtasks.length>0&&(c+=t.subtasks.length);let n=a.tasks.filter(t=>t.dependencies&&t.dependencies.includes(parseInt(e,10)));n.length>0&&l.push(` - Task ${e}: ${n.length} dependent tasks (${n.map(e=>e.id).join(`, `)})`)}}else s.push(`${e} (error finding details)`)}if(s.length>0&&console.warn(B.yellow(`Warning: The following task IDs were not found: ${s.join(`, `)}`)),o.length===0&&(console.log(B.blue(`No existing tasks found to remove.`)),process.exit(0)),!e.yes){console.log(),console.log(B.red.bold(`⚠️ WARNING: This will permanently delete the following ${o.length} item(s):`)),console.log(),o.forEach(({id:e,task:t})=>{t&&(t.isSubtask?(console.log(B.white(` Subtask ${e}: ${t.title||`(no title)`}`)),t.parentTask&&console.log(B.gray(` (Parent: ${t.parentTask.id} - ${t.parentTask.title||`(no title)`})`))):console.log(B.white.bold(` Task ${e}: ${t.title||`(no title)`}`)))}),c>0&&console.log(B.yellow(`⚠️ This will also delete ${c} subtasks associated with the selected main tasks!`)),l.length>0&&(console.log(B.yellow(`⚠️ Warning: Dependencies on the following tasks will be removed:`)),l.forEach(e=>console.log(B.yellow(e)))),console.log();let{confirm:e}=await W.prompt([{type:`confirm`,name:`confirm`,message:B.red.bold(`Are you sure you want to permanently delete these ${o.length} item(s)?`),default:!1}]);e||(console.log(B.blue(`Task deletion cancelled.`)),process.exit(0))}let d=qe(`Removing ${o.length} task(s)/subtask(s)...`),f=o.map(({id:e})=>e).join(`,`),p=await Oe(t.getTasksPath(),f,{projectRoot:t.getProjectRoot(),tag:r});_e(d),p.success?console.log(U(B.green(`Successfully removed ${p.removedTasks.length} task(s)/subtask(s).`)+(p.message?`\n\nDetails:\n${p.message}`:``)+(p.error?`\n\nWarnings:\n${B.yellow(p.error)}`:``),{padding:1,borderColor:`green`,borderStyle:`round`})):(console.error(U(B.red(`Operation completed with errors. Removed ${p.removedTasks.length} task(s)/subtask(s).`)+(p.message?`\n\nDetails:\n${p.message}`:``)+(p.error?`\n\nErrors:\n${B.red(p.error)}`:``),{padding:1,borderColor:`red`,borderStyle:`round`})),process.exit(1)),s.length>0&&(console.warn(B.yellow(`Note: The following IDs were not found initially and were skipped: ${s.join(`, `)}`)),p.removedTasks.length===0&&process.exit(1))}catch(e){console.error(B.red(`Error: ${e.message||`An unknown error occurred`}`)),process.exit(1)}}),t.command(`init`).description(`Initialize a new project with Task Master structure`).option(`-y, --yes`,`Skip prompts and use default values`).option(`-n, --name <name>`,`Project name`).option(`-d, --description <description>`,`Project description`).option(`-v, --version <version>`,`Project version`,`0.1.0`).option(`-a, --author <author>`,`Author name`).option(`-r, --rules <rules...>`,`List of rules to add (roo, windsurf, cursor, ...). Accepts comma or space separated values.`).option(`--skip-install`,`Skip installing dependencies`).option(`--dry-run`,`Show what would be done without making changes`).option(`--with-codex`,`Also initialize Codex longrun assets (AGENTS + Skill)`).option(`--aliases`,`Add shell aliases (tm, taskmaster, hamster, ham)`).option(`--no-aliases`,`Skip shell aliases (tm, taskmaster, hamster, ham)`).option(`--git`,`Initialize Git repository`).option(`--no-git`,`Skip Git repository initialization`).option(`--git-tasks`,`Store tasks in Git`).option(`--no-git-tasks`,`No Git storage of tasks`).action(async e=>{let t=z,n=!1;if(e.rules&&Array.isArray(e.rules)){let r=e.rules.flatMap(e=>e.split(`,`)).map(e=>e.trim()).filter(Boolean);r.length>0&&(t=r,n=!0)}e.rules=t,e.rulesExplicitlyProvided=n;try{await et(e)}catch(e){console.error(B.red(`Error during initialization: ${e.message}`)),process.exit(1)}}),t.command(`models`).description(`Manage AI model configurations`).option(`--set-main <model_id>`,`Set the primary model for task generation/updates`).option(`--set-research <model_id>`,`Set the model for research-backed operations`).option(`--set-fallback <model_id>`,`Set the model to use if the primary fails`).option(`--setup`,`Run interactive setup to configure models`).option(`--openrouter`,`Allow setting a custom OpenRouter model ID (use with --set-*) `).option(`--ollama`,`Allow setting a custom Ollama model ID (use with --set-*) `).option(`--bedrock`,`Allow setting a custom Bedrock model ID (use with --set-*) `).option(`--claude-code`,`Allow setting a Claude Code model ID (use with --set-*)`).option(`--azure`,`Allow setting a custom Azure OpenAI model ID (use with --set-*) `).option(`--vertex`,`Allow setting a custom Vertex AI model ID (use with --set-*) `).option(`--gemini-cli`,`Allow setting a Gemini CLI model ID (use with --set-*)`).option(`--codex-cli`,`Allow setting a Codex CLI model ID (use with --set-*)`).option(`--lmstudio`,`Allow setting a custom LM Studio model ID (use with --set-*)`).option(`--openai-compatible`,`Allow setting a custom OpenAI-compatible model ID (use with --set-*)`).option(`--baseURL <url>`,`Custom base URL for openai-compatible, lmstudio, or ollama providers (e.g., http://localhost:8000/v1)`).addHelpText(`after`,`
|
|
233
233
|
Examples:
|
|
234
234
|
$ task-master models # View current configuration
|
|
235
235
|
$ task-master models --set-main gpt-4o # Set main model (provider inferred)
|
|
@@ -309,7 +309,7 @@ Or specify profiles directly:
|
|
|
309
309
|
`));for(let e of t)console.log(` • ${e.displayName} ${B.gray(`(${e.markerPath})`)}`);console.log(``),e=t.map(e=>e.profileName)}else e=await I(i);if(!e||e.length===0){console.log(B.yellow(`No profiles selected. Exiting.`));return}console.log(B.blue(`Installing ${e.length} selected profile(s)...`));let{allSuccessfulProfiles:t,totalSuccess:n,totalFailed:a}=L(await F(e,i,r.mode));console.log(B.green(`\n✓ Successfully installed ${t.length} profile(s)`)),n>0&&console.log(B.gray(` ${n} files processed, ${a} failed`));return}(!t||t.length===0)&&(console.error(`Please specify at least one rule profile (e.g., windsurf, roo).`),process.exit(1));let a=t.flatMap(e=>e.split(`,`).map(e=>e.trim())).filter(Boolean);if(e===P.REMOVE){let e=!0;if(r.force||(e=Qe(i,a)?await gt(a,Ze(i)):await ht(a)),!e){console.log(B.yellow(`Aborted: No rules were removed.`));return}}let o=[],s=[];for(let t of a){if(!nt(t)){console.warn(`Rule profile for "${t}" not found. Valid profiles: ${z.join(`, `)}. Skipping.`);continue}let a=it(t);if(e===P.ADD){console.log(B.blue(`Adding rules for profile: ${t}...`));let e=ot(i,a,{mode:await n(r.mode)});console.log(B.blue(`Completed adding rules for profile: ${t}`)),s.push({profileName:t,success:e.success,failed:e.failed}),console.log(B.green(rt(t,e)))}else if(e===P.REMOVE){console.log(B.blue(`Removing rules for profile: ${t}...`));let e=tt(i,a);o.push(e),console.log(B.green(st(t,e)))}else console.error(`Unknown action. Use "${P.ADD}" or "${P.REMOVE}".`),process.exit(1)}if(e===P.ADD&&s.length>0){let{allSuccessfulProfiles:e,totalSuccess:t,totalFailed:n}=L(s);e.length>0&&(console.log(B.green(`\nSuccessfully processed profiles: ${e.join(`, `)}`)),t>0?console.log(B.green(`Total: ${t} files processed, ${n} failed.`)):console.log(B.green(`Total: ${e.length} profile(s) set up successfully.`)))}if(e===P.REMOVE&&o.length>0){let{successfulRemovals:e,skippedRemovals:t,failedRemovals:n,removalsWithNotices:r}=at(o);e.length>0&&console.log(B.green(`\nSuccessfully removed profiles for: ${e.join(`, `)}`)),t.length>0&&console.log(B.yellow(`Skipped (default or protected): ${t.join(`, `)}`)),n.length>0&&(console.log(B.red(`
|
|
310
310
|
Errors occurred:`)),n.forEach(e=>{console.log(B.red(` ${e.profileName}: ${e.error}`))})),r.length>0&&(console.log(B.cyan(`
|
|
311
311
|
Notices:`)),r.forEach(e=>{console.log(B.cyan(` ${e.profileName}: ${e.notice}`))}));let i=o.length,a=e.length,s=t.length,c=n.length;console.log(B.blue(`\nTotal: ${i} profile(s) processed - ${a} removed, ${s} skipped, ${c} failed.`))}}),t.command(`migrate`).description(`Migrate existing project to use the new .taskmaster directory structure`).option(`-f, --force`,`Force migration even if .taskmaster directory already exists`).option(`--backup`,`Create backup of old files before migration (default: false)`,!1).option(`--cleanup`,`Remove old files after successful migration (default: true)`,!0).option(`-y, --yes`,`Skip confirmation prompts`).option(`--dry-run`,`Show what would be migrated without actually moving files`).action(async e=>{try{await ie(e)}catch(e){console.error(B.red(`Error during migration:`),e.message),process.exit(1)}}),t.command(`sync-readme`).description(`Sync the current task list to README.md in the project root`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`--with-subtasks`,`Include subtasks in the README output`).option(`-s, --status <status>`,`Show only tasks matching this status (e.g., pending, done)`).option(`-t, --tag <tag>`,`Tag to use for the task list (default: master)`).action(async e=>{let t=J({tasksPath:e.file||!0,tag:e.tag}),n=e.withSubtasks||!1,r=e.status||null,i=t.getCurrentTag();console.log(B.blue(`📝 Syncing tasks to README.md${n?` (with subtasks)`:``}${r?` (status: ${r})`:``}...`)),await syncTasksToReadme(t.getProjectRoot(),{withSubtasks:n,status:r,tasksPath:t.getTasksPath(),tag:i})||(console.error(B.red(`❌ Failed to sync tasks to README.md`)),process.exit(1))}),t.command(`add-tag`).description(`[DEPRECATED] Create a new tag context for organizing tasks (use "tm tags add" instead)`).argument(`[tagName]`,`Name of the new tag to create (optional when using --from-branch)`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`--copy-from-current`,`Copy tasks from the current tag to the new tag`).option(`--copy-from <tag>`,`Copy tasks from the specified tag to the new tag`).option(`--from-branch`,`Create tag name from current git branch (ignores tagName argument)`).option(`-d, --description <text>`,`Optional description for the tag`).action(async(e,t)=>{console.warn(B.yellow(`⚠ Warning: "tm add-tag" is deprecated. Use "tm tags add" instead.`)),console.log(B.gray(` This command will be removed in a future version.
|
|
312
|
-
`));try{let n=J({tasksPath:t.file||!0}),r=n.getTasksPath();V.existsSync(r)||(console.error(B.red(`Error: Tasks file not found at path: ${r}`)),console.log(B.yellow(`Hint: Run task-master init or task-master parse-prd to create tasks.json first`)),process.exit(1)),!e&&!t.fromBranch&&(console.error(B.red(`Error: Either tagName argument or --from-branch option is required.`)),console.log(B.yellow(`Usage examples:`)),console.log(B.cyan(` task-master add-tag my-tag`)),console.log(B.cyan(` task-master add-tag --from-branch`)),process.exit(1));let i={projectRoot:n.getProjectRoot(),commandName:`add-tag`,outputType:`cli`};if(t.fromBranch){let{createTagFromBranch:e}=await import(`./tag-management-
|
|
312
|
+
`));try{let n=J({tasksPath:t.file||!0}),r=n.getTasksPath();V.existsSync(r)||(console.error(B.red(`Error: Tasks file not found at path: ${r}`)),console.log(B.yellow(`Hint: Run task-master init or task-master parse-prd to create tasks.json first`)),process.exit(1)),!e&&!t.fromBranch&&(console.error(B.red(`Error: Either tagName argument or --from-branch option is required.`)),console.log(B.yellow(`Usage examples:`)),console.log(B.cyan(` task-master add-tag my-tag`)),console.log(B.cyan(` task-master add-tag --from-branch`)),process.exit(1));let i={projectRoot:n.getProjectRoot(),commandName:`add-tag`,outputType:`cli`};if(t.fromBranch){let{createTagFromBranch:e}=await import(`./tag-management-DyNjp_pG.js`),r=await import(`./git-utils-PBP1PRVP.js`);await r.isGitRepository(i.projectRoot)||(console.error(B.red(`Error: Not in a git repository. Cannot use --from-branch option.`)),process.exit(1));let a=await r.getCurrentBranch(i.projectRoot);a||(console.error(B.red(`Error: Could not determine current git branch.`)),process.exit(1));let o={copyFromCurrent:t.copyFromCurrent||!1,copyFromTag:t.copyFrom,description:t.description||`Tag created from git branch "${a}"`};await e(n.getTasksPath(),a,o,i,`text`)}else{let r={copyFromCurrent:t.copyFromCurrent||!1,copyFromTag:t.copyFrom,description:t.description};await be(n.getTasksPath(),e,r,i,`text`)}if(t.autoSwitch){let{useTag:r}=await import(`./tag-management-DyNjp_pG.js`),a=t.fromBranch?(await import(`./git-utils-PBP1PRVP.js`)).sanitizeBranchNameForTag(await(await import(`./git-utils-PBP1PRVP.js`)).getCurrentBranch(projectRoot)):e;await r(n.getTasksPath(),a,{},i,`text`)}}catch(e){console.error(B.red(`Error creating tag: ${e.message}`)),l(),process.exit(1)}}).on(`error`,function(e){console.error(B.red(`Error: ${e.message}`)),l(),process.exit(1)}),t.command(`delete-tag`).description(`[DEPRECATED] Delete an existing tag and all its tasks (use "tm tags remove" instead)`).argument(`<tagName>`,`Name of the tag to delete`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-y, --yes`,`Skip confirmation prompts`).action(async(e,t)=>{console.warn(B.yellow(`⚠ Warning: "tm delete-tag" is deprecated. Use "tm tags remove" instead.`)),console.log(B.gray(` This command will be removed in a future version.
|
|
313
313
|
`));try{let n=J({tasksPath:t.file||!0}),r=n.getTasksPath();V.existsSync(r)||(console.error(B.red(`Error: Tasks file not found at path: ${r}`)),process.exit(1));let i={yes:t.yes||!1},a={projectRoot:n.getProjectRoot(),commandName:`delete-tag`,outputType:`cli`};await He(n.getTasksPath(),e,i,a,`text`)}catch(e){console.error(B.red(`Error deleting tag: ${e.message}`)),d(),process.exit(1)}}).on(`error`,function(e){console.error(B.red(`Error: ${e.message}`)),d(),process.exit(1)}),t.command(`use-tag`).description(`[DEPRECATED] Switch to a different tag context (use "tm tags use" instead)`).argument(`<tagName>`,`Name of the tag to switch to`).option(`-f, --file <file>`,`Path to the tasks file`,i).action(async(e,t)=>{console.warn(B.yellow(`⚠ Warning: "tm use-tag" is deprecated. Use "tm tags use" instead.`)),console.log(B.gray(` This command will be removed in a future version.
|
|
314
314
|
`));try{let n=J({tasksPath:t.file||!0}),r=n.getTasksPath();V.existsSync(r)||(console.error(B.red(`Error: Tasks file not found at path: ${r}`)),process.exit(1));let i={projectRoot:n.getProjectRoot(),commandName:`use-tag`,outputType:`cli`};await Te(n.getTasksPath(),e,{},i,`text`)}catch(e){console.error(B.red(`Error switching tag: ${e.message}`)),f(),process.exit(1)}}).on(`error`,function(e){console.error(B.red(`Error: ${e.message}`)),f(),process.exit(1)}),t.command(`rename-tag`).description(`[DEPRECATED] Rename an existing tag (use "tm tags rename" instead)`).argument(`<oldName>`,`Current name of the tag`).argument(`<newName>`,`New name for the tag`).option(`-f, --file <file>`,`Path to the tasks file`,i).action(async(e,t,n)=>{console.warn(B.yellow(`⚠ Warning: "tm rename-tag" is deprecated. Use "tm tags rename" instead.`)),console.log(B.gray(` This command will be removed in a future version.
|
|
315
315
|
`));try{let r=J({tasksPath:n.file||!0}),i=r.getTasksPath();V.existsSync(i)||(console.error(B.red(`Error: Tasks file not found at path: ${i}`)),process.exit(1));let a={projectRoot:r.getProjectRoot(),commandName:`rename-tag`,outputType:`cli`};await xe(r.getTasksPath(),e,t,{},a,`text`)}catch(e){console.error(B.red(`Error renaming tag: ${e.message}`)),process.exit(1)}}).on(`error`,function(e){console.error(B.red(`Error: ${e.message}`)),process.exit(1)}),t.command(`copy-tag`).description(`[DEPRECATED] Copy an existing tag to create a new tag with the same tasks (use "tm tags copy" instead)`).argument(`<sourceName>`,`Name of the source tag to copy from`).argument(`<targetName>`,`Name of the new tag to create`).option(`-f, --file <file>`,`Path to the tasks file`,i).option(`-d, --description <text>`,`Optional description for the new tag`).action(async(e,t,n)=>{console.warn(B.yellow(`⚠ Warning: "tm copy-tag" is deprecated. Use "tm tags copy" instead.`)),console.log(B.gray(` This command will be removed in a future version.
|
|
@@ -1 +1 @@
|
|
|
1
|
-
import{A as e,B as t,C as n,Cn as r,D as i,E as a,F as o,G as s,H as c,I as l,J as u,K as d,L as f,M as p,N as m,O as h,P as g,R as _,S as v,T as y,U as b,V as x,W as S,_ as C,a as w,b as T,bn as E,c as D,d as O,f as k,g as A,h as j,i as M,j as N,k as P,l as F,m as I,n as L,o as R,p as z,q as B,r as V,s as H,t as U,u as W,v as G,w as K,x as q,xn as J,y as Y,z as X}from"./config-manager-
|
|
1
|
+
import{A as e,B as t,C as n,Cn as r,D as i,E as a,F as o,G as s,H as c,I as l,J as u,K as d,L as f,M as p,N as m,O as h,P as g,R as _,S as v,T as y,U as b,V as x,W as S,_ as C,a as w,b as T,bn as E,c as D,d as O,f as k,g as A,h as j,i as M,j as N,k as P,l as F,m as I,n as L,o as R,p as z,q as B,r as V,s as H,t as U,u as W,v as G,w as K,x as q,xn as J,y as Y,z as X}from"./config-manager-sjIWkXoH.js";import"./git-utils-DllbRE35.js";export{E as ALL_PROVIDERS,J as CUSTOM_PROVIDERS,U as ConfigurationError,u as MODEL_MAP,r as VALIDATED_PROVIDERS,L as getAllProviders,V as getAnonymousTelemetryEnabled,M as getAvailableModels,w as getAzureBaseURL,R as getBaseUrlForRole,H as getBedrockBaseURL,D as getClaudeCodeSettings,F as getClaudeCodeSettingsForCommand,W as getCodebaseAnalysisMode,O as getCodexCliSettings,k as getCodexCliSettingsForCommand,z as getConfig,I as getDebugFlag,j as getDefaultNumTasks,A as getDefaultPriority,C as getDefaultSubtasks,G as getFallbackModelId,Y as getFallbackProvider,T as getGrokCliSettings,q as getGrokCliSettingsForCommand,v as getLogLevel,n as getMainModelId,K as getMainProvider,y as getMcpApiKeyStatus,a as getOllamaBaseURL,i as getOperatingMode,h as getParametersForRole,P as getProjectName,e as getProxyEnabled,N as getResearchModelId,p as getResearchProvider,m as getResponseLanguage,g as getSupportedModelsForProvider,o as getUserId,l as getVertexLocation,f as getVertexProjectId,_ as hasCodebaseAnalysis,X as isApiKeySet,t as isConfigFilePresent,x as isConfigWarningSuppressed,c as isProxyEnabled,b as setSuppressConfigWarnings,S as validateClaudeCodeSettings,s as validateCodexCliSettings,d as validateProvider,B as writeConfig};
|
|
@@ -257,7 +257,7 @@ This addendum defines how this upstream skill is integrated with Task Master CLI
|
|
|
257
257
|
1. Task source of truth is Taskmaster tasks data.
|
|
258
258
|
2. Execute exactly one Taskmaster task per \`codex exec\` run.
|
|
259
259
|
3. Load order for prompt context:
|
|
260
|
-
- project AGENTS.md
|
|
260
|
+
- project AGENTS.md
|
|
261
261
|
- .codex/skills/taskmaster-longrun/AGENTS.md
|
|
262
262
|
- .codex/skills/taskmaster-longrun/SKILL.md
|
|
263
263
|
4. Runtime artifacts are managed by the runner:
|
|
@@ -284,7 +284,7 @@ ${nn} {"status":"done|failed","validation":"pass|fail|unknown","summary":"<简
|
|
|
284
284
|
描述: ${e.description||``}
|
|
285
285
|
实现细节: ${e.details||``}
|
|
286
286
|
测试策略: ${e.testStrategy||``}
|
|
287
|
-
依赖: ${o}`}async ensureAgentsHook(e,n,r){let i=`${en}\n## Taskmaster Longrun Hook\nWhen implementation starts, load AGENTS first, then load @.codex/skills/taskmaster-longrun/SKILL.md, then execute one Taskmaster task per Codex run.\n${tn}`;if(!await this.fileExists(e)){await S(e,`${i}\n`,`utf-8`),n.created.push(t.relative(this.projectRoot,e));return}let a=await x(e,`utf-8`),o=a.includes(en),s=a.includes(tn);if(o&&s){n.skipped.push(t.relative(this.projectRoot,e));return}if(o!==s)throw Error(`Invalid AGENTS hook markers in ${t.relative(this.projectRoot,e)}`);if(r===`skip`){n.skipped.push(t.relative(this.projectRoot,e));return}if(r===`fail`)throw Error(`AGENTS hook missing in ${t.relative(this.projectRoot,e)}. Re-run with agentsMode=append to auto-insert.`);await S(e,`${a.trimEnd()}\n\n${i}\n`,`utf-8`),n.updated.push(t.relative(this.projectRoot,e))}async ensureSkillTemplate(e,n){let r=await this.safeRead(e);if(!r){let r=await this.loadRemoteTemplate(cn,an);await S(e,this.mergeSkillWithIntegrationAddon(r),`utf-8`),n.created.push(t.relative(this.projectRoot,e));return}if(this.looksLikeUpstreamTaskmasterSkill(r)&&this.hasSkillIntegrationAddon(r)){n.skipped.push(t.relative(this.projectRoot,e));return}let i=await this.loadRemoteTemplate(cn,an);await S(e,this.mergeSkillWithIntegrationAddon(i),`utf-8`),n.updated.push(t.relative(this.projectRoot,e))}async ensureUpstreamAgentsTemplate(e,n){let r=await this.safeRead(e);if(r&&this.looksLikeUpstreamAgents(r)){n.skipped.push(t.relative(this.projectRoot,e));return}await S(e,await this.loadRemoteTemplate(`https://raw.githubusercontent.com/lili-luo/aicoding-cookbook/refs/heads/main/skills/codex/AGENTS.md`,`# Global Agent Rules
|
|
287
|
+
依赖: ${o}`}async ensureAgentsHook(e,n,r){let i=`${en}\n## Taskmaster Longrun Hook\nWhen implementation starts, load AGENTS first, then load @.codex/skills/taskmaster-longrun/AGENTS.md, then load @.codex/skills/taskmaster-longrun/SKILL.md, then execute one Taskmaster task per Codex run.\n${tn}`;if(!await this.fileExists(e)){await S(e,`${i}\n`,`utf-8`),n.created.push(t.relative(this.projectRoot,e));return}let a=await x(e,`utf-8`),o=a.includes(en),s=a.includes(tn);if(o&&s){let r=en.replace(/[.*+?^${}()|[\]\\]/g,`\\$&`),o=tn.replace(/[.*+?^${}()|[\]\\]/g,`\\$&`),s=RegExp(`${r}[\\s\\S]*?${o}`,`m`),c=a.replace(s,i);if(c===a){n.skipped.push(t.relative(this.projectRoot,e));return}await S(e,`${c.trimEnd()}\n`,`utf-8`),n.updated.push(t.relative(this.projectRoot,e));return}if(o!==s)throw Error(`Invalid AGENTS hook markers in ${t.relative(this.projectRoot,e)}`);if(r===`skip`){n.skipped.push(t.relative(this.projectRoot,e));return}if(r===`fail`)throw Error(`AGENTS hook missing in ${t.relative(this.projectRoot,e)}. Re-run with agentsMode=append to auto-insert.`);await S(e,`${a.trimEnd()}\n\n${i}\n`,`utf-8`),n.updated.push(t.relative(this.projectRoot,e))}async ensureSkillTemplate(e,n){let r=await this.safeRead(e);if(!r){let r=await this.loadRemoteTemplate(cn,an);await S(e,this.mergeSkillWithIntegrationAddon(r),`utf-8`),n.created.push(t.relative(this.projectRoot,e));return}if(this.looksLikeUpstreamTaskmasterSkill(r)&&this.hasSkillIntegrationAddon(r)){n.skipped.push(t.relative(this.projectRoot,e));return}let i=await this.loadRemoteTemplate(cn,an);await S(e,this.mergeSkillWithIntegrationAddon(i),`utf-8`),n.updated.push(t.relative(this.projectRoot,e))}async ensureUpstreamAgentsTemplate(e,n){let r=await this.safeRead(e);if(r&&this.looksLikeUpstreamAgents(r)){n.skipped.push(t.relative(this.projectRoot,e));return}await S(e,await this.loadRemoteTemplate(`https://raw.githubusercontent.com/lili-luo/aicoding-cookbook/refs/heads/main/skills/codex/AGENTS.md`,`# Global Agent Rules
|
|
288
288
|
|
|
289
289
|
## Language
|
|
290
290
|
|
|
@@ -297,7 +297,7 @@ Default to Chinese in user-facing replies unless the user explicitly requests an
|
|
|
297
297
|
`,`utf-8`),r.includes(`!.gitignore`)||await y(n,`!.gitignore
|
|
298
298
|
`,`utf-8`)}async syncTodoAndMap(e,t,n,r){let i=await this.tasksDomain.list({tag:e}),a=[],o=[],s=1;for(let e of i.tasks){a.push(this.toCsvRow(e,String(e.id),t,e.dependencies||[])),o.push({rowId:s,taskId:String(e.id),title:e.title,dependencies:e.dependencies||[]}),s++;for(let n of e.subtasks||[]){let r=`${e.id}.${n.id}`,i=(n.dependencies||[]).map(t=>{let n=String(t);return n.includes(`.`)?n:`${e.id}.${n}`});a.push(this.toCsvRow(n,r,t,i)),o.push({rowId:s,taskId:r,title:n.title||`Subtask ${n.id}`,dependencies:i}),s++}}if(await S(n.todoCsvPath,r===`lite`?this.renderLiteCsv(a):this.renderCsv(a),`utf-8`),r===`full`){let e={generatedAt:F(),rows:o};await S(n.mapPath,JSON.stringify(e,null,2),`utf-8`)}}toCsvRow(e,t,n,r){let i=this.mapStatus(e.status||`pending`,n,t);return{taskId:t,title:e.title||`Task ${t}`,status:i,acceptanceCriteria:e.testStrategy||``,validationCommand:`echo SKIP`,completedAt:i===`DONE`?F():``,retryCount:n.attempts[t]??0,notes:n.blockedTaskIds.includes(t)?`blocked by retry limit`:``,dependencies:r}}mapStatus(e,t,n){return t.doneTaskIds.includes(n)?`DONE`:t.blockedTaskIds.includes(n)?`FAILED`:e===`done`||e===`completed`?`DONE`:e===`in-progress`?`IN_PROGRESS`:e===`blocked`||e===`cancelled`||e===`deferred`?`FAILED`:`TODO`}renderCsv(e){return`id,task,status,acceptance_criteria,validation_command,completed_at,retry_count,notes\n${e.map((e,t)=>[t+1,I(`[${e.taskId}] ${e.title}`),e.status,I(e.acceptanceCriteria),I(e.validationCommand),I(e.completedAt),e.retryCount,I(e.notes)].join(`,`)).join(`
|
|
299
299
|
`)}\n`}renderLiteCsv(e){return`id,task,status,completed_at,notes\n${e.map((e,t)=>[t+1,I(`[${e.taskId}] ${e.title}`),this.mapStatusLite(e.status),I(e.completedAt),I(e.notes)].join(`,`)).join(`
|
|
300
|
-
`)}\n`}mapStatusLite(e){return e===`DONE`?`DONE`:`TODO`}async appendLedger(e,t){await y(e,`${JSON.stringify(t)}\n`,`utf-8`)}async loadCheckpoint(e){if(!await this.fileExists(e))return{updatedAt:F(),attempts:{},doneTaskIds:[],blockedTaskIds:[]};let t=await x(e,`utf-8`),n=JSON.parse(t);return{updatedAt:n.updatedAt||F(),attempts:n.attempts||{},doneTaskIds:n.doneTaskIds||[],blockedTaskIds:n.blockedTaskIds||[],lastTaskId:n.lastTaskId}}async saveCheckpoint(e,t){t.updatedAt=F(),await S(e,JSON.stringify(t,null,2),`utf-8`)}async fileExists(e){try{return await ie(e),!0}catch{return!1}}pushUnique(e,t){return e.includes(t)?e:[...e,t]}resolveDefaultAgentsPath(){let e=t.join(this.projectRoot,`AGENTS.md`);if(g(e))return e;let n=t.join(this.projectRoot,`agent.md`);return g(n)?n:e}},hn=class{service=null;tasksDomain=null;projectRoot;constructor(e){this.projectRoot=e.getProjectRoot()}setTasksDomain(e){this.tasksDomain=e,this.service=new mn(this.projectRoot,this.tasksDomain)}async initAssets(e={}){return this.getService().initAssets(e)}async run(e={}){return this.getService().run(e)}getService(){if(!this.tasksDomain)throw Error(`SkillRunDomain is not initialized with TasksDomain`);return this.service||=new mn(this.projectRoot,this.tasksDomain),this.service}},gn=class{async findBrief(e,t){return e.find(e=>this.matches(e,t))}matches(e,t){let n=e.document?.title||``;return!!(n.toLowerCase()===t.toLowerCase()||n.toLowerCase().includes(t.toLowerCase())||e.id===t||e.id.toLowerCase()===t.toLowerCase()||e.id.slice(-8).toLowerCase()===t.toLowerCase())}async getTagsWithStats(e,t,n,r){let i=await Promise.all(e.map(async e=>{try{let r=await n.getTasks(e.id,{}),i={},a=0,o={totalSubtasks:0,subtasksByStatus:{}};return r.forEach(e=>{let t=e.status||`pending`;i[t]=(i[t]||0)+1,t===`done`&&a++,e.subtasks&&e.subtasks.length>0&&(o.totalSubtasks+=e.subtasks.length,e.subtasks.forEach(e=>{let t=e.status||`pending`;o.subtasksByStatus[t]=(o.subtasksByStatus[t]||0)+1}))}),{name:e.document?.title||e.document?.document_name||e.id,isCurrent:t===e.id,taskCount:r.length,completedTasks:a,statusBreakdown:i,subtaskCounts:o.totalSubtasks>0?o:void 0,created:e.createdAt,description:e.document?.description,status:e.status,briefId:e.id,updatedAt:e.updatedAt}}catch(n){return console.warn(`Failed to get tasks for brief ${e.id}:`,n),{name:e.document?.title||e.document?.document_name||e.id,isCurrent:t===e.id,taskCount:0,completedTasks:0,statusBreakdown:{},created:e.createdAt,description:e.document?.description,status:e.status,briefId:e.id,updatedAt:e.updatedAt}}})),a={delivering:1,aligned:2,refining:3,draft:4,delivered:5,done:6,archived:7},o=i.sort((e,t)=>{let n=(e.status||``).toLowerCase(),r=(t.status||``).toLowerCase(),i=a[n]??999,o=a[r]??999;if(i!==o)return i-o;let s=e.updatedAt?new Date(e.updatedAt).getTime():0;return(t.updatedAt?new Date(t.updatedAt).getTime():0)-s}),s=e.find(e=>e.id===t);return{tags:o,currentTag:s&&(s.document?.title||s.document?.document_name)||null,totalTags:o.length}}validateBriefFound(e,t){if(!e)throw new T(`Brief "${t}" not found in organization`,w.NOT_FOUND)}},_n=class{static parse(e){let t=e?.trim()??``;if(!t)return{orgSlug:null,briefId:null};let n=this.parseAsUrl(t),r=n?n.pathname:t.includes(`/`)?t:null;return r?this.parsePathComponents(r,n):{orgSlug:null,briefId:t}}static extractOrgSlug(e){return this.parse(e).orgSlug}static extractBriefId(e){return this.parse(e).briefId||e.trim()}static parseAsUrl(e){try{return new URL(e)}catch{}try{return new URL(`https://${e}`)}catch{}return null}static parsePathComponents(e,t){let n=e.split(`/`).filter(Boolean),r=n.lastIndexOf(`briefs`),i=null,a=null;if(r>0&&(i=n[r-1]||null),t){let e=t.searchParams.get(`id`)||t.searchParams.get(`briefId`);e&&(a=e)}return!a&&r>=0&&n.length>r+1&&(a=n[r+1]),!a&&n.length>0&&!(r>=0&&r===n.length-1)&&(a=n[n.length-1]),{orgSlug:i,briefId:a}}static validate(e,t={}){if(t.requireOrg&&!e.orgSlug)throw new T(`Organization slug could not be extracted from input`,w.VALIDATION_ERROR);if(t.requireBrief&&!e.briefId)throw new T(`Brief identifier could not be extracted from input`,w.VALIDATION_ERROR)}},vn=class{briefService;authManager;constructor(){this.briefService=new gn,this.authManager=M.getInstance()}async resolveBrief(e,t){let n=_n.parse(e),r=n.briefId||e.trim(),i=t;if(!i&&n.orgSlug)try{let e=(await this.authManager.getOrganizations()).find(e=>e.slug?.toLowerCase()===n.orgSlug?.toLowerCase()||e.name.toLowerCase()===n.orgSlug?.toLowerCase());e&&(i=e.id)}catch{}if(i||=this.authManager.getContext()?.orgId,!i)throw new T(`No organization selected. Run "tm context org" first.`,w.CONFIG_ERROR);let a=await this.authManager.getBriefs(i),o=await this.briefService.findBrief(a,r);return this.briefService.validateBriefFound(o,r),o}async switchBrief(e){let t=await this.resolveBrief(e);await this.authManager.updateContext({briefId:t.id,briefName:t.document?.title||`Brief ${t.id.slice(-8)}`,briefStatus:t.status,briefUpdatedAt:t.updatedAt})}async getBriefsWithStats(e,t){let n=this.authManager.getContext();if(!n?.orgId)throw new T(`No organization context available`,w.MISSING_CONFIGURATION,{operation:`getBriefsWithStats`,userMessage:`No organization selected. Please authenticate first using: tm auth login`});let r=await this.authManager.getBriefs(n.orgId);return this.briefService.getTagsWithStats(r,n.briefId,e,t)}},yn=class{repository;projectId;apiClient;authManager;logger=D(`TaskExpansionService`);constructor(e,t,n,r){this.repository=e,this.projectId=t,this.apiClient=n,this.authManager=r}async expandTask(e,t){try{let n=this.authManager.ensureBriefSelected(`expandTask`),r=await this.repository.getTask(this.projectId,e);if(!r)throw new T(`Task ${e} not found`,w.TASK_NOT_FOUND,{operation:`expandTask`,taskId:e,userMessage:`Task ${e} isn't available in the current project.`});let i=await this.repository.getBrief(n.briefId),a={briefContext:{title:i?.name||n.briefName||n.briefId,description:i?.description||void 0,status:i?.status||`active`},allTasks:await this.repository.getTasks(this.projectId),existingSubtasks:r.subtasks||[],enrichedContext:t?.additionalContext},o=new URLSearchParams;if(t?.numSubtasks!==void 0&&o.set(`numSubtasks`,t.numSubtasks.toString()),t?.useResearch!==void 0&&o.set(`useResearch`,t.useResearch.toString()),t?.force!==void 0&&o.set(`force`,t.force.toString()),!r.databaseId)throw new T(`Task ${e} is missing a database ID. Task expansion requires tasks to be synced with the remote database.`,w.VALIDATION_ERROR,{operation:`expandTask`,taskId:e,userMessage:`This task has not been synced with the remote database. Please ensure the task is saved remotely before attempting expansion.`});if(!C.uuid().safeParse(r.databaseId).success)throw new T(`Task ${e} has an invalid database ID format: ${r.databaseId}`,w.VALIDATION_ERROR,{operation:`expandTask`,taskId:e,databaseId:r.databaseId,userMessage:`The task database ID is not in valid UUID format. This may indicate data corruption.`});let s=r.databaseId,c=`/ai/api/v1/tasks/${s}/subtasks/generate${o.toString()?`?${o.toString()}`:``}`,l=await this.apiClient.post(c,a),u=`${process.env.TM_BASE_DOMAIN||process.env.TM_PUBLIC_BASE_DOMAIN||`http://localhost:8080`}/home/hamster/briefs/${n.briefId}/task/${s}`;return this.logger.info(`✓ Task expansion queued for ${e}`),this.logger.info(` Job ID: ${l.jobId}`),this.logger.info(` ${l.message}`),this.logger.info(` View task: ${u}`),{...l,taskLink:u}}catch(n){throw n instanceof T?n.withContext({operation:`expandTask`,taskId:e,numSubtasks:t?.numSubtasks,useResearch:t?.useResearch}):new T(n instanceof Error?n.message:String(n),w.STORAGE_ERROR,{operation:`expandTask`,taskId:e,numSubtasks:t?.numSubtasks,useResearch:t?.useResearch},n)}}},bn=class{repository;projectId;apiClient;authManager;logger=D(`TaskRetrievalService`);constructor(e,t,n,r){this.repository=e,this.projectId=t,this.apiClient=n,this.authManager=r}async getTask(e){try{this.authManager.ensureBriefSelected(`getTask`);let t=await this.repository.getTask(this.projectId,e);if(!t)throw new T(`Task ${e} not found`,w.TASK_NOT_FOUND,{operation:`getTask`,taskId:e,userMessage:`Task ${e} isn't available in the current project.`});try{let e=`/ai/api/v1/tasks/${t.id}`,n=await this.apiClient.get(e);n.document?.content&&(t.details=n.document.content)}catch(t){this.logger.debug(`Could not fetch document content for task ${e}: ${t}`)}return this.logger.info(`✓ Retrieved task ${e}`),t.details&&this.logger.debug(` Document content available (${t.details.length} chars)`),t}catch(t){throw t instanceof T?t.withContext({operation:`getTask`,taskId:e}):new T(t instanceof Error?t.message:String(t),w.STORAGE_ERROR,{operation:`getTask`,taskId:e},t)}}},xn=class{static mapDatabaseTasksToTasks(e,t){if(!e||e.length===0)return[];let n=t instanceof Map?t:this.groupDependenciesByTaskId(t),r=e.filter(e=>!e.parent_task_id),i=this.groupSubtasksByParentId(e);return r.map(e=>this.mapDatabaseTaskToTask(e,i.get(e.id)||[],n))}static mapDatabaseTaskToTask(e,t,n){let r=t.map((t,r)=>{let i=this.extractImplementationMetadata(t.metadata);return{id:t.display_id||String(r+1),parentId:e.id,title:t.title,description:t.description||``,status:this.mapStatus(t.status),priority:this.mapPriority(t.priority),dependencies:n.get(t.id)||[],details:this.extractMetadataField(t.metadata,`details`,``),testStrategy:this.extractMetadataField(t.metadata,`testStrategy`,``),createdAt:t.created_at,updatedAt:t.updated_at,assignee:t.assignee_id||void 0,complexity:t.complexity??void 0,databaseId:t.id,...this.filterUndefined(i)}}),i=this.extractImplementationMetadata(e.metadata);return{id:e.display_id||e.id,databaseId:e.id,title:e.title,description:e.description||``,status:this.mapStatus(e.status),priority:this.mapPriority(e.priority),dependencies:n.get(e.id)||[],details:this.extractMetadataField(e.metadata,`details`,``),testStrategy:this.extractMetadataField(e.metadata,`testStrategy`,``),subtasks:r,createdAt:e.created_at,updatedAt:e.updated_at,assignee:e.assignee_id||void 0,complexity:e.complexity??void 0,effort:e.estimated_hours||void 0,actualEffort:e.actual_hours||void 0,...this.filterUndefined(i)}}static groupDependenciesByTaskId(e){let t=new Map;if(e)for(let n of e){let e=t.get(n.task_id)||[],r=typeof n.depends_on_task==`object`?n.depends_on_task?.display_id:n.depends_on_task_id;r&&e.push(r),t.set(n.task_id,e)}return t}static groupSubtasksByParentId(e){let t=new Map;for(let n of e)if(n.parent_task_id){let e=t.get(n.parent_task_id)||[];e.push(n),t.set(n.parent_task_id,e)}for(let e of t.values())e.sort((e,t)=>e.subtask_position-t.subtask_position);return t}static mapStatus(e){switch(e){case`todo`:return`pending`;case`in_progress`:return`in-progress`;case`done`:return`done`;default:return`pending`}}static mapPriority(e){switch(e){case`urgent`:return`critical`;default:return e}}static extractMetadataField(e,t,n){if(!e||typeof e!=`object`)return n;let r=e[t];if(r===void 0)return n;let i=typeof n,a=typeof r;return i===a?r:(console.warn(`Type mismatch in metadata field "${t}": expected ${i}, got ${a}. Using default value.`),n)}static extractOptionalString(e,t){if(!e||typeof e!=`object`)return;let n=e[t];return typeof n==`string`?n:void 0}static extractStringArray(e,t){if(!e||typeof e!=`object`)return;let n=e[t];if(!Array.isArray(n))return;let r=n.filter(e=>typeof e==`string`);return r.length>0?r:void 0}static extractRelevantFiles(e){if(!e||typeof e!=`object`)return;let t=e.relevantFiles;if(!Array.isArray(t))return;let n=t.filter(e=>{if(!e||typeof e!=`object`)return!1;let t=e;return typeof t.path==`string`&&typeof t.description==`string`&&(t.action===`create`||t.action===`modify`||t.action===`reference`)});return n.length>0?n:void 0}static extractExistingInfrastructure(e){if(!e||typeof e!=`object`)return;let t=e.existingInfrastructure;if(!Array.isArray(t))return;let n=t.filter(e=>{if(!e||typeof e!=`object`)return!1;let t=e;return typeof t.name==`string`&&typeof t.location==`string`&&typeof t.usage==`string`});return n.length>0?n:void 0}static extractScopeBoundaries(e){if(!e||typeof e!=`object`)return;let t=e.scopeBoundaries;if(!t||typeof t!=`object`)return;let n=t,r={};return typeof n.included==`string`&&(r.included=n.included),typeof n.excluded==`string`&&(r.excluded=n.excluded),r.included||r.excluded?r:void 0}static extractCategory(e){if(!e||typeof e!=`object`)return;let t=e.category;return[`research`,`design`,`development`,`testing`,`documentation`,`review`].includes(t)?t:void 0}static extractImplementationMetadata(e){return{relevantFiles:this.extractRelevantFiles(e),codebasePatterns:this.extractStringArray(e,`codebasePatterns`),existingInfrastructure:this.extractExistingInfrastructure(e),scopeBoundaries:this.extractScopeBoundaries(e),implementationApproach:this.extractOptionalString(e,`implementationApproach`),technicalConstraints:this.extractStringArray(e,`technicalConstraints`),acceptanceCriteria:this.extractStringArray(e,`acceptanceCriteria`),skills:this.extractStringArray(e,`skills`),category:this.extractCategory(e)}}static filterUndefined(e){return Object.fromEntries(Object.entries(e).filter(([e,t])=>t!==void 0))}},Sn=class{constructor(e){this.supabase=e}async fetchDependenciesWithDisplayIds(e){if(!e||e.length===0)return new Map;let{data:t,error:n}=await this.supabase.from(`task_dependencies`).select(`
|
|
300
|
+
`)}\n`}mapStatusLite(e){return e===`DONE`?`DONE`:`TODO`}async appendLedger(e,t){await y(e,`${JSON.stringify(t)}\n`,`utf-8`)}async loadCheckpoint(e){if(!await this.fileExists(e))return{updatedAt:F(),attempts:{},doneTaskIds:[],blockedTaskIds:[]};let t=await x(e,`utf-8`),n=JSON.parse(t);return{updatedAt:n.updatedAt||F(),attempts:n.attempts||{},doneTaskIds:n.doneTaskIds||[],blockedTaskIds:n.blockedTaskIds||[],lastTaskId:n.lastTaskId}}async saveCheckpoint(e,t){t.updatedAt=F(),await S(e,JSON.stringify(t,null,2),`utf-8`)}async fileExists(e){try{return await ie(e),!0}catch{return!1}}pushUnique(e,t){return e.includes(t)?e:[...e,t]}resolveDefaultAgentsPath(){let e=t.join(this.projectRoot,`AGENTS.md`);return g(e),e}},hn=class{service=null;tasksDomain=null;projectRoot;constructor(e){this.projectRoot=e.getProjectRoot()}setTasksDomain(e){this.tasksDomain=e,this.service=new mn(this.projectRoot,this.tasksDomain)}async initAssets(e={}){return this.getService().initAssets(e)}async run(e={}){return this.getService().run(e)}getService(){if(!this.tasksDomain)throw Error(`SkillRunDomain is not initialized with TasksDomain`);return this.service||=new mn(this.projectRoot,this.tasksDomain),this.service}},gn=class{async findBrief(e,t){return e.find(e=>this.matches(e,t))}matches(e,t){let n=e.document?.title||``;return!!(n.toLowerCase()===t.toLowerCase()||n.toLowerCase().includes(t.toLowerCase())||e.id===t||e.id.toLowerCase()===t.toLowerCase()||e.id.slice(-8).toLowerCase()===t.toLowerCase())}async getTagsWithStats(e,t,n,r){let i=await Promise.all(e.map(async e=>{try{let r=await n.getTasks(e.id,{}),i={},a=0,o={totalSubtasks:0,subtasksByStatus:{}};return r.forEach(e=>{let t=e.status||`pending`;i[t]=(i[t]||0)+1,t===`done`&&a++,e.subtasks&&e.subtasks.length>0&&(o.totalSubtasks+=e.subtasks.length,e.subtasks.forEach(e=>{let t=e.status||`pending`;o.subtasksByStatus[t]=(o.subtasksByStatus[t]||0)+1}))}),{name:e.document?.title||e.document?.document_name||e.id,isCurrent:t===e.id,taskCount:r.length,completedTasks:a,statusBreakdown:i,subtaskCounts:o.totalSubtasks>0?o:void 0,created:e.createdAt,description:e.document?.description,status:e.status,briefId:e.id,updatedAt:e.updatedAt}}catch(n){return console.warn(`Failed to get tasks for brief ${e.id}:`,n),{name:e.document?.title||e.document?.document_name||e.id,isCurrent:t===e.id,taskCount:0,completedTasks:0,statusBreakdown:{},created:e.createdAt,description:e.document?.description,status:e.status,briefId:e.id,updatedAt:e.updatedAt}}})),a={delivering:1,aligned:2,refining:3,draft:4,delivered:5,done:6,archived:7},o=i.sort((e,t)=>{let n=(e.status||``).toLowerCase(),r=(t.status||``).toLowerCase(),i=a[n]??999,o=a[r]??999;if(i!==o)return i-o;let s=e.updatedAt?new Date(e.updatedAt).getTime():0;return(t.updatedAt?new Date(t.updatedAt).getTime():0)-s}),s=e.find(e=>e.id===t);return{tags:o,currentTag:s&&(s.document?.title||s.document?.document_name)||null,totalTags:o.length}}validateBriefFound(e,t){if(!e)throw new T(`Brief "${t}" not found in organization`,w.NOT_FOUND)}},_n=class{static parse(e){let t=e?.trim()??``;if(!t)return{orgSlug:null,briefId:null};let n=this.parseAsUrl(t),r=n?n.pathname:t.includes(`/`)?t:null;return r?this.parsePathComponents(r,n):{orgSlug:null,briefId:t}}static extractOrgSlug(e){return this.parse(e).orgSlug}static extractBriefId(e){return this.parse(e).briefId||e.trim()}static parseAsUrl(e){try{return new URL(e)}catch{}try{return new URL(`https://${e}`)}catch{}return null}static parsePathComponents(e,t){let n=e.split(`/`).filter(Boolean),r=n.lastIndexOf(`briefs`),i=null,a=null;if(r>0&&(i=n[r-1]||null),t){let e=t.searchParams.get(`id`)||t.searchParams.get(`briefId`);e&&(a=e)}return!a&&r>=0&&n.length>r+1&&(a=n[r+1]),!a&&n.length>0&&!(r>=0&&r===n.length-1)&&(a=n[n.length-1]),{orgSlug:i,briefId:a}}static validate(e,t={}){if(t.requireOrg&&!e.orgSlug)throw new T(`Organization slug could not be extracted from input`,w.VALIDATION_ERROR);if(t.requireBrief&&!e.briefId)throw new T(`Brief identifier could not be extracted from input`,w.VALIDATION_ERROR)}},vn=class{briefService;authManager;constructor(){this.briefService=new gn,this.authManager=M.getInstance()}async resolveBrief(e,t){let n=_n.parse(e),r=n.briefId||e.trim(),i=t;if(!i&&n.orgSlug)try{let e=(await this.authManager.getOrganizations()).find(e=>e.slug?.toLowerCase()===n.orgSlug?.toLowerCase()||e.name.toLowerCase()===n.orgSlug?.toLowerCase());e&&(i=e.id)}catch{}if(i||=this.authManager.getContext()?.orgId,!i)throw new T(`No organization selected. Run "tm context org" first.`,w.CONFIG_ERROR);let a=await this.authManager.getBriefs(i),o=await this.briefService.findBrief(a,r);return this.briefService.validateBriefFound(o,r),o}async switchBrief(e){let t=await this.resolveBrief(e);await this.authManager.updateContext({briefId:t.id,briefName:t.document?.title||`Brief ${t.id.slice(-8)}`,briefStatus:t.status,briefUpdatedAt:t.updatedAt})}async getBriefsWithStats(e,t){let n=this.authManager.getContext();if(!n?.orgId)throw new T(`No organization context available`,w.MISSING_CONFIGURATION,{operation:`getBriefsWithStats`,userMessage:`No organization selected. Please authenticate first using: tm auth login`});let r=await this.authManager.getBriefs(n.orgId);return this.briefService.getTagsWithStats(r,n.briefId,e,t)}},yn=class{repository;projectId;apiClient;authManager;logger=D(`TaskExpansionService`);constructor(e,t,n,r){this.repository=e,this.projectId=t,this.apiClient=n,this.authManager=r}async expandTask(e,t){try{let n=this.authManager.ensureBriefSelected(`expandTask`),r=await this.repository.getTask(this.projectId,e);if(!r)throw new T(`Task ${e} not found`,w.TASK_NOT_FOUND,{operation:`expandTask`,taskId:e,userMessage:`Task ${e} isn't available in the current project.`});let i=await this.repository.getBrief(n.briefId),a={briefContext:{title:i?.name||n.briefName||n.briefId,description:i?.description||void 0,status:i?.status||`active`},allTasks:await this.repository.getTasks(this.projectId),existingSubtasks:r.subtasks||[],enrichedContext:t?.additionalContext},o=new URLSearchParams;if(t?.numSubtasks!==void 0&&o.set(`numSubtasks`,t.numSubtasks.toString()),t?.useResearch!==void 0&&o.set(`useResearch`,t.useResearch.toString()),t?.force!==void 0&&o.set(`force`,t.force.toString()),!r.databaseId)throw new T(`Task ${e} is missing a database ID. Task expansion requires tasks to be synced with the remote database.`,w.VALIDATION_ERROR,{operation:`expandTask`,taskId:e,userMessage:`This task has not been synced with the remote database. Please ensure the task is saved remotely before attempting expansion.`});if(!C.uuid().safeParse(r.databaseId).success)throw new T(`Task ${e} has an invalid database ID format: ${r.databaseId}`,w.VALIDATION_ERROR,{operation:`expandTask`,taskId:e,databaseId:r.databaseId,userMessage:`The task database ID is not in valid UUID format. This may indicate data corruption.`});let s=r.databaseId,c=`/ai/api/v1/tasks/${s}/subtasks/generate${o.toString()?`?${o.toString()}`:``}`,l=await this.apiClient.post(c,a),u=`${process.env.TM_BASE_DOMAIN||process.env.TM_PUBLIC_BASE_DOMAIN||`http://localhost:8080`}/home/hamster/briefs/${n.briefId}/task/${s}`;return this.logger.info(`✓ Task expansion queued for ${e}`),this.logger.info(` Job ID: ${l.jobId}`),this.logger.info(` ${l.message}`),this.logger.info(` View task: ${u}`),{...l,taskLink:u}}catch(n){throw n instanceof T?n.withContext({operation:`expandTask`,taskId:e,numSubtasks:t?.numSubtasks,useResearch:t?.useResearch}):new T(n instanceof Error?n.message:String(n),w.STORAGE_ERROR,{operation:`expandTask`,taskId:e,numSubtasks:t?.numSubtasks,useResearch:t?.useResearch},n)}}},bn=class{repository;projectId;apiClient;authManager;logger=D(`TaskRetrievalService`);constructor(e,t,n,r){this.repository=e,this.projectId=t,this.apiClient=n,this.authManager=r}async getTask(e){try{this.authManager.ensureBriefSelected(`getTask`);let t=await this.repository.getTask(this.projectId,e);if(!t)throw new T(`Task ${e} not found`,w.TASK_NOT_FOUND,{operation:`getTask`,taskId:e,userMessage:`Task ${e} isn't available in the current project.`});try{let e=`/ai/api/v1/tasks/${t.id}`,n=await this.apiClient.get(e);n.document?.content&&(t.details=n.document.content)}catch(t){this.logger.debug(`Could not fetch document content for task ${e}: ${t}`)}return this.logger.info(`✓ Retrieved task ${e}`),t.details&&this.logger.debug(` Document content available (${t.details.length} chars)`),t}catch(t){throw t instanceof T?t.withContext({operation:`getTask`,taskId:e}):new T(t instanceof Error?t.message:String(t),w.STORAGE_ERROR,{operation:`getTask`,taskId:e},t)}}},xn=class{static mapDatabaseTasksToTasks(e,t){if(!e||e.length===0)return[];let n=t instanceof Map?t:this.groupDependenciesByTaskId(t),r=e.filter(e=>!e.parent_task_id),i=this.groupSubtasksByParentId(e);return r.map(e=>this.mapDatabaseTaskToTask(e,i.get(e.id)||[],n))}static mapDatabaseTaskToTask(e,t,n){let r=t.map((t,r)=>{let i=this.extractImplementationMetadata(t.metadata);return{id:t.display_id||String(r+1),parentId:e.id,title:t.title,description:t.description||``,status:this.mapStatus(t.status),priority:this.mapPriority(t.priority),dependencies:n.get(t.id)||[],details:this.extractMetadataField(t.metadata,`details`,``),testStrategy:this.extractMetadataField(t.metadata,`testStrategy`,``),createdAt:t.created_at,updatedAt:t.updated_at,assignee:t.assignee_id||void 0,complexity:t.complexity??void 0,databaseId:t.id,...this.filterUndefined(i)}}),i=this.extractImplementationMetadata(e.metadata);return{id:e.display_id||e.id,databaseId:e.id,title:e.title,description:e.description||``,status:this.mapStatus(e.status),priority:this.mapPriority(e.priority),dependencies:n.get(e.id)||[],details:this.extractMetadataField(e.metadata,`details`,``),testStrategy:this.extractMetadataField(e.metadata,`testStrategy`,``),subtasks:r,createdAt:e.created_at,updatedAt:e.updated_at,assignee:e.assignee_id||void 0,complexity:e.complexity??void 0,effort:e.estimated_hours||void 0,actualEffort:e.actual_hours||void 0,...this.filterUndefined(i)}}static groupDependenciesByTaskId(e){let t=new Map;if(e)for(let n of e){let e=t.get(n.task_id)||[],r=typeof n.depends_on_task==`object`?n.depends_on_task?.display_id:n.depends_on_task_id;r&&e.push(r),t.set(n.task_id,e)}return t}static groupSubtasksByParentId(e){let t=new Map;for(let n of e)if(n.parent_task_id){let e=t.get(n.parent_task_id)||[];e.push(n),t.set(n.parent_task_id,e)}for(let e of t.values())e.sort((e,t)=>e.subtask_position-t.subtask_position);return t}static mapStatus(e){switch(e){case`todo`:return`pending`;case`in_progress`:return`in-progress`;case`done`:return`done`;default:return`pending`}}static mapPriority(e){switch(e){case`urgent`:return`critical`;default:return e}}static extractMetadataField(e,t,n){if(!e||typeof e!=`object`)return n;let r=e[t];if(r===void 0)return n;let i=typeof n,a=typeof r;return i===a?r:(console.warn(`Type mismatch in metadata field "${t}": expected ${i}, got ${a}. Using default value.`),n)}static extractOptionalString(e,t){if(!e||typeof e!=`object`)return;let n=e[t];return typeof n==`string`?n:void 0}static extractStringArray(e,t){if(!e||typeof e!=`object`)return;let n=e[t];if(!Array.isArray(n))return;let r=n.filter(e=>typeof e==`string`);return r.length>0?r:void 0}static extractRelevantFiles(e){if(!e||typeof e!=`object`)return;let t=e.relevantFiles;if(!Array.isArray(t))return;let n=t.filter(e=>{if(!e||typeof e!=`object`)return!1;let t=e;return typeof t.path==`string`&&typeof t.description==`string`&&(t.action===`create`||t.action===`modify`||t.action===`reference`)});return n.length>0?n:void 0}static extractExistingInfrastructure(e){if(!e||typeof e!=`object`)return;let t=e.existingInfrastructure;if(!Array.isArray(t))return;let n=t.filter(e=>{if(!e||typeof e!=`object`)return!1;let t=e;return typeof t.name==`string`&&typeof t.location==`string`&&typeof t.usage==`string`});return n.length>0?n:void 0}static extractScopeBoundaries(e){if(!e||typeof e!=`object`)return;let t=e.scopeBoundaries;if(!t||typeof t!=`object`)return;let n=t,r={};return typeof n.included==`string`&&(r.included=n.included),typeof n.excluded==`string`&&(r.excluded=n.excluded),r.included||r.excluded?r:void 0}static extractCategory(e){if(!e||typeof e!=`object`)return;let t=e.category;return[`research`,`design`,`development`,`testing`,`documentation`,`review`].includes(t)?t:void 0}static extractImplementationMetadata(e){return{relevantFiles:this.extractRelevantFiles(e),codebasePatterns:this.extractStringArray(e,`codebasePatterns`),existingInfrastructure:this.extractExistingInfrastructure(e),scopeBoundaries:this.extractScopeBoundaries(e),implementationApproach:this.extractOptionalString(e,`implementationApproach`),technicalConstraints:this.extractStringArray(e,`technicalConstraints`),acceptanceCriteria:this.extractStringArray(e,`acceptanceCriteria`),skills:this.extractStringArray(e,`skills`),category:this.extractCategory(e)}}static filterUndefined(e){return Object.fromEntries(Object.entries(e).filter(([e,t])=>t!==void 0))}},Sn=class{constructor(e){this.supabase=e}async fetchDependenciesWithDisplayIds(e){if(!e||e.length===0)return new Map;let{data:t,error:n}=await this.supabase.from(`task_dependencies`).select(`
|
|
301
301
|
task_id,
|
|
302
302
|
depends_on_task:tasks!task_dependencies_depends_on_task_id_fkey (
|
|
303
303
|
display_id
|