cast-code 1.0.0 → 1.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -75,9 +75,6 @@ let ConfigService = class ConfigService {
75
75
  ollamaBaseUrl: frontmatter.ollamaBaseUrl || process.env.OLLAMA_BASE_URL || this.config.ollamaBaseUrl
76
76
  };
77
77
  }
78
- if (this.config.provider === 'openai' && !this.config.apiKey) {
79
- throw new Error('OPENAI_API_KEY not configured. Set it in environment or ~/.cast/config.md');
80
- }
81
78
  }
82
79
  getConfig() {
83
80
  return this.config;
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../src/common/services/config.service.ts"],"sourcesContent":["import { Injectable } from '@nestjs/common';\nimport * as path from 'path';\nimport { MarkdownParserService } from './markdown-parser.service';\nimport { GLOBAL_CONFIG_DIR, DEFAULT_MODEL, DEFAULT_TEMPERATURE } from '../constants';\n\ninterface GlobalConfigFrontmatter {\n model?: string;\n temperature?: number;\n apiKey?: string;\n provider?: string;\n ollamaBaseUrl?: string;\n}\n\nexport interface GlobalConfig {\n model: string;\n temperature: number;\n apiKey: string;\n provider: string;\n ollamaBaseUrl: string;\n}\n\n@Injectable()\nexport class ConfigService {\n private config: GlobalConfig = {\n model: process.env.OLLAMA_MODEL || DEFAULT_MODEL,\n temperature: DEFAULT_TEMPERATURE,\n apiKey: process.env.OPENAI_API_KEY || '',\n provider: process.env.LLM_PROVIDER || 'openai',\n ollamaBaseUrl: process.env.OLLAMA_BASE_URL || 'http://localhost:11434',\n };\n\n constructor(private readonly markdownParser: MarkdownParserService) {}\n\n async loadGlobalConfig() {\n const configPath = path.join(GLOBAL_CONFIG_DIR, 'config.md');\n\n if (await this.markdownParser.exists(configPath)) {\n const { frontmatter } =\n await this.markdownParser.parse<GlobalConfigFrontmatter>(configPath);\n\n this.config = {\n model: frontmatter.model || process.env.OLLAMA_MODEL || this.config.model,\n temperature: frontmatter.temperature ?? this.config.temperature,\n apiKey: frontmatter.apiKey || process.env.OPENAI_API_KEY || '',\n provider: frontmatter.provider || process.env.LLM_PROVIDER || this.config.provider,\n ollamaBaseUrl: frontmatter.ollamaBaseUrl || process.env.OLLAMA_BASE_URL || this.config.ollamaBaseUrl,\n };\n }\n\n if (this.config.provider === 'openai' && !this.config.apiKey) {\n throw new Error(\n 'OPENAI_API_KEY not configured. Set it in environment or ~/.cast/config.md',\n );\n }\n }\n\n getConfig(): GlobalConfig {\n return this.config;\n }\n\n getModel(): string {\n return this.config.model;\n }\n\n getTemperature(): number {\n return this.config.temperature;\n }\n\n getApiKey(): string {\n return this.config.apiKey;\n }\n\n getProvider(): string {\n return this.config.provider;\n }\n\n getOllamaBaseUrl(): string {\n return this.config.ollamaBaseUrl;\n }\n}\n"],"names":["ConfigService","loadGlobalConfig","configPath","path","join","GLOBAL_CONFIG_DIR","markdownParser","exists","frontmatter","parse","config","model","process","env","OLLAMA_MODEL","temperature","apiKey","OPENAI_API_KEY","provider","LLM_PROVIDER","ollamaBaseUrl","OLLAMA_BASE_URL","Error","getConfig","getModel","getTemperature","getApiKey","getProvider","getOllamaBaseUrl","DEFAULT_MODEL","DEFAULT_TEMPERATURE"],"mappings":";;;;+BAsBaA;;;eAAAA;;;wBAtBc;8DACL;uCACgB;2BACgC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAmB/D,IAAA,AAAMA,gBAAN,MAAMA;IAWX,MAAMC,mBAAmB;QACvB,MAAMC,aAAaC,MAAKC,IAAI,CAACC,4BAAiB,EAAE;QAEhD,IAAI,MAAM,IAAI,CAACC,cAAc,CAACC,MAAM,CAACL,aAAa;YAChD,MAAM,EAAEM,WAAW,EAAE,GACnB,MAAM,IAAI,CAACF,cAAc,CAACG,KAAK,CAA0BP;YAE3D,IAAI,CAACQ,MAAM,GAAG;gBACZC,OAAOH,YAAYG,KAAK,IAAIC,QAAQC,GAAG,CAACC,YAAY,IAAI,IAAI,CAACJ,MAAM,CAACC,KAAK;gBACzEI,aAAaP,YAAYO,WAAW,IAAI,IAAI,CAACL,MAAM,CAACK,WAAW;gBAC/DC,QAAQR,YAAYQ,MAAM,IAAIJ,QAAQC,GAAG,CAACI,cAAc,IAAI;gBAC5DC,UAAUV,YAAYU,QAAQ,IAAIN,QAAQC,GAAG,CAACM,YAAY,IAAI,IAAI,CAACT,MAAM,CAACQ,QAAQ;gBAClFE,eAAeZ,YAAYY,aAAa,IAAIR,QAAQC,GAAG,CAACQ,eAAe,IAAI,IAAI,CAACX,MAAM,CAACU,aAAa;YACtG;QACF;QAEA,IAAI,IAAI,CAACV,MAAM,CAACQ,QAAQ,KAAK,YAAY,CAAC,IAAI,CAACR,MAAM,CAACM,MAAM,EAAE;YAC5D,MAAM,IAAIM,MACR;QAEJ;IACF;IAEAC,YAA0B;QACxB,OAAO,IAAI,CAACb,MAAM;IACpB;IAEAc,WAAmB;QACjB,OAAO,IAAI,CAACd,MAAM,CAACC,KAAK;IAC1B;IAEAc,iBAAyB;QACvB,OAAO,IAAI,CAACf,MAAM,CAACK,WAAW;IAChC;IAEAW,YAAoB;QAClB,OAAO,IAAI,CAAChB,MAAM,CAACM,MAAM;IAC3B;IAEAW,cAAsB;QACpB,OAAO,IAAI,CAACjB,MAAM,CAACQ,QAAQ;IAC7B;IAEAU,mBAA2B;QACzB,OAAO,IAAI,CAAClB,MAAM,CAACU,aAAa;IAClC;IA/CA,YAAY,AAAiBd,cAAqC,CAAE;aAAvCA,iBAAAA;aARrBI,SAAuB;YAC7BC,OAAOC,QAAQC,GAAG,CAACC,YAAY,IAAIe,wBAAa;YAChDd,aAAae,8BAAmB;YAChCd,QAAQJ,QAAQC,GAAG,CAACI,cAAc,IAAI;YACtCC,UAAUN,QAAQC,GAAG,CAACM,YAAY,IAAI;YACtCC,eAAeR,QAAQC,GAAG,CAACQ,eAAe,IAAI;QAChD;IAEqE;AAgDvE"}
1
+ {"version":3,"sources":["../../../src/common/services/config.service.ts"],"sourcesContent":["import { Injectable } from '@nestjs/common';\nimport * as path from 'path';\nimport { MarkdownParserService } from './markdown-parser.service';\nimport { GLOBAL_CONFIG_DIR, DEFAULT_MODEL, DEFAULT_TEMPERATURE } from '../constants';\n\ninterface GlobalConfigFrontmatter {\n model?: string;\n temperature?: number;\n apiKey?: string;\n provider?: string;\n ollamaBaseUrl?: string;\n}\n\nexport interface GlobalConfig {\n model: string;\n temperature: number;\n apiKey: string;\n provider: string;\n ollamaBaseUrl: string;\n}\n\n@Injectable()\nexport class ConfigService {\n private config: GlobalConfig = {\n model: process.env.OLLAMA_MODEL || DEFAULT_MODEL,\n temperature: DEFAULT_TEMPERATURE,\n apiKey: process.env.OPENAI_API_KEY || '',\n provider: process.env.LLM_PROVIDER || 'openai',\n ollamaBaseUrl: process.env.OLLAMA_BASE_URL || 'http://localhost:11434',\n };\n\n constructor(private readonly markdownParser: MarkdownParserService) {}\n\n async loadGlobalConfig() {\n const configPath = path.join(GLOBAL_CONFIG_DIR, 'config.md');\n\n if (await this.markdownParser.exists(configPath)) {\n const { frontmatter } =\n await this.markdownParser.parse<GlobalConfigFrontmatter>(configPath);\n\n this.config = {\n model: frontmatter.model || process.env.OLLAMA_MODEL || this.config.model,\n temperature: frontmatter.temperature ?? this.config.temperature,\n apiKey: frontmatter.apiKey || process.env.OPENAI_API_KEY || '',\n provider: frontmatter.provider || process.env.LLM_PROVIDER || this.config.provider,\n ollamaBaseUrl: frontmatter.ollamaBaseUrl || process.env.OLLAMA_BASE_URL || this.config.ollamaBaseUrl,\n };\n }\n\n }\n\n getConfig(): GlobalConfig {\n return this.config;\n }\n\n getModel(): string {\n return this.config.model;\n }\n\n getTemperature(): number {\n return this.config.temperature;\n }\n\n getApiKey(): string {\n return this.config.apiKey;\n }\n\n getProvider(): string {\n return this.config.provider;\n }\n\n getOllamaBaseUrl(): string {\n return this.config.ollamaBaseUrl;\n }\n}\n"],"names":["ConfigService","loadGlobalConfig","configPath","path","join","GLOBAL_CONFIG_DIR","markdownParser","exists","frontmatter","parse","config","model","process","env","OLLAMA_MODEL","temperature","apiKey","OPENAI_API_KEY","provider","LLM_PROVIDER","ollamaBaseUrl","OLLAMA_BASE_URL","getConfig","getModel","getTemperature","getApiKey","getProvider","getOllamaBaseUrl","DEFAULT_MODEL","DEFAULT_TEMPERATURE"],"mappings":";;;;+BAsBaA;;;eAAAA;;;wBAtBc;8DACL;uCACgB;2BACgC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAmB/D,IAAA,AAAMA,gBAAN,MAAMA;IAWX,MAAMC,mBAAmB;QACvB,MAAMC,aAAaC,MAAKC,IAAI,CAACC,4BAAiB,EAAE;QAEhD,IAAI,MAAM,IAAI,CAACC,cAAc,CAACC,MAAM,CAACL,aAAa;YAChD,MAAM,EAAEM,WAAW,EAAE,GACnB,MAAM,IAAI,CAACF,cAAc,CAACG,KAAK,CAA0BP;YAE3D,IAAI,CAACQ,MAAM,GAAG;gBACZC,OAAOH,YAAYG,KAAK,IAAIC,QAAQC,GAAG,CAACC,YAAY,IAAI,IAAI,CAACJ,MAAM,CAACC,KAAK;gBACzEI,aAAaP,YAAYO,WAAW,IAAI,IAAI,CAACL,MAAM,CAACK,WAAW;gBAC/DC,QAAQR,YAAYQ,MAAM,IAAIJ,QAAQC,GAAG,CAACI,cAAc,IAAI;gBAC5DC,UAAUV,YAAYU,QAAQ,IAAIN,QAAQC,GAAG,CAACM,YAAY,IAAI,IAAI,CAACT,MAAM,CAACQ,QAAQ;gBAClFE,eAAeZ,YAAYY,aAAa,IAAIR,QAAQC,GAAG,CAACQ,eAAe,IAAI,IAAI,CAACX,MAAM,CAACU,aAAa;YACtG;QACF;IAEF;IAEAE,YAA0B;QACxB,OAAO,IAAI,CAACZ,MAAM;IACpB;IAEAa,WAAmB;QACjB,OAAO,IAAI,CAACb,MAAM,CAACC,KAAK;IAC1B;IAEAa,iBAAyB;QACvB,OAAO,IAAI,CAACd,MAAM,CAACK,WAAW;IAChC;IAEAU,YAAoB;QAClB,OAAO,IAAI,CAACf,MAAM,CAACM,MAAM;IAC3B;IAEAU,cAAsB;QACpB,OAAO,IAAI,CAAChB,MAAM,CAACQ,QAAQ;IAC7B;IAEAS,mBAA2B;QACzB,OAAO,IAAI,CAACjB,MAAM,CAACU,aAAa;IAClC;IA1CA,YAAY,AAAiBd,cAAqC,CAAE;aAAvCA,iBAAAA;aARrBI,SAAuB;YAC7BC,OAAOC,QAAQC,GAAG,CAACC,YAAY,IAAIc,wBAAa;YAChDb,aAAac,8BAAmB;YAChCb,QAAQJ,QAAQC,GAAG,CAACI,cAAc,IAAI;YACtCC,UAAUN,QAAQC,GAAG,CAACM,YAAY,IAAI;YACtCC,eAAeR,QAAQC,GAAG,CAACQ,eAAe,IAAI;QAChD;IAEqE;AA2CvE"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "cast-code",
3
- "version": "1.0.0",
3
+ "version": "1.0.1",
4
4
  "description": "Multi-agent CLI system powered by DeepAgents",
5
5
  "main": "dist/main.js",
6
6
  "bin": {