coder-config 0.46.0-beta.12 → 0.46.0-beta.13

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/lib/constants.js CHANGED
@@ -2,7 +2,7 @@
2
2
  * Constants and tool path configurations
3
3
  */
4
4
 
5
- const VERSION = '0.46.0-beta.12';
5
+ const VERSION = '0.46.0-beta.13';
6
6
 
7
7
  // Tool-specific path configurations
8
8
  const TOOL_PATHS = {
@@ -746,40 +746,6 @@ const AI_TOOLS = {
746
746
  path.join(os.homedir(), '.local', 'pipx', 'venvs', 'aider-chat', 'bin', 'aider'),
747
747
  ],
748
748
  },
749
- lmstudio: {
750
- name: 'LM Studio',
751
- binary: 'curl',
752
- // LM Studio runs an OpenAI-compatible server at localhost:1234
753
- // We use curl to call the API directly
754
- buildArgs: (prompt, options) => {
755
- const model = options.model || 'default';
756
- const endpoint = options.endpoint || 'http://localhost:1234/v1/chat/completions';
757
- const requestBody = JSON.stringify({
758
- model,
759
- messages: [{ role: 'user', content: prompt }],
760
- temperature: 0.7,
761
- max_tokens: 2048,
762
- stream: false,
763
- });
764
- return [
765
- '-s', // silent
766
- '-X', 'POST',
767
- endpoint,
768
- '-H', 'Content-Type: application/json',
769
- '-d', requestBody,
770
- ];
771
- },
772
- candidates: () => ['/usr/bin/curl', '/opt/homebrew/bin/curl', 'curl'],
773
- // Custom response parser - extract content from OpenAI-style response
774
- parseResponse: (output) => {
775
- try {
776
- const response = JSON.parse(output);
777
- return response.choices?.[0]?.message?.content || output;
778
- } catch (e) {
779
- return output;
780
- }
781
- },
782
- },
783
749
  };
784
750
 
785
751
  /**
@@ -887,18 +853,13 @@ Output markdown suitable for injecting into an AI assistant's context. Keep it c
887
853
 
888
854
  console.log(`Generating context with ${tool.name}...`);
889
855
 
890
- let result = execFileSync(binaryPath, args, {
856
+ const result = execFileSync(binaryPath, args, {
891
857
  cwd: allProjects[0],
892
858
  encoding: 'utf8',
893
859
  timeout: 120000, // 2 minute timeout (some models are slower)
894
860
  maxBuffer: 1024 * 1024, // 1MB buffer
895
861
  });
896
862
 
897
- // Parse response if tool has a custom parser (e.g., LM Studio returns JSON)
898
- if (tool.parseResponse) {
899
- result = tool.parseResponse(result);
900
- }
901
-
902
863
  return result.trim();
903
864
  } catch (error) {
904
865
  console.error(`${tool.name} generation failed:`, error.message);
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "coder-config",
3
- "version": "0.46.0-beta.12",
3
+ "version": "0.46.0-beta.13",
4
4
  "description": "Configuration manager for AI coding tools - Claude Code, Gemini CLI, Codex CLI, Antigravity. Manage MCPs, rules, permissions, memory, and workstreams.",
5
5
  "author": "regression.io",
6
6
  "main": "config-loader.js",