groove-dev 0.24.6 → 0.24.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -11,11 +11,11 @@ export class CodexProvider extends Provider {
11
11
  static authType = 'api-key';
12
12
  static envKey = 'OPENAI_API_KEY';
13
13
  static models = [
14
- { id: 'o3', name: 'o3', tier: 'heavy', pricing: { input: 0.01, output: 0.04 } },
15
- { id: 'o4-mini', name: 'o4-mini', tier: 'medium', pricing: { input: 0.001, output: 0.004 } },
16
- { id: 'gpt-4.1', name: 'GPT-4.1', tier: 'heavy', pricing: { input: 0.002, output: 0.008 } },
17
- { id: 'gpt-4.1-mini', name: 'GPT-4.1 Mini', tier: 'medium', pricing: { input: 0.0004, output: 0.0016 } },
18
- { id: 'gpt-4.1-nano', name: 'GPT-4.1 Nano', tier: 'light', pricing: { input: 0.0001, output: 0.0004 } },
14
+ { id: 'o5-deep', name: 'o5 Deep Reasoning', tier: 'heavy', pricing: { input: 0.015, output: 0.06 } },
15
+ { id: 'o5-mini', name: 'o5 Mini', tier: 'medium', pricing: { input: 0.002, output: 0.008 } },
16
+ { id: 'gpt-5.4-fast', name: 'GPT-5.4 Fast', tier: 'medium', pricing: { input: 0.001, output: 0.004 } },
17
+ { id: 'gpt-5.4-mini', name: 'GPT-5.4 Mini', tier: 'light', pricing: { input: 0.0004, output: 0.0016 } },
18
+ { id: 'gpt-5.4-nano', name: 'GPT-5.4 Nano', tier: 'light', pricing: { input: 0.0001, output: 0.0004 } },
19
19
  ];
20
20
 
21
21
  static isInstalled() {
@@ -9,28 +9,27 @@ import { Provider } from './base.js';
9
9
  // Comprehensive model catalog with RAM requirements
10
10
  const MODEL_CATALOG = [
11
11
  // ── Code-Focused ──────────────────────────────────────────
12
- { id: 'qwen2.5-coder:0.5b', name: 'Qwen 2.5 Coder 0.5B', category: 'code', tier: 'light', ramGb: 1, sizeGb: 0.4, description: 'Ultra-light code completion' },
13
- { id: 'qwen2.5-coder:1.5b', name: 'Qwen 2.5 Coder 1.5B', category: 'code', tier: 'light', ramGb: 2, sizeGb: 1.0, description: 'Fast code assistant' },
14
- { id: 'qwen2.5-coder:3b', name: 'Qwen 2.5 Coder 3B', category: 'code', tier: 'light', ramGb: 4, sizeGb: 1.9, description: 'Balanced speed and quality' },
15
12
  { id: 'qwen2.5-coder:7b', name: 'Qwen 2.5 Coder 7B', category: 'code', tier: 'medium', ramGb: 8, sizeGb: 4.7, description: 'Strong general coding' },
16
13
  { id: 'qwen2.5-coder:14b', name: 'Qwen 2.5 Coder 14B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 9.0, description: 'Advanced code reasoning' },
17
- { id: 'qwen2.5-coder:32b', name: 'Qwen 2.5 Coder 32B', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 20, description: 'Best-in-class code generation' },
18
- { id: 'codellama:7b', name: 'Code Llama 7B', category: 'code', tier: 'medium', ramGb: 8, sizeGb: 3.8, description: 'Meta code model, solid baseline' },
19
- { id: 'codellama:13b', name: 'Code Llama 13B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 7.4, description: 'Stronger code reasoning' },
20
- { id: 'codellama:34b', name: 'Code Llama 34B', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 19, description: 'Largest Code Llama' },
21
- { id: 'deepseek-coder-v2:16b', name: 'DeepSeek Coder V2 16B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 8.9, description: 'Strong multi-language coding' },
22
- { id: 'starcoder2:3b', name: 'StarCoder2 3B', category: 'code', tier: 'light', ramGb: 4, sizeGb: 1.7, description: 'Code completion specialist' },
23
- { id: 'starcoder2:7b', name: 'StarCoder2 7B', category: 'code', tier: 'medium', ramGb: 8, sizeGb: 4.0, description: 'Solid code generation' },
24
- { id: 'starcoder2:15b', name: 'StarCoder2 15B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 9.0, description: 'Best StarCoder variant' },
14
+ { id: 'qwen2.5-coder:32b', name: 'Qwen 2.5 Coder 32B', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 20, description: 'Gold standard for local coding, rivals GPT-4o' },
15
+ { id: 'qwen3-coder-next', name: 'Qwen3 Coder Next', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 18, description: 'Agentic coding optimized — tools, terminal, file systems' },
16
+ { id: 'deepseek-r1:14b', name: 'DeepSeek R1 14B', category: 'code', tier: 'medium', ramGb: 12, sizeGb: 8.5, description: 'Chain-of-thought debugging, finds logic errors' },
17
+ { id: 'deepseek-r1:32b', name: 'DeepSeek R1 32B', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 19, description: 'Deep reasoning powerhouse for complex debugging' },
18
+ { id: 'deepseek-coder-v2:16b', name: 'DeepSeek Coder V2 16B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 8.9, description: 'MoE architecture, fast multi-language coding' },
19
+ { id: 'codestral', name: 'Codestral 25.01 (25B)', category: 'code', tier: 'heavy', ramGb: 18, sizeGb: 14, description: 'Mistral flagship, best-in-class autocomplete/FIM' },
20
+ { id: 'devstral-small-2', name: 'Devstral Small 2 (24B)', category: 'code', tier: 'heavy', ramGb: 18, sizeGb: 14, description: 'Mistral SE agent — multi-file edits, codebase exploration' },
21
+ { id: 'phind-codellama', name: 'Phind CodeLlama 34B', category: 'code', tier: 'heavy', ramGb: 24, sizeGb: 19, description: 'Instruction-tuned for programming, Llama ecosystem' },
22
+ { id: 'starcoder2:15b', name: 'StarCoder2 15B', category: 'code', tier: 'medium', ramGb: 16, sizeGb: 9.0, description: '600+ languages — COBOL, Fortran, DSLs' },
23
+ { id: 'codegemma', name: 'CodeGemma 7B', category: 'code', tier: 'light', ramGb: 8, sizeGb: 5.0, description: 'Google lightweight — scripts, unit tests, laptop-friendly' },
25
24
 
26
25
  // ── General (Strong at Code) ──────────────────────────────
26
+ { id: 'gemma4:26b', name: 'Gemma 4 26B', category: 'general', tier: 'heavy', ramGb: 16, sizeGb: 15, description: 'Google latest — top-tier Python/C++ reasoning' },
27
+ { id: 'gemma4:12b', name: 'Gemma 4 12B', category: 'general', tier: 'medium', ramGb: 10, sizeGb: 7.5, description: 'Google mid-range, strong reasoning per watt' },
27
28
  { id: 'llama3.1:8b', name: 'Llama 3.1 8B', category: 'general', tier: 'medium', ramGb: 8, sizeGb: 4.7, description: 'Meta flagship, great all-rounder' },
28
29
  { id: 'llama3.1:70b', name: 'Llama 3.1 70B', category: 'general', tier: 'heavy', ramGb: 48, sizeGb: 40, description: 'Frontier-class open model' },
30
+ { id: 'deepseek-r1:7b', name: 'DeepSeek R1 7B', category: 'general', tier: 'medium', ramGb: 8, sizeGb: 4.7, description: 'Compact reasoning model, think-before-respond' },
29
31
  { id: 'mistral:7b', name: 'Mistral 7B', category: 'general', tier: 'medium', ramGb: 8, sizeGb: 4.1, description: 'Fast and efficient' },
30
32
  { id: 'mixtral:8x7b', name: 'Mixtral 8x7B', category: 'general', tier: 'heavy', ramGb: 32, sizeGb: 26, description: 'MoE architecture, expert routing' },
31
- { id: 'gemma2:2b', name: 'Gemma 2 2B', category: 'general', tier: 'light', ramGb: 2, sizeGb: 1.6, description: 'Google lightweight model' },
32
- { id: 'gemma2:9b', name: 'Gemma 2 9B', category: 'general', tier: 'medium', ramGb: 8, sizeGb: 5.4, description: 'Google mid-range model' },
33
- { id: 'gemma2:27b', name: 'Gemma 2 27B', category: 'general', tier: 'heavy', ramGb: 20, sizeGb: 16, description: 'Google large model' },
34
33
  { id: 'phi3:mini', name: 'Phi-3 Mini 3.8B', category: 'general', tier: 'light', ramGb: 4, sizeGb: 2.3, description: 'Microsoft small but capable' },
35
34
  { id: 'phi3:medium', name: 'Phi-3 Medium 14B', category: 'general', tier: 'medium', ramGb: 16, sizeGb: 7.9, description: 'Microsoft mid-range' },
36
35
  ];