codemaxxing 0.4.16 → 1.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +41 -8
- package/dist/config.d.ts +11 -0
- package/dist/config.js +48 -4
- package/dist/index.js +55 -15
- package/package.json +1 -1
- package/src/config.ts +56 -1
- package/src/index.tsx +56 -15
package/README.md
CHANGED
|
@@ -191,10 +191,35 @@ Conversations auto-save to SQLite. Pick up where you left off:
|
|
|
191
191
|
- `/session delete` — remove a session
|
|
192
192
|
- `/resume` — interactive session picker
|
|
193
193
|
|
|
194
|
+
### 🔌 MCP Support (Model Context Protocol)
|
|
195
|
+
Connect to external tools via the industry-standard MCP protocol. Databases, GitHub, Slack, browsers — anything with an MCP server.
|
|
196
|
+
- Compatible with `.cursor/mcp.json` and `opencode.json` configs
|
|
197
|
+
- `/mcp` — show connected servers
|
|
198
|
+
- `/mcp add github npx -y @modelcontextprotocol/server-github` — add a server
|
|
199
|
+
- `/mcp tools` — list all available MCP tools
|
|
200
|
+
|
|
201
|
+
### 🖥️ Zero-Setup Local LLM
|
|
202
|
+
First time with no LLM? Codemaxxing walks you through it:
|
|
203
|
+
1. Detects your hardware (CPU, RAM, GPU)
|
|
204
|
+
2. Recommends coding models that fit your machine
|
|
205
|
+
3. Installs Ollama automatically
|
|
206
|
+
4. Downloads the model with a progress bar
|
|
207
|
+
5. Connects and drops you into coding mode
|
|
208
|
+
|
|
209
|
+
No googling, no config files, no decisions. Just run `codemaxxing`.
|
|
210
|
+
|
|
211
|
+
### 🦙 Ollama Management
|
|
212
|
+
Full Ollama control from inside codemaxxing:
|
|
213
|
+
- `/ollama` — status, installed models, GPU usage
|
|
214
|
+
- `/ollama pull` — interactive model picker + download
|
|
215
|
+
- `/ollama delete` — pick and remove models
|
|
216
|
+
- `/ollama start` / `/ollama stop` — server management
|
|
217
|
+
- Exit warning when Ollama is using GPU memory
|
|
218
|
+
|
|
194
219
|
### 🔄 Multi-Provider
|
|
195
|
-
Switch models mid-session
|
|
196
|
-
- `/model
|
|
197
|
-
- `/
|
|
220
|
+
Switch models mid-session with an interactive picker:
|
|
221
|
+
- `/model` — browse and switch models
|
|
222
|
+
- `/model gpt-4o` — switch directly by name
|
|
198
223
|
- Native Anthropic API support (not just OpenAI-compatible)
|
|
199
224
|
|
|
200
225
|
### 🎨 14 Themes
|
|
@@ -216,11 +241,15 @@ Type `/` for autocomplete suggestions. Arrow keys to navigate, Tab or Enter to s
|
|
|
216
241
|
| `/help` | Show all commands |
|
|
217
242
|
| `/connect` | Retry LLM connection |
|
|
218
243
|
| `/login` | Interactive auth setup |
|
|
244
|
+
| `/model` | Browse & switch models (picker) |
|
|
219
245
|
| `/architect` | Toggle architect mode / set model |
|
|
220
246
|
| `/skills` | Browse, install, manage skills |
|
|
221
247
|
| `/lint on/off` | Toggle auto-linting |
|
|
222
|
-
| `/
|
|
223
|
-
| `/
|
|
248
|
+
| `/mcp` | MCP server status & tools |
|
|
249
|
+
| `/ollama` | Ollama status, models & GPU |
|
|
250
|
+
| `/ollama pull` | Download a model (picker) |
|
|
251
|
+
| `/ollama delete` | Remove a model (picker) |
|
|
252
|
+
| `/ollama start/stop` | Server management |
|
|
224
253
|
| `/theme` | Switch color theme |
|
|
225
254
|
| `/map` | Show repository map |
|
|
226
255
|
| `/sessions` | List past sessions |
|
|
@@ -295,14 +324,16 @@ Settings are stored in `~/.codemaxxing/settings.json`:
|
|
|
295
324
|
|
|
296
325
|
## Tools
|
|
297
326
|
|
|
298
|
-
|
|
327
|
+
Built-in tools:
|
|
299
328
|
|
|
300
329
|
- **read_file** — Read file contents (safe)
|
|
301
|
-
- **write_file** — Write/create files (requires approval)
|
|
330
|
+
- **write_file** — Write/create files (requires approval, shows diff)
|
|
302
331
|
- **list_files** — List directory contents (safe)
|
|
303
332
|
- **search_files** — Search for patterns across files (safe)
|
|
304
333
|
- **run_command** — Execute shell commands (requires approval)
|
|
305
334
|
|
|
335
|
+
Plus any tools from connected MCP servers (databases, APIs, GitHub, etc.)
|
|
336
|
+
|
|
306
337
|
## Project Context
|
|
307
338
|
|
|
308
339
|
Drop a `CODEMAXXING.md` file in your project root to give the model extra context about your codebase, conventions, or instructions. It's automatically included in the system prompt.
|
|
@@ -311,8 +342,10 @@ Drop a `CODEMAXXING.md` file in your project root to give the model extra contex
|
|
|
311
342
|
|
|
312
343
|
- **Runtime:** Node.js + TypeScript
|
|
313
344
|
- **TUI:** [Ink](https://github.com/vadimdemedes/ink) (React for the terminal)
|
|
314
|
-
- **LLM
|
|
345
|
+
- **LLM SDKs:** [OpenAI SDK](https://github.com/openai/openai-node) + [Anthropic SDK](https://github.com/anthropics/anthropic-sdk-typescript)
|
|
346
|
+
- **MCP:** [@modelcontextprotocol/sdk](https://github.com/modelcontextprotocol/typescript-sdk)
|
|
315
347
|
- **Sessions:** [better-sqlite3](https://github.com/WiseLibs/better-sqlite3)
|
|
348
|
+
- **Local LLM:** Ollama integration (auto-install, pull, manage)
|
|
316
349
|
- **Zero cloud dependencies** — everything runs locally
|
|
317
350
|
|
|
318
351
|
## Inspired By
|
package/dist/config.d.ts
CHANGED
|
@@ -41,6 +41,17 @@ export declare function getConfigPath(): string;
|
|
|
41
41
|
/**
|
|
42
42
|
* Auto-detect local LLM servers
|
|
43
43
|
*/
|
|
44
|
+
export type DetectionResult = {
|
|
45
|
+
status: "connected";
|
|
46
|
+
provider: ProviderConfig;
|
|
47
|
+
} | {
|
|
48
|
+
status: "no-models";
|
|
49
|
+
serverName: string;
|
|
50
|
+
baseUrl: string;
|
|
51
|
+
} | {
|
|
52
|
+
status: "no-server";
|
|
53
|
+
};
|
|
54
|
+
export declare function detectLocalProviderDetailed(): Promise<DetectionResult>;
|
|
44
55
|
export declare function detectLocalProvider(): Promise<ProviderConfig | null>;
|
|
45
56
|
/**
|
|
46
57
|
* List available models from a provider endpoint
|
package/dist/config.js
CHANGED
|
@@ -149,9 +149,49 @@ export function applyOverrides(config, args) {
|
|
|
149
149
|
export function getConfigPath() {
|
|
150
150
|
return CONFIG_FILE;
|
|
151
151
|
}
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
152
|
+
export async function detectLocalProviderDetailed() {
|
|
153
|
+
const endpoints = [
|
|
154
|
+
{ name: "LM Studio", url: "http://localhost:1234/v1" },
|
|
155
|
+
{ name: "Ollama", url: "http://localhost:11434/v1" },
|
|
156
|
+
{ name: "vLLM", url: "http://localhost:8000/v1" },
|
|
157
|
+
];
|
|
158
|
+
let serverFound = null;
|
|
159
|
+
for (const endpoint of endpoints) {
|
|
160
|
+
try {
|
|
161
|
+
const controller = new AbortController();
|
|
162
|
+
const timeout = setTimeout(() => controller.abort(), 2000);
|
|
163
|
+
const res = await fetch(`${endpoint.url}/models`, {
|
|
164
|
+
signal: controller.signal,
|
|
165
|
+
});
|
|
166
|
+
clearTimeout(timeout);
|
|
167
|
+
if (res.ok) {
|
|
168
|
+
const data = (await res.json());
|
|
169
|
+
const models = data.data ?? [];
|
|
170
|
+
if (models.length === 0) {
|
|
171
|
+
// Server is up but no models — remember it but keep looking
|
|
172
|
+
if (!serverFound)
|
|
173
|
+
serverFound = endpoint;
|
|
174
|
+
continue;
|
|
175
|
+
}
|
|
176
|
+
return {
|
|
177
|
+
status: "connected",
|
|
178
|
+
provider: {
|
|
179
|
+
baseUrl: endpoint.url,
|
|
180
|
+
apiKey: "not-needed",
|
|
181
|
+
model: models[0].id,
|
|
182
|
+
},
|
|
183
|
+
};
|
|
184
|
+
}
|
|
185
|
+
}
|
|
186
|
+
catch {
|
|
187
|
+
// Server not running, try next
|
|
188
|
+
}
|
|
189
|
+
}
|
|
190
|
+
if (serverFound) {
|
|
191
|
+
return { status: "no-models", serverName: serverFound.name, baseUrl: serverFound.url };
|
|
192
|
+
}
|
|
193
|
+
return { status: "no-server" };
|
|
194
|
+
}
|
|
155
195
|
export async function detectLocalProvider() {
|
|
156
196
|
const endpoints = [
|
|
157
197
|
{ name: "LM Studio", url: "http://localhost:1234/v1" },
|
|
@@ -169,7 +209,11 @@ export async function detectLocalProvider() {
|
|
|
169
209
|
if (res.ok) {
|
|
170
210
|
const data = (await res.json());
|
|
171
211
|
const models = data.data ?? [];
|
|
172
|
-
|
|
212
|
+
if (models.length === 0) {
|
|
213
|
+
// Server is up but no models available — don't fake a connection
|
|
214
|
+
continue;
|
|
215
|
+
}
|
|
216
|
+
const model = models[0].id;
|
|
173
217
|
return {
|
|
174
218
|
baseUrl: endpoint.url,
|
|
175
219
|
apiKey: "not-needed",
|
package/dist/index.js
CHANGED
|
@@ -5,7 +5,7 @@ import { render, Box, Text, useInput, useApp, useStdout } from "ink";
|
|
|
5
5
|
import { EventEmitter } from "events";
|
|
6
6
|
import TextInput from "ink-text-input";
|
|
7
7
|
import { CodingAgent } from "./agent.js";
|
|
8
|
-
import { loadConfig, saveConfig, detectLocalProvider, parseCLIArgs, applyOverrides, listModels } from "./config.js";
|
|
8
|
+
import { loadConfig, saveConfig, detectLocalProvider, detectLocalProviderDetailed, parseCLIArgs, applyOverrides, listModels } from "./config.js";
|
|
9
9
|
import { listSessions, getSession, loadMessages, deleteSession } from "./utils/sessions.js";
|
|
10
10
|
import { execSync } from "child_process";
|
|
11
11
|
import { isGitRepo, getBranch, getStatus, getDiff, undoLastCommit } from "./utils/git.js";
|
|
@@ -79,11 +79,37 @@ const SLASH_COMMANDS = [
|
|
|
79
79
|
];
|
|
80
80
|
const SPINNER_FRAMES = ["⣾", "⣽", "⣻", "⢿", "⡿", "⣟", "⣯", "⣷"];
|
|
81
81
|
const SPINNER_MESSAGES = [
|
|
82
|
+
// OG
|
|
82
83
|
"Locking in...", "Cooking...", "Maxxing...", "In the zone...",
|
|
83
84
|
"Yapping...", "Frame mogging...", "Jester gooning...", "Gooning...",
|
|
84
85
|
"Doing back flips...", "Jester maxxing...", "Getting baked...",
|
|
85
86
|
"Blasting tren...", "Pumping...", "Wondering if I should actually do this...",
|
|
86
87
|
"Hacking the main frame...", "Codemaxxing...", "Vibe coding...", "Running a marathon...",
|
|
88
|
+
// Gym/Looksmaxxing
|
|
89
|
+
"Mewing aggressively...", "Looksmaxxing your codebase...", "Hitting a PR on this function...",
|
|
90
|
+
"Eating 4000 calories of code...", "Creatine loading...", "On my bulk arc...",
|
|
91
|
+
"Warming up the deadlift...",
|
|
92
|
+
// Brainrot/Skibidi
|
|
93
|
+
"Going full skibidi...", "Sigma grinding...", "Rizzing up the compiler...",
|
|
94
|
+
"No cap processing...", "Main character coding...", "It's giving implementation...",
|
|
95
|
+
"This code is bussin fr fr...", "Aura check in progress...", "Erm what the sigma...",
|
|
96
|
+
// Deranged/Unhinged
|
|
97
|
+
"Ascending to a higher plane...", "Achieving final form...", "Third eye compiling...",
|
|
98
|
+
"Astral projecting through your repo...", "Becoming one with the codebase...",
|
|
99
|
+
"Having a spiritual awakening...", "Entering the shadow realm...", "Going goblin mode...",
|
|
100
|
+
"Deleting System32... jk...", "Sacrificing tokens to the GPU gods...",
|
|
101
|
+
"Summoning the machine spirit...",
|
|
102
|
+
// Self-aware/Meta
|
|
103
|
+
"Pretending to think really hard...", "Staring at your code judgmentally...",
|
|
104
|
+
"Rethinking my career choices...", "Having an existential crisis...",
|
|
105
|
+
"Hoping this actually works...", "Praying to the stack overflow gods...",
|
|
106
|
+
"Copying from the internet with dignity...",
|
|
107
|
+
// Pure Chaos
|
|
108
|
+
"Doing hot yoga in the terminal...", "Microdosing your dependencies...",
|
|
109
|
+
"Running on 3 hours of sleep...", "Speedrunning your deadline...",
|
|
110
|
+
"Built different rn...", "That's crazy let me cook...",
|
|
111
|
+
"Absolutely feral right now...", "Ong no cap fr fr...",
|
|
112
|
+
"Living rent free in your RAM...", "Ate and left no crumbs...",
|
|
87
113
|
];
|
|
88
114
|
// ── Neon Spinner ──
|
|
89
115
|
function NeonSpinner({ message, colors }) {
|
|
@@ -220,15 +246,21 @@ function App() {
|
|
|
220
246
|
if (provider.model === "auto" || (provider.baseUrl === "http://localhost:1234/v1" && !cliArgs.baseUrl)) {
|
|
221
247
|
info.push("Detecting local LLM server...");
|
|
222
248
|
setConnectionInfo([...info]);
|
|
223
|
-
const
|
|
224
|
-
if (
|
|
249
|
+
const detection = await detectLocalProviderDetailed();
|
|
250
|
+
if (detection.status === "connected") {
|
|
225
251
|
// Keep CLI model override if specified
|
|
226
252
|
if (cliArgs.model)
|
|
227
|
-
|
|
228
|
-
provider =
|
|
253
|
+
detection.provider.model = cliArgs.model;
|
|
254
|
+
provider = detection.provider;
|
|
229
255
|
info.push(`✔ Connected to ${provider.baseUrl} → ${provider.model}`);
|
|
230
256
|
setConnectionInfo([...info]);
|
|
231
257
|
}
|
|
258
|
+
else if (detection.status === "no-models") {
|
|
259
|
+
info.push(`⚠ ${detection.serverName} is running but has no models. Use /ollama pull to download one.`);
|
|
260
|
+
setConnectionInfo([...info]);
|
|
261
|
+
setReady(true);
|
|
262
|
+
return;
|
|
263
|
+
}
|
|
232
264
|
else {
|
|
233
265
|
info.push("✗ No local LLM server found.");
|
|
234
266
|
setConnectionInfo([...info]);
|
|
@@ -903,8 +935,9 @@ function App() {
|
|
|
903
935
|
}
|
|
904
936
|
return;
|
|
905
937
|
}
|
|
906
|
-
if (trimmed === "/model"
|
|
938
|
+
if (trimmed === "/model") {
|
|
907
939
|
// Show picker of available models
|
|
940
|
+
addMsg("info", "Fetching available models...");
|
|
908
941
|
try {
|
|
909
942
|
const ollamaModels = await listInstalledModelsDetailed();
|
|
910
943
|
if (ollamaModels.length > 0) {
|
|
@@ -913,18 +946,25 @@ function App() {
|
|
|
913
946
|
return;
|
|
914
947
|
}
|
|
915
948
|
}
|
|
916
|
-
catch {
|
|
949
|
+
catch (err) {
|
|
950
|
+
// Ollama not available or failed, try provider
|
|
951
|
+
}
|
|
917
952
|
// Fallback: try provider's model list
|
|
918
|
-
|
|
919
|
-
|
|
920
|
-
|
|
921
|
-
|
|
922
|
-
|
|
923
|
-
|
|
953
|
+
if (providerRef.current?.baseUrl && providerRef.current.baseUrl !== "auto") {
|
|
954
|
+
try {
|
|
955
|
+
const providerModels = await listModels(providerRef.current.baseUrl, providerRef.current.apiKey || "");
|
|
956
|
+
if (providerModels.length > 0) {
|
|
957
|
+
setModelPicker(providerModels);
|
|
958
|
+
setModelPickerIndex(0);
|
|
959
|
+
return;
|
|
960
|
+
}
|
|
961
|
+
}
|
|
962
|
+
catch (err) {
|
|
963
|
+
// Provider fetch failed
|
|
924
964
|
}
|
|
925
965
|
}
|
|
926
|
-
|
|
927
|
-
addMsg("
|
|
966
|
+
// No models found anywhere
|
|
967
|
+
addMsg("error", "No models available. Download one with /ollama pull or configure a provider.");
|
|
928
968
|
return;
|
|
929
969
|
}
|
|
930
970
|
if (trimmed.startsWith("/model ")) {
|
package/package.json
CHANGED
package/src/config.ts
CHANGED
|
@@ -193,6 +193,57 @@ export function getConfigPath(): string {
|
|
|
193
193
|
/**
|
|
194
194
|
* Auto-detect local LLM servers
|
|
195
195
|
*/
|
|
196
|
+
export type DetectionResult =
|
|
197
|
+
| { status: "connected"; provider: ProviderConfig }
|
|
198
|
+
| { status: "no-models"; serverName: string; baseUrl: string }
|
|
199
|
+
| { status: "no-server" };
|
|
200
|
+
|
|
201
|
+
export async function detectLocalProviderDetailed(): Promise<DetectionResult> {
|
|
202
|
+
const endpoints = [
|
|
203
|
+
{ name: "LM Studio", url: "http://localhost:1234/v1" },
|
|
204
|
+
{ name: "Ollama", url: "http://localhost:11434/v1" },
|
|
205
|
+
{ name: "vLLM", url: "http://localhost:8000/v1" },
|
|
206
|
+
];
|
|
207
|
+
|
|
208
|
+
let serverFound: { name: string; url: string } | null = null;
|
|
209
|
+
|
|
210
|
+
for (const endpoint of endpoints) {
|
|
211
|
+
try {
|
|
212
|
+
const controller = new AbortController();
|
|
213
|
+
const timeout = setTimeout(() => controller.abort(), 2000);
|
|
214
|
+
const res = await fetch(`${endpoint.url}/models`, {
|
|
215
|
+
signal: controller.signal,
|
|
216
|
+
});
|
|
217
|
+
clearTimeout(timeout);
|
|
218
|
+
|
|
219
|
+
if (res.ok) {
|
|
220
|
+
const data = (await res.json()) as { data?: Array<{ id: string }> };
|
|
221
|
+
const models = data.data ?? [];
|
|
222
|
+
if (models.length === 0) {
|
|
223
|
+
// Server is up but no models — remember it but keep looking
|
|
224
|
+
if (!serverFound) serverFound = endpoint;
|
|
225
|
+
continue;
|
|
226
|
+
}
|
|
227
|
+
return {
|
|
228
|
+
status: "connected",
|
|
229
|
+
provider: {
|
|
230
|
+
baseUrl: endpoint.url,
|
|
231
|
+
apiKey: "not-needed",
|
|
232
|
+
model: models[0]!.id,
|
|
233
|
+
},
|
|
234
|
+
};
|
|
235
|
+
}
|
|
236
|
+
} catch {
|
|
237
|
+
// Server not running, try next
|
|
238
|
+
}
|
|
239
|
+
}
|
|
240
|
+
|
|
241
|
+
if (serverFound) {
|
|
242
|
+
return { status: "no-models", serverName: serverFound.name, baseUrl: serverFound.url };
|
|
243
|
+
}
|
|
244
|
+
return { status: "no-server" };
|
|
245
|
+
}
|
|
246
|
+
|
|
196
247
|
export async function detectLocalProvider(): Promise<ProviderConfig | null> {
|
|
197
248
|
const endpoints = [
|
|
198
249
|
{ name: "LM Studio", url: "http://localhost:1234/v1" },
|
|
@@ -212,7 +263,11 @@ export async function detectLocalProvider(): Promise<ProviderConfig | null> {
|
|
|
212
263
|
if (res.ok) {
|
|
213
264
|
const data = (await res.json()) as { data?: Array<{ id: string }> };
|
|
214
265
|
const models = data.data ?? [];
|
|
215
|
-
|
|
266
|
+
if (models.length === 0) {
|
|
267
|
+
// Server is up but no models available — don't fake a connection
|
|
268
|
+
continue;
|
|
269
|
+
}
|
|
270
|
+
const model = models[0]!.id;
|
|
216
271
|
return {
|
|
217
272
|
baseUrl: endpoint.url,
|
|
218
273
|
apiKey: "not-needed",
|
package/src/index.tsx
CHANGED
|
@@ -5,7 +5,7 @@ import { render, Box, Text, useInput, useApp, useStdout } from "ink";
|
|
|
5
5
|
import { EventEmitter } from "events";
|
|
6
6
|
import TextInput from "ink-text-input";
|
|
7
7
|
import { CodingAgent } from "./agent.js";
|
|
8
|
-
import { loadConfig, saveConfig, detectLocalProvider, parseCLIArgs, applyOverrides, listModels } from "./config.js";
|
|
8
|
+
import { loadConfig, saveConfig, detectLocalProvider, detectLocalProviderDetailed, parseCLIArgs, applyOverrides, listModels } from "./config.js";
|
|
9
9
|
import { listSessions, getSession, loadMessages, deleteSession } from "./utils/sessions.js";
|
|
10
10
|
import { execSync } from "child_process";
|
|
11
11
|
import { isGitRepo, getBranch, getStatus, getDiff, undoLastCommit } from "./utils/git.js";
|
|
@@ -81,11 +81,37 @@ const SLASH_COMMANDS = [
|
|
|
81
81
|
const SPINNER_FRAMES = ["⣾", "⣽", "⣻", "⢿", "⡿", "⣟", "⣯", "⣷"];
|
|
82
82
|
|
|
83
83
|
const SPINNER_MESSAGES = [
|
|
84
|
+
// OG
|
|
84
85
|
"Locking in...", "Cooking...", "Maxxing...", "In the zone...",
|
|
85
86
|
"Yapping...", "Frame mogging...", "Jester gooning...", "Gooning...",
|
|
86
87
|
"Doing back flips...", "Jester maxxing...", "Getting baked...",
|
|
87
88
|
"Blasting tren...", "Pumping...", "Wondering if I should actually do this...",
|
|
88
89
|
"Hacking the main frame...", "Codemaxxing...", "Vibe coding...", "Running a marathon...",
|
|
90
|
+
// Gym/Looksmaxxing
|
|
91
|
+
"Mewing aggressively...", "Looksmaxxing your codebase...", "Hitting a PR on this function...",
|
|
92
|
+
"Eating 4000 calories of code...", "Creatine loading...", "On my bulk arc...",
|
|
93
|
+
"Warming up the deadlift...",
|
|
94
|
+
// Brainrot/Skibidi
|
|
95
|
+
"Going full skibidi...", "Sigma grinding...", "Rizzing up the compiler...",
|
|
96
|
+
"No cap processing...", "Main character coding...", "It's giving implementation...",
|
|
97
|
+
"This code is bussin fr fr...", "Aura check in progress...", "Erm what the sigma...",
|
|
98
|
+
// Deranged/Unhinged
|
|
99
|
+
"Ascending to a higher plane...", "Achieving final form...", "Third eye compiling...",
|
|
100
|
+
"Astral projecting through your repo...", "Becoming one with the codebase...",
|
|
101
|
+
"Having a spiritual awakening...", "Entering the shadow realm...", "Going goblin mode...",
|
|
102
|
+
"Deleting System32... jk...", "Sacrificing tokens to the GPU gods...",
|
|
103
|
+
"Summoning the machine spirit...",
|
|
104
|
+
// Self-aware/Meta
|
|
105
|
+
"Pretending to think really hard...", "Staring at your code judgmentally...",
|
|
106
|
+
"Rethinking my career choices...", "Having an existential crisis...",
|
|
107
|
+
"Hoping this actually works...", "Praying to the stack overflow gods...",
|
|
108
|
+
"Copying from the internet with dignity...",
|
|
109
|
+
// Pure Chaos
|
|
110
|
+
"Doing hot yoga in the terminal...", "Microdosing your dependencies...",
|
|
111
|
+
"Running on 3 hours of sleep...", "Speedrunning your deadline...",
|
|
112
|
+
"Built different rn...", "That's crazy let me cook...",
|
|
113
|
+
"Absolutely feral right now...", "Ong no cap fr fr...",
|
|
114
|
+
"Living rent free in your RAM...", "Ate and left no crumbs...",
|
|
89
115
|
];
|
|
90
116
|
|
|
91
117
|
// ── Neon Spinner ──
|
|
@@ -264,13 +290,18 @@ function App() {
|
|
|
264
290
|
if (provider.model === "auto" || (provider.baseUrl === "http://localhost:1234/v1" && !cliArgs.baseUrl)) {
|
|
265
291
|
info.push("Detecting local LLM server...");
|
|
266
292
|
setConnectionInfo([...info]);
|
|
267
|
-
const
|
|
268
|
-
if (
|
|
293
|
+
const detection = await detectLocalProviderDetailed();
|
|
294
|
+
if (detection.status === "connected") {
|
|
269
295
|
// Keep CLI model override if specified
|
|
270
|
-
if (cliArgs.model)
|
|
271
|
-
provider =
|
|
296
|
+
if (cliArgs.model) detection.provider.model = cliArgs.model;
|
|
297
|
+
provider = detection.provider;
|
|
272
298
|
info.push(`✔ Connected to ${provider.baseUrl} → ${provider.model}`);
|
|
273
299
|
setConnectionInfo([...info]);
|
|
300
|
+
} else if (detection.status === "no-models") {
|
|
301
|
+
info.push(`⚠ ${detection.serverName} is running but has no models. Use /ollama pull to download one.`);
|
|
302
|
+
setConnectionInfo([...info]);
|
|
303
|
+
setReady(true);
|
|
304
|
+
return;
|
|
274
305
|
} else {
|
|
275
306
|
info.push("✗ No local LLM server found.");
|
|
276
307
|
setConnectionInfo([...info]);
|
|
@@ -941,8 +972,9 @@ function App() {
|
|
|
941
972
|
}
|
|
942
973
|
return;
|
|
943
974
|
}
|
|
944
|
-
if (trimmed === "/model"
|
|
975
|
+
if (trimmed === "/model") {
|
|
945
976
|
// Show picker of available models
|
|
977
|
+
addMsg("info", "Fetching available models...");
|
|
946
978
|
try {
|
|
947
979
|
const ollamaModels = await listInstalledModelsDetailed();
|
|
948
980
|
if (ollamaModels.length > 0) {
|
|
@@ -950,17 +982,26 @@ function App() {
|
|
|
950
982
|
setModelPickerIndex(0);
|
|
951
983
|
return;
|
|
952
984
|
}
|
|
953
|
-
} catch {
|
|
985
|
+
} catch (err) {
|
|
986
|
+
// Ollama not available or failed, try provider
|
|
987
|
+
}
|
|
988
|
+
|
|
954
989
|
// Fallback: try provider's model list
|
|
955
|
-
|
|
956
|
-
|
|
957
|
-
|
|
958
|
-
|
|
959
|
-
|
|
960
|
-
|
|
990
|
+
if (providerRef.current?.baseUrl && providerRef.current.baseUrl !== "auto") {
|
|
991
|
+
try {
|
|
992
|
+
const providerModels = await listModels(providerRef.current.baseUrl, providerRef.current.apiKey || "");
|
|
993
|
+
if (providerModels.length > 0) {
|
|
994
|
+
setModelPicker(providerModels);
|
|
995
|
+
setModelPickerIndex(0);
|
|
996
|
+
return;
|
|
997
|
+
}
|
|
998
|
+
} catch (err) {
|
|
999
|
+
// Provider fetch failed
|
|
961
1000
|
}
|
|
962
|
-
}
|
|
963
|
-
|
|
1001
|
+
}
|
|
1002
|
+
|
|
1003
|
+
// No models found anywhere
|
|
1004
|
+
addMsg("error", "No models available. Download one with /ollama pull or configure a provider.");
|
|
964
1005
|
return;
|
|
965
1006
|
}
|
|
966
1007
|
if (trimmed.startsWith("/model ")) {
|