@darksol/terminal 0.15.0 → 0.15.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +2 -2
- package/package.json +1 -1
- package/src/config/keys.js +9 -1
- package/src/llm/engine.js +7 -0
- package/src/llm/models.js +10 -0
package/README.md
CHANGED
|
@@ -498,7 +498,7 @@ darksol ai analyze AERO
|
|
|
498
498
|
darksol ai chat --provider ollama --model llama3
|
|
499
499
|
```
|
|
500
500
|
|
|
501
|
-
**Supported providers:** OpenAI, Anthropic, OpenRouter, Bankr LLM Gateway, Ollama (local = free)
|
|
501
|
+
**Supported providers:** OpenAI, Anthropic, OpenRouter, NVIDIA NIM, Bankr LLM Gateway, MiniMax, Ollama (local = free)
|
|
502
502
|
|
|
503
503
|
The AI gets live market context (prices from DexScreener), knows your config (chain, slippage, wallet), and returns structured intents with confidence scores and risk warnings.
|
|
504
504
|
|
|
@@ -524,7 +524,7 @@ darksol keys remove openai
|
|
|
524
524
|
**Supported services:**
|
|
525
525
|
| Category | Services |
|
|
526
526
|
|----------|----------|
|
|
527
|
-
| LLM | OpenAI, Anthropic, OpenRouter, Bankr LLM Gateway, Ollama |
|
|
527
|
+
| LLM | OpenAI, Anthropic, OpenRouter, NVIDIA NIM, Bankr LLM Gateway, MiniMax, Ollama |
|
|
528
528
|
| Data | CoinGecko Pro, DexScreener, DefiLlama |
|
|
529
529
|
| RPC | Alchemy, Infura, QuickNode |
|
|
530
530
|
| Trading | 1inch, ParaSwap |
|
package/package.json
CHANGED
package/src/config/keys.js
CHANGED
|
@@ -92,6 +92,14 @@ export const SERVICES = {
|
|
|
92
92
|
docsUrl: 'https://platform.minimax.io/docs/guides/models-intro',
|
|
93
93
|
validate: (key) => key.length > 10,
|
|
94
94
|
},
|
|
95
|
+
nvidia: {
|
|
96
|
+
name: 'NVIDIA NIM',
|
|
97
|
+
category: 'llm',
|
|
98
|
+
description: 'NVIDIA NIM — Llama, Nemotron, Mistral via build.nvidia.com',
|
|
99
|
+
envVar: 'NVIDIA_API_KEY',
|
|
100
|
+
docsUrl: 'https://build.nvidia.com',
|
|
101
|
+
validate: (key) => key.startsWith('nvapi-') || key.length > 10,
|
|
102
|
+
},
|
|
95
103
|
ollama: {
|
|
96
104
|
name: 'Ollama (Local)',
|
|
97
105
|
category: 'llm',
|
|
@@ -454,7 +462,7 @@ export function hasKey(service) {
|
|
|
454
462
|
*/
|
|
455
463
|
export function hasAnyLLM() {
|
|
456
464
|
// Cloud providers — need real validated API keys
|
|
457
|
-
if (['openai', 'anthropic', 'openrouter', 'minimax', 'bankr'].some(s => hasKey(s))) return true;
|
|
465
|
+
if (['openai', 'anthropic', 'openrouter', 'minimax', 'nvidia', 'bankr'].some(s => hasKey(s))) return true;
|
|
458
466
|
// Ollama — check if explicitly configured via hasKey (validates URL format)
|
|
459
467
|
if (hasKey('ollama')) return true;
|
|
460
468
|
return false;
|
package/src/llm/engine.js
CHANGED
|
@@ -47,6 +47,13 @@ const PROVIDERS = {
|
|
|
47
47
|
parseResponse: (data) => data.choices?.[0]?.message?.content,
|
|
48
48
|
parseUsage: (data) => data.usage,
|
|
49
49
|
},
|
|
50
|
+
nvidia: {
|
|
51
|
+
url: 'https://integrate.api.nvidia.com/v1/chat/completions',
|
|
52
|
+
defaultModel: getProviderDefaultModel('nvidia'),
|
|
53
|
+
authHeader: (key) => ({ Authorization: `Bearer ${key}` }),
|
|
54
|
+
parseResponse: (data) => data.choices?.[0]?.message?.content,
|
|
55
|
+
parseUsage: (data) => data.usage,
|
|
56
|
+
},
|
|
50
57
|
ollama: {
|
|
51
58
|
url: null,
|
|
52
59
|
defaultModel: getProviderDefaultModel('ollama'),
|
package/src/llm/models.js
CHANGED
|
@@ -39,6 +39,16 @@ export const MODEL_CATALOG = {
|
|
|
39
39
|
{ value: 'MiniMax-M2', label: 'MiniMax-M2', desc: 'agentic, advanced reasoning' },
|
|
40
40
|
],
|
|
41
41
|
},
|
|
42
|
+
nvidia: {
|
|
43
|
+
defaultModel: 'nvidia/llama-3.1-nemotron-70b-instruct',
|
|
44
|
+
choices: [
|
|
45
|
+
{ value: 'nvidia/llama-3.1-nemotron-70b-instruct', label: 'Nemotron 70B', desc: 'high reasoning + tool calling' },
|
|
46
|
+
{ value: 'meta/llama-3.1-8b-instruct', label: 'Llama 3.1 8B', desc: 'fast, free tier friendly' },
|
|
47
|
+
{ value: 'meta/llama-3.1-70b-instruct', label: 'Llama 3.1 70B', desc: 'strong general purpose' },
|
|
48
|
+
{ value: 'mistralai/mistral-large-2-instruct', label: 'Mistral Large 2', desc: 'multilingual, coding' },
|
|
49
|
+
{ value: 'nvidia/nemotron-mini-4b-instruct', label: 'Nemotron Mini 4B', desc: 'ultra-fast, lightweight' },
|
|
50
|
+
],
|
|
51
|
+
},
|
|
42
52
|
ollama: {
|
|
43
53
|
defaultModel: 'llama3.1',
|
|
44
54
|
textInput: true,
|