converse-mcp-server 2.1.0 → 2.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/converse.js +0 -0
- package/docs/PROVIDERS.md +2 -1
- package/package.json +52 -52
- package/src/providers/anthropic.js +19 -3
- package/src/providers/openai.js +18 -4
package/bin/converse.js
CHANGED
|
File without changes
|
package/docs/PROVIDERS.md
CHANGED
|
@@ -40,6 +40,7 @@ This guide documents all supported AI providers in the Converse MCP Server and t
|
|
|
40
40
|
- `claude-opus-4-1-20250805` - Highest intelligence with extended thinking (32K output)
|
|
41
41
|
- `claude-sonnet-4-20250514` - Balanced performance with extended thinking (64K output)
|
|
42
42
|
- `claude-3-7-sonnet-20250219` - Enhanced 3.x generation with thinking (64K output)
|
|
43
|
+
- `claude-haiku-4-5-20251001` - Fast and intelligent with extended thinking (64K output)
|
|
43
44
|
- `claude-3-5-sonnet-20241022` - Fast and intelligent model (8K output)
|
|
44
45
|
- `claude-3-5-haiku-20241022` - Fastest Claude model (8K output)
|
|
45
46
|
- `claude-3-opus-20240229`, `claude-3-sonnet-20240229`, `claude-3-haiku-20240307` - Previous generation
|
|
@@ -165,7 +166,7 @@ CODEX_DEFAULT_MODEL=gpt-5-codex # Default: gpt-5-codex
|
|
|
165
166
|
All providers support streaming responses for real-time output.
|
|
166
167
|
|
|
167
168
|
### Image Support
|
|
168
|
-
- **Full Support**: OpenAI, Google, X.AI (Grok-4), Anthropic (Claude-3.5-Sonnet, Claude-3-Opus)
|
|
169
|
+
- **Full Support**: OpenAI, Google, X.AI (Grok-4), Anthropic (Claude-4 series, Claude-3.5-Sonnet, Claude-3-Opus)
|
|
169
170
|
- **Via OpenRouter**: Depends on the underlying model
|
|
170
171
|
- **No Support**: DeepSeek, Mistral (except Large), Claude-3.5-Haiku, Codex
|
|
171
172
|
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "converse-mcp-server",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.3.0",
|
|
4
4
|
"description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "src/index.js",
|
|
@@ -11,6 +11,56 @@
|
|
|
11
11
|
"engines": {
|
|
12
12
|
"node": ">=20.0.0"
|
|
13
13
|
},
|
|
14
|
+
"keywords": [
|
|
15
|
+
"mcp",
|
|
16
|
+
"server",
|
|
17
|
+
"ai",
|
|
18
|
+
"chat",
|
|
19
|
+
"consensus",
|
|
20
|
+
"openai",
|
|
21
|
+
"google",
|
|
22
|
+
"gemini",
|
|
23
|
+
"grok"
|
|
24
|
+
],
|
|
25
|
+
"author": "Converse MCP Server",
|
|
26
|
+
"license": "MIT",
|
|
27
|
+
"homepage": "https://github.com/FallDownTheSystem/converse#readme",
|
|
28
|
+
"repository": {
|
|
29
|
+
"type": "git",
|
|
30
|
+
"url": "git+https://github.com/FallDownTheSystem/converse.git"
|
|
31
|
+
},
|
|
32
|
+
"bugs": {
|
|
33
|
+
"url": "https://github.com/FallDownTheSystem/converse/issues"
|
|
34
|
+
},
|
|
35
|
+
"files": [
|
|
36
|
+
"src/",
|
|
37
|
+
"bin/",
|
|
38
|
+
"docs/",
|
|
39
|
+
"README.md",
|
|
40
|
+
".env.example"
|
|
41
|
+
],
|
|
42
|
+
"dependencies": {
|
|
43
|
+
"@anthropic-ai/sdk": "^0.65.0",
|
|
44
|
+
"@google/genai": "^1.22.0",
|
|
45
|
+
"@mistralai/mistralai": "^1.10.0",
|
|
46
|
+
"@modelcontextprotocol/sdk": "^1.19.1",
|
|
47
|
+
"@openai/codex-sdk": "^0.45.0",
|
|
48
|
+
"cors": "^2.8.5",
|
|
49
|
+
"dotenv": "^17.2.3",
|
|
50
|
+
"express": "^5.1.0",
|
|
51
|
+
"lru-cache": "^11.2.2",
|
|
52
|
+
"openai": "^6.1.0",
|
|
53
|
+
"p-limit": "^4.0.0",
|
|
54
|
+
"vite": "^7.1.9"
|
|
55
|
+
},
|
|
56
|
+
"devDependencies": {
|
|
57
|
+
"@vitest/coverage-v8": "^3.2.4",
|
|
58
|
+
"cross-env": "^10.1.0",
|
|
59
|
+
"eslint": "^9.36.0",
|
|
60
|
+
"prettier": "^3.6.2",
|
|
61
|
+
"rimraf": "^6.0.1",
|
|
62
|
+
"vitest": "^3.2.4"
|
|
63
|
+
},
|
|
14
64
|
"scripts": {
|
|
15
65
|
"kill-server": "node scripts/kill-server.js",
|
|
16
66
|
"start": "npm run kill-server && node src/index.js",
|
|
@@ -64,55 +114,5 @@
|
|
|
64
114
|
"validate:fix": "node scripts/validate.js --fix",
|
|
65
115
|
"validate:fast": "node scripts/validate.js --skip-tests --skip-lint",
|
|
66
116
|
"precommit": "npm run validate"
|
|
67
|
-
},
|
|
68
|
-
"keywords": [
|
|
69
|
-
"mcp",
|
|
70
|
-
"server",
|
|
71
|
-
"ai",
|
|
72
|
-
"chat",
|
|
73
|
-
"consensus",
|
|
74
|
-
"openai",
|
|
75
|
-
"google",
|
|
76
|
-
"gemini",
|
|
77
|
-
"grok"
|
|
78
|
-
],
|
|
79
|
-
"author": "Converse MCP Server",
|
|
80
|
-
"license": "MIT",
|
|
81
|
-
"homepage": "https://github.com/FallDownTheSystem/converse#readme",
|
|
82
|
-
"repository": {
|
|
83
|
-
"type": "git",
|
|
84
|
-
"url": "git+https://github.com/FallDownTheSystem/converse.git"
|
|
85
|
-
},
|
|
86
|
-
"bugs": {
|
|
87
|
-
"url": "https://github.com/FallDownTheSystem/converse/issues"
|
|
88
|
-
},
|
|
89
|
-
"files": [
|
|
90
|
-
"src/",
|
|
91
|
-
"bin/",
|
|
92
|
-
"docs/",
|
|
93
|
-
"README.md",
|
|
94
|
-
".env.example"
|
|
95
|
-
],
|
|
96
|
-
"dependencies": {
|
|
97
|
-
"@anthropic-ai/sdk": "^0.65.0",
|
|
98
|
-
"@google/genai": "^1.22.0",
|
|
99
|
-
"@mistralai/mistralai": "^1.10.0",
|
|
100
|
-
"@modelcontextprotocol/sdk": "^1.19.1",
|
|
101
|
-
"@openai/codex-sdk": "^0.45.0",
|
|
102
|
-
"cors": "^2.8.5",
|
|
103
|
-
"dotenv": "^17.2.3",
|
|
104
|
-
"express": "^5.1.0",
|
|
105
|
-
"lru-cache": "^11.2.2",
|
|
106
|
-
"openai": "^6.1.0",
|
|
107
|
-
"p-limit": "^4.0.0",
|
|
108
|
-
"vite": "^7.1.9"
|
|
109
|
-
},
|
|
110
|
-
"devDependencies": {
|
|
111
|
-
"@vitest/coverage-v8": "^3.2.4",
|
|
112
|
-
"cross-env": "^10.1.0",
|
|
113
|
-
"eslint": "^9.36.0",
|
|
114
|
-
"prettier": "^3.6.2",
|
|
115
|
-
"rimraf": "^6.0.1",
|
|
116
|
-
"vitest": "^3.2.4"
|
|
117
117
|
}
|
|
118
|
-
}
|
|
118
|
+
}
|
|
@@ -44,7 +44,7 @@ const SUPPORTED_MODELS = {
|
|
|
44
44
|
timeout: 300000,
|
|
45
45
|
supports1MContext: true, // Beta 1M context support
|
|
46
46
|
description: 'Claude Sonnet 4.5 - Latest Sonnet with enhanced intelligence and optional 1M context',
|
|
47
|
-
aliases: ['claude-4.5-sonnet', 'sonnet-4.5', 'claude-sonnet-4.5', 'sonnet4.5', 'claude-sonnet-4-5']
|
|
47
|
+
aliases: ['claude-4.5-sonnet', 'sonnet-4.5', 'claude-sonnet-4.5', 'sonnet4.5', 'claude-sonnet-4-5', 'sonnet', 'claude-sonnet']
|
|
48
48
|
},
|
|
49
49
|
'claude-sonnet-4-20250514': {
|
|
50
50
|
modelName: 'claude-sonnet-4-20250514',
|
|
@@ -61,7 +61,7 @@ const SUPPORTED_MODELS = {
|
|
|
61
61
|
timeout: 300000,
|
|
62
62
|
supports1MContext: true, // Beta 1M context support
|
|
63
63
|
description: 'Claude Sonnet 4 - High intelligence and balanced performance with extended thinking',
|
|
64
|
-
aliases: ['claude-4-sonnet', 'sonnet-4', '
|
|
64
|
+
aliases: ['claude-4-sonnet', 'sonnet-4', 'claude-sonnet-4', 'sonnet4']
|
|
65
65
|
},
|
|
66
66
|
'claude-3-7-sonnet-20250219': {
|
|
67
67
|
modelName: 'claude-3-7-sonnet-20250219',
|
|
@@ -79,6 +79,22 @@ const SUPPORTED_MODELS = {
|
|
|
79
79
|
description: 'Claude 3.7 Sonnet - Enhanced 3.x generation with thinking',
|
|
80
80
|
aliases: ['claude-3.7-sonnet', 'sonnet-3.7', 'claude-3-7-sonnet', 'claude 3.7 sonnet', 'sonnet 3.7']
|
|
81
81
|
},
|
|
82
|
+
'claude-haiku-4-5-20251001': {
|
|
83
|
+
modelName: 'claude-haiku-4-5-20251001',
|
|
84
|
+
friendlyName: 'Claude Haiku 4.5',
|
|
85
|
+
contextWindow: 200000,
|
|
86
|
+
maxOutputTokens: 64000,
|
|
87
|
+
supportsStreaming: true,
|
|
88
|
+
supportsImages: true,
|
|
89
|
+
supportsTemperature: true,
|
|
90
|
+
supportsWebSearch: false,
|
|
91
|
+
supportsThinking: true,
|
|
92
|
+
minThinkingTokens: 1024,
|
|
93
|
+
maxThinkingTokens: 64000,
|
|
94
|
+
timeout: 300000,
|
|
95
|
+
description: 'Claude Haiku 4.5 - Fast and intelligent model with extended thinking',
|
|
96
|
+
aliases: ['claude-haiku-4-5', 'claude-4.5-haiku', 'claude-4-5-haiku', 'haiku-4.5', 'haiku-4-5', 'claude-haiku-4.5', 'haiku4.5', 'claude-haiku-4', 'haiku', 'claude-haiku']
|
|
97
|
+
},
|
|
82
98
|
'claude-3-5-sonnet-20241022': {
|
|
83
99
|
modelName: 'claude-3-5-sonnet-20241022',
|
|
84
100
|
friendlyName: 'Claude 3.5 Sonnet',
|
|
@@ -105,7 +121,7 @@ const SUPPORTED_MODELS = {
|
|
|
105
121
|
supportsThinking: false,
|
|
106
122
|
timeout: 180000,
|
|
107
123
|
description: 'Claude 3.5 Haiku - Fastest model, best for simple queries',
|
|
108
|
-
aliases: ['claude-3.5-haiku', 'claude-3-5-haiku-latest', 'haiku
|
|
124
|
+
aliases: ['claude-3.5-haiku', 'claude-3-5-haiku-latest', 'haiku-3.5', 'claude 3.5 haiku', 'haiku 3.5']
|
|
109
125
|
}
|
|
110
126
|
};
|
|
111
127
|
|
package/src/providers/openai.js
CHANGED
|
@@ -10,9 +10,9 @@ import { debugLog, debugError } from '../utils/console.js';
|
|
|
10
10
|
|
|
11
11
|
// Define supported models with their capabilities
|
|
12
12
|
const SUPPORTED_MODELS = {
|
|
13
|
-
'gpt-5': {
|
|
14
|
-
modelName: 'gpt-5',
|
|
15
|
-
friendlyName: 'OpenAI (GPT-5)',
|
|
13
|
+
'gpt-5.1': {
|
|
14
|
+
modelName: 'gpt-5.1',
|
|
15
|
+
friendlyName: 'OpenAI (GPT-5.1)',
|
|
16
16
|
contextWindow: 400000,
|
|
17
17
|
maxOutputTokens: 128000,
|
|
18
18
|
supportsStreaming: true,
|
|
@@ -22,7 +22,21 @@ const SUPPORTED_MODELS = {
|
|
|
22
22
|
supportsResponsesAPI: true,
|
|
23
23
|
timeout: 3600000, // 1 hour
|
|
24
24
|
description: 'Latest flagship model (400K context, 128K output) - Superior reasoning, code generation, and analysis',
|
|
25
|
-
aliases: ['gpt5', 'gpt 5', 'gpt-5-2025-
|
|
25
|
+
aliases: ['gpt-5', 'gpt5', 'gpt 5', 'gpt-5.1-2025-11-13', 'gpt5.1', 'gpt 5.1']
|
|
26
|
+
},
|
|
27
|
+
'gpt-5-2025-08-07': {
|
|
28
|
+
modelName: 'gpt-5-2025-08-07',
|
|
29
|
+
friendlyName: 'OpenAI (GPT-5.0)',
|
|
30
|
+
contextWindow: 400000,
|
|
31
|
+
maxOutputTokens: 128000,
|
|
32
|
+
supportsStreaming: true,
|
|
33
|
+
supportsImages: true,
|
|
34
|
+
supportsTemperature: false, // GPT-5 doesn't support temperature
|
|
35
|
+
supportsWebSearch: true,
|
|
36
|
+
supportsResponsesAPI: true,
|
|
37
|
+
timeout: 3600000, // 1 hour
|
|
38
|
+
description: 'GPT-5.0 model (400K context, 128K output) - Previous version, accessible via fully qualified name',
|
|
39
|
+
aliases: ['gpt-5.0', 'gpt5.0', 'gpt 5.0']
|
|
26
40
|
},
|
|
27
41
|
'gpt-5-mini': {
|
|
28
42
|
modelName: 'gpt-5-mini',
|