converse-mcp-server 2.1.0 → 2.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/bin/converse.js CHANGED
File without changes
package/docs/PROVIDERS.md CHANGED
@@ -40,6 +40,7 @@ This guide documents all supported AI providers in the Converse MCP Server and t
40
40
  - `claude-opus-4-1-20250805` - Highest intelligence with extended thinking (32K output)
41
41
  - `claude-sonnet-4-20250514` - Balanced performance with extended thinking (64K output)
42
42
  - `claude-3-7-sonnet-20250219` - Enhanced 3.x generation with thinking (64K output)
43
+ - `claude-haiku-4-5-20251001` - Fast and intelligent with extended thinking (64K output)
43
44
  - `claude-3-5-sonnet-20241022` - Fast and intelligent model (8K output)
44
45
  - `claude-3-5-haiku-20241022` - Fastest Claude model (8K output)
45
46
  - `claude-3-opus-20240229`, `claude-3-sonnet-20240229`, `claude-3-haiku-20240307` - Previous generation
@@ -165,7 +166,7 @@ CODEX_DEFAULT_MODEL=gpt-5-codex # Default: gpt-5-codex
165
166
  All providers support streaming responses for real-time output.
166
167
 
167
168
  ### Image Support
168
- - **Full Support**: OpenAI, Google, X.AI (Grok-4), Anthropic (Claude-3.5-Sonnet, Claude-3-Opus)
169
+ - **Full Support**: OpenAI, Google, X.AI (Grok-4), Anthropic (Claude-4 series, Claude-3.5-Sonnet, Claude-3-Opus)
169
170
  - **Via OpenRouter**: Depends on the underlying model
170
171
  - **No Support**: DeepSeek, Mistral (except Large), Claude-3.5-Haiku, Codex
171
172
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.1.0",
3
+ "version": "2.3.0",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -11,6 +11,56 @@
11
11
  "engines": {
12
12
  "node": ">=20.0.0"
13
13
  },
14
+ "keywords": [
15
+ "mcp",
16
+ "server",
17
+ "ai",
18
+ "chat",
19
+ "consensus",
20
+ "openai",
21
+ "google",
22
+ "gemini",
23
+ "grok"
24
+ ],
25
+ "author": "Converse MCP Server",
26
+ "license": "MIT",
27
+ "homepage": "https://github.com/FallDownTheSystem/converse#readme",
28
+ "repository": {
29
+ "type": "git",
30
+ "url": "git+https://github.com/FallDownTheSystem/converse.git"
31
+ },
32
+ "bugs": {
33
+ "url": "https://github.com/FallDownTheSystem/converse/issues"
34
+ },
35
+ "files": [
36
+ "src/",
37
+ "bin/",
38
+ "docs/",
39
+ "README.md",
40
+ ".env.example"
41
+ ],
42
+ "dependencies": {
43
+ "@anthropic-ai/sdk": "^0.65.0",
44
+ "@google/genai": "^1.22.0",
45
+ "@mistralai/mistralai": "^1.10.0",
46
+ "@modelcontextprotocol/sdk": "^1.19.1",
47
+ "@openai/codex-sdk": "^0.45.0",
48
+ "cors": "^2.8.5",
49
+ "dotenv": "^17.2.3",
50
+ "express": "^5.1.0",
51
+ "lru-cache": "^11.2.2",
52
+ "openai": "^6.1.0",
53
+ "p-limit": "^4.0.0",
54
+ "vite": "^7.1.9"
55
+ },
56
+ "devDependencies": {
57
+ "@vitest/coverage-v8": "^3.2.4",
58
+ "cross-env": "^10.1.0",
59
+ "eslint": "^9.36.0",
60
+ "prettier": "^3.6.2",
61
+ "rimraf": "^6.0.1",
62
+ "vitest": "^3.2.4"
63
+ },
14
64
  "scripts": {
15
65
  "kill-server": "node scripts/kill-server.js",
16
66
  "start": "npm run kill-server && node src/index.js",
@@ -64,55 +114,5 @@
64
114
  "validate:fix": "node scripts/validate.js --fix",
65
115
  "validate:fast": "node scripts/validate.js --skip-tests --skip-lint",
66
116
  "precommit": "npm run validate"
67
- },
68
- "keywords": [
69
- "mcp",
70
- "server",
71
- "ai",
72
- "chat",
73
- "consensus",
74
- "openai",
75
- "google",
76
- "gemini",
77
- "grok"
78
- ],
79
- "author": "Converse MCP Server",
80
- "license": "MIT",
81
- "homepage": "https://github.com/FallDownTheSystem/converse#readme",
82
- "repository": {
83
- "type": "git",
84
- "url": "git+https://github.com/FallDownTheSystem/converse.git"
85
- },
86
- "bugs": {
87
- "url": "https://github.com/FallDownTheSystem/converse/issues"
88
- },
89
- "files": [
90
- "src/",
91
- "bin/",
92
- "docs/",
93
- "README.md",
94
- ".env.example"
95
- ],
96
- "dependencies": {
97
- "@anthropic-ai/sdk": "^0.65.0",
98
- "@google/genai": "^1.22.0",
99
- "@mistralai/mistralai": "^1.10.0",
100
- "@modelcontextprotocol/sdk": "^1.19.1",
101
- "@openai/codex-sdk": "^0.45.0",
102
- "cors": "^2.8.5",
103
- "dotenv": "^17.2.3",
104
- "express": "^5.1.0",
105
- "lru-cache": "^11.2.2",
106
- "openai": "^6.1.0",
107
- "p-limit": "^4.0.0",
108
- "vite": "^7.1.9"
109
- },
110
- "devDependencies": {
111
- "@vitest/coverage-v8": "^3.2.4",
112
- "cross-env": "^10.1.0",
113
- "eslint": "^9.36.0",
114
- "prettier": "^3.6.2",
115
- "rimraf": "^6.0.1",
116
- "vitest": "^3.2.4"
117
117
  }
118
- }
118
+ }
@@ -44,7 +44,7 @@ const SUPPORTED_MODELS = {
44
44
  timeout: 300000,
45
45
  supports1MContext: true, // Beta 1M context support
46
46
  description: 'Claude Sonnet 4.5 - Latest Sonnet with enhanced intelligence and optional 1M context',
47
- aliases: ['claude-4.5-sonnet', 'sonnet-4.5', 'claude-sonnet-4.5', 'sonnet4.5', 'claude-sonnet-4-5']
47
+ aliases: ['claude-4.5-sonnet', 'sonnet-4.5', 'claude-sonnet-4.5', 'sonnet4.5', 'claude-sonnet-4-5', 'sonnet', 'claude-sonnet']
48
48
  },
49
49
  'claude-sonnet-4-20250514': {
50
50
  modelName: 'claude-sonnet-4-20250514',
@@ -61,7 +61,7 @@ const SUPPORTED_MODELS = {
61
61
  timeout: 300000,
62
62
  supports1MContext: true, // Beta 1M context support
63
63
  description: 'Claude Sonnet 4 - High intelligence and balanced performance with extended thinking',
64
- aliases: ['claude-4-sonnet', 'sonnet-4', 'sonnet', 'claude-sonnet', 'claude-sonnet-4', 'sonnet4']
64
+ aliases: ['claude-4-sonnet', 'sonnet-4', 'claude-sonnet-4', 'sonnet4']
65
65
  },
66
66
  'claude-3-7-sonnet-20250219': {
67
67
  modelName: 'claude-3-7-sonnet-20250219',
@@ -79,6 +79,22 @@ const SUPPORTED_MODELS = {
79
79
  description: 'Claude 3.7 Sonnet - Enhanced 3.x generation with thinking',
80
80
  aliases: ['claude-3.7-sonnet', 'sonnet-3.7', 'claude-3-7-sonnet', 'claude 3.7 sonnet', 'sonnet 3.7']
81
81
  },
82
+ 'claude-haiku-4-5-20251001': {
83
+ modelName: 'claude-haiku-4-5-20251001',
84
+ friendlyName: 'Claude Haiku 4.5',
85
+ contextWindow: 200000,
86
+ maxOutputTokens: 64000,
87
+ supportsStreaming: true,
88
+ supportsImages: true,
89
+ supportsTemperature: true,
90
+ supportsWebSearch: false,
91
+ supportsThinking: true,
92
+ minThinkingTokens: 1024,
93
+ maxThinkingTokens: 64000,
94
+ timeout: 300000,
95
+ description: 'Claude Haiku 4.5 - Fast and intelligent model with extended thinking',
96
+ aliases: ['claude-haiku-4-5', 'claude-4.5-haiku', 'claude-4-5-haiku', 'haiku-4.5', 'haiku-4-5', 'claude-haiku-4.5', 'haiku4.5', 'claude-haiku-4', 'haiku', 'claude-haiku']
97
+ },
82
98
  'claude-3-5-sonnet-20241022': {
83
99
  modelName: 'claude-3-5-sonnet-20241022',
84
100
  friendlyName: 'Claude 3.5 Sonnet',
@@ -105,7 +121,7 @@ const SUPPORTED_MODELS = {
105
121
  supportsThinking: false,
106
122
  timeout: 180000,
107
123
  description: 'Claude 3.5 Haiku - Fastest model, best for simple queries',
108
- aliases: ['claude-3.5-haiku', 'claude-3-5-haiku-latest', 'haiku', 'claude-haiku', 'haiku-3.5', 'claude 3.5 haiku', 'haiku 3.5']
124
+ aliases: ['claude-3.5-haiku', 'claude-3-5-haiku-latest', 'haiku-3.5', 'claude 3.5 haiku', 'haiku 3.5']
109
125
  }
110
126
  };
111
127
 
@@ -10,9 +10,9 @@ import { debugLog, debugError } from '../utils/console.js';
10
10
 
11
11
  // Define supported models with their capabilities
12
12
  const SUPPORTED_MODELS = {
13
- 'gpt-5': {
14
- modelName: 'gpt-5',
15
- friendlyName: 'OpenAI (GPT-5)',
13
+ 'gpt-5.1': {
14
+ modelName: 'gpt-5.1',
15
+ friendlyName: 'OpenAI (GPT-5.1)',
16
16
  contextWindow: 400000,
17
17
  maxOutputTokens: 128000,
18
18
  supportsStreaming: true,
@@ -22,7 +22,21 @@ const SUPPORTED_MODELS = {
22
22
  supportsResponsesAPI: true,
23
23
  timeout: 3600000, // 1 hour
24
24
  description: 'Latest flagship model (400K context, 128K output) - Superior reasoning, code generation, and analysis',
25
- aliases: ['gpt5', 'gpt 5', 'gpt-5-2025-08-07']
25
+ aliases: ['gpt-5', 'gpt5', 'gpt 5', 'gpt-5.1-2025-11-13', 'gpt5.1', 'gpt 5.1']
26
+ },
27
+ 'gpt-5-2025-08-07': {
28
+ modelName: 'gpt-5-2025-08-07',
29
+ friendlyName: 'OpenAI (GPT-5.0)',
30
+ contextWindow: 400000,
31
+ maxOutputTokens: 128000,
32
+ supportsStreaming: true,
33
+ supportsImages: true,
34
+ supportsTemperature: false, // GPT-5 doesn't support temperature
35
+ supportsWebSearch: true,
36
+ supportsResponsesAPI: true,
37
+ timeout: 3600000, // 1 hour
38
+ description: 'GPT-5.0 model (400K context, 128K output) - Previous version, accessible via fully qualified name',
39
+ aliases: ['gpt-5.0', 'gpt5.0', 'gpt 5.0']
26
40
  },
27
41
  'gpt-5-mini': {
28
42
  modelName: 'gpt-5-mini',