converse-mcp-server 2.2.0 → 2.3.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/bin/converse.js CHANGED
File without changes
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.2.0",
3
+ "version": "2.3.1",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -11,6 +11,56 @@
11
11
  "engines": {
12
12
  "node": ">=20.0.0"
13
13
  },
14
+ "keywords": [
15
+ "mcp",
16
+ "server",
17
+ "ai",
18
+ "chat",
19
+ "consensus",
20
+ "openai",
21
+ "google",
22
+ "gemini",
23
+ "grok"
24
+ ],
25
+ "author": "Converse MCP Server",
26
+ "license": "MIT",
27
+ "homepage": "https://github.com/FallDownTheSystem/converse#readme",
28
+ "repository": {
29
+ "type": "git",
30
+ "url": "git+https://github.com/FallDownTheSystem/converse.git"
31
+ },
32
+ "bugs": {
33
+ "url": "https://github.com/FallDownTheSystem/converse/issues"
34
+ },
35
+ "files": [
36
+ "src/",
37
+ "bin/",
38
+ "docs/",
39
+ "README.md",
40
+ ".env.example"
41
+ ],
42
+ "dependencies": {
43
+ "@anthropic-ai/sdk": "^0.65.0",
44
+ "@google/genai": "^1.22.0",
45
+ "@mistralai/mistralai": "^1.10.0",
46
+ "@modelcontextprotocol/sdk": "^1.19.1",
47
+ "@openai/codex-sdk": "^0.45.0",
48
+ "cors": "^2.8.5",
49
+ "dotenv": "^17.2.3",
50
+ "express": "^5.1.0",
51
+ "lru-cache": "^11.2.2",
52
+ "openai": "^6.1.0",
53
+ "p-limit": "^4.0.0",
54
+ "vite": "^7.1.9"
55
+ },
56
+ "devDependencies": {
57
+ "@vitest/coverage-v8": "^3.2.4",
58
+ "cross-env": "^10.1.0",
59
+ "eslint": "^9.36.0",
60
+ "prettier": "^3.6.2",
61
+ "rimraf": "^6.0.1",
62
+ "vitest": "^3.2.4"
63
+ },
14
64
  "scripts": {
15
65
  "kill-server": "node scripts/kill-server.js",
16
66
  "start": "npm run kill-server && node src/index.js",
@@ -64,55 +114,5 @@
64
114
  "validate:fix": "node scripts/validate.js --fix",
65
115
  "validate:fast": "node scripts/validate.js --skip-tests --skip-lint",
66
116
  "precommit": "npm run validate"
67
- },
68
- "keywords": [
69
- "mcp",
70
- "server",
71
- "ai",
72
- "chat",
73
- "consensus",
74
- "openai",
75
- "google",
76
- "gemini",
77
- "grok"
78
- ],
79
- "author": "Converse MCP Server",
80
- "license": "MIT",
81
- "homepage": "https://github.com/FallDownTheSystem/converse#readme",
82
- "repository": {
83
- "type": "git",
84
- "url": "git+https://github.com/FallDownTheSystem/converse.git"
85
- },
86
- "bugs": {
87
- "url": "https://github.com/FallDownTheSystem/converse/issues"
88
- },
89
- "files": [
90
- "src/",
91
- "bin/",
92
- "docs/",
93
- "README.md",
94
- ".env.example"
95
- ],
96
- "dependencies": {
97
- "@anthropic-ai/sdk": "^0.65.0",
98
- "@google/genai": "^1.22.0",
99
- "@mistralai/mistralai": "^1.10.0",
100
- "@modelcontextprotocol/sdk": "^1.19.1",
101
- "@openai/codex-sdk": "^0.45.0",
102
- "cors": "^2.8.5",
103
- "dotenv": "^17.2.3",
104
- "express": "^5.1.0",
105
- "lru-cache": "^11.2.2",
106
- "openai": "^6.1.0",
107
- "p-limit": "^4.0.0",
108
- "vite": "^7.1.9"
109
- },
110
- "devDependencies": {
111
- "@vitest/coverage-v8": "^3.2.4",
112
- "cross-env": "^10.1.0",
113
- "eslint": "^9.36.0",
114
- "prettier": "^3.6.2",
115
- "rimraf": "^6.0.1",
116
- "vitest": "^3.2.4"
117
117
  }
118
- }
118
+ }
@@ -10,9 +10,9 @@ import { debugLog, debugError } from '../utils/console.js';
10
10
 
11
11
  // Define supported models with their capabilities
12
12
  const SUPPORTED_MODELS = {
13
- 'gpt-5': {
14
- modelName: 'gpt-5',
15
- friendlyName: 'OpenAI (GPT-5)',
13
+ 'gpt-5.1': {
14
+ modelName: 'gpt-5.1',
15
+ friendlyName: 'OpenAI (GPT-5.1)',
16
16
  contextWindow: 400000,
17
17
  maxOutputTokens: 128000,
18
18
  supportsStreaming: true,
@@ -20,9 +20,25 @@ const SUPPORTED_MODELS = {
20
20
  supportsTemperature: false, // GPT-5 doesn't support temperature
21
21
  supportsWebSearch: true,
22
22
  supportsResponsesAPI: true,
23
+ supportsNoneReasoningEffort: true, // GPT-5.1 supports "none" for faster responses
23
24
  timeout: 3600000, // 1 hour
24
- description: 'Latest flagship model (400K context, 128K output) - Superior reasoning, code generation, and analysis',
25
- aliases: ['gpt5', 'gpt 5', 'gpt-5-2025-08-07']
25
+ description: 'Latest flagship model (400K context, 128K output) - Superior reasoning, code generation, analysis. Supports "none" reasoning for faster responses',
26
+ aliases: ['gpt-5', 'gpt5', 'gpt 5', 'gpt-5.1-2025-11-13', 'gpt5.1', 'gpt 5.1']
27
+ },
28
+ 'gpt-5-2025-08-07': {
29
+ modelName: 'gpt-5-2025-08-07',
30
+ friendlyName: 'OpenAI (GPT-5.0)',
31
+ contextWindow: 400000,
32
+ maxOutputTokens: 128000,
33
+ supportsStreaming: true,
34
+ supportsImages: true,
35
+ supportsTemperature: false, // GPT-5 doesn't support temperature
36
+ supportsWebSearch: true,
37
+ supportsResponsesAPI: true,
38
+ supportsNoneReasoningEffort: false, // GPT-5.0 does not support "none" reasoning
39
+ timeout: 3600000, // 1 hour
40
+ description: 'GPT-5.0 model (400K context, 128K output) - Previous version, accessible via fully qualified name',
41
+ aliases: ['gpt-5.0', 'gpt5.0', 'gpt 5.0']
26
42
  },
27
43
  'gpt-5-mini': {
28
44
  modelName: 'gpt-5-mini',
package/src/tools/chat.js CHANGED
@@ -854,8 +854,8 @@ chatTool.inputSchema = {
854
854
  },
855
855
  reasoning_effort: {
856
856
  type: 'string',
857
- enum: ['minimal', 'low', 'medium', 'high', 'max'],
858
- description: 'Reasoning depth for thinking models. Examples: "minimal" (fastest, few reasoning tokens), "low" (light analysis), "medium" (balanced), "high" (complex analysis). Default: "medium"',
857
+ enum: ['none', 'minimal', 'low', 'medium', 'high', 'max'],
858
+ description: 'Reasoning depth for thinking models. Examples: "none" (no reasoning, fastest - GPT-5.1+ only), "minimal" (few reasoning tokens), "low" (light analysis), "medium" (balanced), "high" (complex analysis). Default: "medium"',
859
859
  default: 'medium'
860
860
  },
861
861
  verbosity: {
@@ -1378,8 +1378,8 @@ consensusTool.inputSchema = {
1378
1378
  },
1379
1379
  reasoning_effort: {
1380
1380
  type: 'string',
1381
- enum: ['minimal', 'low', 'medium', 'high', 'max'],
1382
- description: 'Reasoning depth for thinking models. Examples: "low" (light analysis), "medium" (balanced), "high" (complex analysis). Default: "medium"',
1381
+ enum: ['none', 'minimal', 'low', 'medium', 'high', 'max'],
1382
+ description: 'Reasoning depth for thinking models. Examples: "none" (no reasoning, fastest - GPT-5.1+ only), "low" (light analysis), "medium" (balanced), "high" (complex analysis). Default: "medium"',
1383
1383
  default: 'medium'
1384
1384
  },
1385
1385
  use_websearch: {