converse-mcp-server 2.20.0 → 2.20.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.20.0",
3
+ "version": "2.20.2",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -94,28 +94,28 @@
94
94
  ".env.example"
95
95
  ],
96
96
  "dependencies": {
97
- "@anthropic-ai/claude-agent-sdk": "^0.2.50",
97
+ "@anthropic-ai/claude-agent-sdk": "^0.2.63",
98
98
  "@anthropic-ai/sdk": "^0.74.0",
99
- "@github/copilot-sdk": "^0.1.25",
100
- "@google/genai": "^1.42.0",
101
- "@mistralai/mistralai": "^1.14.0",
102
- "@modelcontextprotocol/sdk": "^1.26.0",
103
- "@openai/codex-sdk": "^0.101.0",
104
- "ai": "^6.0.97",
99
+ "@github/copilot-sdk": "^0.1.29",
100
+ "@google/genai": "^1.43.0",
101
+ "@mistralai/mistralai": "^1.14.1",
102
+ "@modelcontextprotocol/sdk": "^1.27.1",
103
+ "@openai/codex-sdk": "^0.110.0",
104
+ "ai": "^6.0.108",
105
105
  "ai-sdk-provider-gemini-cli": "^2.0.1",
106
106
  "cors": "^2.8.6",
107
107
  "dotenv": "^17.3.1",
108
108
  "express": "^5.2.1",
109
109
  "lru-cache": "^11.2.6",
110
110
  "nanoid": "^5.1.6",
111
- "openai": "^6.22.0",
111
+ "openai": "^6.26.0",
112
112
  "p-limit": "^7.3.0",
113
113
  "vite": "^7.3.1"
114
114
  },
115
115
  "devDependencies": {
116
116
  "@vitest/coverage-v8": "^4.0.18",
117
117
  "cross-env": "^10.1.0",
118
- "eslint": "^10.0.1",
118
+ "eslint": "^10.0.2",
119
119
  "prettier": "^3.8.1",
120
120
  "rimraf": "^6.1.3",
121
121
  "vitest": "^4.0.18"
@@ -115,9 +115,9 @@ const SUPPORTED_MODELS = {
115
115
  description: 'OpenAI GPT-5.1 Codex Max via Copilot subscription',
116
116
  aliases: [],
117
117
  },
118
- 'gpt-5.2': {
119
- modelName: 'gpt-5.2',
120
- friendlyName: 'GPT-5.2 (via Copilot)',
118
+ 'gpt-5.4': {
119
+ modelName: 'gpt-5.4',
120
+ friendlyName: 'GPT-5.4 (via Copilot)',
121
121
  contextWindow: 1047576,
122
122
  maxOutputTokens: 32768,
123
123
  supportsStreaming: true,
@@ -126,7 +126,7 @@ const SUPPORTED_MODELS = {
126
126
  supportsWebSearch: false,
127
127
  supportsReasoningEffort: true,
128
128
  timeout: 120000,
129
- description: 'OpenAI GPT-5.2 via Copilot subscription',
129
+ description: 'OpenAI GPT-5.4 via Copilot subscription',
130
130
  aliases: ['gpt-5'],
131
131
  },
132
132
  'gpt-5.2-codex': {
@@ -10,45 +10,28 @@ import { debugLog, debugError } from '../utils/console.js';
10
10
 
11
11
  // Define supported models with their capabilities
12
12
  const SUPPORTED_MODELS = {
13
- 'gpt-5.2': {
14
- modelName: 'gpt-5.2',
15
- friendlyName: 'OpenAI (GPT-5.2)',
16
- contextWindow: 400000,
13
+ 'gpt-5.4': {
14
+ modelName: 'gpt-5.4',
15
+ friendlyName: 'OpenAI (GPT-5.4)',
16
+ contextWindow: 1000000,
17
17
  maxOutputTokens: 128000,
18
18
  supportsStreaming: true,
19
19
  supportsImages: true,
20
- supportsTemperature: false, // GPT-5 doesn't support temperature
20
+ supportsTemperature: false,
21
21
  supportsWebSearch: true,
22
22
  supportsResponsesAPI: true,
23
- supportsNoneReasoningEffort: true, // GPT-5.2 supports "none" for faster responses
23
+ supportsNoneReasoningEffort: true,
24
24
  timeout: 3600000, // 1 hour
25
25
  description:
26
- 'Latest flagship model (400K context, 128K output) - Superior reasoning, code generation, analysis. Supports "none" reasoning for faster responses',
26
+ 'Latest flagship model (1M context, 128K output) - Superior reasoning, coding, agentic workflows, computer use. Most token-efficient reasoning model',
27
27
  aliases: [
28
28
  'gpt-5',
29
29
  'gpt5',
30
30
  'gpt 5',
31
- 'gpt-5.2-2025-12-11',
32
- 'gpt5.2',
33
- 'gpt 5.2',
31
+ 'gpt5.4',
32
+ 'gpt 5.4',
34
33
  ],
35
34
  },
36
- 'gpt-5-2025-08-07': {
37
- modelName: 'gpt-5-2025-08-07',
38
- friendlyName: 'OpenAI (GPT-5.0)',
39
- contextWindow: 400000,
40
- maxOutputTokens: 128000,
41
- supportsStreaming: true,
42
- supportsImages: true,
43
- supportsTemperature: false, // GPT-5 doesn't support temperature
44
- supportsWebSearch: true,
45
- supportsResponsesAPI: true,
46
- supportsNoneReasoningEffort: false, // GPT-5.0 does not support "none" reasoning
47
- timeout: 3600000, // 1 hour
48
- description:
49
- 'GPT-5.0 model (400K context, 128K output) - Previous version, accessible via fully qualified name',
50
- aliases: ['gpt-5.0', 'gpt5.0', 'gpt 5.0'],
51
- },
52
35
  'gpt-5-mini': {
53
36
  modelName: 'gpt-5-mini',
54
37
  friendlyName: 'OpenAI (GPT-5-mini)',
@@ -79,27 +62,26 @@ const SUPPORTED_MODELS = {
79
62
  'Fastest, most cost-efficient GPT-5 (400K context, 128K output) - Summarization, classification',
80
63
  aliases: ['gpt5-nano', 'gpt-5nano', 'gpt 5 nano', 'gpt-5-nano-2025-08-07'],
81
64
  },
82
- 'gpt-5.2-pro': {
83
- modelName: 'gpt-5.2-pro',
84
- friendlyName: 'OpenAI (GPT-5.2 Pro)',
85
- contextWindow: 400000,
65
+ 'gpt-5.4-pro': {
66
+ modelName: 'gpt-5.4-pro',
67
+ friendlyName: 'OpenAI (GPT-5.4 Pro)',
68
+ contextWindow: 1000000,
86
69
  maxOutputTokens: 272000,
87
70
  supportsStreaming: false, // GPT-5 Pro doesn't support streaming
88
71
  supportsImages: true,
89
- supportsTemperature: false, // GPT-5 models don't support temperature
72
+ supportsTemperature: false,
90
73
  supportsWebSearch: true,
91
74
  supportsResponsesAPI: true,
92
- supportsDeepResearch: false, // Not a deep research model
93
- timeout: 3600000, // 60 minutes - some requests may take several minutes
75
+ supportsDeepResearch: false,
76
+ timeout: 3600000, // 60 minutes
94
77
  description:
95
- 'Most advanced reasoning model (400K context, 272K output) - Hardest problems, extended compute time (EXPENSIVE)',
78
+ 'Maximum performance reasoning model (1M context, 272K output) - Most complex tasks, extended compute time (EXPENSIVE)',
96
79
  aliases: [
97
80
  'gpt-5-pro',
98
81
  'gpt5-pro',
99
82
  'gpt-5pro',
100
83
  'gpt 5 pro',
101
84
  'gpt-5 pro',
102
- 'gpt-5.2-pro-2025-12-11',
103
85
  ],
104
86
  },
105
87
  o3: {
@@ -450,7 +432,7 @@ export const openaiProvider = {
450
432
  ) {
451
433
  // GPT-5 Pro only supports 'high' reasoning effort
452
434
  const effectiveEffort =
453
- resolvedModel === 'gpt-5-pro' ? 'high' : reasoning_effort;
435
+ resolvedModel.endsWith('-pro') && resolvedModel.startsWith('gpt-5') ? 'high' : reasoning_effort;
454
436
  requestPayload.reasoning = {
455
437
  effort: effectiveEffort,
456
438
  summary: 'auto', // Enable reasoning summaries
@@ -490,7 +472,7 @@ export const openaiProvider = {
490
472
  ) {
491
473
  // GPT-5 Pro only supports 'high' reasoning effort
492
474
  const effectiveEffort =
493
- resolvedModel === 'gpt-5-pro' ? 'high' : reasoning_effort;
475
+ resolvedModel.endsWith('-pro') && resolvedModel.startsWith('gpt-5') ? 'high' : reasoning_effort;
494
476
  requestPayload.reasoning_effort = effectiveEffort;
495
477
  }
496
478