converse-mcp-server 2.22.2 → 2.22.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/.env.example CHANGED
@@ -77,6 +77,10 @@ OPENROUTER_REFERER=https://github.com/FallDownTheSystem/converse
77
77
  # WARNING: Interactive policies may cause hangs in server/headless mode
78
78
  # CODEX_APPROVAL_POLICY=never
79
79
 
80
+ # Default Codex model (default: gpt-5.5)
81
+ # Options: gpt-5.5, gpt-5.3-codex, gpt-5-codex
82
+ # CODEX_MODEL=gpt-5.5
83
+
80
84
  # ============================================
81
85
  # Server Configuration
82
86
  # ============================================
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.22.2",
3
+ "version": "2.22.4",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -94,14 +94,14 @@
94
94
  ".env.example"
95
95
  ],
96
96
  "dependencies": {
97
- "@anthropic-ai/claude-agent-sdk": "^0.2.110",
98
- "@anthropic-ai/sdk": "^0.86.1",
97
+ "@anthropic-ai/claude-agent-sdk": "^0.2.119",
98
+ "@anthropic-ai/sdk": "^0.91.0",
99
99
  "@github/copilot-sdk": "^0.2.2",
100
100
  "@google/genai": "^1.50.1",
101
- "@mistralai/mistralai": "^2.2.0",
101
+ "@mistralai/mistralai": "^2.2.1",
102
102
  "@modelcontextprotocol/sdk": "^1.29.0",
103
- "@openai/codex-sdk": "^0.118.0",
104
- "ai": "^6.0.164",
103
+ "@openai/codex-sdk": "^0.124.0",
104
+ "ai": "^6.0.168",
105
105
  "ai-sdk-provider-gemini-cli": "^2.0.1",
106
106
  "cors": "^2.8.6",
107
107
  "dotenv": "^17.4.2",
@@ -110,14 +110,14 @@
110
110
  "nanoid": "^5.1.9",
111
111
  "openai": "^6.34.0",
112
112
  "p-limit": "^7.3.0",
113
- "vite": "^8.0.8"
113
+ "vite": "^8.0.10"
114
114
  },
115
115
  "devDependencies": {
116
- "@vitest/coverage-v8": "^4.1.4",
116
+ "@vitest/coverage-v8": "^4.1.5",
117
117
  "cross-env": "^10.1.0",
118
- "eslint": "^10.2.0",
118
+ "eslint": "^10.2.1",
119
119
  "prettier": "^3.8.3",
120
120
  "rimraf": "^6.1.3",
121
- "vitest": "^4.1.4"
121
+ "vitest": "^4.1.5"
122
122
  }
123
123
  }
package/src/config.js CHANGED
@@ -276,6 +276,12 @@ const CONFIG_SCHEMA = {
276
276
  description:
277
277
  'Approval policy (never | untrusted | on-failure | on-request)',
278
278
  },
279
+ CODEX_MODEL: {
280
+ type: 'string',
281
+ default: 'gpt-5.5',
282
+ description:
283
+ 'Default Codex model (e.g., gpt-5.5, gpt-5.3-codex, gpt-5-codex)',
284
+ },
279
285
 
280
286
  // Copilot configuration
281
287
  COPILOT_TOOL_ACCESS: {
@@ -21,7 +21,7 @@ import { normalizeExtendedPath } from '../utils/pathUtils.js';
21
21
  const SUPPORTED_MODELS = {
22
22
  codex: {
23
23
  modelName: 'codex',
24
- friendlyName: 'OpenAI Codex (GPT-5.3)',
24
+ friendlyName: 'OpenAI Codex (GPT-5.5)',
25
25
  contextWindow: 400000,
26
26
  maxOutputTokens: 128000,
27
27
  supportsStreaming: true,
@@ -30,13 +30,15 @@ const SUPPORTED_MODELS = {
30
30
  supportsWebSearch: false, // Codex accesses files directly, not web
31
31
  timeout: 600000, // 10 minutes
32
32
  description:
33
- 'OpenAI Codex agentic coding assistant with local file access and tool execution (GPT-5.3-Codex)',
33
+ 'OpenAI Codex agentic coding assistant with local file access and tool execution (GPT-5.5)',
34
34
  aliases: [
35
35
  'gpt-5-codex',
36
36
  'gpt5-codex',
37
37
  'gpt-5.2-codex',
38
38
  'gpt-5.3-codex',
39
39
  'gpt5.3-codex',
40
+ 'gpt-5.5',
41
+ 'gpt5.5',
40
42
  ],
41
43
  },
42
44
  };
@@ -344,6 +346,7 @@ export const codexProvider = {
344
346
 
345
347
  // Create or resume thread
346
348
  const threadOptions = {
349
+ model: config.providers?.codexmodel,
347
350
  workingDirectory,
348
351
  sandboxMode,
349
352
  skipGitRepoCheck,