converse-mcp-server 2.14.0 → 2.15.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.14.0",
3
+ "version": "2.15.1",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -94,14 +94,14 @@
94
94
  ".env.example"
95
95
  ],
96
96
  "dependencies": {
97
- "@anthropic-ai/claude-agent-sdk": "^0.2.44",
97
+ "@anthropic-ai/claude-agent-sdk": "^0.2.49",
98
98
  "@anthropic-ai/sdk": "^0.74.0",
99
- "@github/copilot-sdk": "^0.1.24",
100
- "@google/genai": "^1.41.0",
99
+ "@github/copilot-sdk": "^0.1.25",
100
+ "@google/genai": "^1.42.0",
101
101
  "@mistralai/mistralai": "^1.14.0",
102
102
  "@modelcontextprotocol/sdk": "^1.26.0",
103
103
  "@openai/codex-sdk": "^0.101.0",
104
- "ai": "^6.0.87",
104
+ "ai": "^6.0.94",
105
105
  "ai-sdk-provider-gemini-cli": "^2.0.1",
106
106
  "cors": "^2.8.6",
107
107
  "dotenv": "^17.3.1",
@@ -26,7 +26,7 @@ import { ProviderError, ErrorCodes, StopReasons } from './interface.js';
26
26
  const SUPPORTED_MODELS = {
27
27
  gemini: {
28
28
  modelName: 'gemini',
29
- friendlyName: 'Gemini 3.0 Pro Preview (via CLI)',
29
+ friendlyName: 'Gemini 3.1 Pro Preview (via CLI)',
30
30
  contextWindow: 1048576, // 1M tokens
31
31
  maxOutputTokens: 64000,
32
32
  supportsStreaming: true,
@@ -36,10 +36,10 @@ const SUPPORTED_MODELS = {
36
36
  supportsWebSearch: true,
37
37
  timeout: 300000, // 5 minutes
38
38
  description:
39
- 'Gemini 3.0 Pro Preview via OAuth - requires Gemini CLI authentication',
39
+ 'Gemini 3.1 Pro Preview via OAuth - requires Gemini CLI authentication',
40
40
  aliases: ['gemini-cli'],
41
- // Internal SDK model name (user-facing "gemini" maps to SDK's "gemini-3-pro-preview")
42
- sdkModelName: 'gemini-3-pro-preview',
41
+ // Internal SDK model name (user-facing "gemini" maps to SDK's "gemini-3.1-pro-preview")
42
+ sdkModelName: 'gemini-3.1-pro-preview',
43
43
  },
44
44
  };
45
45
 
@@ -76,28 +76,34 @@ const SUPPORTED_MODELS = {
76
76
  'Deep reasoning + thinking mode (1M context) - Complex problems, architecture, deep analysis',
77
77
  aliases: ['pro 2.5', 'gemini pro 2.5', 'gemini-2.5-pro-latest'],
78
78
  },
79
- 'gemini-3-pro-preview': {
80
- modelName: 'gemini-3-pro-preview',
81
- friendlyName: 'Gemini (Pro 3.0)',
79
+ 'gemini-3.1-pro-preview': {
80
+ modelName: 'gemini-3.1-pro-preview',
81
+ friendlyName: 'Gemini (Pro 3.1)',
82
82
  contextWindow: 1048576, // 1M tokens
83
83
  maxOutputTokens: 64000,
84
84
  supportsStreaming: true,
85
85
  supportsImages: true,
86
86
  supportsTemperature: true,
87
- supportsThinking: true, // Always enabled for Gemini 3.0
87
+ supportsThinking: true,
88
88
  supportsWebSearch: true,
89
- thinkingMode: 'level', // Distinguishes from 2.5's budget mode
89
+ thinkingMode: 'level',
90
+ thinkingLevels: ['minimal', 'low', 'medium', 'high'],
90
91
  timeout: 300000,
91
92
  description:
92
- 'Gemini 3.0 Pro - Enhanced reasoning with dynamic thinking levels (1M context)',
93
+ 'Gemini 3.1 Pro - Most advanced reasoning with expanded thinking levels (1M context)',
93
94
  aliases: [
94
95
  'gemini-3',
95
96
  'gemini3',
96
97
  'gemini-3-pro',
98
+ 'gemini-3-pro-preview',
97
99
  '3-pro',
98
- 'gemini pro', // Moving from 2.5 Pro
99
- 'gemini-pro', // Moving from 2.5 Pro
100
- 'pro', // Moving from 2.5 Pro
100
+ 'gemini-3.1',
101
+ 'gemini3.1',
102
+ 'gemini-3.1-pro',
103
+ '3.1-pro',
104
+ 'gemini pro',
105
+ 'gemini-pro',
106
+ 'pro',
101
107
  ],
102
108
  },
103
109
  };
@@ -484,10 +490,30 @@ export const googleProvider = {
484
490
  // Add thinking configuration for models that support it
485
491
  if (modelConfig.supportsThinking && reasoning_effort) {
486
492
  if (modelConfig.thinkingMode === 'level') {
487
- // Gemini 3.0: Use thinking level (low/high)
488
- const thinkingLevel = ['minimal', 'low'].includes(reasoning_effort)
489
- ? 'low'
490
- : 'high';
493
+ let thinkingLevel;
494
+ if (modelConfig.thinkingLevels) {
495
+ // Model supports specific levels (e.g., Gemini 3.1 Pro: minimal/low/medium/high)
496
+ const levelMap = {
497
+ none: 'minimal',
498
+ minimal: 'minimal',
499
+ low: 'low',
500
+ medium: 'medium',
501
+ high: 'high',
502
+ max: 'high',
503
+ };
504
+ thinkingLevel = levelMap[reasoning_effort] || 'high';
505
+ if (!modelConfig.thinkingLevels.includes(thinkingLevel)) {
506
+ thinkingLevel =
507
+ modelConfig.thinkingLevels[
508
+ modelConfig.thinkingLevels.length - 1
509
+ ];
510
+ }
511
+ } else {
512
+ // Binary levels only (Gemini 3.0 Pro: low/high)
513
+ thinkingLevel = ['minimal', 'low'].includes(reasoning_effort)
514
+ ? 'low'
515
+ : 'high';
516
+ }
491
517
  generationConfig.thinkingConfig = { thinkingLevel };
492
518
  } else {
493
519
  // Gemini 2.5: Use thinking budget (token count)