converse-mcp-server 2.4.0 → 2.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/docs/API.md
CHANGED
|
@@ -101,9 +101,9 @@ MCP_TRANSPORT=stdio npm start
|
|
|
101
101
|
},
|
|
102
102
|
"media_resolution": {
|
|
103
103
|
"type": "string",
|
|
104
|
-
"enum": ["
|
|
105
|
-
"default": "
|
|
106
|
-
"description": "Control image/PDF/video processing quality (Gemini 3.0
|
|
104
|
+
"enum": ["MEDIA_RESOLUTION_LOW", "MEDIA_RESOLUTION_MEDIUM", "MEDIA_RESOLUTION_HIGH", "MEDIA_RESOLUTION_UNSPECIFIED"],
|
|
105
|
+
"default": "MEDIA_RESOLUTION_HIGH",
|
|
106
|
+
"description": "Control image/PDF/video processing quality (Gemini 3.0). Defaults to 'MEDIA_RESOLUTION_HIGH' for Gemini 3.0. Examples: 'MEDIA_RESOLUTION_LOW' (faster, less detail), 'MEDIA_RESOLUTION_MEDIUM' (balanced), 'MEDIA_RESOLUTION_HIGH' (maximum detail)"
|
|
107
107
|
},
|
|
108
108
|
"async": {
|
|
109
109
|
"type": "boolean",
|
package/package.json
CHANGED
|
@@ -95,7 +95,7 @@ export function generateHelpContent(config = null) {
|
|
|
95
95
|
return `\`\`\`json
|
|
96
96
|
{
|
|
97
97
|
"prompt": "Should we use microservices architecture for our new project?",
|
|
98
|
-
"models": ["gpt-5", "gemini-
|
|
98
|
+
"models": ["gpt-5", "gemini-pro", "grok-4"],
|
|
99
99
|
"files": ["./requirements.md", "C:\\\\Users\\\\username\\\\architecture.md"],
|
|
100
100
|
"enable_cross_feedback": true,
|
|
101
101
|
"temperature": 0.3
|
|
@@ -138,12 +138,12 @@ ${formatProviderModels('OpenRouter', allModels.openrouter)}
|
|
|
138
138
|
## Model Selection Tips
|
|
139
139
|
|
|
140
140
|
### For Complex Reasoning Tasks
|
|
141
|
-
- **Most Intelligent**: gpt-5, gpt-5-pro, gemini-
|
|
142
|
-
- **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini,
|
|
141
|
+
- **Most Intelligent**: gpt-5, gpt-5-pro, gemini-pro, grok-4
|
|
142
|
+
- **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini, flash
|
|
143
143
|
- **Budget-Friendly**: gpt-5-nano, gpt-4o-mini, gemini-2.0-flash-lite
|
|
144
144
|
|
|
145
145
|
### For Quick Responses
|
|
146
|
-
- **Ultra-Fast**: gpt-5-nano,
|
|
146
|
+
- **Ultra-Fast**: gpt-5-nano, flash, gemini-2.0-flash, gpt-4o-mini
|
|
147
147
|
- **Good Balance**: gpt-5-mini, o4-mini, grok-code-fast-1
|
|
148
148
|
|
|
149
149
|
### For Large Context Windows
|
package/src/providers/google.js
CHANGED
|
@@ -553,14 +553,28 @@ export const googleProvider = {
|
|
|
553
553
|
|
|
554
554
|
// Add media resolution for Gemini 3.0 models
|
|
555
555
|
if (modelConfig.thinkingMode === 'level') {
|
|
556
|
-
// Default to
|
|
557
|
-
const resolution = media_resolution || '
|
|
558
|
-
|
|
556
|
+
// Default to MEDIA_RESOLUTION_HIGH for Gemini 3.0 if not specified
|
|
557
|
+
const resolution = media_resolution || 'MEDIA_RESOLUTION_HIGH';
|
|
558
|
+
const validResolutions = [
|
|
559
|
+
'MEDIA_RESOLUTION_LOW',
|
|
560
|
+
'MEDIA_RESOLUTION_MEDIUM',
|
|
561
|
+
'MEDIA_RESOLUTION_HIGH',
|
|
562
|
+
'MEDIA_RESOLUTION_UNSPECIFIED',
|
|
563
|
+
];
|
|
564
|
+
if (validResolutions.includes(resolution)) {
|
|
559
565
|
generationConfig.mediaResolution = resolution;
|
|
560
566
|
}
|
|
561
|
-
} else if (media_resolution
|
|
567
|
+
} else if (media_resolution) {
|
|
562
568
|
// For other models, only add if explicitly specified
|
|
563
|
-
|
|
569
|
+
const validResolutions = [
|
|
570
|
+
'MEDIA_RESOLUTION_LOW',
|
|
571
|
+
'MEDIA_RESOLUTION_MEDIUM',
|
|
572
|
+
'MEDIA_RESOLUTION_HIGH',
|
|
573
|
+
'MEDIA_RESOLUTION_UNSPECIFIED',
|
|
574
|
+
];
|
|
575
|
+
if (validResolutions.includes(media_resolution)) {
|
|
576
|
+
generationConfig.mediaResolution = media_resolution;
|
|
577
|
+
}
|
|
564
578
|
}
|
|
565
579
|
|
|
566
580
|
// Add web search grounding if requested and model supports it
|
|
@@ -17,7 +17,7 @@ const logger = createLogger('summarization');
|
|
|
17
17
|
// Default fast models for summarization tasks (prioritize GPT-5-nano for speed)
|
|
18
18
|
const FAST_MODELS = {
|
|
19
19
|
openai: 'gpt-5-nano', // Fastest GPT-5 model with minimal reasoning
|
|
20
|
-
google: '
|
|
20
|
+
google: 'flash',
|
|
21
21
|
xai: 'grok-4',
|
|
22
22
|
anthropic: 'claude-3-5-haiku-latest',
|
|
23
23
|
mistral: 'mistral-small-latest',
|
package/src/tools/chat.js
CHANGED
|
@@ -418,7 +418,7 @@ function resolveAutoModel(model, providerName) {
|
|
|
418
418
|
const defaults = {
|
|
419
419
|
openai: 'gpt-5',
|
|
420
420
|
xai: 'grok-4-0709',
|
|
421
|
-
google: 'gemini-
|
|
421
|
+
google: 'gemini-pro',
|
|
422
422
|
anthropic: 'claude-sonnet-4-20250514',
|
|
423
423
|
mistral: 'magistral-medium-2506',
|
|
424
424
|
deepseek: 'deepseek-reasoner',
|
|
@@ -934,7 +934,7 @@ chatTool.inputSchema = {
|
|
|
934
934
|
model: {
|
|
935
935
|
type: 'string',
|
|
936
936
|
description:
|
|
937
|
-
'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-
|
|
937
|
+
'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-pro", "grok-4-0709". Defaults to auto-selection.',
|
|
938
938
|
},
|
|
939
939
|
files: {
|
|
940
940
|
type: 'array',
|
package/src/tools/consensus.js
CHANGED
|
@@ -655,7 +655,7 @@ function getDefaultModelForProvider(providerName) {
|
|
|
655
655
|
const defaults = {
|
|
656
656
|
openai: 'gpt-5',
|
|
657
657
|
xai: 'grok-4-0709',
|
|
658
|
-
google: 'gemini-
|
|
658
|
+
google: 'gemini-pro',
|
|
659
659
|
anthropic: 'claude-sonnet-4-20250514',
|
|
660
660
|
mistral: 'magistral-medium-2506',
|
|
661
661
|
deepseek: 'deepseek-reasoner',
|
|
@@ -1500,7 +1500,7 @@ consensusTool.inputSchema = {
|
|
|
1500
1500
|
items: { type: 'string' },
|
|
1501
1501
|
minItems: 1,
|
|
1502
1502
|
description:
|
|
1503
|
-
'List of models to consult. Example: ["gpt-5", "gemini-
|
|
1503
|
+
'List of models to consult. Example: ["gpt-5", "gemini-pro", "grok-4-0709"]',
|
|
1504
1504
|
},
|
|
1505
1505
|
files: {
|
|
1506
1506
|
type: 'array',
|