converse-mcp-server 2.4.1 → 2.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
|
@@ -95,7 +95,7 @@ export function generateHelpContent(config = null) {
|
|
|
95
95
|
return `\`\`\`json
|
|
96
96
|
{
|
|
97
97
|
"prompt": "Should we use microservices architecture for our new project?",
|
|
98
|
-
"models": ["gpt-5", "gemini-
|
|
98
|
+
"models": ["gpt-5", "gemini-pro", "grok-4"],
|
|
99
99
|
"files": ["./requirements.md", "C:\\\\Users\\\\username\\\\architecture.md"],
|
|
100
100
|
"enable_cross_feedback": true,
|
|
101
101
|
"temperature": 0.3
|
|
@@ -138,12 +138,12 @@ ${formatProviderModels('OpenRouter', allModels.openrouter)}
|
|
|
138
138
|
## Model Selection Tips
|
|
139
139
|
|
|
140
140
|
### For Complex Reasoning Tasks
|
|
141
|
-
- **Most Intelligent**: gpt-5, gpt-5-pro, gemini-
|
|
142
|
-
- **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini,
|
|
141
|
+
- **Most Intelligent**: gpt-5, gpt-5-pro, gemini-pro, grok-4
|
|
142
|
+
- **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini, flash
|
|
143
143
|
- **Budget-Friendly**: gpt-5-nano, gpt-4o-mini, gemini-2.0-flash-lite
|
|
144
144
|
|
|
145
145
|
### For Quick Responses
|
|
146
|
-
- **Ultra-Fast**: gpt-5-nano,
|
|
146
|
+
- **Ultra-Fast**: gpt-5-nano, flash, gemini-2.0-flash, gpt-4o-mini
|
|
147
147
|
- **Good Balance**: gpt-5-mini, o4-mini, grok-code-fast-1
|
|
148
148
|
|
|
149
149
|
### For Large Context Windows
|
|
@@ -17,7 +17,7 @@ const logger = createLogger('summarization');
|
|
|
17
17
|
// Default fast models for summarization tasks (prioritize GPT-5-nano for speed)
|
|
18
18
|
const FAST_MODELS = {
|
|
19
19
|
openai: 'gpt-5-nano', // Fastest GPT-5 model with minimal reasoning
|
|
20
|
-
google: '
|
|
20
|
+
google: 'flash',
|
|
21
21
|
xai: 'grok-4',
|
|
22
22
|
anthropic: 'claude-3-5-haiku-latest',
|
|
23
23
|
mistral: 'mistral-small-latest',
|
package/src/tools/chat.js
CHANGED
|
@@ -418,7 +418,7 @@ function resolveAutoModel(model, providerName) {
|
|
|
418
418
|
const defaults = {
|
|
419
419
|
openai: 'gpt-5',
|
|
420
420
|
xai: 'grok-4-0709',
|
|
421
|
-
google: 'gemini-
|
|
421
|
+
google: 'gemini-pro',
|
|
422
422
|
anthropic: 'claude-sonnet-4-20250514',
|
|
423
423
|
mistral: 'magistral-medium-2506',
|
|
424
424
|
deepseek: 'deepseek-reasoner',
|
|
@@ -934,7 +934,7 @@ chatTool.inputSchema = {
|
|
|
934
934
|
model: {
|
|
935
935
|
type: 'string',
|
|
936
936
|
description:
|
|
937
|
-
'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-
|
|
937
|
+
'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-pro", "grok-4-0709". Defaults to auto-selection.',
|
|
938
938
|
},
|
|
939
939
|
files: {
|
|
940
940
|
type: 'array',
|
package/src/tools/consensus.js
CHANGED
|
@@ -655,7 +655,7 @@ function getDefaultModelForProvider(providerName) {
|
|
|
655
655
|
const defaults = {
|
|
656
656
|
openai: 'gpt-5',
|
|
657
657
|
xai: 'grok-4-0709',
|
|
658
|
-
google: 'gemini-
|
|
658
|
+
google: 'gemini-pro',
|
|
659
659
|
anthropic: 'claude-sonnet-4-20250514',
|
|
660
660
|
mistral: 'magistral-medium-2506',
|
|
661
661
|
deepseek: 'deepseek-reasoner',
|
|
@@ -1500,7 +1500,7 @@ consensusTool.inputSchema = {
|
|
|
1500
1500
|
items: { type: 'string' },
|
|
1501
1501
|
minItems: 1,
|
|
1502
1502
|
description:
|
|
1503
|
-
'List of models to consult. Example: ["gpt-5", "gemini-
|
|
1503
|
+
'List of models to consult. Example: ["gpt-5", "gemini-pro", "grok-4-0709"]',
|
|
1504
1504
|
},
|
|
1505
1505
|
files: {
|
|
1506
1506
|
type: 'array',
|