converse-mcp-server 2.4.1 → 2.4.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "converse-mcp-server",
3
- "version": "2.4.1",
3
+ "version": "2.4.2",
4
4
  "description": "Converse MCP Server - Converse with other LLMs with chat and consensus tools",
5
5
  "type": "module",
6
6
  "main": "src/index.js",
@@ -95,7 +95,7 @@ export function generateHelpContent(config = null) {
95
95
  return `\`\`\`json
96
96
  {
97
97
  "prompt": "Should we use microservices architecture for our new project?",
98
- "models": ["gpt-5", "gemini-2.5-pro", "grok-4"],
98
+ "models": ["gpt-5", "gemini-pro", "grok-4"],
99
99
  "files": ["./requirements.md", "C:\\\\Users\\\\username\\\\architecture.md"],
100
100
  "enable_cross_feedback": true,
101
101
  "temperature": 0.3
@@ -138,12 +138,12 @@ ${formatProviderModels('OpenRouter', allModels.openrouter)}
138
138
  ## Model Selection Tips
139
139
 
140
140
  ### For Complex Reasoning Tasks
141
- - **Most Intelligent**: gpt-5, gpt-5-pro, gemini-2.5-pro, grok-4
142
- - **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini, gemini-2.5-flash
141
+ - **Most Intelligent**: gpt-5, gpt-5-pro, gemini-pro, grok-4
142
+ - **Fast & Smart**: gpt-5-mini, gpt-5-mini, o4-mini, flash
143
143
  - **Budget-Friendly**: gpt-5-nano, gpt-4o-mini, gemini-2.0-flash-lite
144
144
 
145
145
  ### For Quick Responses
146
- - **Ultra-Fast**: gpt-5-nano, gemini-2.5-flash, gemini-2.0-flash, gpt-4o-mini
146
+ - **Ultra-Fast**: gpt-5-nano, flash, gemini-2.0-flash, gpt-4o-mini
147
147
  - **Good Balance**: gpt-5-mini, o4-mini, grok-code-fast-1
148
148
 
149
149
  ### For Large Context Windows
@@ -17,7 +17,7 @@ const logger = createLogger('summarization');
17
17
  // Default fast models for summarization tasks (prioritize GPT-5-nano for speed)
18
18
  const FAST_MODELS = {
19
19
  openai: 'gpt-5-nano', // Fastest GPT-5 model with minimal reasoning
20
- google: 'gemini-2.5-flash',
20
+ google: 'flash',
21
21
  xai: 'grok-4',
22
22
  anthropic: 'claude-3-5-haiku-latest',
23
23
  mistral: 'mistral-small-latest',
package/src/tools/chat.js CHANGED
@@ -418,7 +418,7 @@ function resolveAutoModel(model, providerName) {
418
418
  const defaults = {
419
419
  openai: 'gpt-5',
420
420
  xai: 'grok-4-0709',
421
- google: 'gemini-2.5-pro',
421
+ google: 'gemini-pro',
422
422
  anthropic: 'claude-sonnet-4-20250514',
423
423
  mistral: 'magistral-medium-2506',
424
424
  deepseek: 'deepseek-reasoner',
@@ -934,7 +934,7 @@ chatTool.inputSchema = {
934
934
  model: {
935
935
  type: 'string',
936
936
  description:
937
- 'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-2.5-pro", "grok-4-0709". Defaults to auto-selection.',
937
+ 'AI model to use. Examples: "auto" (recommended), "gpt-5", "gemini-pro", "grok-4-0709". Defaults to auto-selection.',
938
938
  },
939
939
  files: {
940
940
  type: 'array',
@@ -655,7 +655,7 @@ function getDefaultModelForProvider(providerName) {
655
655
  const defaults = {
656
656
  openai: 'gpt-5',
657
657
  xai: 'grok-4-0709',
658
- google: 'gemini-2.5-pro',
658
+ google: 'gemini-pro',
659
659
  anthropic: 'claude-sonnet-4-20250514',
660
660
  mistral: 'magistral-medium-2506',
661
661
  deepseek: 'deepseek-reasoner',
@@ -1500,7 +1500,7 @@ consensusTool.inputSchema = {
1500
1500
  items: { type: 'string' },
1501
1501
  minItems: 1,
1502
1502
  description:
1503
- 'List of models to consult. Example: ["gpt-5", "gemini-2.5-pro", "grok-4-0709"]',
1503
+ 'List of models to consult. Example: ["gpt-5", "gemini-pro", "grok-4-0709"]',
1504
1504
  },
1505
1505
  files: {
1506
1506
  type: 'array',