polydev-ai 1.8.39 → 1.8.41

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/lib/cliManager.js CHANGED
@@ -584,10 +584,10 @@ This is a known issue with @google/gemini-cli@0.3.4 and older Node.js versions.`
584
584
  // Claude Code uses --model flag
585
585
  args = ['--model', model, ...args, prompt];
586
586
  } else if (providerId === 'gemini_cli') {
587
- // Gemini CLI: -s for sandbox (no tools), -m for model, -p for prompt (headless mode)
588
- // Format: gemini -s -m gemini-2.0-flash -p "prompt"
589
- // -s flag disables agentic tool use which causes truncated "I will search..." outputs
590
- args = ['-s', '-m', model, '-p', prompt];
587
+ // Gemini CLI: -m for model, -p for prompt (headless mode)
588
+ // Add prompt prefix to prevent tool planning in non-interactive mode
589
+ const geminiPrompt = `Answer directly without using any tools, file operations, or searches. Do not say "I will search" or "I will look up". Provide your analysis immediately.\n\n${prompt}`;
590
+ args = ['-m', model, '-p', geminiPrompt];
591
591
  } else {
592
592
  // Default: just append prompt
593
593
  args = [...args, prompt];
@@ -595,8 +595,10 @@ This is a known issue with @google/gemini-cli@0.3.4 and older Node.js versions.`
595
595
  } else {
596
596
  // No model specified
597
597
  if (providerId === 'gemini_cli') {
598
- // Gemini CLI: -s for sandbox (no tools), -p for headless mode
599
- args = ['-s', '-p', prompt];
598
+ // Gemini CLI: -p for headless mode
599
+ // Add prompt prefix to prevent tool planning in non-interactive mode
600
+ const geminiPrompt = `Answer directly without using any tools, file operations, or searches. Do not say "I will search" or "I will look up". Provide your analysis immediately.\n\n${prompt}`;
601
+ args = ['-p', geminiPrompt];
600
602
  } else {
601
603
  args = [...args, prompt];
602
604
  }
package/mcp/README.md CHANGED
@@ -4,7 +4,7 @@ Get diverse AI perspectives from multiple LLMs via Model Context Protocol (MCP).
4
4
 
5
5
  ## Features
6
6
 
7
- - 🤖 **Multi-Model AI Perspectives**: Get responses from GPT-4, Claude, Gemini, and more
7
+ - 🤖 **Multi-Model AI Perspectives**: Get responses from multiple AI models simultaneously
8
8
  - 🔧 **Local CLI Detection**: Automatically detect and use Claude Code, Codex CLI, Gemini CLI
9
9
  - ⚡ **Smart Caching**: Intelligent refresh intervals based on CLI status
10
10
  - 🔄 **Fallback Support**: Local CLI + remote perspectives for comprehensive responses
@@ -99,7 +99,7 @@ session_timeout = 600
99
99
 
100
100
  ## Available Tools
101
101
 
102
- - `get_perspectives` - Get AI responses from multiple models
102
+ - `get_perspectives` - Get AI responses from multiple models (models managed by Polydev)
103
103
  - `force_cli_detection` - Force detection of local CLI tools
104
104
  - `get_cli_status` - Check status of CLI tools (Claude Code, Codex, Gemini)
105
105
  - `send_cli_prompt` - Send prompts to local CLI with perspectives fallback
@@ -111,8 +111,7 @@ session_timeout = 600
111
111
  ```typescript
112
112
  // Use the get_perspectives tool
113
113
  {
114
- "prompt": "How do I optimize React performance?",
115
- "models": ["gpt-4", "claude-3-sonnet", "gemini-pro"]
114
+ "prompt": "How do I optimize React performance?"
116
115
  }
117
116
  ```
118
117
 
package/mcp/manifest.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "polydev-perspectives",
3
- "version": "1.2.0",
3
+ "version": "1.3.0",
4
4
  "description": "Agentic workflow assistant - get diverse perspectives from multiple LLMs when stuck or need enhanced reasoning",
5
5
  "author": "Polydev AI",
6
6
  "license": "MIT",
@@ -23,22 +23,6 @@
23
23
  "description": "The prompt to get perspectives on",
24
24
  "minLength": 1
25
25
  },
26
- "models": {
27
- "type": "array",
28
- "description": "Array of model names to query (optional, defaults to [gpt-4, claude-3-sonnet, gemini-pro])",
29
- "items": {
30
- "type": "string",
31
- "enum": [
32
- "gpt-4",
33
- "gpt-3.5-turbo",
34
- "claude-3-sonnet",
35
- "claude-3-haiku",
36
- "gemini-pro",
37
- "gemini-1.5-pro"
38
- ]
39
- },
40
- "default": ["gpt-4", "claude-3-sonnet", "gemini-pro"]
41
- },
42
26
  "user_token": {
43
27
  "type": "string",
44
28
  "description": "Polydev user authentication token (get from dashboard)",
@@ -103,14 +87,10 @@
103
87
  "properties": {
104
88
  "responses": {
105
89
  "type": "array",
106
- "description": "Array of responses from different models",
90
+ "description": "Array of responses from different AI perspectives",
107
91
  "items": {
108
92
  "type": "object",
109
93
  "properties": {
110
- "model": {
111
- "type": "string",
112
- "description": "Model name"
113
- },
114
94
  "content": {
115
95
  "type": "string",
116
96
  "description": "Response content"
@@ -128,12 +108,12 @@
128
108
  "description": "Error message if the request failed"
129
109
  }
130
110
  },
131
- "required": ["model", "content"]
111
+ "required": ["content"]
132
112
  }
133
113
  },
134
114
  "total_tokens": {
135
115
  "type": "integer",
136
- "description": "Total tokens used across all models"
116
+ "description": "Total tokens used across all perspectives"
137
117
  },
138
118
  "total_latency_ms": {
139
119
  "type": "integer",
@@ -152,25 +132,6 @@
152
132
  "input": {
153
133
  "prompt": "I'm building a new API and considering TypeScript vs JavaScript. I need multiple perspectives to make the right choice.",
154
134
  "user_token": "poly_abc123..."
155
- },
156
- "output": {
157
- "responses": [
158
- {
159
- "model": "gpt-4",
160
- "content": "TypeScript offers several key advantages over JavaScript: 1) Static type checking catches errors at compile time...",
161
- "tokens_used": 150,
162
- "latency_ms": 1200
163
- },
164
- {
165
- "model": "claude-3-sonnet",
166
- "content": "TypeScript provides enhanced developer experience through: Type safety, better IDE support...",
167
- "tokens_used": 140,
168
- "latency_ms": 1100
169
- }
170
- ],
171
- "total_tokens": 290,
172
- "total_latency_ms": 1300,
173
- "cached": false
174
135
  }
175
136
  },
176
137
  {
@@ -178,7 +139,6 @@
178
139
  "input": {
179
140
  "prompt": "My React app is slow and I can't figure out why. Help me identify the bottlenecks.",
180
141
  "user_token": "poly_abc123...",
181
- "models": ["gpt-4", "claude-3-sonnet"],
182
142
  "project_memory": "full",
183
143
  "project_context": {
184
144
  "root_path": "/workspace/my-react-app",
@@ -191,8 +151,7 @@
191
151
  "description": "Agent needs diverse architectural opinions",
192
152
  "input": {
193
153
  "prompt": "I'm designing a microservices architecture for a fintech app. What are the key considerations and potential pitfalls?",
194
- "user_token": "poly_abc123...",
195
- "models": ["gpt-4", "claude-3-sonnet", "gemini-pro"]
154
+ "user_token": "poly_abc123..."
196
155
  }
197
156
  }
198
157
  ]
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "polydev-ai",
3
- "version": "1.8.39",
3
+ "version": "1.8.41",
4
4
  "description": "Agentic workflow assistant with CLI integration - get diverse perspectives from multiple LLMs when stuck or need enhanced reasoning",
5
5
  "keywords": [
6
6
  "mcp",
@@ -67,7 +67,7 @@
67
67
  "lucide-react": "^0.542.0",
68
68
  "marked": "^16.2.1",
69
69
  "next": "^15.5.7",
70
- "polydev-ai": "^1.8.38",
70
+ "polydev-ai": "^1.8.40",
71
71
  "posthog-js": "^1.157.2",
72
72
  "prismjs": "^1.30.0",
73
73
  "react": "^18.3.1",