@tuteliq/mcp 3.3.5 → 3.4.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"detection.d.ts","sourceRoot":"","sources":["../../../src/tools/detection.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,yCAAyC,CAAC;AAEzE,OAAO,KAAK,EAAE,OAAO,EAAE,MAAM,cAAc,CAAC;AA6C5C,wBAAgB,sBAAsB,CAAC,MAAM,EAAE,SAAS,EAAE,MAAM,EAAE,OAAO,GAAG,IAAI,
|
|
1
|
+
{"version":3,"file":"detection.d.ts","sourceRoot":"","sources":["../../../src/tools/detection.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,yCAAyC,CAAC;AAEzE,OAAO,KAAK,EAAE,OAAO,EAAE,MAAM,cAAc,CAAC;AA6C5C,wBAAgB,sBAAsB,CAAC,MAAM,EAAE,SAAS,EAAE,MAAM,EAAE,OAAO,GAAG,IAAI,CA+N/E"}
|
|
@@ -201,10 +201,11 @@ ${result.rationale}
|
|
|
201
201
|
try {
|
|
202
202
|
const result = await client.analyze({ content, include });
|
|
203
203
|
const emoji = riskEmoji[result.risk_level] || '\u26AA';
|
|
204
|
-
|
|
204
|
+
let text = `## Safety Analysis Results
|
|
205
205
|
|
|
206
206
|
**Overall Risk:** ${emoji} ${result.risk_level.charAt(0).toUpperCase() + result.risk_level.slice(1)}
|
|
207
207
|
**Risk Score:** ${(result.risk_score * 100).toFixed(0)}%
|
|
208
|
+
**Confidence:** ${(result.confidence * 100).toFixed(0)}%
|
|
208
209
|
|
|
209
210
|
### Summary
|
|
210
211
|
${result.summary}
|
|
@@ -214,6 +215,9 @@ ${result.summary}
|
|
|
214
215
|
|
|
215
216
|
---
|
|
216
217
|
${result.bullying ? `\n**Bullying Check:** ${result.bullying.is_bullying ? '\u26A0\uFE0F Detected' : '\u2705 Clear'}\n` : ''}${result.unsafe ? `\n**Unsafe Content:** ${result.unsafe.unsafe ? '\u26A0\uFE0F Detected' : '\u2705 Clear'}\n` : ''}`;
|
|
218
|
+
// Show support resources from the unsafe sub-result if available
|
|
219
|
+
if (result.unsafe?.support)
|
|
220
|
+
text += formatSupportText(result.unsafe.support);
|
|
217
221
|
return {
|
|
218
222
|
structuredContent: { toolName: 'analyze', result, branding: { appName: 'Tuteliq' } },
|
|
219
223
|
content: [{ type: 'text', text }],
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@tuteliq/mcp",
|
|
3
|
-
"version": "3.
|
|
3
|
+
"version": "3.4.0",
|
|
4
4
|
"description": "MCP server for Tuteliq — 41 AI-powered tools for child safety, fraud detection, grooming, bullying, sextortion, and content moderation. Interactive UI widgets for Claude, Cursor, and MCP-compatible AI assistants.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "./dist/src/index.js",
|
|
@@ -73,7 +73,7 @@
|
|
|
73
73
|
"dependencies": {
|
|
74
74
|
"@modelcontextprotocol/ext-apps": "^1.1.0",
|
|
75
75
|
"@modelcontextprotocol/sdk": "^1.27.0",
|
|
76
|
-
"@tuteliq/sdk": "^2.
|
|
76
|
+
"@tuteliq/sdk": "^2.4.0",
|
|
77
77
|
"cors": "^2.8.5",
|
|
78
78
|
"express": "^4.21.0",
|
|
79
79
|
"zod": "^4.3.6"
|