trace.ai-cli 1.1.7 → 1.1.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "trace.ai-cli",
3
- "version": "1.1.7",
3
+ "version": "1.1.8",
4
4
  "description": "A powerful AI-powered CLI tool",
5
5
  "main": "index.js",
6
6
  "bin": {
@@ -37,7 +37,7 @@
37
37
  "ora": "^5.4.1",
38
38
  "os-utils": "^0.0.14",
39
39
  "systeminformation": "^5.27.1",
40
- "trace.ai-cli": "^1.1.6"
40
+ "trace.ai-cli": "^1.1.7"
41
41
  },
42
42
  "engines": {
43
43
  "node": ">=14.0.0"
@@ -518,7 +518,7 @@ function formatSystemInfoManually(sysInfo, prompt) {
518
518
  response += `\n`;
519
519
  }
520
520
 
521
- response += `*Query processed for user: m0v0dga_walmart*`;
521
+ response ;
522
522
  return response;
523
523
  }
524
524
 
@@ -575,7 +575,6 @@ async function processWithAI(prompt, context = '') {
575
575
 
576
576
  if (isSystemQuery) {
577
577
  try {
578
- console.log(`[${new Date().toISOString()}] Processing system information query...`);
579
578
 
580
579
  // First get basic system info to provide context
581
580
  const basicInfo = await getSystemInfo('basic');
@@ -589,7 +588,6 @@ async function processWithAI(prompt, context = '') {
589
588
  // Format the response using AI
590
589
  const formattedResponse = await formatSystemInfoResponse(systemInfo, prompt);
591
590
 
592
- console.log(`[${new Date().toISOString()}] System info query completed successfully`);
593
591
  return formattedResponse;
594
592
  } catch (sysError) {
595
593
  console.error(`[${new Date().toISOString()}] ❌ System info error:`, sysError.message);
@@ -599,7 +597,6 @@ async function processWithAI(prompt, context = '') {
599
597
  }
600
598
 
601
599
  // Regular AI processing for non-system queries
602
- console.log(`[${new Date().toISOString()}] Processing query...`);
603
600
  const models = ['kimi', 'mvrk', 'gma3', 'dsv3', 'qw32b', 'ms24b', 'll70b', 'qw3', 'nlm3'];
604
601
 
605
602
  const modelRequests = models.map(async (model) => {
@@ -647,7 +644,6 @@ async function processWithAI(prompt, context = '') {
647
644
  const encryptedResult = await finalResponse.text();
648
645
  const decryptedResult = decryptData(encryptedResult);
649
646
 
650
- console.log(`[${new Date().toISOString()}] Query completed successfully`);
651
647
  return decryptedResult.text || 'No response generated';
652
648
  } catch (error) {
653
649
  console.error(`[${new Date().toISOString()}] ❌ Processing error:`, error.message);