universal-agent-memory 1.0.6 → 1.0.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "universal-agent-memory",
3
- "version": "1.0.6",
3
+ "version": "1.0.8",
4
4
  "description": "Universal AI agent memory system - CLAUDE.md templates, memory, worktrees for Claude Code, Factory.AI, VSCode, OpenCode",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
@@ -1,7 +1,28 @@
1
1
  <!--
2
- CLAUDE.md Universal Template - v10.5
2
+ CLAUDE.md Universal Template - v10.7
3
3
 
4
4
  CHANGES IN THIS VERSION:
5
+ - Added Verification Patterns (P27-P31) from Terminal-Bench failure analysis
6
+ - Pattern 27: Output Directory Cleanup (ODC) - remove non-required files
7
+ - Pattern 28: Service Smoke Test (SST) - test services before completing
8
+ - Pattern 29: Multi-Solution Discovery (MSD) - find ALL valid solutions
9
+ - Pattern 30: Performance Threshold Tuning (PTT) - iterate to meet thresholds
10
+ - Pattern 31: Round-Trip Verification (RTV) - verify compress/decompress works
11
+ - Updated Pattern Router to include verification patterns
12
+ - Added 5 completion gates (output, tests, constraints, cleanup, services)
13
+
14
+ PREVIOUS (v10.6):
15
+ - Added Domain-Specific Patterns (P21-P26)
16
+ - Pattern 21: Chess Engine Integration (CEI) - use Stockfish, not reasoning
17
+ - Pattern 22: Git Recovery Forensics (GRF) - backup first, forensic approach
18
+ - Pattern 23: Compression Impossibility Detection (CID) - refuse impossible tasks
19
+ - Pattern 24: Polyglot Code Construction (PCC) - search for existing examples
20
+ - Pattern 25: Service Configuration Pipeline (SCP) - ordered setup, test each
21
+ - Pattern 26: Near-Miss Iteration (NMI) - iterate on partial success tasks
22
+ - Updated Pattern Router to include domain patterns
23
+ - Added 30% time budget reservation for iteration
24
+
25
+ PREVIOUS (v10.5):
5
26
  - STRENGTHENED Pattern Router: Now requires explicit analysis block output
6
27
  - STRENGTHENED Constraint Extraction: Mandatory checklist with verification
7
28
  - STRENGTHENED Adversarial Thinking: Explicit attack vector enumeration
@@ -424,6 +445,261 @@ ADVERSARIAL: "How do I craft input that bypasses the filter?"
424
445
 
425
446
  ---
426
447
 
448
+ ## 🎯 DOMAIN-SPECIFIC PATTERNS
449
+
450
+ **Task-specific guidance for common failure categories.**
451
+
452
+ ### Pattern 21: Chess Engine Integration (CEI)
453
+ Chess tasks require external engines - reasoning alone cannot find optimal moves.
454
+
455
+ **Detection**: Keywords "chess", "best move", "FEN", "PGN", "checkmate"
456
+
457
+ **Protocol**:
458
+ 1. Check for Stockfish: `which stockfish || apt-get install -y stockfish`
459
+ 2. Parse board state to FEN format
460
+ 3. Use Stockfish for analysis: `echo "position fen $FEN" | stockfish`
461
+ 4. Extract best move from engine output
462
+
463
+ ```bash
464
+ # Stockfish one-liner for best move
465
+ echo -e "position fen rnbqkbnr/pppppppp/8/8/8/8/PPPPPPPP/RNBQKBNR w KQkq - 0 1\ngo depth 20" | stockfish | grep "bestmove"
466
+ ```
467
+
468
+ **NEVER**: Try to calculate chess moves through reasoning - use the engine.
469
+
470
+ ### Pattern 22: Git Recovery Forensics (GRF)
471
+ Git recovery tasks require forensic approach, not standard git commands.
472
+
473
+ **Detection**: Keywords "recover", "corrupted", "lost commit", "fix git", "reflog"
474
+
475
+ **Protocol**:
476
+ 1. **BACKUP FIRST**: `cp -r .git .git.bak`
477
+ 2. Check integrity: `git fsck --full --no-dangling`
478
+ 3. Check reflog: `git reflog` (commits survive even after reset)
479
+ 4. Check loose objects: `find .git/objects -type f`
480
+ 5. Recover from pack files if needed
481
+
482
+ ```bash
483
+ # Recovery sequence
484
+ git fsck --full 2>&1 | tee /tmp/git-fsck.log
485
+ git reflog --all
486
+ git log --all --oneline --graph
487
+ ```
488
+
489
+ **NEVER**: Run destructive commands without backup. Use `--dry-run` first.
490
+
491
+ ### Pattern 23: Compression Impossibility Detection (CID)
492
+ Some compression tasks are mathematically impossible.
493
+
494
+ **Detection**: Keywords "compress", "codegolf", "under X bytes", "minimal size"
495
+
496
+ **Impossibility Markers**:
497
+ - ML model weights (GPT-2 = 500MB+, cannot be <1MB without losing function)
498
+ - Random/encrypted data (incompressible by definition)
499
+ - Asking for compression ratio beyond information-theoretic limits
500
+
501
+ ```bash
502
+ # If task asks for impossible compression:
503
+ echo "IMPOSSIBLE: Task requires compressing ${ORIGINAL_SIZE} to ${TARGET_SIZE}.
504
+ Shannon's source coding theorem proves this is impossible for this data type.
505
+ Minimum achievable size is approximately ${MIN_POSSIBLE}." > /app/IMPOSSIBLE.md
506
+ ```
507
+
508
+ **NEVER**: Attempt to compress ML weights to <1% of original size.
509
+
510
+ ### Pattern 24: Polyglot Code Construction (PCC)
511
+ Polyglot tasks require specific compiler/interpreter quirks.
512
+
513
+ **Detection**: Keywords "polyglot", "works in both", "compile as X and Y"
514
+
515
+ **Known Polyglot Techniques**:
516
+ - C/Python: Use `#if 0` blocks and string literals
517
+ - Rust/C: Careful use of macros and comments
518
+ - Shell/Python: Shebang and triple-quote tricks
519
+
520
+ ```c
521
+ // C and Python polyglot example
522
+ #if 0
523
+ """
524
+ #endif
525
+ #include <stdio.h>
526
+ int main() { printf("Hello from C\n"); return 0; }
527
+ #if 0
528
+ """
529
+ print("Hello from Python")
530
+ #endif
531
+ ```
532
+
533
+ **Protocol**: Search for existing polyglot examples before implementing.
534
+
535
+ ### Pattern 25: Service Configuration Pipeline (SCP)
536
+ Multi-service configuration requires ordered setup.
537
+
538
+ **Detection**: Keywords "configure", "server", "webserver", "service", "daemon"
539
+
540
+ **Protocol**:
541
+ 1. **Identify all services** needed (nginx, git, ssh, etc.)
542
+ 2. **Check service status**: `systemctl status <service>`
543
+ 3. **Configure in dependency order** (base → dependent)
544
+ 4. **Test each service** before moving to next
545
+ 5. **Verify end-to-end** after all configured
546
+
547
+ ```bash
548
+ # Service configuration pattern
549
+ for svc in nginx git-daemon ssh; do
550
+ systemctl status $svc || systemctl start $svc
551
+ systemctl is-active $svc || echo "FAILED: $svc"
552
+ done
553
+ ```
554
+
555
+ ### Pattern 26: Near-Miss Iteration (NMI)
556
+ When tests show >50% passing, focus on specific failing tests.
557
+
558
+ **Detection**: Test results show partial success (e.g., 8/9, 6/7, 5/6)
559
+
560
+ **Protocol**:
561
+ 1. Run tests with verbose output: `pytest -v 2>&1 | tee results.txt`
562
+ 2. Extract ONLY failing test names
563
+ 3. Read failing test code to understand exact requirement
564
+ 4. Fix specific issue without breaking passing tests
565
+ 5. Re-run ONLY failing tests first: `pytest test_file.py::test_name -v`
566
+ 6. Then run full suite to verify no regressions
567
+
568
+ ```bash
569
+ # Near-miss iteration loop
570
+ while true; do
571
+ pytest -v 2>&1 | tee /tmp/results.txt
572
+ FAILED=$(grep "FAILED" /tmp/results.txt | head -1)
573
+ [ -z "$FAILED" ] && echo "ALL PASS" && break
574
+ echo "Fixing: $FAILED"
575
+ # ... fix specific test ...
576
+ done
577
+ ```
578
+
579
+ **Reserve 30% of time budget for near-miss iteration.**
580
+
581
+ ### Pattern 27: Output Directory Cleanup (ODC)
582
+ Tests often check for ONLY specific files in output directories.
583
+
584
+ **Detection**: Tasks mentioning "single file", "only", constraints on output directory contents
585
+
586
+ **Protocol**:
587
+ 1. **Before completing**, list output directory: `ls /app/output/`
588
+ 2. **Remove non-required files**: compiled binaries, temp files, backups
589
+ 3. **Keep ONLY the required outputs** as specified in task
590
+
591
+ ```bash
592
+ # Clean output directory - keep only required file
593
+ cd /app/polyglot
594
+ ls -la # Check what's there
595
+ rm -f *.o *.out main cmain # Remove compiled artifacts
596
+ ls -la # Verify only main.rs remains
597
+ ```
598
+
599
+ **Common mistakes**: Leaving compiled binaries, test files, or backup files in output directories.
600
+
601
+ ### Pattern 28: Service Smoke Test (SST)
602
+ Services must be tested BEFORE claiming completion.
603
+
604
+ **Detection**: Tasks involving servers, daemons, APIs, web services
605
+
606
+ **Protocol**:
607
+ 1. Start the service
608
+ 2. **Immediately test it** with curl/wget/client
609
+ 3. If test fails: debug, fix, restart, re-test
610
+ 4. Only complete when test PASSES
611
+
612
+ ```bash
613
+ # Service verification pattern
614
+ # 1. Start service
615
+ systemctl start nginx
616
+
617
+ # 2. IMMEDIATELY test it
618
+ curl -v http://localhost:80/ 2>&1 | grep -q "200 OK" || echo "FAIL: nginx not serving"
619
+
620
+ # 3. For PyPI server
621
+ python -m http.server 8080 &
622
+ curl -s http://localhost:8080/simple/ | grep -q "href" || echo "FAIL: PyPI not serving"
623
+
624
+ # 4. For git server
625
+ git ls-remote git@localhost:/repo.git 2>&1 | grep -q "refs" || echo "FAIL: git not serving"
626
+ ```
627
+
628
+ **NEVER complete without a successful smoke test.**
629
+
630
+ ### Pattern 29: Multi-Solution Discovery (MSD)
631
+ Some tasks require finding ALL valid solutions, not just one.
632
+
633
+ **Detection**: Keywords "all moves", "both solutions", "list all", "find every"
634
+
635
+ **Protocol**:
636
+ 1. **Parse task for multi-solution requirement**
637
+ 2. For chess: Find ALL checkmate moves, not just the first one
638
+ 3. For puzzles: Enumerate all valid answers
639
+ 4. **Output ALL solutions** in required format
640
+
641
+ ```bash
642
+ # Chess: Find ALL checkmate-in-one moves
643
+ # Don't stop at first bestmove, analyze ALL legal moves
644
+ echo -e "position fen $FEN\ngo depth 20" | stockfish | grep "bestmove"
645
+ # Then check if there are MULTIPLE winning moves
646
+ echo -e "position fen $FEN\ngo movetime 5000 multipv 5" | stockfish
647
+ # Extract ALL moves that lead to checkmate
648
+ ```
649
+
650
+ ### Pattern 30: Performance Threshold Tuning (PTT)
651
+ Tasks with numeric thresholds require iterative tuning.
652
+
653
+ **Detection**: Keywords "win rate", "accuracy", "percentage", "threshold", "at least X%"
654
+
655
+ **Protocol**:
656
+ 1. Run initial implementation
657
+ 2. Check if thresholds are met
658
+ 3. If not: analyze what's missing, tune parameters
659
+ 4. Re-run until thresholds pass
660
+
661
+ ```bash
662
+ # Performance tuning loop for Core Wars
663
+ while true; do
664
+ # Run tests
665
+ pytest -v 2>&1 | tee results.txt
666
+
667
+ # Check for threshold failures
668
+ if grep -q "need.*%+" results.txt; then
669
+ echo "Threshold not met - tuning..."
670
+ # Analyze which opponent needs better strategy
671
+ # Modify warrior code
672
+ # Re-run
673
+ else
674
+ break
675
+ fi
676
+ done
677
+ ```
678
+
679
+ **For compression tasks**: If decompression fails, verify round-trip: `compress | decompress == original`
680
+
681
+ ### Pattern 31: Round-Trip Verification (RTV)
682
+ For transform/encode/compress tasks, verify the reverse operation.
683
+
684
+ **Detection**: Keywords "compress", "encode", "serialize", "encrypt", and task mentions "decompress", "decode", etc.
685
+
686
+ **Protocol**:
687
+ 1. Create test data
688
+ 2. Apply forward transform (compress)
689
+ 3. **Immediately apply reverse** (decompress)
690
+ 4. **Verify original == result**
691
+ 5. Fix if not matching
692
+
693
+ ```bash
694
+ # Round-trip verification for compression
695
+ echo "original data" > /tmp/original.txt
696
+ python compress.py /tmp/original.txt /tmp/compressed.bin
697
+ python decompress.py /tmp/compressed.bin /tmp/recovered.txt
698
+ diff /tmp/original.txt /tmp/recovered.txt && echo "PASS" || echo "FAIL: round-trip broken"
699
+ ```
700
+
701
+ ---
702
+
427
703
  ## ⚡ SESSION START PROTOCOL
428
704
 
429
705
  **EXECUTE IMMEDIATELY before any response:**