newtype-profile 1.0.33 → 1.0.34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli/index.js CHANGED
@@ -2253,7 +2253,7 @@ var require_picocolors = __commonJS((exports, module) => {
2253
2253
  var require_package = __commonJS((exports, module) => {
2254
2254
  module.exports = {
2255
2255
  name: "newtype-profile",
2256
- version: "1.0.33",
2256
+ version: "1.0.34",
2257
2257
  description: "AI Agent Collaboration System for Content Creation - Based on oh-my-opencode",
2258
2258
  main: "dist/index.js",
2259
2259
  types: "dist/index.d.ts",
@@ -22178,9 +22178,7 @@ var BuiltinAgentNameSchema = exports_external.enum([
22178
22178
  "editor"
22179
22179
  ]);
22180
22180
  var BuiltinSkillNameSchema = exports_external.enum([
22181
- "playwright",
22182
- "frontend-ui-ux",
22183
- "git-master"
22181
+ "playwright"
22184
22182
  ]);
22185
22183
  var OverridableAgentNameSchema = exports_external.enum([
22186
22184
  "build",
@@ -22223,12 +22221,13 @@ var HookNameSchema = exports_external.enum([
22223
22221
  "auto-slash-command",
22224
22222
  "edit-error-recovery",
22225
22223
  "prometheus-md-only",
22226
- "start-work",
22227
22224
  "chief-orchestrator"
22228
22225
  ]);
22229
22226
  var BuiltinCommandNameSchema = exports_external.enum([
22230
22227
  "init-deep",
22231
- "start-work"
22228
+ "ralph-loop",
22229
+ "cancel-ralph",
22230
+ "switch"
22232
22231
  ]);
22233
22232
  var AgentOverrideConfigSchema = exports_external.object({
22234
22233
  model: exports_external.string().optional(),
@@ -11,8 +11,6 @@ export declare const BuiltinAgentNameSchema: z.ZodEnum<{
11
11
  }>;
12
12
  export declare const BuiltinSkillNameSchema: z.ZodEnum<{
13
13
  playwright: "playwright";
14
- "frontend-ui-ux": "frontend-ui-ux";
15
- "git-master": "git-master";
16
14
  }>;
17
15
  export declare const OverridableAgentNameSchema: z.ZodEnum<{
18
16
  chief: "chief";
@@ -64,12 +62,13 @@ export declare const HookNameSchema: z.ZodEnum<{
64
62
  "auto-slash-command": "auto-slash-command";
65
63
  "edit-error-recovery": "edit-error-recovery";
66
64
  "prometheus-md-only": "prometheus-md-only";
67
- "start-work": "start-work";
68
65
  "chief-orchestrator": "chief-orchestrator";
69
66
  }>;
70
67
  export declare const BuiltinCommandNameSchema: z.ZodEnum<{
71
- "start-work": "start-work";
68
+ "ralph-loop": "ralph-loop";
72
69
  "init-deep": "init-deep";
70
+ "cancel-ralph": "cancel-ralph";
71
+ switch: "switch";
73
72
  }>;
74
73
  export declare const AgentOverrideConfigSchema: z.ZodObject<{
75
74
  model: z.ZodOptional<z.ZodString>;
@@ -848,8 +847,6 @@ export declare const OhMyOpenCodeConfigSchema: z.ZodObject<{
848
847
  }>>>;
849
848
  disabled_skills: z.ZodOptional<z.ZodArray<z.ZodEnum<{
850
849
  playwright: "playwright";
851
- "frontend-ui-ux": "frontend-ui-ux";
852
- "git-master": "git-master";
853
850
  }>>>;
854
851
  disabled_hooks: z.ZodOptional<z.ZodArray<z.ZodEnum<{
855
852
  "todo-continuation-enforcer": "todo-continuation-enforcer";
@@ -881,12 +878,13 @@ export declare const OhMyOpenCodeConfigSchema: z.ZodObject<{
881
878
  "auto-slash-command": "auto-slash-command";
882
879
  "edit-error-recovery": "edit-error-recovery";
883
880
  "prometheus-md-only": "prometheus-md-only";
884
- "start-work": "start-work";
885
881
  "chief-orchestrator": "chief-orchestrator";
886
882
  }>>>;
887
883
  disabled_commands: z.ZodOptional<z.ZodArray<z.ZodEnum<{
888
- "start-work": "start-work";
884
+ "ralph-loop": "ralph-loop";
889
885
  "init-deep": "init-deep";
886
+ "cancel-ralph": "cancel-ralph";
887
+ switch: "switch";
890
888
  }>>>;
891
889
  agents: z.ZodOptional<z.ZodObject<{
892
890
  build: z.ZodOptional<z.ZodObject<{
@@ -1 +1 @@
1
- export declare const INIT_DEEP_TEMPLATE = "# /init-deep\n\nGenerate hierarchical AGENTS.md files. Root + complexity-scored subdirectories.\n\n## Usage\n\n```\n/init-deep # Update mode: modify existing + create new where warranted\n/init-deep --create-new # Read existing \u2192 remove all \u2192 regenerate from scratch\n/init-deep --max-depth=2 # Limit directory depth (default: 3)\n```\n\n---\n\n## Workflow (High-Level)\n\n1. **Discovery + Analysis** (concurrent)\n - Delegate to Deputy for codebase exploration\n - Main session: bash structure + LSP codemap + read existing AGENTS.md\n2. **Score & Decide** - Determine AGENTS.md locations from merged findings\n3. **Generate** - Root first, then subdirs in parallel\n4. **Review** - Deduplicate, trim, validate\n\n<critical>\n**TodoWrite ALL phases. Mark in_progress \u2192 completed in real-time.**\n```\nTodoWrite([\n { id: \"discovery\", content: \"Delegate to Deputy for exploration + LSP codemap + read existing\", status: \"pending\", priority: \"high\" },\n { id: \"scoring\", content: \"Score directories, determine locations\", status: \"pending\", priority: \"high\" },\n { id: \"generate\", content: \"Generate AGENTS.md files (root + subdirs)\", status: \"pending\", priority: \"high\" },\n { id: \"review\", content: \"Deduplicate, validate, trim\", status: \"pending\", priority: \"medium\" }\n])\n```\n</critical>\n\n---\n\n## Phase 1: Discovery + Analysis (Concurrent)\n\n**Mark \"discovery\" as in_progress.**\n\n### Delegate Exploration to Deputy\n\nDelegate comprehensive codebase analysis to Deputy:\n\n```\nchief_task(\n subagent_type=\"deputy\",\n prompt=\"Analyze codebase for AGENTS.md generation. Research:\n 1. Project structure: standard patterns vs deviations\n 2. Entry points: main files and non-standard organization\n 3. Conventions: config files (.eslintrc, pyproject.toml, .editorconfig)\n 4. Anti-patterns: 'DO NOT', 'NEVER', 'ALWAYS', 'DEPRECATED' comments\n 5. Build/CI: .github/workflows, Makefile patterns\n 6. Test patterns: test configs and structure\n \n Return structured findings for documentation generation.\",\n run_in_background=true,\n skills=[]\n)\n```\n\n<dynamic-analysis>\n**DYNAMIC DEPTH**: For large projects, Deputy will automatically dispatch multiple researcher tasks:\n\n| Factor | Threshold | Deputy Action |\n|--------|-----------|---------------|\n| **Total files** | >100 | More thorough file analysis |\n| **Directory depth** | \u22654 | Deep module exploration |\n| **Large files (>500 lines)** | >10 files | Complexity hotspot analysis |\n| **Monorepo** | detected | Per-package analysis |\n\n```bash\n# Measure project scale first (optional - Deputy will assess)\ntotal_files=$(find . -type f -not -path '*/node_modules/*' -not -path '*/.git/*' | wc -l)\nmax_depth=$(find . -type d -not -path '*/node_modules/*' -not -path '*/.git/*' | awk -F/ '{print NF}' | sort -rn | head -1)\n```\n\nFor very large projects, send additional Deputy tasks:\n```\nchief_task(\n subagent_type=\"deputy\",\n prompt=\"Deep analysis for large codebase: Find files >500 lines, complexity hotspots, cross-cutting concerns.\",\n run_in_background=true,\n skills=[]\n)\n```\n</dynamic-analysis>\n\n### Main Session: Concurrent Analysis\n\n**While background agents run**, main session does:\n\n#### 1. Bash Structural Analysis\n```bash\n# Directory depth + file counts\nfind . -type d -not -path '*/\\.*' -not -path '*/node_modules/*' -not -path '*/venv/*' -not -path '*/dist/*' -not -path '*/build/*' | awk -F/ '{print NF-1}' | sort -n | uniq -c\n\n# Files per directory (top 30)\nfind . -type f -not -path '*/\\.*' -not -path '*/node_modules/*' | sed 's|/[^/]*$||' | sort | uniq -c | sort -rn | head -30\n\n# Code concentration by extension\nfind . -type f \\( -name \"*.py\" -o -name \"*.ts\" -o -name \"*.tsx\" -o -name \"*.js\" -o -name \"*.go\" -o -name \"*.rs\" \\) -not -path '*/node_modules/*' | sed 's|/[^/]*$||' | sort | uniq -c | sort -rn | head -20\n\n# Existing AGENTS.md / CLAUDE.md\nfind . -type f \\( -name \"AGENTS.md\" -o -name \"CLAUDE.md\" \\) -not -path '*/node_modules/*' 2>/dev/null\n```\n\n#### 2. Read Existing AGENTS.md\n```\nFor each existing file found:\n Read(filePath=file)\n Extract: key insights, conventions, anti-patterns\n Store in EXISTING_AGENTS map\n```\n\nIf `--create-new`: Read all existing first (preserve context) \u2192 then delete all \u2192 regenerate.\n\n#### 3. LSP Codemap (if available)\n```\nlsp_servers() # Check availability\n\n# Entry points (parallel)\nlsp_document_symbols(filePath=\"src/index.ts\")\nlsp_document_symbols(filePath=\"main.py\")\n\n# Key symbols (parallel)\nlsp_workspace_symbols(filePath=\".\", query=\"class\")\nlsp_workspace_symbols(filePath=\".\", query=\"interface\")\nlsp_workspace_symbols(filePath=\".\", query=\"function\")\n\n# Centrality for top exports\nlsp_find_references(filePath=\"...\", line=X, character=Y)\n```\n\n**LSP Fallback**: If unavailable, rely on Deputy + AST-grep.\n\n### Collect Background Results\n\n```\n// After main session analysis done, collect all task results\nfor each task_id: background_output(task_id=\"...\")\n```\n\n**Merge: bash + LSP + existing + Deputy findings. Mark \"discovery\" as completed.**\n\n---\n\n## Phase 2: Scoring & Location Decision\n\n**Mark \"scoring\" as in_progress.**\n\n### Scoring Matrix\n\n| Factor | Weight | High Threshold | Source |\n|--------|--------|----------------|--------|\n| File count | 3x | >20 | bash |\n| Subdir count | 2x | >5 | bash |\n| Code ratio | 2x | >70% | bash |\n| Unique patterns | 1x | Has own config | Deputy |\n| Module boundary | 2x | Has index.ts/__init__.py | bash |\n| Symbol density | 2x | >30 symbols | LSP |\n| Export count | 2x | >10 exports | LSP |\n| Reference centrality | 3x | >20 refs | LSP |\n\n### Decision Rules\n\n| Score | Action |\n|-------|--------|\n| **Root (.)** | ALWAYS create |\n| **>15** | Create AGENTS.md |\n| **8-15** | Create if distinct domain |\n| **<8** | Skip (parent covers) |\n\n### Output\n```\nAGENTS_LOCATIONS = [\n { path: \".\", type: \"root\" },\n { path: \"src/hooks\", score: 18, reason: \"high complexity\" },\n { path: \"src/api\", score: 12, reason: \"distinct domain\" }\n]\n```\n\n**Mark \"scoring\" as completed.**\n\n---\n\n## Phase 3: Generate AGENTS.md\n\n**Mark \"generate\" as in_progress.**\n\n### Root AGENTS.md (Full Treatment)\n\n```markdown\n# PROJECT KNOWLEDGE BASE\n\n**Generated:** {TIMESTAMP}\n**Commit:** {SHORT_SHA}\n**Branch:** {BRANCH}\n\n## OVERVIEW\n{1-2 sentences: what + core stack}\n\n## STRUCTURE\n\\`\\`\\`\n{root}/\n\u251C\u2500\u2500 {dir}/ # {non-obvious purpose only}\n\u2514\u2500\u2500 {entry}\n\\`\\`\\`\n\n## WHERE TO LOOK\n| Task | Location | Notes |\n|------|----------|-------|\n\n## CODE MAP\n{From LSP - skip if unavailable or project <10 files}\n\n| Symbol | Type | Location | Refs | Role |\n|--------|------|----------|------|------|\n\n## CONVENTIONS\n{ONLY deviations from standard}\n\n## ANTI-PATTERNS (THIS PROJECT)\n{Explicitly forbidden here}\n\n## UNIQUE STYLES\n{Project-specific}\n\n## COMMANDS\n\\`\\`\\`bash\n{dev/test/build}\n\\`\\`\\`\n\n## NOTES\n{Gotchas}\n```\n\n**Quality gates**: 50-150 lines, no generic advice, no obvious info.\n\n### Subdirectory AGENTS.md (Parallel)\n\nLaunch document-writer agents for each location:\n\n```\nfor loc in AGENTS_LOCATIONS (except root):\n sisyphus_task(agent=\"document-writer\", prompt=\\`\n Generate AGENTS.md for: ${loc.path}\n - Reason: ${loc.reason}\n - 30-80 lines max\n - NEVER repeat parent content\n - Sections: OVERVIEW (1 line), STRUCTURE (if >5 subdirs), WHERE TO LOOK, CONVENTIONS (if different), ANTI-PATTERNS\n \\`)\n```\n\n**Wait for all. Mark \"generate\" as completed.**\n\n---\n\n## Phase 4: Review & Deduplicate\n\n**Mark \"review\" as in_progress.**\n\nFor each generated file:\n- Remove generic advice\n- Remove parent duplicates\n- Trim to size limits\n- Verify telegraphic style\n\n**Mark \"review\" as completed.**\n\n---\n\n## Final Report\n\n```\n=== init-deep Complete ===\n\nMode: {update | create-new}\n\nFiles:\n \u2713 ./AGENTS.md (root, {N} lines)\n \u2713 ./src/hooks/AGENTS.md ({N} lines)\n\nDirs Analyzed: {N}\nAGENTS.md Created: {N}\nAGENTS.md Updated: {N}\n\nHierarchy:\n ./AGENTS.md\n \u2514\u2500\u2500 src/hooks/AGENTS.md\n```\n\n---\n\n## Anti-Patterns\n\n- **Static approach**: MUST adapt depth of analysis based on project size\n- **Sequential execution**: MUST parallel (Deputy + LSP concurrent)\n- **Ignoring existing**: ALWAYS read existing first, even with --create-new\n- **Over-documenting**: Not every dir needs AGENTS.md\n- **Redundancy**: Child never repeats parent\n- **Generic content**: Remove anything that applies to ALL projects\n- **Verbose style**: Telegraphic or die";
1
+ export declare const INIT_DEEP_TEMPLATE = "# /init-deep\n\nGenerate a KNOWLEDGE.md file to help AI understand a document repository or knowledge base.\n\n## Usage\n\n```\n/init-deep # Analyze and generate/update KNOWLEDGE.md\n/init-deep --create-new # Delete existing and regenerate from scratch\n/init-deep --max-depth=3 # Limit directory scan depth (default: 5)\n```\n\n---\n\n## What This Does\n\nScans a document repository (any structure) and generates a knowledge index that helps AI:\n- Understand what content exists\n- Know where to find specific types of documents\n- Recognize the organizational pattern (even if non-standard)\n\n---\n\n## Workflow\n\n<critical>\n**TodoWrite ALL phases. Mark in_progress \u2192 completed in real-time.**\n```\nTodoWrite([\n { id: \"scan\", content: \"Scan directory structure and file types\", status: \"pending\", priority: \"high\" },\n { id: \"analyze\", content: \"Analyze content and extract summaries\", status: \"pending\", priority: \"high\" },\n { id: \"generate\", content: \"Generate KNOWLEDGE.md\", status: \"pending\", priority: \"high\" },\n { id: \"review\", content: \"Review and refine output\", status: \"pending\", priority: \"medium\" }\n])\n```\n</critical>\n\n---\n\n## Phase 1: Directory Scan\n\n**Mark \"scan\" as in_progress.**\n\n### 1.1 Discover Structure\n\n```bash\n# Get directory tree (exclude hidden, node_modules, etc.)\nfind . -type d -not -path '*/\\.*' -not -path '*/node_modules/*' -not -path '*/__pycache__/*' | head -100\n\n# Count files by type\nfind . -type f -not -path '*/\\.*' | sed 's/.*\\.//' | sort | uniq -c | sort -rn | head -20\n\n# List document files\nfind . -type f \\( -name \"*.md\" -o -name \"*.pdf\" -o -name \"*.docx\" -o -name \"*.txt\" -o -name \"*.rtf\" \\) -not -path '*/\\.*' | head -100\n\n# List media files\nfind . -type f \\( -name \"*.png\" -o -name \"*.jpg\" -o -name \"*.jpeg\" -o -name \"*.gif\" -o -name \"*.svg\" -o -name \"*.mp4\" -o -name \"*.mp3\" \\) -not -path '*/\\.*' | head -50\n\n# Check for existing KNOWLEDGE.md or similar\nfind . -type f \\( -name \"KNOWLEDGE.md\" -o -name \"README.md\" -o -name \"INDEX.md\" \\) -not -path '*/\\.*'\n\n# File count per directory (top 30)\nfind . -type f -not -path '*/\\.*' | sed 's|/[^/]*$||' | sort | uniq -c | sort -rn | head -30\n```\n\n### 1.2 Identify File Type Distribution\n\nBuild a mental model:\n```\nCONTENT_PROFILE:\n documents: N (md, pdf, docx, txt)\n images: N (png, jpg, svg)\n data: N (json, csv, xlsx)\n code: N (if any - py, js, ts)\n other: N\n \nSTRUCTURE_TYPE:\n - flat (all files in root)\n - shallow (1-2 levels)\n - deep (3+ levels)\n - mixed\n```\n\n**Mark \"scan\" as completed.**\n\n---\n\n## Phase 2: Content Analysis\n\n**Mark \"analyze\" as in_progress.**\n\n### 2.1 Read Key Documents\n\nFor each directory with documents:\n\n```\n# Priority reading order:\n1. README.md / INDEX.md (if exists)\n2. Largest markdown files (likely main content)\n3. Recently modified files\n4. Files with descriptive names\n```\n\nUse `Read` tool on markdown files to understand content.\n\n### 2.2 Extract PDF/Document Summaries\n\nFor PDFs and other binary documents:\n```\nUse look_at tool to extract:\n- Title\n- First page summary\n- Key topics\n```\n\n### 2.3 Infer Topics and Categories\n\nFrom file names and content, extract:\n- Main topics/themes\n- Naming patterns\n- Organizational logic (by date? by topic? by project?)\n\n**Mark \"analyze\" as completed.**\n\n---\n\n## Phase 3: Generate KNOWLEDGE.md\n\n**Mark \"generate\" as in_progress.**\n\n### Output Template\n\n```markdown\n# KNOWLEDGE BASE INDEX\n\n**Generated:** {TIMESTAMP}\n**Last Updated:** {DATE}\n**Total Files:** {N} documents, {N} images, {N} other\n\n---\n\n## OVERVIEW\n\n{1-3 sentences describing what this knowledge base contains and its purpose}\n\n---\n\n## STRUCTURE\n\n\\`\\`\\`\n{directory}/\n\u251C\u2500\u2500 {folder1}/ # {what's in here}\n\u2502 \u2514\u2500\u2500 ...\n\u251C\u2500\u2500 {folder2}/ # {what's in here}\n\u2514\u2500\u2500 {file.md} # {brief description}\n\\`\\`\\`\n\n---\n\n## KEY DOCUMENTS\n\n| File | Description | Topics |\n|------|-------------|--------|\n| {path} | {what it contains} | {tags} |\n| ... | ... | ... |\n\n---\n\n## TOPICS & TAGS\n\n- **{Topic 1}**: {related files or folders}\n- **{Topic 2}**: {related files or folders}\n- ...\n\n---\n\n## FOLDER GUIDE\n\n| Folder | Purpose | File Count |\n|--------|---------|------------|\n| {path} | {what to find here} | {N} |\n| ... | ... | ... |\n\n---\n\n## FILE TYPES\n\n| Type | Count | Location |\n|------|-------|----------|\n| Markdown (.md) | {N} | {where they are} |\n| PDF (.pdf) | {N} | {where they are} |\n| Images | {N} | {where they are} |\n| ... | ... | ... |\n\n---\n\n## NOTES\n\n{Any discovered patterns, conventions, or important observations}\n- {Naming convention if any}\n- {Organization pattern}\n- {Special files or folders}\n```\n\n### Writing Guidelines\n\n1. **Be specific** - Don't say \"various documents\", say what they actually are\n2. **Be concise** - One line per item, no verbose descriptions\n3. **Be accurate** - Only include what you actually found\n4. **Be helpful** - Focus on what helps AI find relevant content\n\n**Mark \"generate\" as completed.**\n\n---\n\n## Phase 4: Review\n\n**Mark \"review\" as in_progress.**\n\n1. Check KNOWLEDGE.md is not too long (50-200 lines ideal)\n2. Remove any generic/unhelpful entries\n3. Ensure all paths are correct\n4. Verify descriptions are accurate\n\nWrite the final KNOWLEDGE.md to the repository root.\n\n**Mark \"review\" as completed.**\n\n---\n\n## Final Report\n\n```\n=== init-deep Complete ===\n\nRepository: {path}\nFiles Analyzed: {N}\nFolders Scanned: {N}\n\nGenerated: ./KNOWLEDGE.md ({N} lines)\n\nKey Findings:\n- {Main content type}\n- {Organization pattern}\n- {Notable files}\n```\n\n---\n\n## Anti-Patterns\n\n- **Don't assume structure** - Every repository is different\n- **Don't skip binary files** - PDFs often contain key content\n- **Don't be verbose** - Keep entries concise\n- **Don't list everything** - Focus on key/representative files\n- **Don't ignore patterns** - Document any discovered conventions";
@@ -1,5 +1,5 @@
1
1
  import type { CommandDefinition } from "../claude-code-command-loader";
2
- export type BuiltinCommandName = "init-deep" | "ralph-loop" | "cancel-ralph" | "refactor" | "start-work" | "switch";
2
+ export type BuiltinCommandName = "init-deep" | "ralph-loop" | "cancel-ralph" | "switch";
3
3
  export interface BuiltinCommandConfig {
4
4
  disabled_commands?: BuiltinCommandName[];
5
5
  }
@@ -26,5 +26,4 @@ export { createAutoSlashCommandHook } from "./auto-slash-command";
26
26
  export { createEditErrorRecoveryHook } from "./edit-error-recovery";
27
27
  export { createPrometheusMdOnlyHook } from "./prometheus-md-only";
28
28
  export { createTaskResumeInfoHook } from "./task-resume-info";
29
- export { createStartWorkHook } from "./start-work";
30
29
  export { createChiefOrchestratorHook } from "./chief-orchestrator";