omni-context-cli 0.0.69 → 0.0.71

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (61) hide show
  1. package/README.md +165 -2
  2. package/README.zh-CN.md +166 -0
  3. package/dist/bin/aarch64-apple-darwin/rg +0 -0
  4. package/dist/bin/x86_64-pc-windows-msvc/rg.exe +0 -0
  5. package/dist/bin/x86_64-unknown-linux-musl/rg +0 -0
  6. package/dist/cli.js +7 -7
  7. package/dist/clients/extension.vsix +0 -0
  8. package/dist/clients/web/assets/{_baseUniq-BdJJXNOW.js → _baseUniq-BFeEqFFb.js} +1 -1
  9. package/dist/clients/web/assets/{arc-CpfZ_DMQ.js → arc-C9Tr_R9S.js} +1 -1
  10. package/dist/clients/web/assets/{architectureDiagram-VXUJARFQ-CYTA7LQw.js → architectureDiagram-VXUJARFQ-DC6ztb2L.js} +1 -1
  11. package/dist/clients/web/assets/{blockDiagram-VD42YOAC-B3vgAk0t.js → blockDiagram-VD42YOAC-C2M_nJ2J.js} +1 -1
  12. package/dist/clients/web/assets/{c4Diagram-YG6GDRKO-1msHiZOI.js → c4Diagram-YG6GDRKO-B3QAoUTS.js} +1 -1
  13. package/dist/clients/web/assets/channel-B67y3J8I.js +1 -0
  14. package/dist/clients/web/assets/{chunk-4BX2VUAB-DvFObXE1.js → chunk-4BX2VUAB-CzqdeDCq.js} +1 -1
  15. package/dist/clients/web/assets/{chunk-55IACEB6-BNULYKvK.js → chunk-55IACEB6-s0-JRHl8.js} +1 -1
  16. package/dist/clients/web/assets/{chunk-B4BG7PRW-Gv_eP507.js → chunk-B4BG7PRW--mh2YYmd.js} +1 -1
  17. package/dist/clients/web/assets/{chunk-DI55MBZ5-wg4ctBQb.js → chunk-DI55MBZ5-TrICfnS-.js} +1 -1
  18. package/dist/clients/web/assets/{chunk-FMBD7UC4-Cz7PBt22.js → chunk-FMBD7UC4-C3oWsxDp.js} +1 -1
  19. package/dist/clients/web/assets/{chunk-QN33PNHL-CWhF6fiQ.js → chunk-QN33PNHL-BgMWLCf9.js} +1 -1
  20. package/dist/clients/web/assets/{chunk-QZHKN3VN-Be3bjn3u.js → chunk-QZHKN3VN-CFfsjT73.js} +1 -1
  21. package/dist/clients/web/assets/{chunk-TZMSLE5B-CyGYH1Au.js → chunk-TZMSLE5B-Dqbd8fEC.js} +1 -1
  22. package/dist/clients/web/assets/classDiagram-2ON5EDUG-B-txL_jH.js +1 -0
  23. package/dist/clients/web/assets/classDiagram-v2-WZHVMYZB-B-txL_jH.js +1 -0
  24. package/dist/clients/web/assets/clone-Ceucd3vE.js +1 -0
  25. package/dist/clients/web/assets/{cose-bilkent-S5V4N54A-DPWJGdbi.js → cose-bilkent-S5V4N54A-C-Kk0fEV.js} +1 -1
  26. package/dist/clients/web/assets/{dagre-6UL2VRFP-B53H6oGW.js → dagre-6UL2VRFP-v9VuT4Bu.js} +1 -1
  27. package/dist/clients/web/assets/{diagram-PSM6KHXK-CtUiylLq.js → diagram-PSM6KHXK-CL1lUjek.js} +1 -1
  28. package/dist/clients/web/assets/{diagram-QEK2KX5R-Bz54QBzS.js → diagram-QEK2KX5R-DDeA1tBD.js} +1 -1
  29. package/dist/clients/web/assets/{diagram-S2PKOQOG-BZ7N6I4Y.js → diagram-S2PKOQOG-CusffA32.js} +1 -1
  30. package/dist/clients/web/assets/{erDiagram-Q2GNP2WA-DsImZvNn.js → erDiagram-Q2GNP2WA-5uQ63pcd.js} +1 -1
  31. package/dist/clients/web/assets/{flowDiagram-NV44I4VS-B44cWQVu.js → flowDiagram-NV44I4VS-QR_zLHDZ.js} +1 -1
  32. package/dist/clients/web/assets/{ganttDiagram-JELNMOA3-DojhesoU.js → ganttDiagram-JELNMOA3-CTdvH7sX.js} +1 -1
  33. package/dist/clients/web/assets/{gitGraphDiagram-NY62KEGX-Bu0pwZnP.js → gitGraphDiagram-NY62KEGX-DuIORY6P.js} +1 -1
  34. package/dist/clients/web/assets/{graph-C9dyAS6N.js → graph-BRerRg5c.js} +1 -1
  35. package/dist/clients/web/assets/{index-mH7Mq7n7.css → index-BP79FsyI.css} +1 -1
  36. package/dist/clients/web/assets/index-cCvkX8q-.js +27 -0
  37. package/dist/clients/web/assets/{infoDiagram-WHAUD3N6-DJRulTfk.js → infoDiagram-WHAUD3N6-BgAzOEuY.js} +1 -1
  38. package/dist/clients/web/assets/{journeyDiagram-XKPGCS4Q-B1fbxy9E.js → journeyDiagram-XKPGCS4Q-2xfxZX3j.js} +1 -1
  39. package/dist/clients/web/assets/{kanban-definition-3W4ZIXB7-BiNR_wyT.js → kanban-definition-3W4ZIXB7-COrfmGnm.js} +1 -1
  40. package/dist/clients/web/assets/{layout-hVFdLOZ2.js → layout-Cvp52J2g.js} +1 -1
  41. package/dist/clients/web/assets/{linear-C1edgSmu.js → linear-Ders2TTm.js} +1 -1
  42. package/dist/clients/web/assets/{min-D7HqCeJW.js → min-Doap8eF2.js} +1 -1
  43. package/dist/clients/web/assets/{mindmap-definition-VGOIOE7T-Dpar1qlD.js → mindmap-definition-VGOIOE7T-B7wAlBzN.js} +1 -1
  44. package/dist/clients/web/assets/{pieDiagram-ADFJNKIX-DbkzXPM_.js → pieDiagram-ADFJNKIX-BV9gbcw_.js} +1 -1
  45. package/dist/clients/web/assets/{quadrantDiagram-AYHSOK5B-BoosvYcl.js → quadrantDiagram-AYHSOK5B-wZpTwLN1.js} +1 -1
  46. package/dist/clients/web/assets/{requirementDiagram-UZGBJVZJ-PnQeKXVd.js → requirementDiagram-UZGBJVZJ-CmHVFwI3.js} +1 -1
  47. package/dist/clients/web/assets/{sankeyDiagram-TZEHDZUN-B9HCbz2x.js → sankeyDiagram-TZEHDZUN-BPug29IB.js} +1 -1
  48. package/dist/clients/web/assets/{sequenceDiagram-WL72ISMW-BYbDEn0a.js → sequenceDiagram-WL72ISMW-B7qu9W_m.js} +1 -1
  49. package/dist/clients/web/assets/{stateDiagram-FKZM4ZOC-pgV59aNp.js → stateDiagram-FKZM4ZOC-Bkb1h2bL.js} +1 -1
  50. package/dist/clients/web/assets/stateDiagram-v2-4FDKWEC3-BBL6VW3E.js +1 -0
  51. package/dist/clients/web/assets/{timeline-definition-IT6M3QCI-aGhd2XOg.js → timeline-definition-IT6M3QCI-CItajHZm.js} +1 -1
  52. package/dist/clients/web/assets/{treemap-KMMF4GRG-D51M4sLQ.js → treemap-KMMF4GRG-Dzl3uQ1Q.js} +1 -1
  53. package/dist/clients/web/assets/{xychartDiagram-PRI3JC2R-B82leH3j.js → xychartDiagram-PRI3JC2R-Cpapi6ab.js} +1 -1
  54. package/dist/clients/web/index.html +2 -2
  55. package/package.json +1 -1
  56. package/dist/clients/web/assets/channel-DcMGaTgy.js +0 -1
  57. package/dist/clients/web/assets/classDiagram-2ON5EDUG-f5pkm7Az.js +0 -1
  58. package/dist/clients/web/assets/classDiagram-v2-WZHVMYZB-f5pkm7Az.js +0 -1
  59. package/dist/clients/web/assets/clone-DWrt8dNG.js +0 -1
  60. package/dist/clients/web/assets/index-CA6Bo1wO.js +0 -27
  61. package/dist/clients/web/assets/stateDiagram-v2-4FDKWEC3-BLVAstvD.js +0 -1
package/README.md CHANGED
@@ -1,3 +1,166 @@
1
- # Omni Context CLI
1
+ # OmniContext CLI
2
2
 
3
- Omx is a small, helpful, zero-telemetry coding assistant.
3
+ **Precision context. Minimal cost.**
4
+
5
+ OmniContext CLI is a terminal-native coding assistant that treats context as a first-class resource. Lean system prompts keep overhead low. Specialist delegation routes grunt work to cheaper models while keeping your main context clean. Zero telemetry means your code never leaves your machine. And it extends into VS Code, Office, the browser, Figma, Obsidian, and Zed.
6
+
7
+ ```bash
8
+ npm install -g omni-context-cli && omx
9
+ ```
10
+
11
+ ## How It Works
12
+
13
+ Traditional assistants call basic tools one at a time, resending your entire context with every round. OmniContext CLI delegates multi-step operations to agentic sub-agents running on a cheaper model -- your expensive model stays focused on reasoning, not file I/O.
14
+
15
+ **Task: "Find the definition of `handleAuth`"**
16
+
17
+ Traditional approach:
18
+
19
+ | Round | Call | Result |
20
+ |-------|------|--------|
21
+ | R1 | `glob("src/**/*.ts")` | 43 files returned |
22
+ | R2 | `grep("handleAuth", ...)` | 7 matches in 4 files |
23
+ | R3 | `read("src/middleware/auth.ts")` | 186 lines -- wrong file |
24
+ | R4 | `read("src/routes/login.ts")` | 124 lines -- still looking |
25
+ | R5 | `read("src/services/auth.ts", 40-90)` | Found it -- 50 more lines |
26
+
27
+ > 5 rounds, ~12K context added, all on main model
28
+
29
+ Specialist mode:
30
+
31
+ | Round | Call | Result |
32
+ |-------|------|--------|
33
+ | R1 | `pluck("handleAuth definition")` | Sub-agent (cheap model): glob -> grep -> read -> locate -> extract |
34
+
35
+ > 1 round, ~1K context added, grunt work on cheap model
36
+
37
+ ## Agentic Tools
38
+
39
+ Each tool runs as an autonomous sub-agent on a cheaper model. It handles file I/O, error recovery, and retries internally -- keeping intermediate output out of your main context and your token bill down.
40
+
41
+ | Tool | Purpose |
42
+ |------|---------|
43
+ | **explore** | Survey project architecture -- directory layout, key files, and how the codebase is organized |
44
+ | **spark** | Run shell commands with automatic error detection and retry |
45
+ | **sculpt** | Edit files with surgical precision, find the right location, make the change, validate the result |
46
+ | **weave** | Write entire files from scratch with auto-validation |
47
+ | **sweep** | Find files matching complex criteria by name, content, or structure |
48
+ | **pluck** | Extract specific code segments -- functions, classes, or blocks you need |
49
+ | **ripple** | Trace symbol references across your codebase |
50
+ | **slice** | Answer targeted code questions by reading only the relevant parts |
51
+ | **quest** | Research topics via web search |
52
+ | **glance** | Preview multiple files at once with brief summaries |
53
+
54
+ ## Workflow Presets
55
+
56
+ Switch how OmniContext CLI behaves with a single command. Each preset changes the tools available, the system prompt, and the response style.
57
+
58
+ | Preset | Description |
59
+ |--------|-------------|
60
+ | **Specialist** (default) | Your main model reasons, a cheaper agent model executes. Fewer rounds, cleaner context, lower cost. |
61
+ | **Explorer** | Research-first mode. Launches multiple web searches before answering. Great for current events, docs, and fact-checking. |
62
+ | **Artist** | Visual-first responses. Prioritizes image generation when the model supports it. Ideal for design exploration and mockups. |
63
+ | **Assistant** | Personal assistant for app integrations. Controls browser tabs, Office documents, and Figma designs through natural language. |
64
+ | **Normal** | Basic tools with manual orchestration. Direct read, write, edit, and bash access. Full control, no abstraction. |
65
+
66
+ ## Native Multi-Protocol
67
+
68
+ Most tools funnel everything through a single API format and hope for the best. OmniContext CLI has a dedicated request builder and stream handler for each protocol. Prompt caching, extended thinking, and provider-specific features work exactly as the vendor intended -- no lossy translation layer in between.
69
+
70
+ | Protocol | Description |
71
+ |----------|-------------|
72
+ | **Anthropic** | Native Messages API with prompt caching, extended thinking, and streaming. Token-level cache control via custom TTL. |
73
+ | **OpenAI** | Native Chat Completions API. Compatible with any endpoint that speaks the OpenAI format. |
74
+ | **Gemini** | Native generateContent API with Gemini-specific streaming. Tools and function calling use Gemini's own schema. |
75
+ | **Responses API** | OpenAI's newer Responses API with built-in tool orchestration. Separate path from Chat Completions. |
76
+
77
+ ## Cost Optimization
78
+
79
+ Every API call resends your full conversation history. Fewer rounds means fewer cache reads. Cleaner context means fewer tokens written. Specialist mode cuts both -- and offloads the grunt work to a cheaper model.
80
+
81
+ - **Fewer API rounds** -- Traditional tools need 5 rounds to find a function definition. Specialist mode does it in 1. That's 4 fewer full-context resends -- saving cache read costs on every skipped round.
82
+ - **Smaller context growth** -- Basic tools dump ~10KB of intermediate output into your conversation. Agentic tools return only the final result. Context editing automatically trims old tool payloads and thinking blocks, keeping growth in check even over long sessions.
83
+ - **Cheap model for execution** -- Sub-agents run on a low-cost model while your main model handles only planning and decisions. The expensive model never does file I/O.
84
+ - **1-hour cache for deep work** -- The default 5-minute prompt cache expires if you pause to think. Switch to 1-hour for debugging, refactoring, or research -- it eliminates repeated cache rebuilds across a session.
85
+
86
+ **Simulated cost comparison: "Find the definition of handleAuth"**
87
+
88
+ | | Traditional | Specialist | Saved |
89
+ |---|---|---|---|
90
+ | API rounds | 5 | 1 | -4 rounds |
91
+ | Cache read per round | ~20K tokens x 5 | ~20K tokens x 1 | -80K tokens |
92
+ | New context added | ~10KB | ~3KB | -70% |
93
+ | Cache write (new tokens) | ~2.5K tokens | ~1K tokens | -60% |
94
+ | Execution model | Expensive model only | Expensive + cheap | ~30% cheaper |
95
+
96
+ *Based on a 20K-token conversation finding a function across a TypeScript project. Actual savings depend on project size and model pricing.*
97
+
98
+ ## Model Providers
99
+
100
+ One command to add all your models. OmniContext CLI ships with built-in provider presets -- pick one, paste your API key, and every model from that service is ready to use.
101
+
102
+ ```bash
103
+ # List available providers
104
+ $ omx --list-providers
105
+
106
+ # Add all models from a provider in one go
107
+ $ omx --add-provider zenmux --api-key zmx-...
108
+
109
+ # Remove a provider just as easily
110
+ $ omx --remove-provider zenmux
111
+ ```
112
+
113
+ Built-in providers: **Zenmux**, **DeepSeek**, **OpenRouter**, **Zhipu (GLM)**, **MiniMax**
114
+
115
+ ## Cross-Session Memory
116
+
117
+ OmniContext CLI remembers your coding style, project patterns, and past decisions across sessions. Key points are extracted from every conversation and injected into future sessions. Helpful points gain score, harmful ones drop fast, unused ones decay naturally. Each project has its own memory file -- edit it directly if you want full control.
118
+
119
+ ## Integrations
120
+
121
+ Terminal is home base, but OmniContext CLI reaches into every tool you use. One AI, consistent context, zero context switching.
122
+
123
+ - **VS Code Extension** -- full IDE integration with file context, diagnostics, and diff views
124
+ - **Desktop App** -- standalone GUI that acts as the local hub connecting Office, browser, and Figma extensions
125
+ - **Chrome Extension** -- sidebar on any webpage for summarization, data extraction, and browser automation
126
+ - **Office Add-in** -- AI panel inside Word, Excel, and PowerPoint
127
+ - **Figma Plugin** -- inspect layouts, create shapes, modify nodes, and export assets through chat
128
+ - **Zed Editor** -- external agent via Agent Client Protocol with full tool access
129
+ - **Web Client** -- browser UI with LaTeX, Mermaid diagrams, file attachments, and drag-and-drop
130
+ - **Mobile Access** -- run `omx --serve` and connect from your phone
131
+
132
+ ## Extensibility
133
+
134
+ Custom agents, skills, slash commands, and MCP servers. Everything is a markdown file or JSON config.
135
+
136
+ - **Custom SubAgents** -- write a markdown file with a prompt template and tool permissions. It becomes a new agentic tool instantly. Add `OMX-AGENTS.md` for global agent instructions.
137
+ - **Custom Skills** -- teach OmniContext CLI domain-specific knowledge and workflows. Skills inject instructions into the current conversation.
138
+ - **Slash Commands** -- create shortcuts for common prompts with Handlebars templating.
139
+ - **MCP Servers** -- connect external tools and data sources via Model Context Protocol. Stdio and HTTP transports supported.
140
+
141
+ ## The Details
142
+
143
+ - **Lean system prompts** -- minimal, focused instructions and concise tool descriptions. Your tokens go toward actual work, not bloated framework overhead.
144
+ - **Zero telemetry** -- no usage tracking, no analytics, no data collection.
145
+ - **Context editing** -- automatically trims old tool call payloads and thinking blocks from your conversation history.
146
+ - **Extended thinking** -- enable deeper reasoning for complex tasks with configurable budget limits.
147
+ - **CLAUDE.md compatible** -- already have a CLAUDE.md in your repo? OmniContext CLI reads it automatically.
148
+ - **Auto-compaction** -- when context hits 80% capacity, the conversation is compacted, key memories are extracted, and a fresh session picks up where you left off.
149
+ - **Native prompt caching** -- automatic cache control for Anthropic and Gemini with custom TTL settings.
150
+ - **Project instructions** -- drop an `OMX.md` in your repo root and everyone on the team gets the same conventions and context.
151
+
152
+ ## Build & Release
153
+
154
+ ```bash
155
+ npm run release
156
+ ```
157
+
158
+ One command builds the CLI, all clients, packages release zips, and builds the desktop app for the current platform. Artifacts go to `release/`.
159
+
160
+ ## Documentation
161
+
162
+ **https://bluenoah1991.github.io/omni-context-cli-landing/docs/**
163
+
164
+ ## License
165
+
166
+ MIT
@@ -0,0 +1,166 @@
1
+ # OmniContext CLI
2
+
3
+ **精准上下文,最小成本。**
4
+
5
+ OmniContext CLI 是一个终端原生的编程助手,把上下文当作一等资源来管理。精简的系统提示词控制开销。专家委派机制把脏活路由给便宜的模型,同时保持主上下文的干净。零遥测意味着你的代码不会离开你的机器。它还能延伸到 VS Code、Office、浏览器、Figma、Obsidian 和 Zed。
6
+
7
+ ```bash
8
+ npm install -g omni-context-cli && omx
9
+ ```
10
+
11
+ ## 工作原理
12
+
13
+ 传统助手逐个调用基础工具,每一轮都重新发送完整上下文。OmniContext CLI 把多步操作委派给运行在便宜模型上的 Agentic 子代理——贵价模型专注推理,不做文件 I/O。
14
+
15
+ **任务:"找到 `handleAuth` 的定义"**
16
+
17
+ 传统模式:
18
+
19
+ | 轮次 | 调用 | 结果 |
20
+ |------|------|------|
21
+ | R1 | `glob("src/**/*.ts")` | 返回 43 个文件 |
22
+ | R2 | `grep("handleAuth", ...)` | 4 个文件中有 7 处匹配 |
23
+ | R3 | `read("src/middleware/auth.ts")` | 186 行——找错文件了 |
24
+ | R4 | `read("src/routes/login.ts")` | 124 行——还在找 |
25
+ | R5 | `read("src/services/auth.ts", 40-90)` | 找到了——又多 50 行 |
26
+
27
+ > 5 轮,新增 ~12K 上下文,全部在主模型上执行
28
+
29
+ 专家模式:
30
+
31
+ | 轮次 | 调用 | 结果 |
32
+ |------|------|------|
33
+ | R1 | `pluck("handleAuth definition")` | 子代理(便宜模型):glob -> grep -> read -> locate -> extract |
34
+
35
+ > 1 轮,新增 ~1K 上下文,脏活在便宜模型上完成
36
+
37
+ ## Agentic 工具
38
+
39
+ 每个工具作为自主子代理运行在便宜模型上,内部处理文件 I/O、错误恢复和重试——中间输出不会进入你的主上下文,token 账单也不会膨胀。
40
+
41
+ | 工具 | 用途 |
42
+ |------|------|
43
+ | **explore** | 勘察项目架构——目录布局、关键文件和代码组织方式 |
44
+ | **spark** | 执行 shell 命令,自动检测错误并重试 |
45
+ | **sculpt** | 精准编辑文件,定位正确位置,修改并验证结果 |
46
+ | **weave** | 从头写入完整文件,自动验证 |
47
+ | **sweep** | 按名称、内容或结构查找匹配的文件 |
48
+ | **pluck** | 提取特定代码片段——函数、类或你需要的代码块 |
49
+ | **ripple** | 追踪符号在代码库中的所有引用 |
50
+ | **slice** | 只读取相关部分来回答针对性的代码问题 |
51
+ | **quest** | 通过网络搜索调研主题 |
52
+ | **glance** | 一次预览多个文件,附带简要摘要 |
53
+
54
+ ## 工作流预设
55
+
56
+ 一条命令切换 OmniContext CLI 的行为模式。每个预设改变可用工具、系统提示词和响应风格。
57
+
58
+ | 预设 | 说明 |
59
+ |------|------|
60
+ | **Specialist**(默认) | 主模型负责推理,便宜的代理模型负责执行。更少轮次,更干净的上下文,更低的成本。 |
61
+ | **Explorer** | 调研优先模式。先发起多次网络搜索再回答。适合时事、文档查阅和事实核查。 |
62
+ | **Artist** | 视觉优先响应。在模型支持时优先生成图像。适合设计探索和原型。 |
63
+ | **Assistant** | 应用集成的个人助理。通过自然语言控制浏览器标签页、Office 文档和 Figma 设计。 |
64
+ | **Normal** | 基础工具加手动编排。直接使用 read、write、edit 和 bash。完全控制,没有抽象。 |
65
+
66
+ ## 原生多协议
67
+
68
+ 大多数工具把所有请求转换成单一 API 格式。OmniContext CLI 为每种协议提供专用的请求构建器和流处理器。提示词缓存、扩展思考和供应商专属特性按原厂设计工作——没有有损的转换层。
69
+
70
+ | 协议 | 说明 |
71
+ |------|------|
72
+ | **Anthropic** | 原生 Messages API,支持提示词缓存、扩展思考和流式传输。通过自定义 TTL 实现 token 级缓存控制。 |
73
+ | **OpenAI** | 原生 Chat Completions API。兼容任何 OpenAI 格式的接口。 |
74
+ | **Gemini** | 原生 generateContent API,Gemini 专用流式传输。工具和函数调用使用 Gemini 自己的 schema。 |
75
+ | **Responses API** | OpenAI 新一代 Responses API,内置工具编排。独立于 Chat Completions 的路径。 |
76
+
77
+ ## 成本优化
78
+
79
+ 每次 API 调用都会重新发送完整的对话历史。更少的轮次意味着更少的缓存读取。更干净的上下文意味着更少的 token 写入。专家模式两者兼省——并且把脏活卸载给便宜的模型。
80
+
81
+ - **更少的 API 轮次** ——传统工具需要 5 轮才能找到一个函数定义,专家模式只要 1 轮。省掉 4 次完整上下文重传,每省一轮都节省缓存读取成本。
82
+ - **更小的上下文增长** ——基础工具往对话里塞 ~10KB 中间输出,Agentic 工具只返回最终结果。上下文编辑自动裁剪旧的工具负载和思考块,长会话也能控制增长。
83
+ - **便宜模型做执行** ——子代理运行在低成本模型上,主模型只负责规划和决策。贵价模型永远不做文件 I/O。
84
+ - **1 小时缓存应对深度工作** ——默认 5 分钟提示词缓存在你暂停思考时就会过期。切换到 1 小时缓存适合调试、重构或调研——消除会话中反复重建缓存的开销。
85
+
86
+ **模拟成本对比:"找到 handleAuth 的定义"**
87
+
88
+ | | 传统模式 | 专家模式 | 节省 |
89
+ |---|---|---|---|
90
+ | API 轮次 | 5 | 1 | -4 轮 |
91
+ | 每轮缓存读取 | ~20K tokens x 5 | ~20K tokens x 1 | -80K tokens |
92
+ | 新增上下文 | ~10KB | ~3KB | -70% |
93
+ | 缓存写入(新 token) | ~2.5K tokens | ~1K tokens | -60% |
94
+ | 执行模型 | 仅贵价模型 | 贵价 + 便宜 | 便宜 ~30% |
95
+
96
+ *基于在 TypeScript 项目中查找函数的 20K token 对话。实际节省取决于项目规模和模型定价。*
97
+
98
+ ## 模型供应商
99
+
100
+ 一条命令添加所有模型。OmniContext CLI 内置供应商预设——选一个,粘贴 API key,该服务的所有模型就可以使用了。
101
+
102
+ ```bash
103
+ # 列出可用供应商
104
+ $ omx --list-providers
105
+
106
+ # 一次性添加供应商的所有模型
107
+ $ omx --add-provider zenmux --api-key zmx-...
108
+
109
+ # 移除同样简单
110
+ $ omx --remove-provider zenmux
111
+ ```
112
+
113
+ 内置供应商:**Zenmux**、**DeepSeek**、**OpenRouter**、**Zhipu (GLM)**、**MiniMax**
114
+
115
+ ## 跨会话记忆
116
+
117
+ OmniContext CLI 跨会话记住你的编码风格、项目模式和历史决策。关键要点从每次对话中提取并注入未来的会话。有用的要点加分(+1),有害的快速扣分(-3),不再使用的自然衰减。每个项目有自己的记忆文件——想要完全控制可以直接编辑。
118
+
119
+ ## 集成
120
+
121
+ 终端是大本营,但 OmniContext CLI 延伸到你使用的每个工具。一个 AI,一致的上下文,零切换成本。
122
+
123
+ - **VS Code 扩展** ——完整的 IDE 集成,感知打开文件、诊断信息和 diff 视图
124
+ - **桌面应用** ——独立 GUI,作为本地中枢连接 Office、浏览器和 Figma 扩展
125
+ - **Chrome 扩展** ——任意网页上的侧边栏,支持摘要、数据提取和浏览器自动化
126
+ - **Office 插件** ——Word、Excel 和 PowerPoint 内的 AI 面板
127
+ - **Figma 插件** ——通过聊天面板检查布局、创建图形、修改节点和导出资源
128
+ - **Zed 编辑器** ——通过 Agent Client Protocol 作为外部代理接入,拥有完整工具访问
129
+ - **Web 客户端** ——浏览器 UI,支持 LaTeX、Mermaid 图表、文件附件和拖拽
130
+ - **移动端访问** ——运行 `omx --serve` 后从手机连接
131
+
132
+ ## 可扩展性
133
+
134
+ 自定义 Agent、技能、斜杠命令和 MCP 服务器。一切都是 Markdown 文件或 JSON 配置。
135
+
136
+ - **自定义子代理** ——写一个带提示词模板和工具权限的 Markdown 文件,它立刻成为新的 Agentic 工具。添加 `OMX-AGENTS.md` 作为全局代理指令。
137
+ - **自定义技能** ——教 OmniContext CLI 领域知识和工作流。技能会注入当前对话。
138
+ - **斜杠命令** ——为常用提示词创建快捷方式,支持 Handlebars 模板。
139
+ - **MCP 服务器** ——通过 Model Context Protocol 接入外部工具和数据源。支持 stdio 和 HTTP 传输。
140
+
141
+ ## 细节
142
+
143
+ - **精简的系统提示词** ——最小化、聚焦的指令和简洁的工具描述。你的 token 用在实际工作上,而不是臃肿的框架开销。
144
+ - **零遥测** ——没有使用追踪,没有数据分析,没有数据收集。
145
+ - **上下文编辑** ——自动裁剪对话历史中旧的工具调用负载和思考块。
146
+ - **扩展思考** ——为复杂任务启用深度推理,支持可配置的预算限制。
147
+ - **兼容 CLAUDE.md** ——仓库里已经有 CLAUDE.md?OmniContext CLI 会自动读取。
148
+ - **自动压缩** ——上下文达到 80% 容量时,对话被压缩,关键记忆被提取,新会话无缝接续。
149
+ - **原生提示词缓存** ——Anthropic 和 Gemini 的自动缓存控制,支持自定义 TTL 设置。
150
+ - **项目指令** ——在仓库根目录放一个 `OMX.md`,团队里每个人都能得到相同的约定和上下文。
151
+
152
+ ## 构建与发布
153
+
154
+ ```bash
155
+ npm run release
156
+ ```
157
+
158
+ 一条命令构建 CLI 和所有客户端,打包发布 zip,并为当前平台构建桌面应用。产物输出到 `release/`。
159
+
160
+ ## 文档
161
+
162
+ **https://bluenoah1991.github.io/omni-context-cli-landing/docs/zh-Hans/**
163
+
164
+ ## 许可证
165
+
166
+ MIT
File without changes
File without changes
File without changes