@just-every/code 0.1.7 → 0.2.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (2) hide show
  1. package/README.md +193 -644
  2. package/package.json +3 -2
package/README.md CHANGED
@@ -1,736 +1,285 @@
1
- <h1 align="center">OpenAI Codex CLI</h1>
2
- <p align="center">Lightweight coding agent that runs in your terminal</p>
1
+ # JustEvery_ Code
3
2
 
4
- <p align="center"><code>npm i -g @openai/codex</code></p>
3
+ &ensp;
5
4
 
6
- > [!IMPORTANT]
7
- > This is the documentation for the _legacy_ TypeScript implementation of the Codex CLI. It has been superseded by the _Rust_ implementation. See the [README in the root of the Codex repository](https://github.com/openai/codex/blob/main/README.md) for details.
5
+ <p align="center">
6
+ <img src="docs/logo.png" alt="Code Logo" width="400">
7
+ </p>
8
8
 
9
- ![Codex demo GIF using: codex "explain this codebase to me"](../.github/demo.gif)
9
+ &ensp;
10
10
 
11
- ---
12
-
13
- <details>
14
- <summary><strong>Table of contents</strong></summary>
15
-
16
- <!-- Begin ToC -->
17
-
18
- - [Experimental technology disclaimer](#experimental-technology-disclaimer)
19
- - [Quickstart](#quickstart)
20
- - [Why Codex?](#why-codex)
21
- - [Security model & permissions](#security-model--permissions)
22
- - [Platform sandboxing details](#platform-sandboxing-details)
23
- - [System requirements](#system-requirements)
24
- - [CLI reference](#cli-reference)
25
- - [Memory & project docs](#memory--project-docs)
26
- - [Non-interactive / CI mode](#non-interactive--ci-mode)
27
- - [Tracing / verbose logging](#tracing--verbose-logging)
28
- - [Recipes](#recipes)
29
- - [Installation](#installation)
30
- - [Configuration guide](#configuration-guide)
31
- - [Basic configuration parameters](#basic-configuration-parameters)
32
- - [Custom AI provider configuration](#custom-ai-provider-configuration)
33
- - [History configuration](#history-configuration)
34
- - [Configuration examples](#configuration-examples)
35
- - [Full configuration example](#full-configuration-example)
36
- - [Custom instructions](#custom-instructions)
37
- - [Environment variables setup](#environment-variables-setup)
38
- - [FAQ](#faq)
39
- - [Zero data retention (ZDR) usage](#zero-data-retention-zdr-usage)
40
- - [Codex open source fund](#codex-open-source-fund)
41
- - [Contributing](#contributing)
42
- - [Development workflow](#development-workflow)
43
- - [Git hooks with Husky](#git-hooks-with-husky)
44
- - [Debugging](#debugging)
45
- - [Writing high-impact code changes](#writing-high-impact-code-changes)
46
- - [Opening a pull request](#opening-a-pull-request)
47
- - [Review process](#review-process)
48
- - [Community values](#community-values)
49
- - [Getting help](#getting-help)
50
- - [Contributor license agreement (CLA)](#contributor-license-agreement-cla)
51
- - [Quick fixes](#quick-fixes)
52
- - [Releasing `codex`](#releasing-codex)
53
- - [Alternative build options](#alternative-build-options)
54
- - [Nix flake development](#nix-flake-development)
55
- - [Security & responsible AI](#security--responsible-ai)
56
- - [License](#license)
57
-
58
- <!-- End ToC -->
59
-
60
- </details>
11
+ **Code** is a fast, local coding agent for your terminal. It's a community-driven fork of `openai/codex` focused on real developer ergonomics: Browser integration, multi-agents, theming, and reasoning control — all while staying compatible with upstream.
61
12
 
62
- ---
13
+ &ensp;
14
+ ## Why Code
63
15
 
64
- ## Experimental technology disclaimer
16
+ - 🌐 **Browser Integration** - CDP support, headless browsing, screenshots
17
+ - 📝 **Diff Viewer** - Side-by-side diffs with syntax highlighting
18
+ - 🤖 **Multi-Agent Commands** - /plan, /solve, /code with agent panels
19
+ - 🎨 **Theme System** - /themes with live preview and accessibility
20
+ - 🧠 **Reasoning Control** - /reasoning for dynamic effort adjustment
21
+ - 🔌 **MCP support** – Extend with filesystem, DBs, APIs, or your own tools.
22
+ - 🔒 **Safety modes** – Read-only, approvals, and workspace sandboxing.
23
+ - 🔁 **Backwards compatible** – Supports `~/.codex/*` or default `~/.coder/*`
65
24
 
66
- Codex CLI is an experimental project under active development. It is not yet stable, may contain bugs, incomplete features, or undergo breaking changes. We're building it in the open with the community and welcome:
25
+ &ensp;
26
+ | <img src="docs/screenshots/simple.png" alt="Simple interface" width="100%"><br>Simple interface | <img src="docs/screenshots/diff.png" alt="Unified diff viewer" width="100%"><br>Unified diffs |
27
+ |:--:|:--:|
67
28
 
68
- - Bug reports
69
- - Feature requests
70
- - Pull requests
71
- - Good vibes
29
+ | <br><img src="docs/screenshots/browser.png" alt="Browser control" width="100%"><br>Browser control | <br><img src="docs/screenshots/agents.png" alt="Assist with Claude & Gemini" width="100%"><br>Assist with Claude & Gemini |
30
+ |:--:|:--:|
72
31
 
73
- Help us improve by filing issues or submitting PRs (see the section below for how to contribute)!
74
32
 
33
+ &ensp;
75
34
  ## Quickstart
76
35
 
77
- Install globally:
78
-
79
- ```shell
80
- npm install -g @openai/codex
81
- ```
82
-
83
- Next, set your OpenAI API key as an environment variable:
84
-
85
- ```shell
86
- export OPENAI_API_KEY="your-api-key-here"
87
- ```
88
-
89
- > **Note:** This command sets the key only for your current terminal session. You can add the `export` line to your shell's configuration file (e.g., `~/.zshrc`) but we recommend setting for the session. **Tip:** You can also place your API key into a `.env` file at the root of your project:
90
- >
91
- > ```env
92
- > OPENAI_API_KEY=your-api-key-here
93
- > ```
94
- >
95
- > The CLI will automatically load variables from `.env` (via `dotenv/config`).
96
-
97
- <details>
98
- <summary><strong>Use <code>--provider</code> to use other models</strong></summary>
99
-
100
- > Codex also allows you to use other providers that support the OpenAI Chat Completions API. You can set the provider in the config file or use the `--provider` flag. The possible options for `--provider` are:
101
- >
102
- > - openai (default)
103
- > - openrouter
104
- > - azure
105
- > - gemini
106
- > - ollama
107
- > - mistral
108
- > - deepseek
109
- > - xai
110
- > - groq
111
- > - arceeai
112
- > - any other provider that is compatible with the OpenAI API
113
- >
114
- > If you use a provider other than OpenAI, you will need to set the API key for the provider in the config file or in the environment variable as:
115
- >
116
- > ```shell
117
- > export <provider>_API_KEY="your-api-key-here"
118
- > ```
119
- >
120
- > If you use a provider not listed above, you must also set the base URL for the provider:
121
- >
122
- > ```shell
123
- > export <provider>_BASE_URL="https://your-provider-api-base-url"
124
- > ```
125
-
126
- </details>
127
- <br />
128
-
129
- Run interactively:
36
+ ### Install Code
130
37
 
131
- ```shell
132
- codex
133
- ```
134
-
135
- Or, run with a prompt as input (and optionally in `Full Auto` mode):
136
-
137
- ```shell
138
- codex "explain this codebase to me"
139
- ```
140
-
141
- ```shell
142
- codex --approval-mode full-auto "create the fanciest todo-list app"
38
+ ```bash
39
+ npm install -g @just-every/code
40
+ code
143
41
  ```
144
42
 
145
- That's it - Codex will scaffold a file, run it inside a sandbox, install any
146
- missing dependencies, and show you the live result. Approve the changes and
147
- they'll be committed to your working directory.
148
-
149
- ---
150
-
151
- ## Why Codex?
43
+ Note: If another tool already provides a `code` command (e.g., VS Code), our CLI is also installed as `coder`. Use `coder` to avoid conflicts. Check which command runs with `which -a code` (or `type -a code`). You can also run our CLI explicitly via `npx -y @just-every/code` or `$(npm bin -g)/code`.
152
44
 
153
- Codex CLI is built for developers who already **live in the terminal** and want
154
- ChatGPT-level reasoning **plus** the power to actually run code, manipulate
155
- files, and iterate - all under version control. In short, it's _chat-driven
156
- development_ that understands and executes your repo.
157
-
158
- - **Zero setup** - bring your OpenAI API key and it just works!
159
- - **Full auto-approval, while safe + secure** by running network-disabled and directory-sandboxed
160
- - **Multimodal** - pass in screenshots or diagrams to implement features ✨
161
-
162
- And it's **fully open-source** so you can see and contribute to how it develops!
163
-
164
- ---
165
-
166
- ## Security model & permissions
167
-
168
- Codex lets you decide _how much autonomy_ the agent receives and auto-approval policy via the
169
- `--approval-mode` flag (or the interactive onboarding prompt):
170
-
171
- | Mode | What the agent may do without asking | Still requires approval |
172
- | ------------------------- | --------------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- |
173
- | **Suggest** <br>(default) | <li>Read any file in the repo | <li>**All** file writes/patches<li> **Any** arbitrary shell commands (aside from reading files) |
174
- | **Auto Edit** | <li>Read **and** apply-patch writes to files | <li>**All** shell commands |
175
- | **Full Auto** | <li>Read/write files <li> Execute shell commands (network disabled, writes limited to your workdir) | - |
176
-
177
- In **Full Auto** every command is run **network-disabled** and confined to the
178
- current working directory (plus temporary files) for defense-in-depth. Codex
179
- will also show a warning/confirmation if you start in **auto-edit** or
180
- **full-auto** while the directory is _not_ tracked by Git, so you always have a
181
- safety net.
182
-
183
- Coming soon: you'll be able to whitelist specific commands to auto-execute with
184
- the network enabled, once we're confident in additional safeguards.
185
-
186
- ### Platform sandboxing details
187
-
188
- The hardening mechanism Codex uses depends on your OS:
189
-
190
- - **macOS 12+** - commands are wrapped with **Apple Seatbelt** (`sandbox-exec`).
191
-
192
- - Everything is placed in a read-only jail except for a small set of
193
- writable roots (`$PWD`, `$TMPDIR`, `~/.codex`, etc.).
194
- - Outbound network is _fully blocked_ by default - even if a child process
195
- tries to `curl` somewhere it will fail.
196
-
197
- - **Linux** - there is no sandboxing by default.
198
- We recommend using Docker for sandboxing, where Codex launches itself inside a **minimal
199
- container image** and mounts your repo _read/write_ at the same path. A
200
- custom `iptables`/`ipset` firewall script denies all egress except the
201
- OpenAI API. This gives you deterministic, reproducible runs without needing
202
- root on the host. You can use the [`run_in_container.sh`](../codex-cli/scripts/run_in_container.sh) script to set up the sandbox.
203
-
204
- ---
205
-
206
- ## System requirements
207
-
208
- | Requirement | Details |
209
- | --------------------------- | --------------------------------------------------------------- |
210
- | Operating systems | macOS 12+, Ubuntu 20.04+/Debian 10+, or Windows 11 **via WSL2** |
211
- | Node.js | **22 or newer** (LTS recommended) |
212
- | Git (optional, recommended) | 2.23+ for built-in PR helpers |
213
- | RAM | 4-GB minimum (8-GB recommended) |
214
-
215
- > Never run `sudo npm install -g`; fix npm permissions instead.
216
-
217
- ---
218
-
219
- ## CLI reference
45
+ **Authenticate** (one of the following):
46
+ - **Sign in with ChatGPT** (Plus/Pro/Team; uses models available to your plan)
47
+ - Run `code` and pick "Sign in with ChatGPT"
48
+ - Stores creds locally at `~/.coder/auth.json` (also reads legacy `~/.codex/auth.json`)
49
+ - **API key** (usage-based)
50
+ - Set `export OPENAI_API_KEY=xyz` and run `code`
220
51
 
221
- | Command | Purpose | Example |
222
- | ------------------------------------ | ----------------------------------- | ------------------------------------ |
223
- | `codex` | Interactive REPL | `codex` |
224
- | `codex "..."` | Initial prompt for interactive REPL | `codex "fix lint errors"` |
225
- | `codex -q "..."` | Non-interactive "quiet mode" | `codex -q --json "explain utils.ts"` |
226
- | `codex completion <bash\|zsh\|fish>` | Print shell completion script | `codex completion bash` |
52
+ ### Install Claude & Gemini (optional)
227
53
 
228
- Key flags: `--model/-m`, `--approval-mode/-a`, `--quiet/-q`, and `--notify`.
54
+ Code supports orchestrating other AI CLI tools. Install these and config to use alongside Code.
229
55
 
230
- ---
231
-
232
- ## Memory & project docs
233
-
234
- You can give Codex extra instructions and guidance using `AGENTS.md` files. Codex looks for `AGENTS.md` files in the following places, and merges them top-down:
235
-
236
- 1. `~/.codex/AGENTS.md` - personal global guidance
237
- 2. `AGENTS.md` at repo root - shared project notes
238
- 3. `AGENTS.md` in the current working directory - sub-folder/feature specifics
239
-
240
- Disable loading of these files with `--no-project-doc` or the environment variable `CODEX_DISABLE_PROJECT_DOC=1`.
241
-
242
- ---
243
-
244
- ## Non-interactive / CI mode
245
-
246
- Run Codex head-less in pipelines. Example GitHub Action step:
247
-
248
- ```yaml
249
- - name: Update changelog via Codex
250
- run: |
251
- npm install -g @openai/codex
252
- export OPENAI_API_KEY="${{ secrets.OPENAI_KEY }}"
253
- codex -a auto-edit --quiet "update CHANGELOG for next release"
254
- ```
255
-
256
- Set `CODEX_QUIET_MODE=1` to silence interactive UI noise.
257
-
258
- ## Tracing / verbose logging
259
-
260
- Setting the environment variable `DEBUG=true` prints full API request and response details:
56
+ ```bash
261
57
 
262
- ```shell
263
- DEBUG=true codex
58
+ npm install -g @anthropic-ai/claude-code @google/gemini-cli && claude "Just checking you're working! Let me know how I can exit." && gemini -i "Just checking you're working! Let me know how I can exit."
264
59
  ```
265
60
 
266
- ---
267
-
268
- ## Recipes
269
-
270
- Below are a few bite-size examples you can copy-paste. Replace the text in quotes with your own task. See the [prompting guide](https://github.com/openai/codex/blob/main/codex-cli/examples/prompting_guide.md) for more tips and usage patterns.
271
-
272
- | ✨ | What you type | What happens |
273
- | --- | ------------------------------------------------------------------------------- | -------------------------------------------------------------------------- |
274
- | 1 | `codex "Refactor the Dashboard component to React Hooks"` | Codex rewrites the class component, runs `npm test`, and shows the diff. |
275
- | 2 | `codex "Generate SQL migrations for adding a users table"` | Infers your ORM, creates migration files, and runs them in a sandboxed DB. |
276
- | 3 | `codex "Write unit tests for utils/date.ts"` | Generates tests, executes them, and iterates until they pass. |
277
- | 4 | `codex "Bulk-rename *.jpeg -> *.jpg with git mv"` | Safely renames files and updates imports/usages. |
278
- | 5 | `codex "Explain what this regex does: ^(?=.*[A-Z]).{8,}$"` | Outputs a step-by-step human explanation. |
279
- | 6 | `codex "Carefully review this repo, and propose 3 high impact well-scoped PRs"` | Suggests impactful PRs in the current codebase. |
280
- | 7 | `codex "Look for vulnerabilities and create a security review report"` | Finds and explains security bugs. |
281
-
282
- ---
283
-
284
- ## Installation
285
-
286
- <details open>
287
- <summary><strong>From npm (Recommended)</strong></summary>
61
+ &ensp;
62
+ ## Commands
288
63
 
64
+ ### Browser
289
65
  ```bash
290
- npm install -g @openai/codex
291
- # or
292
- yarn global add @openai/codex
293
- # or
294
- bun install -g @openai/codex
295
- # or
296
- pnpm add -g @openai/codex
297
- ```
298
-
299
- </details>
66
+ # Connect code to external Chrome browser (running CDP)
67
+ /chrome # Connect with auto-detect port
68
+ /chrome 9222 # Connect to specific port
300
69
 
301
- <details>
302
- <summary><strong>Build from source</strong></summary>
70
+ # Switch to internal browser mode
71
+ /browser # Use internal headless browser
72
+ /browser https://example.com # Open URL in internal browser
73
+ ```
303
74
 
75
+ ### Agents
304
76
  ```bash
305
- # Clone the repository and navigate to the CLI package
306
- git clone https://github.com/openai/codex.git
307
- cd codex/codex-cli
77
+ # Plan code changes (Claude, Gemini and GPT-5 consensus)
78
+ # All agents review task and create a consolidated plan
79
+ /plan "Stop the AI from ordering pizza at 3AM"
308
80
 
309
- # Enable corepack
310
- corepack enable
81
+ # Solve complex problems (Claude, Gemini and GPT-5 race)
82
+ # Fastest preferred (see https://arxiv.org/abs/2505.17813)
83
+ /solve "Why does deleting one user drop the whole database?"
311
84
 
312
- # Install dependencies and build
313
- pnpm install
314
- pnpm build
315
-
316
- # Linux-only: download prebuilt sandboxing binaries (requires gh and zstd).
317
- ./scripts/install_native_deps.sh
318
-
319
- # Get the usage and the options
320
- node ./dist/cli.js --help
321
-
322
- # Run the locally-built CLI directly
323
- node ./dist/cli.js
324
-
325
- # Or link the command globally for convenience
326
- pnpm link
85
+ # Write code! (Claude, Gemini and GPT-5 consensus)
86
+ # Creates multiple worktrees then implements the optimal solution
87
+ /code "Show dark mode when I feel cranky"
327
88
  ```
328
89
 
329
- </details>
330
-
331
- ---
332
-
333
- ## Configuration guide
334
-
335
- Codex configuration files can be placed in the `~/.codex/` directory, supporting both YAML and JSON formats.
336
-
337
- ### Basic configuration parameters
338
-
339
- | Parameter | Type | Default | Description | Available Options |
340
- | ------------------- | ------- | ---------- | -------------------------------- | ---------------------------------------------------------------------------------------------- |
341
- | `model` | string | `o4-mini` | AI model to use | Any model name supporting OpenAI API |
342
- | `approvalMode` | string | `suggest` | AI assistant's permission mode | `suggest` (suggestions only)<br>`auto-edit` (automatic edits)<br>`full-auto` (fully automatic) |
343
- | `fullAutoErrorMode` | string | `ask-user` | Error handling in full-auto mode | `ask-user` (prompt for user input)<br>`ignore-and-continue` (ignore and proceed) |
344
- | `notify` | boolean | `true` | Enable desktop notifications | `true`/`false` |
345
-
346
- ### Custom AI provider configuration
347
-
348
- In the `providers` object, you can configure multiple AI service providers. Each provider requires the following parameters:
349
-
350
- | Parameter | Type | Description | Example |
351
- | --------- | ------ | --------------------------------------- | ----------------------------- |
352
- | `name` | string | Display name of the provider | `"OpenAI"` |
353
- | `baseURL` | string | API service URL | `"https://api.openai.com/v1"` |
354
- | `envKey` | string | Environment variable name (for API key) | `"OPENAI_API_KEY"` |
355
-
356
- ### History configuration
357
-
358
- In the `history` object, you can configure conversation history settings:
359
-
360
- | Parameter | Type | Description | Example Value |
361
- | ------------------- | ------- | ------------------------------------------------------ | ------------- |
362
- | `maxSize` | number | Maximum number of history entries to save | `1000` |
363
- | `saveHistory` | boolean | Whether to save history | `true` |
364
- | `sensitivePatterns` | array | Patterns of sensitive information to filter in history | `[]` |
365
-
366
- ### Configuration examples
90
+ ### General
91
+ ```bash
92
+ # Try a new theme!
93
+ /themes
367
94
 
368
- 1. YAML format (save as `~/.codex/config.yaml`):
95
+ # Change reasoning level
96
+ /reasoning low|medium|high
369
97
 
370
- ```yaml
371
- model: o4-mini
372
- approvalMode: suggest
373
- fullAutoErrorMode: ask-user
374
- notify: true
98
+ # Start new conversation
99
+ /new
375
100
  ```
376
101
 
377
- 2. JSON format (save as `~/.codex/config.json`):
378
-
379
- ```json
380
- {
381
- "model": "o4-mini",
382
- "approvalMode": "suggest",
383
- "fullAutoErrorMode": "ask-user",
384
- "notify": true
385
- }
386
- ```
102
+ ## CLI reference
387
103
 
388
- ### Full configuration example
389
-
390
- Below is a comprehensive example of `config.json` with multiple custom providers:
391
-
392
- ```json
393
- {
394
- "model": "o4-mini",
395
- "provider": "openai",
396
- "providers": {
397
- "openai": {
398
- "name": "OpenAI",
399
- "baseURL": "https://api.openai.com/v1",
400
- "envKey": "OPENAI_API_KEY"
401
- },
402
- "azure": {
403
- "name": "AzureOpenAI",
404
- "baseURL": "https://YOUR_PROJECT_NAME.openai.azure.com/openai",
405
- "envKey": "AZURE_OPENAI_API_KEY"
406
- },
407
- "openrouter": {
408
- "name": "OpenRouter",
409
- "baseURL": "https://openrouter.ai/api/v1",
410
- "envKey": "OPENROUTER_API_KEY"
411
- },
412
- "gemini": {
413
- "name": "Gemini",
414
- "baseURL": "https://generativelanguage.googleapis.com/v1beta/openai",
415
- "envKey": "GEMINI_API_KEY"
416
- },
417
- "ollama": {
418
- "name": "Ollama",
419
- "baseURL": "http://localhost:11434/v1",
420
- "envKey": "OLLAMA_API_KEY"
421
- },
422
- "mistral": {
423
- "name": "Mistral",
424
- "baseURL": "https://api.mistral.ai/v1",
425
- "envKey": "MISTRAL_API_KEY"
426
- },
427
- "deepseek": {
428
- "name": "DeepSeek",
429
- "baseURL": "https://api.deepseek.com",
430
- "envKey": "DEEPSEEK_API_KEY"
431
- },
432
- "xai": {
433
- "name": "xAI",
434
- "baseURL": "https://api.x.ai/v1",
435
- "envKey": "XAI_API_KEY"
436
- },
437
- "groq": {
438
- "name": "Groq",
439
- "baseURL": "https://api.groq.com/openai/v1",
440
- "envKey": "GROQ_API_KEY"
441
- },
442
- "arceeai": {
443
- "name": "ArceeAI",
444
- "baseURL": "https://conductor.arcee.ai/v1",
445
- "envKey": "ARCEEAI_API_KEY"
446
- }
447
- },
448
- "history": {
449
- "maxSize": 1000,
450
- "saveHistory": true,
451
- "sensitivePatterns": []
452
- }
453
- }
104
+ ```shell
105
+ code [options] [prompt]
106
+
107
+ Options:
108
+ --model <name> Override the model (gpt-5, claude-opus, etc.)
109
+ --read-only Prevent file modifications
110
+ --no-approval Skip approval prompts (use with caution)
111
+ --config <key=val> Override config values
112
+ --oss Use local open source models
113
+ --sandbox <mode> Set sandbox level (read-only, workspace-write, etc.)
114
+ --help Show help information
115
+ --debug Log API requests and responses to file
116
+ --version Show version number
454
117
  ```
455
118
 
456
- ### Custom instructions
119
+ &ensp;
120
+ ## Memory & project docs
457
121
 
458
- You can create a `~/.codex/AGENTS.md` file to define custom guidance for the agent:
122
+ Code can remember context across sessions:
459
123
 
124
+ 1. **Create an `AGENTS.md` or `CLAUDE.md` file** in your project root:
460
125
  ```markdown
461
- - Always respond with emojis
462
- - Only use git commands when explicitly requested
126
+ # Project Context
127
+ This is a React TypeScript application with:
128
+ - Authentication via JWT
129
+ - PostgreSQL database
130
+ - Express.js backend
131
+
132
+ ## Key files:
133
+ - `/src/auth/` - Authentication logic
134
+ - `/src/api/` - API client code
135
+ - `/server/` - Backend services
463
136
  ```
464
137
 
465
- ### Environment variables setup
138
+ 2. **Session memory**: Code maintains conversation history
139
+ 3. **Codebase analysis**: Automatically understands project structure
466
140
 
467
- For each AI provider, you need to set the corresponding API key in your environment variables. For example:
141
+ &ensp;
142
+ ## Non-interactive / CI mode
468
143
 
469
- ```bash
470
- # OpenAI
471
- export OPENAI_API_KEY="your-api-key-here"
144
+ For automation and CI/CD:
472
145
 
473
- # Azure OpenAI
474
- export AZURE_OPENAI_API_KEY="your-azure-api-key-here"
475
- export AZURE_OPENAI_API_VERSION="2025-04-01-preview" (Optional)
146
+ ```shell
147
+ # Run a specific task
148
+ code --no-approval "run tests and fix any failures"
476
149
 
477
- # OpenRouter
478
- export OPENROUTER_API_KEY="your-openrouter-key-here"
150
+ # Generate reports
151
+ code --read-only "analyze code quality and generate report"
479
152
 
480
- # Similarly for other providers
153
+ # Batch processing
154
+ code --config output_format=json "list all TODO comments"
481
155
  ```
482
156
 
483
- ---
157
+ &ensp;
158
+ ## Model Context Protocol (MCP)
484
159
 
485
- ## FAQ
160
+ Code supports MCP for extended capabilities:
486
161
 
487
- <details>
488
- <summary>OpenAI released a model called Codex in 2021 - is this related?</summary>
162
+ - **File operations**: Advanced file system access
163
+ - **Database connections**: Query and modify databases
164
+ - **API integrations**: Connect to external services
165
+ - **Custom tools**: Build your own extensions
489
166
 
490
- In 2021, OpenAI released Codex, an AI system designed to generate code from natural language prompts. That original Codex model was deprecated as of March 2023 and is separate from the CLI tool.
167
+ Configure MCP in `~/.codex/config.toml`:
491
168
 
492
- </details>
493
-
494
- <details>
495
- <summary>Which models are supported?</summary>
496
-
497
- Any model available with [Responses API](https://platform.openai.com/docs/api-reference/responses). The default is `o4-mini`, but pass `--model gpt-4.1` or set `model: gpt-4.1` in your config file to override.
498
-
499
- </details>
500
- <details>
501
- <summary>Why does <code>o3</code> or <code>o4-mini</code> not work for me?</summary>
502
-
503
- It's possible that your [API account needs to be verified](https://help.openai.com/en/articles/10910291-api-organization-verification) in order to start streaming responses and seeing chain of thought summaries from the API. If you're still running into issues, please let us know!
504
-
505
- </details>
506
-
507
- <details>
508
- <summary>How do I stop Codex from editing my files?</summary>
509
-
510
- Codex runs model-generated commands in a sandbox. If a proposed command or file change doesn't look right, you can simply type **n** to deny the command or give the model feedback.
169
+ ```toml
170
+ [[mcp_servers]]
171
+ name = "filesystem"
172
+ command = "npx"
173
+ args = ["-y", "@modelcontextprotocol/server-filesystem", "/path/to/project"]
174
+ ```
511
175
 
512
- </details>
513
- <details>
514
- <summary>Does it work on Windows?</summary>
176
+ &ensp;
177
+ ## Configuration
515
178
 
516
- Not directly. It requires [Windows Subsystem for Linux (WSL2)](https://learn.microsoft.com/en-us/windows/wsl/install) - Codex has been tested on macOS and Linux with Node 22.
179
+ Main config file: `~/.codex/config.toml`
517
180
 
518
- </details>
181
+ ```toml
182
+ # Model settings
183
+ model = "gpt-5"
184
+ model_provider = "openai"
519
185
 
520
- ---
186
+ # Behavior
187
+ approval_policy = "on_request" # untrusted | on-failure | on-request | never
188
+ model_reasoning_effort = "medium" # low | medium | high
189
+ sandbox_mode = "workspace_write"
521
190
 
522
- ## Zero data retention (ZDR) usage
191
+ # UI preferences see THEME_CONFIG.md
192
+ [tui.theme]
193
+ name = "light-photon"
523
194
 
524
- Codex CLI **does** support OpenAI organizations with [Zero Data Retention (ZDR)](https://platform.openai.com/docs/guides/your-data#zero-data-retention) enabled. If your OpenAI organization has Zero Data Retention enabled and you still encounter errors such as:
525
-
526
- ```
527
- OpenAI rejected the request. Error details: Status: 400, Code: unsupported_parameter, Type: invalid_request_error, Message: 400 Previous response cannot be used for this organization due to Zero Data Retention.
195
+ # Add config for specific models
196
+ [profiles.gpt-5]
197
+ model = "gpt-5"
198
+ model_provider = "openai"
199
+ approval_policy = "never"
200
+ model_reasoning_effort = "high"
201
+ model_reasoning_summary = "detailed"
528
202
  ```
529
203
 
530
- You may need to upgrade to a more recent version with: `npm i -g @openai/codex@latest`
204
+ ### Environment variables
531
205
 
532
- ---
206
+ - `CODEX_HOME`: Override config directory location
207
+ - `OPENAI_API_KEY`: Use API key instead of ChatGPT auth
208
+ - `OPENAI_BASE_URL`: Use alternative API endpoints
533
209
 
534
- ## Codex open source fund
210
+ &ensp;
211
+ ## FAQ
535
212
 
536
- We're excited to launch a **$1 million initiative** supporting open source projects that use Codex CLI and other OpenAI models.
213
+ **How is this different from the original?**
214
+ > This fork adds browser integration, multi-agent commands (`/plan`, `/solve`, `/code`), theme system, and enhanced reasoning controls while maintaining full compatibility.
537
215
 
538
- - Grants are awarded up to **$25,000** API credits.
539
- - Applications are reviewed **on a rolling basis**.
216
+ **Can I use my existing Codex configuration?**
217
+ > Yes! This fork is fully backward compatible with existing `~/.codex/` configurations.
540
218
 
541
- **Interested? [Apply here](https://openai.com/form/codex-open-source-fund/).**
219
+ **Does this work with ChatGPT Plus?**
220
+ > Absolutely. Use the same "Sign in with ChatGPT" flow as the original.
542
221
 
543
- ---
222
+ **Is my data secure?**
223
+ > Yes. Authentication stays on your machine, and we don't proxy your credentials or conversations.
544
224
 
225
+ &ensp;
545
226
  ## Contributing
546
227
 
547
- This project is under active development and the code will likely change pretty significantly. We'll update this message once that's complete!
548
-
549
- More broadly we welcome contributions - whether you are opening your very first pull request or you're a seasoned maintainer. At the same time we care about reliability and long-term maintainability, so the bar for merging code is intentionally **high**. The guidelines below spell out what "high-quality" means in practice and should make the whole process transparent and friendly.
228
+ We welcome contributions! This fork maintains compatibility with upstream while adding community-requested features.
550
229
 
551
230
  ### Development workflow
552
231
 
553
- - Create a _topic branch_ from `main` - e.g. `feat/interactive-prompt`.
554
- - Keep your changes focused. Multiple unrelated fixes should be opened as separate PRs.
555
- - Use `pnpm test:watch` during development for super-fast feedback.
556
- - We use **Vitest** for unit tests, **ESLint** + **Prettier** for style, and **TypeScript** for type-checking.
557
- - Before pushing, run the full test/type/lint suite:
558
-
559
- ### Git hooks with Husky
560
-
561
- This project uses [Husky](https://typicode.github.io/husky/) to enforce code quality checks:
562
-
563
- - **Pre-commit hook**: Automatically runs lint-staged to format and lint files before committing
564
- - **Pre-push hook**: Runs tests and type checking before pushing to the remote
565
-
566
- These hooks help maintain code quality and prevent pushing code with failing tests. For more details, see [HUSKY.md](./HUSKY.md).
567
-
568
- ```bash
569
- pnpm test && pnpm run lint && pnpm run typecheck
570
- ```
571
-
572
- - If you have **not** yet signed the Contributor License Agreement (CLA), add a PR comment containing the exact text
573
-
574
- ```text
575
- I have read the CLA Document and I hereby sign the CLA
576
- ```
577
-
578
- The CLA-Assistant bot will turn the PR status green once all authors have signed.
579
-
580
232
  ```bash
581
- # Watch mode (tests rerun on change)
582
- pnpm test:watch
233
+ # Clone and setup
234
+ git clone https://github.com/just-every/code.git
235
+ cd code
236
+ npm install
583
237
 
584
- # Type-check without emitting files
585
- pnpm typecheck
238
+ # Build (use fast build for development)
239
+ ./build-fast.sh
586
240
 
587
- # Automatically fix lint + prettier issues
588
- pnpm lint:fix
589
- pnpm format:fix
241
+ # Run locally
242
+ ./codex-rs/target/dev-fast/code
590
243
  ```
591
244
 
592
- ### Debugging
593
-
594
- To debug the CLI with a visual debugger, do the following in the `codex-cli` folder:
595
-
596
- - Run `pnpm run build` to build the CLI, which will generate `cli.js.map` alongside `cli.js` in the `dist` folder.
597
- - Run the CLI with `node --inspect-brk ./dist/cli.js` The program then waits until a debugger is attached before proceeding. Options:
598
- - In VS Code, choose **Debug: Attach to Node Process** from the command palette and choose the option in the dropdown with debug port `9229` (likely the first option)
599
- - Go to <chrome://inspect> in Chrome and find **localhost:9229** and click **trace**
600
-
601
- ### Writing high-impact code changes
602
-
603
- 1. **Start with an issue.** Open a new one or comment on an existing discussion so we can agree on the solution before code is written.
604
- 2. **Add or update tests.** Every new feature or bug-fix should come with test coverage that fails before your change and passes afterwards. 100% coverage is not required, but aim for meaningful assertions.
605
- 3. **Document behaviour.** If your change affects user-facing behaviour, update the README, inline help (`codex --help`), or relevant example projects.
606
- 4. **Keep commits atomic.** Each commit should compile and the tests should pass. This makes reviews and potential rollbacks easier.
607
-
608
245
  ### Opening a pull request
609
246
 
610
- - Fill in the PR template (or include similar information) - **What? Why? How?**
611
- - Run **all** checks locally (`npm test && npm run lint && npm run typecheck`). CI failures that could have been caught locally slow down the process.
612
- - Make sure your branch is up-to-date with `main` and that you have resolved merge conflicts.
613
- - Mark the PR as **Ready for review** only when you believe it is in a merge-able state.
614
-
615
- ### Review process
616
-
617
- 1. One maintainer will be assigned as a primary reviewer.
618
- 2. We may ask for changes - please do not take this personally. We value the work, we just also value consistency and long-term maintainability.
619
- 3. When there is consensus that the PR meets the bar, a maintainer will squash-and-merge.
247
+ 1. Fork the repository
248
+ 2. Create a feature branch: `git checkout -b feature/amazing-feature`
249
+ 3. Make your changes
250
+ 4. Run tests: `cargo test`
251
+ 5. Build successfully: `./build-fast.sh`
252
+ 6. Submit a pull request
620
253
 
621
- ### Community values
622
254
 
623
- - **Be kind and inclusive.** Treat others with respect; we follow the [Contributor Covenant](https://www.contributor-covenant.org/).
624
- - **Assume good intent.** Written communication is hard - err on the side of generosity.
625
- - **Teach & learn.** If you spot something confusing, open an issue or PR with improvements.
255
+ &ensp;
256
+ ## Legal & Use
626
257
 
627
- ### Getting help
258
+ ### License & attribution
259
+ - This project is a community fork of `openai/codex` under **Apache-2.0**. We preserve upstream LICENSE and NOTICE files.
260
+ - **Code** is **not** affiliated with, sponsored by, or endorsed by OpenAI.
628
261
 
629
- If you run into problems setting up the project, would like feedback on an idea, or just want to say _hi_ - please open a Discussion or jump into the relevant issue. We are happy to help.
262
+ ### Your responsibilities
263
+ Using OpenAI, Anthropic or Google services through Code means you agree to **their Terms and policies**. In particular:
264
+ - **Don't** programmatically scrape/extract content outside intended flows.
265
+ - **Don't** bypass or interfere with rate limits, quotas, or safety mitigations.
266
+ - Use your **own** account; don't share or rotate accounts to evade limits.
267
+ - If you configure other model providers, you're responsible for their terms.
630
268
 
631
- Together we can make Codex CLI an incredible tool. **Happy hacking!** :rocket:
269
+ ### Privacy
270
+ - Your auth file lives at `~/.codex/auth.json`.
271
+ - Inputs/outputs you send to AI providers are handled under their Terms and Privacy Policy; consult those documents (and any org-level data-sharing settings).
632
272
 
633
- ### Contributor license agreement (CLA)
273
+ ### Subject to change
274
+ AI providers can change eligibility, limits, models, or authentication flows. Code supports **both** ChatGPT sign-in and API-key modes so you can pick what fits (local/hobby vs CI/automation).
634
275
 
635
- All contributors **must** accept the CLA. The process is lightweight:
636
-
637
- 1. Open your pull request.
638
- 2. Paste the following comment (or reply `recheck` if you've signed before):
639
-
640
- ```text
641
- I have read the CLA Document and I hereby sign the CLA
642
- ```
643
-
644
- 3. The CLA-Assistant bot records your signature in the repo and marks the status check as passed.
645
-
646
- No special Git commands, email attachments, or commit footers required.
647
-
648
- #### Quick fixes
649
-
650
- | Scenario | Command |
651
- | ----------------- | ------------------------------------------------ |
652
- | Amend last commit | `git commit --amend -s --no-edit && git push -f` |
653
-
654
- The **DCO check** blocks merges until every commit in the PR carries the footer (with squash this is just the one).
655
-
656
- ### Releasing `codex`
657
-
658
- To publish a new version of the CLI you first need to stage the npm package. A
659
- helper script in `codex-cli/scripts/` does all the heavy lifting. Inside the
660
- `codex-cli` folder run:
661
-
662
- ```bash
663
- # Classic, JS implementation that includes small, native binaries for Linux sandboxing.
664
- pnpm stage-release
665
-
666
- # Optionally specify the temp directory to reuse between runs.
667
- RELEASE_DIR=$(mktemp -d)
668
- pnpm stage-release --tmp "$RELEASE_DIR"
669
-
670
- # "Fat" package that additionally bundles the native Rust CLI binaries for
671
- # Linux. End-users can then opt-in at runtime by setting CODEX_RUST=1.
672
- pnpm stage-release --native
673
- ```
674
-
675
- Go to the folder where the release is staged and verify that it works as intended. If so, run the following from the temp folder:
676
-
677
- ```
678
- cd "$RELEASE_DIR"
679
- npm publish
680
- ```
681
-
682
- ### Alternative build options
683
-
684
- #### Nix flake development
685
-
686
- Prerequisite: Nix >= 2.4 with flakes enabled (`experimental-features = nix-command flakes` in `~/.config/nix/nix.conf`).
687
-
688
- Enter a Nix development shell:
689
-
690
- ```bash
691
- # Use either one of the commands according to which implementation you want to work with
692
- nix develop .#codex-cli # For entering codex-cli specific shell
693
- nix develop .#codex-rs # For entering codex-rs specific shell
694
- ```
695
-
696
- This shell includes Node.js, installs dependencies, builds the CLI, and provides a `codex` command alias.
697
-
698
- Build and run the CLI directly:
699
-
700
- ```bash
701
- # Use either one of the commands according to which implementation you want to work with
702
- nix build .#codex-cli # For building codex-cli
703
- nix build .#codex-rs # For building codex-rs
704
- ./result/bin/codex --help
705
- ```
706
-
707
- Run the CLI via the flake app:
708
-
709
- ```bash
710
- # Use either one of the commands according to which implementation you want to work with
711
- nix run .#codex-cli # For running codex-cli
712
- nix run .#codex-rs # For running codex-rs
713
- ```
714
-
715
- Use direnv with flakes
716
-
717
- If you have direnv installed, you can use the following `.envrc` to automatically enter the Nix shell when you `cd` into the project directory:
718
-
719
- ```bash
720
- cd codex-rs
721
- echo "use flake ../flake.nix#codex-cli" >> .envrc && direnv allow
722
- cd codex-cli
723
- echo "use flake ../flake.nix#codex-rs" >> .envrc && direnv allow
724
- ```
725
-
726
- ---
276
+ &ensp;
277
+ ## License
727
278
 
728
- ## Security & responsible AI
279
+ Apache 2.0 - See [LICENSE](LICENSE) file for details.
729
280
 
730
- Have you discovered a vulnerability or have concerns about model output? Please e-mail **security@openai.com** and we will respond promptly.
281
+ This project is a community fork of the original Codex CLI. We maintain compatibility while adding enhanced features requested by the developer community.
731
282
 
283
+ &ensp;
732
284
  ---
733
-
734
- ## License
735
-
736
- This repository is licensed under the [Apache-2.0 License](LICENSE).
285
+ **Need help?** Open an issue on [GitHub](https://github.com/just-every/code/issues) or check our documentation.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@just-every/code",
3
- "version": "0.1.7",
3
+ "version": "0.2.1",
4
4
  "license": "Apache-2.0",
5
5
  "description": "Lightweight coding agent that runs in your terminal - fork of OpenAI Codex",
6
6
  "bin": {
@@ -18,7 +18,8 @@
18
18
  "dist"
19
19
  ],
20
20
  "scripts": {
21
- "postinstall": "node postinstall.js"
21
+ "postinstall": "node postinstall.js",
22
+ "prepublishOnly": "node -e \"const fs=require('fs'),path=require('path'); const repoGit=path.join(__dirname,'..','.git'); const inCi=process.env.GITHUB_ACTIONS==='true'||process.env.CI==='true'; if(fs.existsSync(repoGit) && !inCi){ console.error('Refusing to publish from codex-cli. Use codex-cli/scripts/stage_release.sh to stage a release.'); process.exit(1);} else { console.log(inCi ? 'CI publish detected.' : 'Publishing staged package...'); }\""
22
23
  },
23
24
  "repository": {
24
25
  "type": "git",