@eminent337/aery 0.67.68

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (148) hide show
  1. package/CHANGELOG.md +3768 -0
  2. package/README.md +623 -0
  3. package/docs/compaction.md +394 -0
  4. package/docs/custom-provider.md +637 -0
  5. package/docs/development.md +71 -0
  6. package/docs/extensions.md +2368 -0
  7. package/docs/images/doom-extension.png +0 -0
  8. package/docs/images/exy.png +0 -0
  9. package/docs/images/interactive-mode.png +0 -0
  10. package/docs/images/tree-view.png +0 -0
  11. package/docs/json.md +82 -0
  12. package/docs/keybindings.md +197 -0
  13. package/docs/models.md +395 -0
  14. package/docs/packages.md +218 -0
  15. package/docs/prompt-templates.md +88 -0
  16. package/docs/providers.md +195 -0
  17. package/docs/rpc.md +1407 -0
  18. package/docs/sdk.md +1149 -0
  19. package/docs/session.md +412 -0
  20. package/docs/settings.md +247 -0
  21. package/docs/shell-aliases.md +13 -0
  22. package/docs/skills.md +232 -0
  23. package/docs/terminal-setup.md +106 -0
  24. package/docs/termux.md +127 -0
  25. package/docs/themes.md +295 -0
  26. package/docs/tmux.md +61 -0
  27. package/docs/tree.md +233 -0
  28. package/docs/tui.md +918 -0
  29. package/docs/windows.md +17 -0
  30. package/examples/README.md +25 -0
  31. package/examples/extensions/README.md +208 -0
  32. package/examples/extensions/antigravity-image-gen.ts +418 -0
  33. package/examples/extensions/auto-commit-on-exit.ts +49 -0
  34. package/examples/extensions/bash-spawn-hook.ts +30 -0
  35. package/examples/extensions/bookmark.ts +50 -0
  36. package/examples/extensions/built-in-tool-renderer.ts +249 -0
  37. package/examples/extensions/claude-rules.ts +86 -0
  38. package/examples/extensions/commands.ts +72 -0
  39. package/examples/extensions/confirm-destructive.ts +59 -0
  40. package/examples/extensions/custom-compaction.ts +127 -0
  41. package/examples/extensions/custom-footer.ts +64 -0
  42. package/examples/extensions/custom-header.ts +73 -0
  43. package/examples/extensions/custom-provider-anthropic/index.ts +604 -0
  44. package/examples/extensions/custom-provider-anthropic/package-lock.json +24 -0
  45. package/examples/extensions/custom-provider-anthropic/package.json +19 -0
  46. package/examples/extensions/custom-provider-gitlab-duo/index.ts +349 -0
  47. package/examples/extensions/custom-provider-gitlab-duo/package.json +16 -0
  48. package/examples/extensions/custom-provider-gitlab-duo/test.ts +82 -0
  49. package/examples/extensions/custom-provider-qwen-cli/index.ts +345 -0
  50. package/examples/extensions/custom-provider-qwen-cli/package.json +16 -0
  51. package/examples/extensions/dirty-repo-guard.ts +56 -0
  52. package/examples/extensions/doom-overlay/README.md +46 -0
  53. package/examples/extensions/doom-overlay/doom/build/doom.js +21 -0
  54. package/examples/extensions/doom-overlay/doom/build/doom.wasm +0 -0
  55. package/examples/extensions/doom-overlay/doom/build.sh +152 -0
  56. package/examples/extensions/doom-overlay/doom/doomgeneric_pi.c +72 -0
  57. package/examples/extensions/doom-overlay/doom-component.ts +132 -0
  58. package/examples/extensions/doom-overlay/doom-engine.ts +173 -0
  59. package/examples/extensions/doom-overlay/doom-keys.ts +104 -0
  60. package/examples/extensions/doom-overlay/index.ts +74 -0
  61. package/examples/extensions/doom-overlay/wad-finder.ts +51 -0
  62. package/examples/extensions/dynamic-resources/SKILL.md +8 -0
  63. package/examples/extensions/dynamic-resources/dynamic.json +79 -0
  64. package/examples/extensions/dynamic-resources/dynamic.md +5 -0
  65. package/examples/extensions/dynamic-resources/index.ts +15 -0
  66. package/examples/extensions/dynamic-tools.ts +74 -0
  67. package/examples/extensions/event-bus.ts +43 -0
  68. package/examples/extensions/file-trigger.ts +41 -0
  69. package/examples/extensions/git-checkpoint.ts +53 -0
  70. package/examples/extensions/handoff.ts +153 -0
  71. package/examples/extensions/hello.ts +26 -0
  72. package/examples/extensions/hidden-thinking-label.ts +53 -0
  73. package/examples/extensions/inline-bash.ts +94 -0
  74. package/examples/extensions/input-transform.ts +43 -0
  75. package/examples/extensions/interactive-shell.ts +196 -0
  76. package/examples/extensions/mac-system-theme.ts +47 -0
  77. package/examples/extensions/message-renderer.ts +59 -0
  78. package/examples/extensions/minimal-mode.ts +426 -0
  79. package/examples/extensions/modal-editor.ts +85 -0
  80. package/examples/extensions/model-status.ts +31 -0
  81. package/examples/extensions/notify.ts +55 -0
  82. package/examples/extensions/overlay-qa-tests.ts +1348 -0
  83. package/examples/extensions/overlay-test.ts +150 -0
  84. package/examples/extensions/permission-gate.ts +34 -0
  85. package/examples/extensions/pirate.ts +47 -0
  86. package/examples/extensions/plan-mode/README.md +65 -0
  87. package/examples/extensions/plan-mode/index.ts +340 -0
  88. package/examples/extensions/plan-mode/utils.ts +168 -0
  89. package/examples/extensions/preset.ts +430 -0
  90. package/examples/extensions/protected-paths.ts +30 -0
  91. package/examples/extensions/provider-payload.ts +18 -0
  92. package/examples/extensions/qna.ts +122 -0
  93. package/examples/extensions/question.ts +264 -0
  94. package/examples/extensions/questionnaire.ts +427 -0
  95. package/examples/extensions/rainbow-editor.ts +88 -0
  96. package/examples/extensions/reload-runtime.ts +37 -0
  97. package/examples/extensions/rpc-demo.ts +118 -0
  98. package/examples/extensions/sandbox/index.ts +321 -0
  99. package/examples/extensions/sandbox/package-lock.json +92 -0
  100. package/examples/extensions/sandbox/package.json +19 -0
  101. package/examples/extensions/send-user-message.ts +97 -0
  102. package/examples/extensions/session-name.ts +27 -0
  103. package/examples/extensions/shutdown-command.ts +63 -0
  104. package/examples/extensions/snake.ts +343 -0
  105. package/examples/extensions/space-invaders.ts +560 -0
  106. package/examples/extensions/ssh.ts +220 -0
  107. package/examples/extensions/status-line.ts +32 -0
  108. package/examples/extensions/subagent/README.md +172 -0
  109. package/examples/extensions/subagent/agents/planner.md +37 -0
  110. package/examples/extensions/subagent/agents/reviewer.md +35 -0
  111. package/examples/extensions/subagent/agents/scout.md +50 -0
  112. package/examples/extensions/subagent/agents/worker.md +24 -0
  113. package/examples/extensions/subagent/agents.ts +126 -0
  114. package/examples/extensions/subagent/index.ts +987 -0
  115. package/examples/extensions/subagent/prompts/implement-and-review.md +10 -0
  116. package/examples/extensions/subagent/prompts/implement.md +10 -0
  117. package/examples/extensions/subagent/prompts/scout-and-plan.md +9 -0
  118. package/examples/extensions/summarize.ts +206 -0
  119. package/examples/extensions/system-prompt-header.ts +17 -0
  120. package/examples/extensions/tic-tac-toe.ts +1008 -0
  121. package/examples/extensions/timed-confirm.ts +70 -0
  122. package/examples/extensions/titlebar-spinner.ts +58 -0
  123. package/examples/extensions/todo.ts +297 -0
  124. package/examples/extensions/tool-override.ts +144 -0
  125. package/examples/extensions/tools.ts +141 -0
  126. package/examples/extensions/trigger-compact.ts +50 -0
  127. package/examples/extensions/truncated-tool.ts +195 -0
  128. package/examples/extensions/widget-placement.ts +9 -0
  129. package/examples/extensions/with-deps/index.ts +32 -0
  130. package/examples/extensions/with-deps/package-lock.json +31 -0
  131. package/examples/extensions/with-deps/package.json +22 -0
  132. package/examples/extensions/working-indicator.ts +123 -0
  133. package/examples/rpc-extension-ui.ts +632 -0
  134. package/examples/sdk/01-minimal.ts +22 -0
  135. package/examples/sdk/02-custom-model.ts +49 -0
  136. package/examples/sdk/03-custom-prompt.ts +62 -0
  137. package/examples/sdk/04-skills.ts +55 -0
  138. package/examples/sdk/05-tools.ts +44 -0
  139. package/examples/sdk/06-extensions.ts +90 -0
  140. package/examples/sdk/07-context-files.ts +42 -0
  141. package/examples/sdk/08-prompt-templates.ts +51 -0
  142. package/examples/sdk/09-api-keys-and-oauth.ts +48 -0
  143. package/examples/sdk/10-settings.ts +53 -0
  144. package/examples/sdk/11-sessions.ts +48 -0
  145. package/examples/sdk/12-full-control.ts +73 -0
  146. package/examples/sdk/13-session-runtime.ts +67 -0
  147. package/examples/sdk/README.md +147 -0
  148. package/package.json +102 -0
@@ -0,0 +1,637 @@
1
+ # Custom Providers
2
+
3
+ Extensions can register custom model providers via `pi.registerProvider()`. This enables:
4
+
5
+ - **Proxies** - Route requests through corporate proxies or API gateways
6
+ - **Custom endpoints** - Use self-hosted or private model deployments
7
+ - **OAuth/SSO** - Add authentication flows for enterprise providers
8
+ - **Custom APIs** - Implement streaming for non-standard LLM APIs
9
+
10
+ ## Example Extensions
11
+
12
+ See these complete provider examples:
13
+
14
+ - [`examples/extensions/custom-provider-anthropic/`](../examples/extensions/custom-provider-anthropic/)
15
+ - [`examples/extensions/custom-provider-gitlab-duo/`](../examples/extensions/custom-provider-gitlab-duo/)
16
+ - [`examples/extensions/custom-provider-qwen-cli/`](../examples/extensions/custom-provider-qwen-cli/)
17
+
18
+ ## Table of Contents
19
+
20
+ - [Example Extensions](#example-extensions)
21
+ - [Quick Reference](#quick-reference)
22
+ - [Override Existing Provider](#override-existing-provider)
23
+ - [Register New Provider](#register-new-provider)
24
+ - [Unregister Provider](#unregister-provider)
25
+ - [OAuth Support](#oauth-support)
26
+ - [Custom Streaming API](#custom-streaming-api)
27
+ - [Testing Your Implementation](#testing-your-implementation)
28
+ - [Config Reference](#config-reference)
29
+ - [Model Definition Reference](#model-definition-reference)
30
+
31
+ ## Quick Reference
32
+
33
+ ```typescript
34
+ import type { ExtensionAPI } from "@mariozechner/pi-coding-agent";
35
+
36
+ export default function (pi: ExtensionAPI) {
37
+ // Override baseUrl for existing provider
38
+ pi.registerProvider("anthropic", {
39
+ baseUrl: "https://proxy.example.com"
40
+ });
41
+
42
+ // Register new provider with models
43
+ pi.registerProvider("my-provider", {
44
+ baseUrl: "https://api.example.com",
45
+ apiKey: "MY_API_KEY",
46
+ api: "openai-completions",
47
+ models: [
48
+ {
49
+ id: "my-model",
50
+ name: "My Model",
51
+ reasoning: false,
52
+ input: ["text", "image"],
53
+ cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
54
+ contextWindow: 128000,
55
+ maxTokens: 4096
56
+ }
57
+ ]
58
+ });
59
+ }
60
+ ```
61
+
62
+ The extension factory can also be `async`. For dynamic model discovery, fetch and register models in the factory instead of `session_start`. pi waits for the factory before startup continues, so the provider is available during interactive startup and to `pi --list-models`.
63
+
64
+ ## Override Existing Provider
65
+
66
+ The simplest use case: redirect an existing provider through a proxy.
67
+
68
+ ```typescript
69
+ // All Anthropic requests now go through your proxy
70
+ pi.registerProvider("anthropic", {
71
+ baseUrl: "https://proxy.example.com"
72
+ });
73
+
74
+ // Add custom headers to OpenAI requests
75
+ pi.registerProvider("openai", {
76
+ headers: {
77
+ "X-Custom-Header": "value"
78
+ }
79
+ });
80
+
81
+ // Both baseUrl and headers
82
+ pi.registerProvider("google", {
83
+ baseUrl: "https://ai-gateway.corp.com/google",
84
+ headers: {
85
+ "X-Corp-Auth": "CORP_AUTH_TOKEN" // env var or literal
86
+ }
87
+ });
88
+ ```
89
+
90
+ When only `baseUrl` and/or `headers` are provided (no `models`), all existing models for that provider are preserved with the new endpoint.
91
+
92
+ ## Register New Provider
93
+
94
+ To add a completely new provider, specify `models` along with the required configuration.
95
+
96
+ If the model list comes from a remote endpoint, use an async extension factory:
97
+
98
+ ```typescript
99
+ import type { ExtensionAPI } from "@mariozechner/pi-coding-agent";
100
+
101
+ export default async function (pi: ExtensionAPI) {
102
+ const response = await fetch("http://localhost:1234/v1/models");
103
+ const payload = (await response.json()) as {
104
+ data: Array<{
105
+ id: string;
106
+ name?: string;
107
+ context_window?: number;
108
+ max_tokens?: number;
109
+ }>;
110
+ };
111
+
112
+ pi.registerProvider("local-openai", {
113
+ baseUrl: "http://localhost:1234/v1",
114
+ apiKey: "LOCAL_OPENAI_API_KEY",
115
+ api: "openai-completions",
116
+ models: payload.data.map((model) => ({
117
+ id: model.id,
118
+ name: model.name ?? model.id,
119
+ reasoning: false,
120
+ input: ["text"],
121
+ cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 },
122
+ contextWindow: model.context_window ?? 128000,
123
+ maxTokens: model.max_tokens ?? 4096,
124
+ })),
125
+ });
126
+ }
127
+ ```
128
+
129
+ This registers the fetched models before startup finishes.
130
+
131
+ ```typescript
132
+ pi.registerProvider("my-llm", {
133
+ baseUrl: "https://api.my-llm.com/v1",
134
+ apiKey: "MY_LLM_API_KEY", // env var name or literal value
135
+ api: "openai-completions", // which streaming API to use
136
+ models: [
137
+ {
138
+ id: "my-llm-large",
139
+ name: "My LLM Large",
140
+ reasoning: true, // supports extended thinking
141
+ input: ["text", "image"],
142
+ cost: {
143
+ input: 3.0, // $/million tokens
144
+ output: 15.0,
145
+ cacheRead: 0.3,
146
+ cacheWrite: 3.75
147
+ },
148
+ contextWindow: 200000,
149
+ maxTokens: 16384
150
+ }
151
+ ]
152
+ });
153
+ ```
154
+
155
+ When `models` is provided, it **replaces** all existing models for that provider.
156
+
157
+ ## Unregister Provider
158
+
159
+ Use `pi.unregisterProvider(name)` to remove a provider that was previously registered via `pi.registerProvider(name, ...)`:
160
+
161
+ ```typescript
162
+ // Register
163
+ pi.registerProvider("my-llm", {
164
+ baseUrl: "https://api.my-llm.com/v1",
165
+ apiKey: "MY_LLM_API_KEY",
166
+ api: "openai-completions",
167
+ models: [
168
+ {
169
+ id: "my-llm-large",
170
+ name: "My LLM Large",
171
+ reasoning: true,
172
+ input: ["text", "image"],
173
+ cost: { input: 3.0, output: 15.0, cacheRead: 0.3, cacheWrite: 3.75 },
174
+ contextWindow: 200000,
175
+ maxTokens: 16384
176
+ }
177
+ ]
178
+ });
179
+
180
+ // Later, remove it
181
+ pi.unregisterProvider("my-llm");
182
+ ```
183
+
184
+ Unregistering removes that provider's dynamic models, API key fallback, OAuth provider registration, and custom stream handler registrations. Any built-in models or provider behavior that were overridden are restored.
185
+
186
+ Calls made after the initial extension load phase are applied immediately, so no `/reload` is required.
187
+
188
+ ### API Types
189
+
190
+ The `api` field determines which streaming implementation is used:
191
+
192
+ | API | Use for |
193
+ |-----|---------|
194
+ | `anthropic-messages` | Anthropic Claude API and compatibles |
195
+ | `openai-completions` | OpenAI Chat Completions API and compatibles |
196
+ | `openai-responses` | OpenAI Responses API |
197
+ | `azure-openai-responses` | Azure OpenAI Responses API |
198
+ | `openai-codex-responses` | OpenAI Codex Responses API |
199
+ | `mistral-conversations` | Mistral SDK Conversations/Chat streaming |
200
+ | `google-generative-ai` | Google Generative AI API |
201
+ | `google-gemini-cli` | Google Cloud Code Assist API |
202
+ | `google-vertex` | Google Vertex AI API |
203
+ | `bedrock-converse-stream` | Amazon Bedrock Converse API |
204
+
205
+ Most OpenAI-compatible providers work with `openai-completions`. Use `compat` for quirks:
206
+
207
+ ```typescript
208
+ models: [{
209
+ id: "custom-model",
210
+ // ...
211
+ compat: {
212
+ supportsDeveloperRole: false, // use "system" instead of "developer"
213
+ supportsReasoningEffort: true,
214
+ reasoningEffortMap: { // map pi-ai levels to provider values
215
+ minimal: "default",
216
+ low: "default",
217
+ medium: "default",
218
+ high: "default",
219
+ xhigh: "default"
220
+ },
221
+ maxTokensField: "max_tokens", // instead of "max_completion_tokens"
222
+ requiresToolResultName: true, // tool results need name field
223
+ thinkingFormat: "qwen", // top-level enable_thinking: true
224
+ cacheControlFormat: "anthropic" // Anthropic-style cache_control markers
225
+ }
226
+ }]
227
+ ```
228
+
229
+ Use `qwen-chat-template` instead for local Qwen-compatible servers that read `chat_template_kwargs.enable_thinking`.
230
+ Use `cacheControlFormat: "anthropic"` for OpenAI-compatible providers that expose Anthropic-style prompt caching via `cache_control` on the system prompt, last tool definition, and last user/assistant text content.
231
+
232
+ > Migration note: Mistral moved from `openai-completions` to `mistral-conversations`.
233
+ > Use `mistral-conversations` for native Mistral models.
234
+ > If you intentionally route Mistral-compatible/custom endpoints through `openai-completions`, set `compat` flags explicitly as needed.
235
+
236
+ ### Auth Header
237
+
238
+ If your provider expects `Authorization: Bearer <key>` but doesn't use a standard API, set `authHeader: true`:
239
+
240
+ ```typescript
241
+ pi.registerProvider("custom-api", {
242
+ baseUrl: "https://api.example.com",
243
+ apiKey: "MY_API_KEY",
244
+ authHeader: true, // adds Authorization: Bearer header
245
+ api: "openai-completions",
246
+ models: [...]
247
+ });
248
+ ```
249
+
250
+ ## OAuth Support
251
+
252
+ Add OAuth/SSO authentication that integrates with `/login`:
253
+
254
+ ```typescript
255
+ import type { OAuthCredentials, OAuthLoginCallbacks } from "@mariozechner/pi-ai";
256
+
257
+ pi.registerProvider("corporate-ai", {
258
+ baseUrl: "https://ai.corp.com/v1",
259
+ api: "openai-responses",
260
+ models: [...],
261
+ oauth: {
262
+ name: "Corporate AI (SSO)",
263
+
264
+ async login(callbacks: OAuthLoginCallbacks): Promise<OAuthCredentials> {
265
+ // Option 1: Browser-based OAuth
266
+ callbacks.onAuth({ url: "https://sso.corp.com/authorize?..." });
267
+
268
+ // Option 2: Device code flow
269
+ callbacks.onDeviceCode({
270
+ userCode: "ABCD-1234",
271
+ verificationUri: "https://sso.corp.com/device"
272
+ });
273
+
274
+ // Option 3: Prompt for token/code
275
+ const code = await callbacks.onPrompt({ message: "Enter SSO code:" });
276
+
277
+ // Exchange for tokens (your implementation)
278
+ const tokens = await exchangeCodeForTokens(code);
279
+
280
+ return {
281
+ refresh: tokens.refreshToken,
282
+ access: tokens.accessToken,
283
+ expires: Date.now() + tokens.expiresIn * 1000
284
+ };
285
+ },
286
+
287
+ async refreshToken(credentials: OAuthCredentials): Promise<OAuthCredentials> {
288
+ const tokens = await refreshAccessToken(credentials.refresh);
289
+ return {
290
+ refresh: tokens.refreshToken ?? credentials.refresh,
291
+ access: tokens.accessToken,
292
+ expires: Date.now() + tokens.expiresIn * 1000
293
+ };
294
+ },
295
+
296
+ getApiKey(credentials: OAuthCredentials): string {
297
+ return credentials.access;
298
+ },
299
+
300
+ // Optional: modify models based on user's subscription
301
+ modifyModels(models, credentials) {
302
+ const region = decodeRegionFromToken(credentials.access);
303
+ return models.map(m => ({
304
+ ...m,
305
+ baseUrl: `https://${region}.ai.corp.com/v1`
306
+ }));
307
+ }
308
+ }
309
+ });
310
+ ```
311
+
312
+ After registration, users can authenticate via `/login corporate-ai`.
313
+
314
+ ### OAuthLoginCallbacks
315
+
316
+ The `callbacks` object provides three ways to authenticate:
317
+
318
+ ```typescript
319
+ interface OAuthLoginCallbacks {
320
+ // Open URL in browser (for OAuth redirects)
321
+ onAuth(params: { url: string }): void;
322
+
323
+ // Show device code (for device authorization flow)
324
+ onDeviceCode(params: { userCode: string; verificationUri: string }): void;
325
+
326
+ // Prompt user for input (for manual token entry)
327
+ onPrompt(params: { message: string }): Promise<string>;
328
+ }
329
+ ```
330
+
331
+ ### OAuthCredentials
332
+
333
+ Credentials are persisted in `~/.pi/agent/auth.json`:
334
+
335
+ ```typescript
336
+ interface OAuthCredentials {
337
+ refresh: string; // Refresh token (for refreshToken())
338
+ access: string; // Access token (returned by getApiKey())
339
+ expires: number; // Expiration timestamp in milliseconds
340
+ }
341
+ ```
342
+
343
+ ## Custom Streaming API
344
+
345
+ For providers with non-standard APIs, implement `streamSimple`. Study the existing provider implementations before writing your own:
346
+
347
+ **Reference implementations:**
348
+ - [anthropic.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/anthropic.ts) - Anthropic Messages API
349
+ - [mistral.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/mistral.ts) - Mistral Conversations API
350
+ - [openai-completions.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/openai-completions.ts) - OpenAI Chat Completions
351
+ - [openai-responses.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/openai-responses.ts) - OpenAI Responses API
352
+ - [google.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/google.ts) - Google Generative AI
353
+ - [amazon-bedrock.ts](https://github.com/badlogic/pi-mono/blob/main/packages/ai/src/providers/amazon-bedrock.ts) - AWS Bedrock
354
+
355
+ ### Stream Pattern
356
+
357
+ All providers follow the same pattern:
358
+
359
+ ```typescript
360
+ import {
361
+ type AssistantMessage,
362
+ type AssistantMessageEventStream,
363
+ type Context,
364
+ type Model,
365
+ type SimpleStreamOptions,
366
+ calculateCost,
367
+ createAssistantMessageEventStream,
368
+ } from "@mariozechner/pi-ai";
369
+
370
+ function streamMyProvider(
371
+ model: Model<any>,
372
+ context: Context,
373
+ options?: SimpleStreamOptions
374
+ ): AssistantMessageEventStream {
375
+ const stream = createAssistantMessageEventStream();
376
+
377
+ (async () => {
378
+ // Initialize output message
379
+ const output: AssistantMessage = {
380
+ role: "assistant",
381
+ content: [],
382
+ api: model.api,
383
+ provider: model.provider,
384
+ model: model.id,
385
+ usage: {
386
+ input: 0,
387
+ output: 0,
388
+ cacheRead: 0,
389
+ cacheWrite: 0,
390
+ totalTokens: 0,
391
+ cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0, total: 0 },
392
+ },
393
+ stopReason: "stop",
394
+ timestamp: Date.now(),
395
+ };
396
+
397
+ try {
398
+ // Push start event
399
+ stream.push({ type: "start", partial: output });
400
+
401
+ // Make API request and process response...
402
+ // Push content events as they arrive...
403
+
404
+ // Push done event
405
+ stream.push({
406
+ type: "done",
407
+ reason: output.stopReason as "stop" | "length" | "toolUse",
408
+ message: output
409
+ });
410
+ stream.end();
411
+ } catch (error) {
412
+ output.stopReason = options?.signal?.aborted ? "aborted" : "error";
413
+ output.errorMessage = error instanceof Error ? error.message : String(error);
414
+ stream.push({ type: "error", reason: output.stopReason, error: output });
415
+ stream.end();
416
+ }
417
+ })();
418
+
419
+ return stream;
420
+ }
421
+ ```
422
+
423
+ ### Event Types
424
+
425
+ Push events via `stream.push()` in this order:
426
+
427
+ 1. `{ type: "start", partial: output }` - Stream started
428
+
429
+ 2. Content events (repeatable, track `contentIndex` for each block):
430
+ - `{ type: "text_start", contentIndex, partial }` - Text block started
431
+ - `{ type: "text_delta", contentIndex, delta, partial }` - Text chunk
432
+ - `{ type: "text_end", contentIndex, content, partial }` - Text block ended
433
+ - `{ type: "thinking_start", contentIndex, partial }` - Thinking started
434
+ - `{ type: "thinking_delta", contentIndex, delta, partial }` - Thinking chunk
435
+ - `{ type: "thinking_end", contentIndex, content, partial }` - Thinking ended
436
+ - `{ type: "toolcall_start", contentIndex, partial }` - Tool call started
437
+ - `{ type: "toolcall_delta", contentIndex, delta, partial }` - Tool call JSON chunk
438
+ - `{ type: "toolcall_end", contentIndex, toolCall, partial }` - Tool call ended
439
+
440
+ 3. `{ type: "done", reason, message }` or `{ type: "error", reason, error }` - Stream ended
441
+
442
+ The `partial` field in each event contains the current `AssistantMessage` state. Update `output.content` as you receive data, then include `output` as the `partial`.
443
+
444
+ ### Content Blocks
445
+
446
+ Add content blocks to `output.content` as they arrive:
447
+
448
+ ```typescript
449
+ // Text block
450
+ output.content.push({ type: "text", text: "" });
451
+ stream.push({ type: "text_start", contentIndex: output.content.length - 1, partial: output });
452
+
453
+ // As text arrives
454
+ const block = output.content[contentIndex];
455
+ if (block.type === "text") {
456
+ block.text += delta;
457
+ stream.push({ type: "text_delta", contentIndex, delta, partial: output });
458
+ }
459
+
460
+ // When block completes
461
+ stream.push({ type: "text_end", contentIndex, content: block.text, partial: output });
462
+ ```
463
+
464
+ ### Tool Calls
465
+
466
+ Tool calls require accumulating JSON and parsing:
467
+
468
+ ```typescript
469
+ // Start tool call
470
+ output.content.push({
471
+ type: "toolCall",
472
+ id: toolCallId,
473
+ name: toolName,
474
+ arguments: {}
475
+ });
476
+ stream.push({ type: "toolcall_start", contentIndex: output.content.length - 1, partial: output });
477
+
478
+ // Accumulate JSON
479
+ let partialJson = "";
480
+ partialJson += jsonDelta;
481
+ try {
482
+ block.arguments = JSON.parse(partialJson);
483
+ } catch {}
484
+ stream.push({ type: "toolcall_delta", contentIndex, delta: jsonDelta, partial: output });
485
+
486
+ // Complete
487
+ stream.push({
488
+ type: "toolcall_end",
489
+ contentIndex,
490
+ toolCall: { type: "toolCall", id, name, arguments: block.arguments },
491
+ partial: output
492
+ });
493
+ ```
494
+
495
+ ### Usage and Cost
496
+
497
+ Update usage from API response and calculate cost:
498
+
499
+ ```typescript
500
+ output.usage.input = response.usage.input_tokens;
501
+ output.usage.output = response.usage.output_tokens;
502
+ output.usage.cacheRead = response.usage.cache_read_tokens ?? 0;
503
+ output.usage.cacheWrite = response.usage.cache_write_tokens ?? 0;
504
+ output.usage.totalTokens = output.usage.input + output.usage.output +
505
+ output.usage.cacheRead + output.usage.cacheWrite;
506
+ calculateCost(model, output.usage);
507
+ ```
508
+
509
+ ### Registration
510
+
511
+ Register your stream function:
512
+
513
+ ```typescript
514
+ pi.registerProvider("my-provider", {
515
+ baseUrl: "https://api.example.com",
516
+ apiKey: "MY_API_KEY",
517
+ api: "my-custom-api",
518
+ models: [...],
519
+ streamSimple: streamMyProvider
520
+ });
521
+ ```
522
+
523
+ ## Testing Your Implementation
524
+
525
+ Test your provider against the same test suites used by built-in providers. Copy and adapt these test files from [packages/ai/test/](https://github.com/badlogic/pi-mono/tree/main/packages/ai/test):
526
+
527
+ | Test | Purpose |
528
+ |------|---------|
529
+ | `stream.test.ts` | Basic streaming, text output |
530
+ | `tokens.test.ts` | Token counting and usage |
531
+ | `abort.test.ts` | AbortSignal handling |
532
+ | `empty.test.ts` | Empty/minimal responses |
533
+ | `context-overflow.test.ts` | Context window limits |
534
+ | `image-limits.test.ts` | Image input handling |
535
+ | `unicode-surrogate.test.ts` | Unicode edge cases |
536
+ | `tool-call-without-result.test.ts` | Tool call edge cases |
537
+ | `image-tool-result.test.ts` | Images in tool results |
538
+ | `total-tokens.test.ts` | Total token calculation |
539
+ | `cross-provider-handoff.test.ts` | Context handoff between providers |
540
+
541
+ Run tests with your provider/model pairs to verify compatibility.
542
+
543
+ ## Config Reference
544
+
545
+ ```typescript
546
+ interface ProviderConfig {
547
+ /** API endpoint URL. Required when defining models. */
548
+ baseUrl?: string;
549
+
550
+ /** API key or environment variable name. Required when defining models (unless oauth). */
551
+ apiKey?: string;
552
+
553
+ /** API type for streaming. Required at provider or model level when defining models. */
554
+ api?: Api;
555
+
556
+ /** Custom streaming implementation for non-standard APIs. */
557
+ streamSimple?: (
558
+ model: Model<Api>,
559
+ context: Context,
560
+ options?: SimpleStreamOptions
561
+ ) => AssistantMessageEventStream;
562
+
563
+ /** Custom headers to include in requests. Values can be env var names. */
564
+ headers?: Record<string, string>;
565
+
566
+ /** If true, adds Authorization: Bearer header with the resolved API key. */
567
+ authHeader?: boolean;
568
+
569
+ /** Models to register. If provided, replaces all existing models for this provider. */
570
+ models?: ProviderModelConfig[];
571
+
572
+ /** OAuth provider for /login support. */
573
+ oauth?: {
574
+ name: string;
575
+ login(callbacks: OAuthLoginCallbacks): Promise<OAuthCredentials>;
576
+ refreshToken(credentials: OAuthCredentials): Promise<OAuthCredentials>;
577
+ getApiKey(credentials: OAuthCredentials): string;
578
+ modifyModels?(models: Model<Api>[], credentials: OAuthCredentials): Model<Api>[];
579
+ };
580
+ }
581
+ ```
582
+
583
+ ## Model Definition Reference
584
+
585
+ ```typescript
586
+ interface ProviderModelConfig {
587
+ /** Model ID (e.g., "claude-sonnet-4-20250514"). */
588
+ id: string;
589
+
590
+ /** Display name (e.g., "Claude 4 Sonnet"). */
591
+ name: string;
592
+
593
+ /** API type override for this specific model. */
594
+ api?: Api;
595
+
596
+ /** Whether the model supports extended thinking. */
597
+ reasoning: boolean;
598
+
599
+ /** Supported input types. */
600
+ input: ("text" | "image")[];
601
+
602
+ /** Cost per million tokens (for usage tracking). */
603
+ cost: {
604
+ input: number;
605
+ output: number;
606
+ cacheRead: number;
607
+ cacheWrite: number;
608
+ };
609
+
610
+ /** Maximum context window size in tokens. */
611
+ contextWindow: number;
612
+
613
+ /** Maximum output tokens. */
614
+ maxTokens: number;
615
+
616
+ /** Custom headers for this specific model. */
617
+ headers?: Record<string, string>;
618
+
619
+ /** OpenAI compatibility settings for openai-completions API. */
620
+ compat?: {
621
+ supportsStore?: boolean;
622
+ supportsDeveloperRole?: boolean;
623
+ supportsReasoningEffort?: boolean;
624
+ reasoningEffortMap?: Partial<Record<"minimal" | "low" | "medium" | "high" | "xhigh", string>>;
625
+ supportsUsageInStreaming?: boolean;
626
+ maxTokensField?: "max_completion_tokens" | "max_tokens";
627
+ requiresToolResultName?: boolean;
628
+ requiresAssistantAfterToolResult?: boolean;
629
+ requiresThinkingAsText?: boolean;
630
+ thinkingFormat?: "openai" | "zai" | "qwen" | "qwen-chat-template";
631
+ cacheControlFormat?: "anthropic";
632
+ };
633
+ }
634
+ ```
635
+
636
+ `qwen` is for DashScope-style top-level `enable_thinking`. Use `qwen-chat-template` for local Qwen-compatible servers that read `chat_template_kwargs.enable_thinking`.
637
+ `cacheControlFormat: "anthropic"` applies Anthropic-style `cache_control` markers to the system prompt, last tool definition, and last user/assistant text content.
@@ -0,0 +1,71 @@
1
+ # Development
2
+
3
+ See [AGENTS.md](../../../AGENTS.md) for additional guidelines.
4
+
5
+ ## Setup
6
+
7
+ ```bash
8
+ git clone https://github.com/badlogic/pi-mono
9
+ cd pi-mono
10
+ npm install
11
+ npm run build
12
+ ```
13
+
14
+ Run from source:
15
+
16
+ ```bash
17
+ /path/to/pi-mono/pi-test.sh
18
+ ```
19
+
20
+ The script can be run from any directory. Pi keeps the caller's current working directory.
21
+
22
+ ## Forking / Rebranding
23
+
24
+ Configure via `package.json`:
25
+
26
+ ```json
27
+ {
28
+ "piConfig": {
29
+ "name": "pi",
30
+ "configDir": ".pi"
31
+ }
32
+ }
33
+ ```
34
+
35
+ Change `name`, `configDir`, and `bin` field for your fork. Affects CLI banner, config paths, and environment variable names.
36
+
37
+ ## Path Resolution
38
+
39
+ Three execution modes: npm install, standalone binary, tsx from source.
40
+
41
+ **Always use `src/config.ts`** for package assets:
42
+
43
+ ```typescript
44
+ import { getPackageDir, getThemeDir } from "./config.js";
45
+ ```
46
+
47
+ Never use `__dirname` directly for package assets.
48
+
49
+ ## Debug Command
50
+
51
+ `/debug` (hidden) writes to `~/.pi/agent/pi-debug.log`:
52
+ - Rendered TUI lines with ANSI codes
53
+ - Last messages sent to the LLM
54
+
55
+ ## Testing
56
+
57
+ ```bash
58
+ ./test.sh # Run non-LLM tests (no API keys needed)
59
+ npm test # Run all tests
60
+ npm test -- test/specific.test.ts # Run specific test
61
+ ```
62
+
63
+ ## Project Structure
64
+
65
+ ```
66
+ packages/
67
+ ai/ # LLM provider abstraction
68
+ agent/ # Agent loop and message types
69
+ tui/ # Terminal UI components
70
+ coding-agent/ # CLI and interactive mode
71
+ ```