omegon 0.10.4 → 0.10.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/omegon.mjs +5 -0
- package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts +68 -0
- package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts.map +1 -1
- package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js +68 -0
- package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js.map +1 -1
- package/node_modules/@styrene-lab/pi-coding-agent/dist/modes/interactive/interactive-mode.d.ts.map +1 -1
- package/node_modules/@styrene-lab/pi-coding-agent/dist/modes/interactive/interactive-mode.js +5 -0
- package/node_modules/@styrene-lab/pi-coding-agent/dist/modes/interactive/interactive-mode.js.map +1 -1
- package/package.json +1 -1
package/bin/omegon.mjs
CHANGED
|
@@ -82,6 +82,11 @@ if (process.argv.includes("--where")) {
|
|
|
82
82
|
}
|
|
83
83
|
|
|
84
84
|
process.env.PI_CODING_AGENT_DIR = stateDir;
|
|
85
|
+
|
|
86
|
+
// Suppress the upstream runtime's version check and changelog display.
|
|
87
|
+
// Omegon has its own /update command and version-check extension —
|
|
88
|
+
// the pi-coding-agent's built-in checks leak upstream version numbers.
|
|
89
|
+
process.env.PI_SKIP_VERSION_CHECK = "1";
|
|
85
90
|
migrateLegacyStatePath("auth.json");
|
|
86
91
|
migrateLegacyStatePath("settings.json");
|
|
87
92
|
migrateLegacyStatePath("sessions", "directory");
|
|
@@ -9071,6 +9071,40 @@ export declare const MODELS: {
|
|
|
9071
9071
|
contextWindow: number;
|
|
9072
9072
|
maxTokens: number;
|
|
9073
9073
|
};
|
|
9074
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9075
|
+
id: string;
|
|
9076
|
+
name: string;
|
|
9077
|
+
api: "openai-completions";
|
|
9078
|
+
provider: string;
|
|
9079
|
+
baseUrl: string;
|
|
9080
|
+
reasoning: true;
|
|
9081
|
+
input: ("image" | "text")[];
|
|
9082
|
+
cost: {
|
|
9083
|
+
input: number;
|
|
9084
|
+
output: number;
|
|
9085
|
+
cacheRead: number;
|
|
9086
|
+
cacheWrite: number;
|
|
9087
|
+
};
|
|
9088
|
+
contextWindow: number;
|
|
9089
|
+
maxTokens: number;
|
|
9090
|
+
};
|
|
9091
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9092
|
+
id: string;
|
|
9093
|
+
name: string;
|
|
9094
|
+
api: "openai-completions";
|
|
9095
|
+
provider: string;
|
|
9096
|
+
baseUrl: string;
|
|
9097
|
+
reasoning: true;
|
|
9098
|
+
input: ("image" | "text")[];
|
|
9099
|
+
cost: {
|
|
9100
|
+
input: number;
|
|
9101
|
+
output: number;
|
|
9102
|
+
cacheRead: number;
|
|
9103
|
+
cacheWrite: number;
|
|
9104
|
+
};
|
|
9105
|
+
contextWindow: number;
|
|
9106
|
+
maxTokens: number;
|
|
9107
|
+
};
|
|
9074
9108
|
readonly "openai/gpt-5.4-pro": {
|
|
9075
9109
|
id: string;
|
|
9076
9110
|
name: string;
|
|
@@ -12439,6 +12473,40 @@ export declare const MODELS: {
|
|
|
12439
12473
|
contextWindow: number;
|
|
12440
12474
|
maxTokens: number;
|
|
12441
12475
|
};
|
|
12476
|
+
readonly "openai/gpt-5.4-mini": {
|
|
12477
|
+
id: string;
|
|
12478
|
+
name: string;
|
|
12479
|
+
api: "anthropic-messages";
|
|
12480
|
+
provider: string;
|
|
12481
|
+
baseUrl: string;
|
|
12482
|
+
reasoning: true;
|
|
12483
|
+
input: ("image" | "text")[];
|
|
12484
|
+
cost: {
|
|
12485
|
+
input: number;
|
|
12486
|
+
output: number;
|
|
12487
|
+
cacheRead: number;
|
|
12488
|
+
cacheWrite: number;
|
|
12489
|
+
};
|
|
12490
|
+
contextWindow: number;
|
|
12491
|
+
maxTokens: number;
|
|
12492
|
+
};
|
|
12493
|
+
readonly "openai/gpt-5.4-nano": {
|
|
12494
|
+
id: string;
|
|
12495
|
+
name: string;
|
|
12496
|
+
api: "anthropic-messages";
|
|
12497
|
+
provider: string;
|
|
12498
|
+
baseUrl: string;
|
|
12499
|
+
reasoning: true;
|
|
12500
|
+
input: ("image" | "text")[];
|
|
12501
|
+
cost: {
|
|
12502
|
+
input: number;
|
|
12503
|
+
output: number;
|
|
12504
|
+
cacheRead: number;
|
|
12505
|
+
cacheWrite: number;
|
|
12506
|
+
};
|
|
12507
|
+
contextWindow: number;
|
|
12508
|
+
maxTokens: number;
|
|
12509
|
+
};
|
|
12442
12510
|
readonly "openai/gpt-5.4-pro": {
|
|
12443
12511
|
id: string;
|
|
12444
12512
|
name: string;
|