@pushpalsdev/cli 1.0.69 → 1.0.70
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/pushpals-cli.js +9 -3
- package/monitor-ui/+not-found.html +1 -1
- package/monitor-ui/_expo/static/js/web/{entry-6fc2a4ed369a881b7b810e1a9df3ce7b.js → entry-d69cc703f72afa383e4108efd6e0f726.js} +1138 -1138
- package/monitor-ui/_expo/static/js/web/{index-071ae9ae3713008b86319979f7148137.js → index-6013f9ebc87a963a55bb9137af1a5a06.js} +6 -6
- package/monitor-ui/_sitemap.html +1 -1
- package/monitor-ui/index.html +1 -1
- package/monitor-ui/modal.html +1 -1
- package/package.json +1 -1
- package/runtime/configs/default.toml +1 -1
- package/runtime/configs/local.example.toml +4 -4
- package/runtime/sandbox/.pushpals-remotebuddy-fallback.js +19 -7
- package/runtime/sandbox/apps/workerpals/src/backends/openai_codex/openai_codex_executor.py +4 -3
- package/runtime/sandbox/apps/workerpals/src/backends/openai_codex/test_openai_codex_runtime_config.py +7 -1
- package/runtime/sandbox/apps/workerpals/src/execute_job.ts +2 -1
- package/runtime/sandbox/configs/default.toml +1 -1
- package/runtime/sandbox/configs/local.example.toml +4 -4
- package/runtime/sandbox/packages/shared/src/config.ts +14 -6
package/dist/pushpals-cli.js
CHANGED
|
@@ -456,6 +456,8 @@ var DEFAULT_REMOTEBUDDY_MEMORY_MAX_RECALL_ITEMS = 12;
|
|
|
456
456
|
var DEFAULT_REMOTEBUDDY_MEMORY_MAX_RECALL_CHARS = 2400;
|
|
457
457
|
var DEFAULT_REMOTEBUDDY_MEMORY_MAX_SUMMARY_CHARS = 420;
|
|
458
458
|
var DEFAULT_REMOTEBUDDY_MEMORY_RETENTION_DAYS = 30;
|
|
459
|
+
var DEFAULT_OPENAI_CODEX_MODEL = "gpt-5.5";
|
|
460
|
+
var DEFAULT_OPENAI_CODEX_REASONING_EFFORT = "xhigh";
|
|
459
461
|
var cachedConfig = null;
|
|
460
462
|
var cachedConfigKey = "";
|
|
461
463
|
function firstNonEmpty(...values) {
|
|
@@ -486,7 +488,7 @@ function parseIntEnv(name) {
|
|
|
486
488
|
function parseTomlFile(path) {
|
|
487
489
|
if (!existsSync2(path))
|
|
488
490
|
return {};
|
|
489
|
-
const raw = readFileSync(path, "utf-8");
|
|
491
|
+
const raw = readFileSync(path, "utf-8").replace(/^\uFEFF/, "");
|
|
490
492
|
const parsed = Bun.TOML.parse(raw);
|
|
491
493
|
if (!parsed || typeof parsed !== "object" || Array.isArray(parsed))
|
|
492
494
|
return {};
|
|
@@ -651,10 +653,14 @@ function resolveLlmConfig(serviceNode, envPrefix, defaults, globalSessionId) {
|
|
|
651
653
|
const llmNode = getObject(serviceNode, "llm");
|
|
652
654
|
const backend = normalizeBackend(firstNonEmpty(process.env[`${envPrefix}_LLM_BACKEND`], asString(llmNode.backend, defaults.backend), defaults.backend));
|
|
653
655
|
const endpoint = firstNonEmpty(process.env[`${envPrefix}_LLM_ENDPOINT`], asString(llmNode.endpoint, defaults.endpoint), defaults.endpoint);
|
|
654
|
-
const
|
|
656
|
+
const envModel = firstNonEmpty(process.env[`${envPrefix}_LLM_MODEL`]);
|
|
657
|
+
const configuredFileModel = firstNonEmpty(asString(llmNode.model, ""));
|
|
658
|
+
const configuredModel = firstNonEmpty(envModel, configuredFileModel);
|
|
659
|
+
const modelFallback = backend === "openai_codex" ? DEFAULT_OPENAI_CODEX_MODEL : defaults.model;
|
|
660
|
+
const model = backend === "openai_codex" && !envModel && (!configuredFileModel || configuredFileModel === defaults.model) ? DEFAULT_OPENAI_CODEX_MODEL : firstNonEmpty(configuredModel, modelFallback) ?? modelFallback;
|
|
655
661
|
const sessionId = firstNonEmpty(process.env[`${envPrefix}_LLM_SESSION_ID`], asString(llmNode.session_id, defaults.sessionId), process.env.PUSHPALS_LLM_SESSION_ID, globalSessionId);
|
|
656
662
|
const apiKey = firstNonEmpty(process.env[`${envPrefix}_LLM_API_KEY`], defaultApiKeyForBackend(backend, endpoint));
|
|
657
|
-
const reasoningEffort = firstNonEmpty(process.env[`${envPrefix}_LLM_REASONING_EFFORT`], asString(llmNode.reasoning_effort, ""));
|
|
663
|
+
const reasoningEffort = firstNonEmpty(process.env[`${envPrefix}_LLM_REASONING_EFFORT`], asString(llmNode.reasoning_effort, ""), backend === "openai_codex" ? DEFAULT_OPENAI_CODEX_REASONING_EFFORT : "");
|
|
658
664
|
const codexAuthMode = firstNonEmpty(process.env[`${envPrefix}_LLM_CODEX_AUTH_MODE`], asString(llmNode.codex_auth_mode, ""));
|
|
659
665
|
const codexBin = firstNonEmpty(process.env[`${envPrefix}_LLM_CODEX_BIN`], asString(llmNode.codex_bin, ""));
|
|
660
666
|
const codexTimeoutMs = Math.max(1e4, asInt(parseIntEnv(`${envPrefix}_LLM_CODEX_TIMEOUT_MS`) ?? llmNode.codex_timeout_ms, 120000));
|
|
@@ -435,5 +435,5 @@ input::-webkit-search-cancel-button,input::-webkit-search-decoration,input::-web
|
|
|
435
435
|
@keyframes r-1pzkwqh{0%{transform:translateY(100%);}100%{transform:translateY(0%);}}
|
|
436
436
|
@keyframes r-imtty0{0%{opacity:0;}100%{opacity:1;}}
|
|
437
437
|
@keyframes r-q67da2{0%{transform:translateX(-100%);}100%{transform:translateX(400%);}}
|
|
438
|
-
@keyframes r-t2lo5v{0%{opacity:1;}100%{opacity:0;}}</style><script type="module">globalThis.__EXPO_ROUTER_HYDRATE__=true;</script><link rel="icon" href="/favicon.ico" /></head><body><div id="root"><div class="css-g5y9jx r-13awgt0"></div></div><script src="/_expo/static/js/web/entry-
|
|
438
|
+
@keyframes r-t2lo5v{0%{opacity:1;}100%{opacity:0;}}</style><script type="module">globalThis.__EXPO_ROUTER_HYDRATE__=true;</script><link rel="icon" href="/favicon.ico" /></head><body><div id="root"><div class="css-g5y9jx r-13awgt0"></div></div><script src="/_expo/static/js/web/entry-d69cc703f72afa383e4108efd6e0f726.js" defer></script>
|
|
439
439
|
</body></html>
|