@rungate/llmrouter 0.1.0 → 0.1.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +4 -7
- package/dist/src/proxy/server.js +1 -1
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -44,6 +44,8 @@ X402_NETWORK=eip155:8453
|
|
|
44
44
|
|
|
45
45
|
`llm_router` remains local-only in this setup. OpenClaw still talks to the local proxy at `http://127.0.0.1:3000/v1`; the remote upstream is configured by environment.
|
|
46
46
|
|
|
47
|
+
If `INFERENCE_PROVIDER_BASE_URL` is unset, the router defaults to `https://api.rungate.ai`.
|
|
48
|
+
|
|
47
49
|
## Local OpenClaw Install
|
|
48
50
|
|
|
49
51
|
For local unpublished development, install from a packaged tarball:
|
|
@@ -63,6 +65,8 @@ openclaw plugins install /staging/rungate-llmrouter-0.1.0.tgz
|
|
|
63
65
|
|
|
64
66
|
## Environment
|
|
65
67
|
|
|
68
|
+
For local development, override the production default upstream:
|
|
69
|
+
|
|
66
70
|
```bash
|
|
67
71
|
LLM_ROUTER_HOST=127.0.0.1
|
|
68
72
|
LLM_ROUTER_PORT=3000
|
|
@@ -70,13 +74,6 @@ INFERENCE_PROVIDER_BASE_URL=http://127.0.0.1:8787
|
|
|
70
74
|
X402_NETWORK=eip155:84532
|
|
71
75
|
```
|
|
72
76
|
|
|
73
|
-
For production, point the upstream at Rungate and use Base mainnet:
|
|
74
|
-
|
|
75
|
-
```bash
|
|
76
|
-
INFERENCE_PROVIDER_BASE_URL=https://api.rungate.ai
|
|
77
|
-
X402_NETWORK=eip155:8453
|
|
78
|
-
```
|
|
79
|
-
|
|
80
77
|
## OpenClaw Models
|
|
81
78
|
|
|
82
79
|
- `llmrouter/auto`
|
package/dist/src/proxy/server.js
CHANGED
|
@@ -146,7 +146,7 @@ async function handleChat(req, res, upstreamBaseUrl, payFetch) {
|
|
|
146
146
|
export async function startProxyServer(options = {}) {
|
|
147
147
|
const host = options.host ?? process.env.LLM_ROUTER_HOST ?? '127.0.0.1';
|
|
148
148
|
const port = options.port ?? Number(process.env.LLM_ROUTER_PORT ?? 3000);
|
|
149
|
-
const upstreamBaseUrl = options.upstreamBaseUrl ?? process.env.INFERENCE_PROVIDER_BASE_URL ?? '
|
|
149
|
+
const upstreamBaseUrl = options.upstreamBaseUrl ?? process.env.INFERENCE_PROVIDER_BASE_URL ?? 'https://api.rungate.ai';
|
|
150
150
|
const payFetch = createPaymentFetch(fetch, process.env);
|
|
151
151
|
const server = createServer(async (req, res) => {
|
|
152
152
|
try {
|