lcclaude 1.0.0 → 1.0.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/.env.example
CHANGED
|
@@ -1,18 +1,15 @@
|
|
|
1
|
-
# Default example:
|
|
2
|
-
ANTHROPIC_AUTH_TOKEN=
|
|
3
|
-
ANTHROPIC_BASE_URL=
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
1
|
+
# Default example: local Qwen / OpenAI-compatible backend via LiteLLM
|
|
2
|
+
ANTHROPIC_AUTH_TOKEN=dummy
|
|
3
|
+
ANTHROPIC_BASE_URL=http://127.0.0.1:4000/anthropic
|
|
4
|
+
ANTHROPIC_MODEL=qwen-local
|
|
5
|
+
ANTHROPIC_DEFAULT_SONNET_MODEL=qwen-local
|
|
6
|
+
ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen-local
|
|
7
|
+
ANTHROPIC_DEFAULT_OPUS_MODEL=qwen-local
|
|
8
|
+
CLAUDE_CODE_MAX_OUTPUT_TOKENS=4096
|
|
9
|
+
CLAUDE_CODE_MAX_CONTEXT_TOKENS=65536
|
|
10
|
+
CLAUDE_CODE_TEMPERATURE=0
|
|
8
11
|
|
|
9
|
-
#
|
|
10
|
-
# ANTHROPIC_AUTH_TOKEN=dummy
|
|
11
|
-
# ANTHROPIC_BASE_URL=http://127.0.0.1:4000/anthropic
|
|
12
|
-
# ANTHROPIC_MODEL=qwen-local
|
|
13
|
-
# ANTHROPIC_DEFAULT_SONNET_MODEL=qwen-local
|
|
14
|
-
# ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen-local
|
|
15
|
-
# ANTHROPIC_DEFAULT_OPUS_MODEL=qwen-local
|
|
12
|
+
# If your OpenAI-compatible backend requires extra auth headers, you can add:
|
|
16
13
|
# ANTHROPIC_CUSTOM_HEADERS=Authorization: Bearer your-token
|
|
17
14
|
|
|
18
15
|
API_TIMEOUT_MS=3000000
|
|
@@ -201,6 +201,17 @@ import {
|
|
|
201
201
|
is529Error,
|
|
202
202
|
withRetry
|
|
203
203
|
} from "./withRetry.js";
|
|
204
|
+
function getDefaultTemperature() {
|
|
205
|
+
const raw = process.env.CLAUDE_CODE_TEMPERATURE;
|
|
206
|
+
if (raw === undefined) {
|
|
207
|
+
return 1;
|
|
208
|
+
}
|
|
209
|
+
const parsed = Number(raw);
|
|
210
|
+
if (!Number.isFinite(parsed)) {
|
|
211
|
+
return 1;
|
|
212
|
+
}
|
|
213
|
+
return Math.min(1, Math.max(0, parsed));
|
|
214
|
+
}
|
|
204
215
|
export function getExtraBodyParams(betaHeaders) {
|
|
205
216
|
const extraBodyStr = process.env.CLAUDE_CODE_EXTRA_BODY;
|
|
206
217
|
let result = {};
|
|
@@ -861,7 +872,7 @@ ${deferredToolList}
|
|
|
861
872
|
betasParams.push(cacheEditingBetaHeader);
|
|
862
873
|
logForDebugging("Cache editing beta header enabled for cached microcompact");
|
|
863
874
|
}
|
|
864
|
-
const temperature = !hasThinking ? options.temperatureOverride ??
|
|
875
|
+
const temperature = !hasThinking ? options.temperatureOverride ?? getDefaultTemperature() : undefined;
|
|
865
876
|
lastRequestBetas = betasParams;
|
|
866
877
|
return {
|
|
867
878
|
model: normalizeModelStringForAPI(options.model),
|
|
@@ -897,7 +908,7 @@ ${deferredToolList}
|
|
|
897
908
|
logAPIQuery({
|
|
898
909
|
model: options.model,
|
|
899
910
|
messagesLength: logMessagesLength,
|
|
900
|
-
temperature: options.temperatureOverride ??
|
|
911
|
+
temperature: options.temperatureOverride ?? getDefaultTemperature(),
|
|
901
912
|
betas: logBetas,
|
|
902
913
|
permissionMode: permissionContext.mode,
|
|
903
914
|
querySource: options.querySource,
|
package/dist/src/utils/config.js
CHANGED
|
@@ -62,7 +62,20 @@ function createDefaultGlobalConfig() {
|
|
|
62
62
|
approved: [],
|
|
63
63
|
rejected: []
|
|
64
64
|
},
|
|
65
|
-
env: {
|
|
65
|
+
env: {
|
|
66
|
+
ANTHROPIC_AUTH_TOKEN: "dummy",
|
|
67
|
+
ANTHROPIC_BASE_URL: "http://127.0.0.1:4000/anthropic",
|
|
68
|
+
ANTHROPIC_MODEL: "qwen-local",
|
|
69
|
+
ANTHROPIC_DEFAULT_SONNET_MODEL: "qwen-local",
|
|
70
|
+
ANTHROPIC_DEFAULT_HAIKU_MODEL: "qwen-local",
|
|
71
|
+
ANTHROPIC_DEFAULT_OPUS_MODEL: "qwen-local",
|
|
72
|
+
CLAUDE_CODE_MAX_OUTPUT_TOKENS: "4096",
|
|
73
|
+
CLAUDE_CODE_MAX_CONTEXT_TOKENS: "65536",
|
|
74
|
+
CLAUDE_CODE_TEMPERATURE: "0",
|
|
75
|
+
CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC: "1",
|
|
76
|
+
CLAUDE_CODE_DISABLE_EXPERIMENTAL_BETAS: "1",
|
|
77
|
+
DISABLE_TELEMETRY: "1"
|
|
78
|
+
},
|
|
66
79
|
tipsHistory: {},
|
|
67
80
|
memoryUsageCount: 0,
|
|
68
81
|
promptQueueUseCount: 0,
|
|
@@ -26,7 +26,7 @@ export function modelSupports1M(model) {
|
|
|
26
26
|
return canonical.includes("claude-sonnet-4") || canonical.includes("opus-4-6");
|
|
27
27
|
}
|
|
28
28
|
export function getContextWindowForModel(model, betas) {
|
|
29
|
-
if (process.env.
|
|
29
|
+
if (process.env.CLAUDE_CODE_MAX_CONTEXT_TOKENS) {
|
|
30
30
|
const override = parseInt(process.env.CLAUDE_CODE_MAX_CONTEXT_TOKENS, 10);
|
|
31
31
|
if (!isNaN(override) && override > 0) {
|
|
32
32
|
return override;
|
|
@@ -85,11 +85,13 @@ export const SAFE_ENV_VARS = new Set([
|
|
|
85
85
|
"CLAUDE_CODE_ENABLE_TELEMETRY",
|
|
86
86
|
"CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS",
|
|
87
87
|
"CLAUDE_CODE_IDE_SKIP_AUTO_INSTALL",
|
|
88
|
+
"CLAUDE_CODE_MAX_CONTEXT_TOKENS",
|
|
88
89
|
"CLAUDE_CODE_MAX_OUTPUT_TOKENS",
|
|
89
90
|
"CLAUDE_CODE_SKIP_BEDROCK_AUTH",
|
|
90
91
|
"CLAUDE_CODE_SKIP_FOUNDRY_AUTH",
|
|
91
92
|
"CLAUDE_CODE_SKIP_VERTEX_AUTH",
|
|
92
93
|
"CLAUDE_CODE_SUBAGENT_MODEL",
|
|
94
|
+
"CLAUDE_CODE_TEMPERATURE",
|
|
93
95
|
"CLAUDE_CODE_USE_BEDROCK",
|
|
94
96
|
"CLAUDE_CODE_USE_FOUNDRY",
|
|
95
97
|
"CLAUDE_CODE_USE_VERTEX",
|
package/litellm.yaml
CHANGED
|
@@ -1,9 +1,11 @@
|
|
|
1
1
|
model_list:
|
|
2
2
|
- model_name: qwen-local
|
|
3
3
|
litellm_params:
|
|
4
|
-
model: openai/
|
|
5
|
-
api_base: http://
|
|
4
|
+
model: openai/Qwen3-Next-80B-A3B-Instruct-FP8
|
|
5
|
+
api_base: http://xinf.lmisp.devops.zgcbank:13002/v1
|
|
6
6
|
api_key: dummy
|
|
7
|
+
temperature: 0
|
|
8
|
+
max_tokens: 4096
|
|
7
9
|
|
|
8
10
|
general_settings:
|
|
9
11
|
master_key: dummy
|