lcclaude 1.0.0 → 1.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/.env.example CHANGED
@@ -1,18 +1,15 @@
1
- # Default example: Anthropic-compatible gateway (e.g. MiniMax / OpenRouter / proxy)
2
- ANTHROPIC_AUTH_TOKEN=your_token_here
3
- ANTHROPIC_BASE_URL=https://api.minimaxi.com/anthropic
4
- ANTHROPIC_DEFAULT_HAIKU_MODEL=MiniMax-M2.7-highspeed
5
- ANTHROPIC_DEFAULT_OPUS_MODEL=MiniMax-M2.7-highspeed
6
- ANTHROPIC_DEFAULT_SONNET_MODEL=MiniMax-M2.7-highspeed
7
- ANTHROPIC_MODEL=MiniMax-M2.7-highspeed
1
+ # Default example: local Qwen / OpenAI-compatible backend via LiteLLM
2
+ ANTHROPIC_AUTH_TOKEN=dummy
3
+ ANTHROPIC_BASE_URL=http://127.0.0.1:4000/anthropic
4
+ ANTHROPIC_MODEL=qwen-local
5
+ ANTHROPIC_DEFAULT_SONNET_MODEL=qwen-local
6
+ ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen-local
7
+ ANTHROPIC_DEFAULT_OPUS_MODEL=qwen-local
8
+ CLAUDE_CODE_MAX_OUTPUT_TOKENS=4096
9
+ CLAUDE_CODE_MAX_CONTEXT_TOKENS=65536
10
+ CLAUDE_CODE_TEMPERATURE=0
8
11
 
9
- # Local Qwen / OpenAI-compatible example via LiteLLM:
10
- # ANTHROPIC_AUTH_TOKEN=dummy
11
- # ANTHROPIC_BASE_URL=http://127.0.0.1:4000/anthropic
12
- # ANTHROPIC_MODEL=qwen-local
13
- # ANTHROPIC_DEFAULT_SONNET_MODEL=qwen-local
14
- # ANTHROPIC_DEFAULT_HAIKU_MODEL=qwen-local
15
- # ANTHROPIC_DEFAULT_OPUS_MODEL=qwen-local
12
+ # If your OpenAI-compatible backend requires extra auth headers, you can add:
16
13
  # ANTHROPIC_CUSTOM_HEADERS=Authorization: Bearer your-token
17
14
 
18
15
  API_TIMEOUT_MS=3000000
@@ -201,6 +201,17 @@ import {
201
201
  is529Error,
202
202
  withRetry
203
203
  } from "./withRetry.js";
204
+ function getDefaultTemperature() {
205
+ const raw = process.env.CLAUDE_CODE_TEMPERATURE;
206
+ if (raw === undefined) {
207
+ return 1;
208
+ }
209
+ const parsed = Number(raw);
210
+ if (!Number.isFinite(parsed)) {
211
+ return 1;
212
+ }
213
+ return Math.min(1, Math.max(0, parsed));
214
+ }
204
215
  export function getExtraBodyParams(betaHeaders) {
205
216
  const extraBodyStr = process.env.CLAUDE_CODE_EXTRA_BODY;
206
217
  let result = {};
@@ -861,7 +872,7 @@ ${deferredToolList}
861
872
  betasParams.push(cacheEditingBetaHeader);
862
873
  logForDebugging("Cache editing beta header enabled for cached microcompact");
863
874
  }
864
- const temperature = !hasThinking ? options.temperatureOverride ?? 1 : undefined;
875
+ const temperature = !hasThinking ? options.temperatureOverride ?? getDefaultTemperature() : undefined;
865
876
  lastRequestBetas = betasParams;
866
877
  return {
867
878
  model: normalizeModelStringForAPI(options.model),
@@ -897,7 +908,7 @@ ${deferredToolList}
897
908
  logAPIQuery({
898
909
  model: options.model,
899
910
  messagesLength: logMessagesLength,
900
- temperature: options.temperatureOverride ?? 1,
911
+ temperature: options.temperatureOverride ?? getDefaultTemperature(),
901
912
  betas: logBetas,
902
913
  permissionMode: permissionContext.mode,
903
914
  querySource: options.querySource,
@@ -62,7 +62,20 @@ function createDefaultGlobalConfig() {
62
62
  approved: [],
63
63
  rejected: []
64
64
  },
65
- env: {},
65
+ env: {
66
+ ANTHROPIC_AUTH_TOKEN: "dummy",
67
+ ANTHROPIC_BASE_URL: "http://127.0.0.1:4000/anthropic",
68
+ ANTHROPIC_MODEL: "qwen-local",
69
+ ANTHROPIC_DEFAULT_SONNET_MODEL: "qwen-local",
70
+ ANTHROPIC_DEFAULT_HAIKU_MODEL: "qwen-local",
71
+ ANTHROPIC_DEFAULT_OPUS_MODEL: "qwen-local",
72
+ CLAUDE_CODE_MAX_OUTPUT_TOKENS: "4096",
73
+ CLAUDE_CODE_MAX_CONTEXT_TOKENS: "65536",
74
+ CLAUDE_CODE_TEMPERATURE: "0",
75
+ CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC: "1",
76
+ CLAUDE_CODE_DISABLE_EXPERIMENTAL_BETAS: "1",
77
+ DISABLE_TELEMETRY: "1"
78
+ },
66
79
  tipsHistory: {},
67
80
  memoryUsageCount: 0,
68
81
  promptQueueUseCount: 0,
@@ -26,7 +26,7 @@ export function modelSupports1M(model) {
26
26
  return canonical.includes("claude-sonnet-4") || canonical.includes("opus-4-6");
27
27
  }
28
28
  export function getContextWindowForModel(model, betas) {
29
- if (process.env.USER_TYPE === "ant" && process.env.CLAUDE_CODE_MAX_CONTEXT_TOKENS) {
29
+ if (process.env.CLAUDE_CODE_MAX_CONTEXT_TOKENS) {
30
30
  const override = parseInt(process.env.CLAUDE_CODE_MAX_CONTEXT_TOKENS, 10);
31
31
  if (!isNaN(override) && override > 0) {
32
32
  return override;
@@ -85,11 +85,13 @@ export const SAFE_ENV_VARS = new Set([
85
85
  "CLAUDE_CODE_ENABLE_TELEMETRY",
86
86
  "CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS",
87
87
  "CLAUDE_CODE_IDE_SKIP_AUTO_INSTALL",
88
+ "CLAUDE_CODE_MAX_CONTEXT_TOKENS",
88
89
  "CLAUDE_CODE_MAX_OUTPUT_TOKENS",
89
90
  "CLAUDE_CODE_SKIP_BEDROCK_AUTH",
90
91
  "CLAUDE_CODE_SKIP_FOUNDRY_AUTH",
91
92
  "CLAUDE_CODE_SKIP_VERTEX_AUTH",
92
93
  "CLAUDE_CODE_SUBAGENT_MODEL",
94
+ "CLAUDE_CODE_TEMPERATURE",
93
95
  "CLAUDE_CODE_USE_BEDROCK",
94
96
  "CLAUDE_CODE_USE_FOUNDRY",
95
97
  "CLAUDE_CODE_USE_VERTEX",
package/litellm.yaml CHANGED
@@ -1,9 +1,11 @@
1
1
  model_list:
2
2
  - model_name: qwen-local
3
3
  litellm_params:
4
- model: openai/Qwen2.5-Coder-32B-Instruct
5
- api_base: http://127.0.0.1:8000/v1
4
+ model: openai/Qwen3-Next-80B-A3B-Instruct-FP8
5
+ api_base: http://xinf.lmisp.devops.zgcbank:13002/v1
6
6
  api_key: dummy
7
+ temperature: 0
8
+ max_tokens: 4096
7
9
 
8
10
  general_settings:
9
11
  master_key: dummy
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "lcclaude",
3
- "version": "1.0.0",
3
+ "version": "1.0.1",
4
4
  "private": false,
5
5
  "type": "module",
6
6
  "imports": {