@juspay/neurolink 9.50.0 → 9.50.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/dist/browser/neurolink.min.js +292 -292
- package/dist/context/contextCompactor.js +2 -2
- package/dist/context/stages/slidingWindowTruncator.d.ts +1 -1
- package/dist/context/stages/slidingWindowTruncator.js +3 -3
- package/dist/core/modules/Utilities.d.ts +5 -0
- package/dist/core/modules/Utilities.js +29 -18
- package/dist/lib/context/contextCompactor.js +2 -2
- package/dist/lib/context/stages/slidingWindowTruncator.d.ts +1 -1
- package/dist/lib/context/stages/slidingWindowTruncator.js +3 -3
- package/dist/lib/core/modules/Utilities.d.ts +5 -0
- package/dist/lib/core/modules/Utilities.js +29 -18
- package/dist/lib/mcp/externalServerManager.d.ts +5 -0
- package/dist/lib/mcp/externalServerManager.js +24 -2
- package/dist/lib/neurolink.js +37 -3
- package/dist/lib/providers/litellm.js +2 -2
- package/dist/lib/proxy/proxyTracer.d.ts +14 -0
- package/dist/lib/proxy/proxyTracer.js +43 -0
- package/dist/lib/server/routes/claudeProxyRoutes.js +112 -33
- package/dist/lib/services/server/ai/observability/instrumentation.js +39 -1
- package/dist/lib/types/externalMcp.d.ts +7 -0
- package/dist/mcp/externalServerManager.d.ts +5 -0
- package/dist/mcp/externalServerManager.js +24 -2
- package/dist/neurolink.js +37 -3
- package/dist/providers/litellm.js +2 -2
- package/dist/proxy/proxyTracer.d.ts +14 -0
- package/dist/proxy/proxyTracer.js +43 -0
- package/dist/server/routes/claudeProxyRoutes.js +112 -33
- package/dist/services/server/ai/observability/instrumentation.js +39 -1
- package/dist/types/externalMcp.d.ts +7 -0
- package/package.json +1 -1
|
@@ -23,7 +23,7 @@ const DEFAULT_CONFIG = {
|
|
|
23
23
|
enableSummarize: true,
|
|
24
24
|
enableTruncate: true,
|
|
25
25
|
pruneProtectTokens: 40_000,
|
|
26
|
-
pruneMinimumSavings:
|
|
26
|
+
pruneMinimumSavings: 500,
|
|
27
27
|
pruneProtectedTools: ["skill"],
|
|
28
28
|
summarizationProvider: "vertex",
|
|
29
29
|
summarizationModel: "gemini-2.5-flash",
|
|
@@ -151,7 +151,7 @@ export class ContextCompactor {
|
|
|
151
151
|
targetTokens: targetTokens,
|
|
152
152
|
provider: provider,
|
|
153
153
|
adaptiveBuffer: 0.15,
|
|
154
|
-
maxIterations:
|
|
154
|
+
maxIterations: 6,
|
|
155
155
|
});
|
|
156
156
|
if (truncResult.truncated) {
|
|
157
157
|
currentMessages = truncResult.messages;
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
*
|
|
8
8
|
* Features:
|
|
9
9
|
* - Adaptive truncation (PERF-001): calculates fraction from actual overage
|
|
10
|
-
* instead of fixed 50%, with iterative refinement up to
|
|
10
|
+
* instead of fixed 50%, with iterative refinement up to 6 passes.
|
|
11
11
|
* - Small conversation handling (BUG-005): for <= 4 messages, truncates
|
|
12
12
|
* message content proportionally instead of returning no-op.
|
|
13
13
|
*/
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
*
|
|
8
8
|
* Features:
|
|
9
9
|
* - Adaptive truncation (PERF-001): calculates fraction from actual overage
|
|
10
|
-
* instead of fixed 50%, with iterative refinement up to
|
|
10
|
+
* instead of fixed 50%, with iterative refinement up to 6 passes.
|
|
11
11
|
* - Small conversation handling (BUG-005): for <= 4 messages, truncates
|
|
12
12
|
* message content proportionally instead of returning no-op.
|
|
13
13
|
*/
|
|
@@ -153,8 +153,8 @@ export function truncateWithSlidingWindow(messages, config) {
|
|
|
153
153
|
messagesRemoved: evenRemoveCount,
|
|
154
154
|
};
|
|
155
155
|
}
|
|
156
|
-
// Not enough -- increase fraction by
|
|
157
|
-
currentFraction = Math.min(0.95, currentFraction + 0.
|
|
156
|
+
// Not enough -- increase fraction by 10% for finer-grained escalation
|
|
157
|
+
currentFraction = Math.min(0.95, currentFraction + 0.1);
|
|
158
158
|
continue;
|
|
159
159
|
}
|
|
160
160
|
// No token targets -- single-pass with calculated fraction
|
|
@@ -57,6 +57,11 @@ export declare class Utilities {
|
|
|
57
57
|
* Supports number or string formats (e.g., '30s', '2m', '1h')
|
|
58
58
|
*/
|
|
59
59
|
getTimeout(options: TextGenerationOptions | StreamOptions): number;
|
|
60
|
+
/**
|
|
61
|
+
* Get timeout scaled by estimated input token count.
|
|
62
|
+
* For large contexts (>100K tokens), increase timeout proportionally.
|
|
63
|
+
*/
|
|
64
|
+
getContextAwareTimeout(options: TextGenerationOptions | StreamOptions, estimatedTokens?: number): number;
|
|
60
65
|
/**
|
|
61
66
|
* Check if a schema is a Zod schema
|
|
62
67
|
*/
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
import { z } from "zod";
|
|
20
20
|
import { logger } from "../../utils/logger.js";
|
|
21
21
|
import { getSafeMaxTokens } from "../../utils/tokenLimits.js";
|
|
22
|
-
import { TimeoutError } from "../../utils/timeout.js";
|
|
22
|
+
import { TimeoutError, getDefaultTimeout, parseTimeout, } from "../../utils/timeout.js";
|
|
23
23
|
import { validateStreamOptions as validateStreamOpts, validateTextGenerationOptions, ValidationError, createValidationSummary, } from "../../utils/parameterValidation.js";
|
|
24
24
|
import { STEP_LIMITS } from "../constants.js";
|
|
25
25
|
/**
|
|
@@ -157,25 +157,36 @@ export class Utilities {
|
|
|
157
157
|
* Supports number or string formats (e.g., '30s', '2m', '1h')
|
|
158
158
|
*/
|
|
159
159
|
getTimeout(options) {
|
|
160
|
-
|
|
161
|
-
|
|
162
|
-
|
|
163
|
-
|
|
164
|
-
|
|
165
|
-
|
|
166
|
-
// Parse string timeout (e.g., '30s', '2m', '1h')
|
|
167
|
-
const timeoutStr = options.timeout.toLowerCase();
|
|
168
|
-
const value = parseInt(timeoutStr);
|
|
169
|
-
if (timeoutStr.includes("h")) {
|
|
170
|
-
return value * 60 * 60 * 1000;
|
|
171
|
-
}
|
|
172
|
-
else if (timeoutStr.includes("m")) {
|
|
173
|
-
return value * 60 * 1000;
|
|
160
|
+
// If caller specified a timeout, use it (supports number ms and string formats)
|
|
161
|
+
if (options.timeout !== undefined && options.timeout !== null) {
|
|
162
|
+
const parsed = parseTimeout(options.timeout);
|
|
163
|
+
if (parsed !== undefined) {
|
|
164
|
+
return parsed;
|
|
165
|
+
}
|
|
174
166
|
}
|
|
175
|
-
|
|
176
|
-
|
|
167
|
+
// Use per-provider default (e.g., vertex=60s, ollama=5m) instead of global 30s.
|
|
168
|
+
// Always use "generate" operation here — streaming operations have their own
|
|
169
|
+
// longer timeout (DEFAULT_TIMEOUTS.streaming = 2m) applied by the streaming
|
|
170
|
+
// infrastructure in BaseProvider.stream(). Both TextGenerationOptions and
|
|
171
|
+
// StreamOptions share the same `input` property, so there is no reliable
|
|
172
|
+
// discriminator to detect streaming at this level.
|
|
173
|
+
const providerDefault = parseTimeout(getDefaultTimeout(this.providerName, "generate"));
|
|
174
|
+
return providerDefault ?? this.defaultTimeout;
|
|
175
|
+
}
|
|
176
|
+
/**
|
|
177
|
+
* Get timeout scaled by estimated input token count.
|
|
178
|
+
* For large contexts (>100K tokens), increase timeout proportionally.
|
|
179
|
+
*/
|
|
180
|
+
getContextAwareTimeout(options, estimatedTokens) {
|
|
181
|
+
const baseTimeout = this.getTimeout(options);
|
|
182
|
+
if (!estimatedTokens || estimatedTokens <= 100_000) {
|
|
183
|
+
return baseTimeout;
|
|
177
184
|
}
|
|
178
|
-
|
|
185
|
+
// Scale: >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x)
|
|
186
|
+
// Use (estimatedTokens - 1) so exact multiples stay in the lower tier
|
|
187
|
+
// (e.g., 100_000 → 1x, 100_001 → 1.5x)
|
|
188
|
+
const scale = 1 + Math.floor((estimatedTokens - 1) / 100_000) * 0.5;
|
|
189
|
+
return Math.round(baseTimeout * Math.min(scale, 4));
|
|
179
190
|
}
|
|
180
191
|
/**
|
|
181
192
|
* Check if a schema is a Zod schema
|
|
@@ -23,7 +23,7 @@ const DEFAULT_CONFIG = {
|
|
|
23
23
|
enableSummarize: true,
|
|
24
24
|
enableTruncate: true,
|
|
25
25
|
pruneProtectTokens: 40_000,
|
|
26
|
-
pruneMinimumSavings:
|
|
26
|
+
pruneMinimumSavings: 500,
|
|
27
27
|
pruneProtectedTools: ["skill"],
|
|
28
28
|
summarizationProvider: "vertex",
|
|
29
29
|
summarizationModel: "gemini-2.5-flash",
|
|
@@ -151,7 +151,7 @@ export class ContextCompactor {
|
|
|
151
151
|
targetTokens: targetTokens,
|
|
152
152
|
provider: provider,
|
|
153
153
|
adaptiveBuffer: 0.15,
|
|
154
|
-
maxIterations:
|
|
154
|
+
maxIterations: 6,
|
|
155
155
|
});
|
|
156
156
|
if (truncResult.truncated) {
|
|
157
157
|
currentMessages = truncResult.messages;
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
*
|
|
8
8
|
* Features:
|
|
9
9
|
* - Adaptive truncation (PERF-001): calculates fraction from actual overage
|
|
10
|
-
* instead of fixed 50%, with iterative refinement up to
|
|
10
|
+
* instead of fixed 50%, with iterative refinement up to 6 passes.
|
|
11
11
|
* - Small conversation handling (BUG-005): for <= 4 messages, truncates
|
|
12
12
|
* message content proportionally instead of returning no-op.
|
|
13
13
|
*/
|
|
@@ -7,7 +7,7 @@
|
|
|
7
7
|
*
|
|
8
8
|
* Features:
|
|
9
9
|
* - Adaptive truncation (PERF-001): calculates fraction from actual overage
|
|
10
|
-
* instead of fixed 50%, with iterative refinement up to
|
|
10
|
+
* instead of fixed 50%, with iterative refinement up to 6 passes.
|
|
11
11
|
* - Small conversation handling (BUG-005): for <= 4 messages, truncates
|
|
12
12
|
* message content proportionally instead of returning no-op.
|
|
13
13
|
*/
|
|
@@ -153,8 +153,8 @@ export function truncateWithSlidingWindow(messages, config) {
|
|
|
153
153
|
messagesRemoved: evenRemoveCount,
|
|
154
154
|
};
|
|
155
155
|
}
|
|
156
|
-
// Not enough -- increase fraction by
|
|
157
|
-
currentFraction = Math.min(0.95, currentFraction + 0.
|
|
156
|
+
// Not enough -- increase fraction by 10% for finer-grained escalation
|
|
157
|
+
currentFraction = Math.min(0.95, currentFraction + 0.1);
|
|
158
158
|
continue;
|
|
159
159
|
}
|
|
160
160
|
// No token targets -- single-pass with calculated fraction
|
|
@@ -57,6 +57,11 @@ export declare class Utilities {
|
|
|
57
57
|
* Supports number or string formats (e.g., '30s', '2m', '1h')
|
|
58
58
|
*/
|
|
59
59
|
getTimeout(options: TextGenerationOptions | StreamOptions): number;
|
|
60
|
+
/**
|
|
61
|
+
* Get timeout scaled by estimated input token count.
|
|
62
|
+
* For large contexts (>100K tokens), increase timeout proportionally.
|
|
63
|
+
*/
|
|
64
|
+
getContextAwareTimeout(options: TextGenerationOptions | StreamOptions, estimatedTokens?: number): number;
|
|
60
65
|
/**
|
|
61
66
|
* Check if a schema is a Zod schema
|
|
62
67
|
*/
|
|
@@ -19,7 +19,7 @@
|
|
|
19
19
|
import { z } from "zod";
|
|
20
20
|
import { logger } from "../../utils/logger.js";
|
|
21
21
|
import { getSafeMaxTokens } from "../../utils/tokenLimits.js";
|
|
22
|
-
import { TimeoutError } from "../../utils/timeout.js";
|
|
22
|
+
import { TimeoutError, getDefaultTimeout, parseTimeout, } from "../../utils/timeout.js";
|
|
23
23
|
import { validateStreamOptions as validateStreamOpts, validateTextGenerationOptions, ValidationError, createValidationSummary, } from "../../utils/parameterValidation.js";
|
|
24
24
|
import { STEP_LIMITS } from "../constants.js";
|
|
25
25
|
/**
|
|
@@ -157,25 +157,36 @@ export class Utilities {
|
|
|
157
157
|
* Supports number or string formats (e.g., '30s', '2m', '1h')
|
|
158
158
|
*/
|
|
159
159
|
getTimeout(options) {
|
|
160
|
-
|
|
161
|
-
|
|
162
|
-
|
|
163
|
-
|
|
164
|
-
|
|
165
|
-
|
|
166
|
-
// Parse string timeout (e.g., '30s', '2m', '1h')
|
|
167
|
-
const timeoutStr = options.timeout.toLowerCase();
|
|
168
|
-
const value = parseInt(timeoutStr);
|
|
169
|
-
if (timeoutStr.includes("h")) {
|
|
170
|
-
return value * 60 * 60 * 1000;
|
|
171
|
-
}
|
|
172
|
-
else if (timeoutStr.includes("m")) {
|
|
173
|
-
return value * 60 * 1000;
|
|
160
|
+
// If caller specified a timeout, use it (supports number ms and string formats)
|
|
161
|
+
if (options.timeout !== undefined && options.timeout !== null) {
|
|
162
|
+
const parsed = parseTimeout(options.timeout);
|
|
163
|
+
if (parsed !== undefined) {
|
|
164
|
+
return parsed;
|
|
165
|
+
}
|
|
174
166
|
}
|
|
175
|
-
|
|
176
|
-
|
|
167
|
+
// Use per-provider default (e.g., vertex=60s, ollama=5m) instead of global 30s.
|
|
168
|
+
// Always use "generate" operation here — streaming operations have their own
|
|
169
|
+
// longer timeout (DEFAULT_TIMEOUTS.streaming = 2m) applied by the streaming
|
|
170
|
+
// infrastructure in BaseProvider.stream(). Both TextGenerationOptions and
|
|
171
|
+
// StreamOptions share the same `input` property, so there is no reliable
|
|
172
|
+
// discriminator to detect streaming at this level.
|
|
173
|
+
const providerDefault = parseTimeout(getDefaultTimeout(this.providerName, "generate"));
|
|
174
|
+
return providerDefault ?? this.defaultTimeout;
|
|
175
|
+
}
|
|
176
|
+
/**
|
|
177
|
+
* Get timeout scaled by estimated input token count.
|
|
178
|
+
* For large contexts (>100K tokens), increase timeout proportionally.
|
|
179
|
+
*/
|
|
180
|
+
getContextAwareTimeout(options, estimatedTokens) {
|
|
181
|
+
const baseTimeout = this.getTimeout(options);
|
|
182
|
+
if (!estimatedTokens || estimatedTokens <= 100_000) {
|
|
183
|
+
return baseTimeout;
|
|
177
184
|
}
|
|
178
|
-
|
|
185
|
+
// Scale: >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x)
|
|
186
|
+
// Use (estimatedTokens - 1) so exact multiples stay in the lower tier
|
|
187
|
+
// (e.g., 100_000 → 1x, 100_001 → 1.5x)
|
|
188
|
+
const scale = 1 + Math.floor((estimatedTokens - 1) / 100_000) * 0.5;
|
|
189
|
+
return Math.round(baseTimeout * Math.min(scale, 4));
|
|
179
190
|
}
|
|
180
191
|
/**
|
|
181
192
|
* Check if a schema is a Zod schema
|
|
@@ -36,6 +36,11 @@ export declare class ExternalServerManager extends EventEmitter {
|
|
|
36
36
|
* Get current HITL manager
|
|
37
37
|
*/
|
|
38
38
|
getHITLManager(): HITLManager | undefined;
|
|
39
|
+
/**
|
|
40
|
+
* Resolve the human-readable server name for an event payload.
|
|
41
|
+
* Falls back to serverId if the instance or config.name isn't available.
|
|
42
|
+
*/
|
|
43
|
+
getServerName(serverId: string): string;
|
|
39
44
|
/**
|
|
40
45
|
* Load MCP server configurations from .mcp-config.json file with parallel loading support
|
|
41
46
|
* Automatically registers servers found in the configuration
|
|
@@ -194,10 +194,16 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
194
194
|
this.toolDiscovery = new ToolDiscoveryService();
|
|
195
195
|
// Forward tool discovery events
|
|
196
196
|
this.toolDiscovery.on("toolRegistered", (event) => {
|
|
197
|
-
this.emit("toolDiscovered",
|
|
197
|
+
this.emit("toolDiscovered", {
|
|
198
|
+
...event,
|
|
199
|
+
serverName: this.getServerName(event.serverId),
|
|
200
|
+
});
|
|
198
201
|
});
|
|
199
202
|
this.toolDiscovery.on("toolUnregistered", (event) => {
|
|
200
|
-
this.emit("toolRemoved",
|
|
203
|
+
this.emit("toolRemoved", {
|
|
204
|
+
...event,
|
|
205
|
+
serverName: this.getServerName(event.serverId),
|
|
206
|
+
});
|
|
201
207
|
});
|
|
202
208
|
// Handle process cleanup
|
|
203
209
|
process.on("SIGINT", () => this.shutdown());
|
|
@@ -223,6 +229,14 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
223
229
|
getHITLManager() {
|
|
224
230
|
return this.hitlManager;
|
|
225
231
|
}
|
|
232
|
+
/**
|
|
233
|
+
* Resolve the human-readable server name for an event payload.
|
|
234
|
+
* Falls back to serverId if the instance or config.name isn't available.
|
|
235
|
+
*/
|
|
236
|
+
getServerName(serverId) {
|
|
237
|
+
const instance = this.servers.get(serverId);
|
|
238
|
+
return instance?.config?.name || serverId;
|
|
239
|
+
}
|
|
226
240
|
/**
|
|
227
241
|
* Load MCP server configurations from .mcp-config.json file with parallel loading support
|
|
228
242
|
* Automatically registers servers found in the configuration
|
|
@@ -712,6 +726,8 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
712
726
|
};
|
|
713
727
|
}
|
|
714
728
|
mcpLogger.info(`[ExternalServerManager] Removing server: ${serverId}`);
|
|
729
|
+
// Capture name before deletion removes the instance
|
|
730
|
+
const serverName = this.getServerName(serverId);
|
|
715
731
|
// Stop the server
|
|
716
732
|
await this.stopServer(serverId);
|
|
717
733
|
// Remove from registry
|
|
@@ -719,6 +735,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
719
735
|
// Emit event
|
|
720
736
|
this.emit("disconnected", {
|
|
721
737
|
serverId,
|
|
738
|
+
serverName,
|
|
722
739
|
reason: "Manually removed",
|
|
723
740
|
timestamp: new Date(),
|
|
724
741
|
});
|
|
@@ -816,6 +833,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
816
833
|
// Emit connected event
|
|
817
834
|
this.emit("connected", {
|
|
818
835
|
serverId,
|
|
836
|
+
serverName: this.getServerName(serverId),
|
|
819
837
|
toolCount: instance.toolsMap.size,
|
|
820
838
|
timestamp: new Date(),
|
|
821
839
|
});
|
|
@@ -921,6 +939,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
921
939
|
// Emit status change event
|
|
922
940
|
this.emit("statusChanged", {
|
|
923
941
|
serverId,
|
|
942
|
+
serverName: this.getServerName(serverId),
|
|
924
943
|
oldStatus,
|
|
925
944
|
newStatus,
|
|
926
945
|
timestamp: new Date(),
|
|
@@ -941,6 +960,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
941
960
|
// Emit failed event
|
|
942
961
|
this.emit("failed", {
|
|
943
962
|
serverId,
|
|
963
|
+
serverName: this.getServerName(serverId),
|
|
944
964
|
error: error.message,
|
|
945
965
|
timestamp: new Date(),
|
|
946
966
|
});
|
|
@@ -965,6 +985,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
965
985
|
// Emit disconnected event
|
|
966
986
|
this.emit("disconnected", {
|
|
967
987
|
serverId,
|
|
988
|
+
serverName: this.getServerName(serverId),
|
|
968
989
|
reason,
|
|
969
990
|
timestamp: new Date(),
|
|
970
991
|
});
|
|
@@ -1078,6 +1099,7 @@ export class ExternalServerManager extends EventEmitter {
|
|
|
1078
1099
|
// Emit health check event
|
|
1079
1100
|
this.emit("healthCheck", {
|
|
1080
1101
|
serverId,
|
|
1102
|
+
serverName: this.getServerName(serverId),
|
|
1081
1103
|
health,
|
|
1082
1104
|
timestamp: new Date(),
|
|
1083
1105
|
});
|
package/dist/lib/neurolink.js
CHANGED
|
@@ -50,7 +50,7 @@ import { createMemoryRetrievalTools } from "./memory/memoryRetrievalTools.js";
|
|
|
50
50
|
import { getMetricsAggregator, MetricsAggregator, } from "./observability/metricsAggregator.js";
|
|
51
51
|
import { SpanStatus, SpanType } from "./observability/types/spanTypes.js";
|
|
52
52
|
import { SpanSerializer } from "./observability/utils/spanSerializer.js";
|
|
53
|
-
import { flushOpenTelemetry, getLangfuseHealthStatus, initializeOpenTelemetry, isOpenTelemetryInitialized, setLangfuseContext, shutdownOpenTelemetry, } from "./services/server/ai/observability/instrumentation.js";
|
|
53
|
+
import { flushOpenTelemetry, getLangfuseHealthStatus, initializeOpenTelemetry, isOpenTelemetryInitialized, runWithCurrentLangfuseContext, setLangfuseContext, shutdownOpenTelemetry, } from "./services/server/ai/observability/instrumentation.js";
|
|
54
54
|
import { TaskManager } from "./tasks/taskManager.js";
|
|
55
55
|
import { createTaskTools } from "./tasks/tools/taskTools.js";
|
|
56
56
|
import { ATTR } from "./telemetry/attributes.js";
|
|
@@ -1129,7 +1129,10 @@ Current user's request: ${currentInput}`;
|
|
|
1129
1129
|
* Supports additional users with per-user prompt and maxWords overrides.
|
|
1130
1130
|
*/
|
|
1131
1131
|
storeMemoryInBackground(originalPrompt, responseContent, userId, additionalUsers) {
|
|
1132
|
-
setImmediate
|
|
1132
|
+
// Preserve AsyncLocalStorage context across setImmediate boundary so that
|
|
1133
|
+
// memory writes appear under the originating Langfuse trace instead of
|
|
1134
|
+
// becoming orphan spans.
|
|
1135
|
+
const wrappedMemoryWrite = runWithCurrentLangfuseContext(async () => {
|
|
1133
1136
|
try {
|
|
1134
1137
|
const client = this.ensureMemoryReady();
|
|
1135
1138
|
if (!client) {
|
|
@@ -1145,12 +1148,18 @@ Current user's request: ${currentInput}`;
|
|
|
1145
1148
|
: undefined;
|
|
1146
1149
|
writeOps.push(client.add(user.userId, content, addOptions));
|
|
1147
1150
|
}
|
|
1148
|
-
|
|
1151
|
+
// withTimeout races against Promise.all — if the timeout fires, the
|
|
1152
|
+
// await resolves with an error but the underlying client.add() calls
|
|
1153
|
+
// may still complete in the background. This is acceptable: the memory
|
|
1154
|
+
// client API (Mem0) doesn't support AbortSignal, and these are
|
|
1155
|
+
// fire-and-forget background writes where a stale completion is harmless.
|
|
1156
|
+
await withTimeout(Promise.all(writeOps), 30_000, new Error("Background memory write timed out after 30s"));
|
|
1149
1157
|
}
|
|
1150
1158
|
catch (error) {
|
|
1151
1159
|
logger.warn("Memory storage failed:", error);
|
|
1152
1160
|
}
|
|
1153
1161
|
});
|
|
1162
|
+
setImmediate(wrappedMemoryWrite);
|
|
1154
1163
|
}
|
|
1155
1164
|
/**
|
|
1156
1165
|
* Set up HITL event forwarding to main emitter
|
|
@@ -3723,6 +3732,21 @@ Current user's request: ${currentInput}`;
|
|
|
3723
3732
|
conversationMessageCount: conversationMessages.length,
|
|
3724
3733
|
shouldCompact: budgetResult.shouldCompact,
|
|
3725
3734
|
});
|
|
3735
|
+
// Scale timeout for large contexts if caller didn't set one explicitly.
|
|
3736
|
+
// Providers read options.timeout via getTimeout(), so setting it here
|
|
3737
|
+
// propagates to any downstream provider call.
|
|
3738
|
+
if (options.timeout === undefined &&
|
|
3739
|
+
budgetResult.estimatedInputTokens > 100_000) {
|
|
3740
|
+
// >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x) of 60s base
|
|
3741
|
+
const scale = 1 + Math.floor((budgetResult.estimatedInputTokens - 1) / 100_000) * 0.5;
|
|
3742
|
+
const scaledMs = Math.round(60_000 * Math.min(scale, 4));
|
|
3743
|
+
options.timeout = scaledMs;
|
|
3744
|
+
logger.info("[TokenBudget] Scaled timeout for large context", {
|
|
3745
|
+
requestId,
|
|
3746
|
+
estimatedTokens: budgetResult.estimatedInputTokens,
|
|
3747
|
+
scaledTimeoutMs: scaledMs,
|
|
3748
|
+
});
|
|
3749
|
+
}
|
|
3726
3750
|
const compactionSessionId = this.getCompactionSessionId(options);
|
|
3727
3751
|
const lastCompactionCount = this.lastCompactionMessageCount.get(compactionSessionId) ?? 0;
|
|
3728
3752
|
if (!budgetResult.shouldCompact ||
|
|
@@ -3798,6 +3822,8 @@ Current user's request: ${currentInput}`;
|
|
|
3798
3822
|
toolDefinitions: availableTools,
|
|
3799
3823
|
});
|
|
3800
3824
|
if (!finalBudget.withinBudget) {
|
|
3825
|
+
// Clear watermark so handleContextOverflow recovery can re-compact
|
|
3826
|
+
this.lastCompactionMessageCount.delete(compactionSessionId);
|
|
3801
3827
|
throw new ContextBudgetExceededError(`Context exceeds model budget after all compaction stages. ` +
|
|
3802
3828
|
`Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
|
|
3803
3829
|
`Budget: ${finalBudget.availableInputTokens} tokens. ` +
|
|
@@ -3993,6 +4019,8 @@ Current user's request: ${currentInput}`;
|
|
|
3993
4019
|
: undefined,
|
|
3994
4020
|
});
|
|
3995
4021
|
if (!finalBudget.withinBudget) {
|
|
4022
|
+
// Clear watermark so handleContextOverflow recovery can re-compact
|
|
4023
|
+
this.lastCompactionMessageCount.delete(dpgCompactionSessionId);
|
|
3996
4024
|
throw new ContextBudgetExceededError(`Context exceeds model budget after all compaction stages. ` +
|
|
3997
4025
|
`Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
|
|
3998
4026
|
`Budget: ${finalBudget.availableInputTokens} tokens.`, {
|
|
@@ -5016,6 +5044,8 @@ Current user's request: ${currentInput}`;
|
|
|
5016
5044
|
toolDefinitions: availableTools,
|
|
5017
5045
|
});
|
|
5018
5046
|
if (!finalBudget.withinBudget) {
|
|
5047
|
+
// Clear watermark so handleContextOverflow recovery can re-compact
|
|
5048
|
+
this.lastCompactionMessageCount.delete(streamCompactionSessionId);
|
|
5019
5049
|
throw new ContextBudgetExceededError(`Stream context exceeds model budget after all compaction stages. ` +
|
|
5020
5050
|
`Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
|
|
5021
5051
|
`Budget: ${finalBudget.availableInputTokens} tokens.`, {
|
|
@@ -7508,6 +7538,7 @@ Current user's request: ${currentInput}`;
|
|
|
7508
7538
|
// Emit server added event
|
|
7509
7539
|
this.emitter.emit("externalMCP:serverAdded", {
|
|
7510
7540
|
serverId,
|
|
7541
|
+
serverName: config.name || serverId,
|
|
7511
7542
|
config,
|
|
7512
7543
|
toolCount: result.metadata?.toolsDiscovered || 0,
|
|
7513
7544
|
timestamp: Date.now(),
|
|
@@ -7535,12 +7566,15 @@ Current user's request: ${currentInput}`;
|
|
|
7535
7566
|
this.invalidateToolCache(); // Invalidate cache when an external server is removed
|
|
7536
7567
|
try {
|
|
7537
7568
|
mcpLogger.info(`[NeuroLink] Removing external MCP server: ${serverId}`);
|
|
7569
|
+
// Capture the configured name before removal destroys the instance
|
|
7570
|
+
const serverName = this.externalServerManager.getServerName(serverId);
|
|
7538
7571
|
const result = await this.externalServerManager.removeServer(serverId);
|
|
7539
7572
|
if (result.success) {
|
|
7540
7573
|
mcpLogger.info(`[NeuroLink] External MCP server removed successfully: ${serverId}`);
|
|
7541
7574
|
// Emit server removed event
|
|
7542
7575
|
this.emitter.emit("externalMCP:serverRemoved", {
|
|
7543
7576
|
serverId,
|
|
7577
|
+
serverName,
|
|
7544
7578
|
timestamp: Date.now(),
|
|
7545
7579
|
});
|
|
7546
7580
|
}
|
|
@@ -356,8 +356,8 @@ export class LiteLLMProvider extends BaseProvider {
|
|
|
356
356
|
}
|
|
357
357
|
catch (streamError) {
|
|
358
358
|
if (NoOutputGeneratedError.isInstance(streamError)) {
|
|
359
|
-
logger.warn("LiteLLM: Stream produced no output (NoOutputGeneratedError)");
|
|
360
|
-
|
|
359
|
+
logger.warn("LiteLLM: Stream produced no output (NoOutputGeneratedError) — propagating to fallback chain");
|
|
360
|
+
throw streamError;
|
|
361
361
|
}
|
|
362
362
|
throw streamError;
|
|
363
363
|
}
|
|
@@ -90,6 +90,13 @@ declare class ProxyTracer {
|
|
|
90
90
|
* Sets span attributes and increments the substitution metric counter.
|
|
91
91
|
*/
|
|
92
92
|
setModelSubstitution(requestedModel: string, actualModel: string): void;
|
|
93
|
+
setFallbackInfo(info: {
|
|
94
|
+
triggered: boolean;
|
|
95
|
+
provider?: string;
|
|
96
|
+
model?: string;
|
|
97
|
+
attemptCount: number;
|
|
98
|
+
reason: string;
|
|
99
|
+
}): void;
|
|
93
100
|
/** Log the incoming client request body (redacted). */
|
|
94
101
|
logRequestBody(body: string): void;
|
|
95
102
|
/** Log the incoming client request headers (redacted). */
|
|
@@ -129,5 +136,12 @@ declare class ProxyTracer {
|
|
|
129
136
|
*/
|
|
130
137
|
getTraceHeaders(): Record<string, string>;
|
|
131
138
|
}
|
|
139
|
+
export declare function recordFallbackAttempt(attrs: {
|
|
140
|
+
provider: string;
|
|
141
|
+
model: string;
|
|
142
|
+
status: "success" | "failure";
|
|
143
|
+
errorMessage?: string;
|
|
144
|
+
durationMs: number;
|
|
145
|
+
}): void;
|
|
132
146
|
export { ProxyTracer };
|
|
133
147
|
export type { ProxyRequestContext, AccountSelectionContext, UpstreamAttemptContext, UsageContext, };
|
|
@@ -79,6 +79,18 @@ function getProxyMetrics() {
|
|
|
79
79
|
description: "Response body size in bytes received from upstream",
|
|
80
80
|
unit: "By",
|
|
81
81
|
}),
|
|
82
|
+
fallbackAttemptsTotal: meter.createCounter("proxy_fallback_attempts_total", {
|
|
83
|
+
description: "Total fallback provider attempts",
|
|
84
|
+
unit: "{attempt}",
|
|
85
|
+
}),
|
|
86
|
+
fallbackSuccessTotal: meter.createCounter("proxy_fallback_success_total", {
|
|
87
|
+
description: "Total successful fallback provider responses",
|
|
88
|
+
unit: "{success}",
|
|
89
|
+
}),
|
|
90
|
+
fallbackFailureTotal: meter.createCounter("proxy_fallback_failure_total", {
|
|
91
|
+
description: "Total failed fallback provider responses",
|
|
92
|
+
unit: "{failure}",
|
|
93
|
+
}),
|
|
82
94
|
};
|
|
83
95
|
_metrics = createdMetrics;
|
|
84
96
|
return createdMetrics;
|
|
@@ -396,6 +408,18 @@ class ProxyTracer {
|
|
|
396
408
|
actual_model: actualModel,
|
|
397
409
|
});
|
|
398
410
|
}
|
|
411
|
+
setFallbackInfo(info) {
|
|
412
|
+
if (!this.rootSpan) {
|
|
413
|
+
return;
|
|
414
|
+
}
|
|
415
|
+
this.rootSpan.setAttributes({
|
|
416
|
+
"proxy.fallback.triggered": info.triggered,
|
|
417
|
+
...(info.provider ? { "proxy.fallback.provider": info.provider } : {}),
|
|
418
|
+
...(info.model ? { "proxy.fallback.model": info.model } : {}),
|
|
419
|
+
"proxy.fallback.attempt_count": info.attemptCount,
|
|
420
|
+
"proxy.fallback.reason": info.reason,
|
|
421
|
+
});
|
|
422
|
+
}
|
|
399
423
|
// -------------------------------------------------------------------------
|
|
400
424
|
// Log payloads as span events
|
|
401
425
|
// -------------------------------------------------------------------------
|
|
@@ -641,5 +665,24 @@ class ProxyTracer {
|
|
|
641
665
|
return this.bridge.injectContext({}, trace.setSpan(context.active(), this.rootSpan));
|
|
642
666
|
}
|
|
643
667
|
}
|
|
668
|
+
export function recordFallbackAttempt(attrs) {
|
|
669
|
+
try {
|
|
670
|
+
const m = getProxyMetrics();
|
|
671
|
+
const labels = { provider: attrs.provider, model: attrs.model };
|
|
672
|
+
m.fallbackAttemptsTotal.add(1, labels);
|
|
673
|
+
if (attrs.status === "success") {
|
|
674
|
+
m.fallbackSuccessTotal.add(1, labels);
|
|
675
|
+
}
|
|
676
|
+
else {
|
|
677
|
+
m.fallbackFailureTotal.add(1, {
|
|
678
|
+
...labels,
|
|
679
|
+
error: attrs.errorMessage?.slice(0, 100) ?? "unknown",
|
|
680
|
+
});
|
|
681
|
+
}
|
|
682
|
+
}
|
|
683
|
+
catch {
|
|
684
|
+
// metrics are best-effort
|
|
685
|
+
}
|
|
686
|
+
}
|
|
644
687
|
export { ProxyTracer };
|
|
645
688
|
//# sourceMappingURL=proxyTracer.js.map
|