@juspay/neurolink 9.50.1 → 9.50.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -23,7 +23,7 @@ const DEFAULT_CONFIG = {
23
23
  enableSummarize: true,
24
24
  enableTruncate: true,
25
25
  pruneProtectTokens: 40_000,
26
- pruneMinimumSavings: 20_000,
26
+ pruneMinimumSavings: 500,
27
27
  pruneProtectedTools: ["skill"],
28
28
  summarizationProvider: "vertex",
29
29
  summarizationModel: "gemini-2.5-flash",
@@ -151,7 +151,7 @@ export class ContextCompactor {
151
151
  targetTokens: targetTokens,
152
152
  provider: provider,
153
153
  adaptiveBuffer: 0.15,
154
- maxIterations: 3,
154
+ maxIterations: 6,
155
155
  });
156
156
  if (truncResult.truncated) {
157
157
  currentMessages = truncResult.messages;
@@ -7,7 +7,7 @@
7
7
  *
8
8
  * Features:
9
9
  * - Adaptive truncation (PERF-001): calculates fraction from actual overage
10
- * instead of fixed 50%, with iterative refinement up to 3 passes.
10
+ * instead of fixed 50%, with iterative refinement up to 6 passes.
11
11
  * - Small conversation handling (BUG-005): for <= 4 messages, truncates
12
12
  * message content proportionally instead of returning no-op.
13
13
  */
@@ -7,7 +7,7 @@
7
7
  *
8
8
  * Features:
9
9
  * - Adaptive truncation (PERF-001): calculates fraction from actual overage
10
- * instead of fixed 50%, with iterative refinement up to 3 passes.
10
+ * instead of fixed 50%, with iterative refinement up to 6 passes.
11
11
  * - Small conversation handling (BUG-005): for <= 4 messages, truncates
12
12
  * message content proportionally instead of returning no-op.
13
13
  */
@@ -153,8 +153,8 @@ export function truncateWithSlidingWindow(messages, config) {
153
153
  messagesRemoved: evenRemoveCount,
154
154
  };
155
155
  }
156
- // Not enough -- increase fraction by 25% for next iteration
157
- currentFraction = Math.min(0.95, currentFraction + 0.25);
156
+ // Not enough -- increase fraction by 10% for finer-grained escalation
157
+ currentFraction = Math.min(0.95, currentFraction + 0.1);
158
158
  continue;
159
159
  }
160
160
  // No token targets -- single-pass with calculated fraction
@@ -57,6 +57,11 @@ export declare class Utilities {
57
57
  * Supports number or string formats (e.g., '30s', '2m', '1h')
58
58
  */
59
59
  getTimeout(options: TextGenerationOptions | StreamOptions): number;
60
+ /**
61
+ * Get timeout scaled by estimated input token count.
62
+ * For large contexts (>100K tokens), increase timeout proportionally.
63
+ */
64
+ getContextAwareTimeout(options: TextGenerationOptions | StreamOptions, estimatedTokens?: number): number;
60
65
  /**
61
66
  * Check if a schema is a Zod schema
62
67
  */
@@ -19,7 +19,7 @@
19
19
  import { z } from "zod";
20
20
  import { logger } from "../../utils/logger.js";
21
21
  import { getSafeMaxTokens } from "../../utils/tokenLimits.js";
22
- import { TimeoutError } from "../../utils/timeout.js";
22
+ import { TimeoutError, getDefaultTimeout, parseTimeout, } from "../../utils/timeout.js";
23
23
  import { validateStreamOptions as validateStreamOpts, validateTextGenerationOptions, ValidationError, createValidationSummary, } from "../../utils/parameterValidation.js";
24
24
  import { STEP_LIMITS } from "../constants.js";
25
25
  /**
@@ -157,25 +157,36 @@ export class Utilities {
157
157
  * Supports number or string formats (e.g., '30s', '2m', '1h')
158
158
  */
159
159
  getTimeout(options) {
160
- if (!options.timeout) {
161
- return this.defaultTimeout;
162
- }
163
- if (typeof options.timeout === "number") {
164
- return options.timeout;
165
- }
166
- // Parse string timeout (e.g., '30s', '2m', '1h')
167
- const timeoutStr = options.timeout.toLowerCase();
168
- const value = parseInt(timeoutStr);
169
- if (timeoutStr.includes("h")) {
170
- return value * 60 * 60 * 1000;
171
- }
172
- else if (timeoutStr.includes("m")) {
173
- return value * 60 * 1000;
160
+ // If caller specified a timeout, use it (supports number ms and string formats)
161
+ if (options.timeout !== undefined && options.timeout !== null) {
162
+ const parsed = parseTimeout(options.timeout);
163
+ if (parsed !== undefined) {
164
+ return parsed;
165
+ }
174
166
  }
175
- else if (timeoutStr.includes("s")) {
176
- return value * 1000;
167
+ // Use per-provider default (e.g., vertex=60s, ollama=5m) instead of global 30s.
168
+ // Always use "generate" operation here — streaming operations have their own
169
+ // longer timeout (DEFAULT_TIMEOUTS.streaming = 2m) applied by the streaming
170
+ // infrastructure in BaseProvider.stream(). Both TextGenerationOptions and
171
+ // StreamOptions share the same `input` property, so there is no reliable
172
+ // discriminator to detect streaming at this level.
173
+ const providerDefault = parseTimeout(getDefaultTimeout(this.providerName, "generate"));
174
+ return providerDefault ?? this.defaultTimeout;
175
+ }
176
+ /**
177
+ * Get timeout scaled by estimated input token count.
178
+ * For large contexts (>100K tokens), increase timeout proportionally.
179
+ */
180
+ getContextAwareTimeout(options, estimatedTokens) {
181
+ const baseTimeout = this.getTimeout(options);
182
+ if (!estimatedTokens || estimatedTokens <= 100_000) {
183
+ return baseTimeout;
177
184
  }
178
- return this.defaultTimeout;
185
+ // Scale: >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x)
186
+ // Use (estimatedTokens - 1) so exact multiples stay in the lower tier
187
+ // (e.g., 100_000 → 1x, 100_001 → 1.5x)
188
+ const scale = 1 + Math.floor((estimatedTokens - 1) / 100_000) * 0.5;
189
+ return Math.round(baseTimeout * Math.min(scale, 4));
179
190
  }
180
191
  /**
181
192
  * Check if a schema is a Zod schema
@@ -23,7 +23,7 @@ const DEFAULT_CONFIG = {
23
23
  enableSummarize: true,
24
24
  enableTruncate: true,
25
25
  pruneProtectTokens: 40_000,
26
- pruneMinimumSavings: 20_000,
26
+ pruneMinimumSavings: 500,
27
27
  pruneProtectedTools: ["skill"],
28
28
  summarizationProvider: "vertex",
29
29
  summarizationModel: "gemini-2.5-flash",
@@ -151,7 +151,7 @@ export class ContextCompactor {
151
151
  targetTokens: targetTokens,
152
152
  provider: provider,
153
153
  adaptiveBuffer: 0.15,
154
- maxIterations: 3,
154
+ maxIterations: 6,
155
155
  });
156
156
  if (truncResult.truncated) {
157
157
  currentMessages = truncResult.messages;
@@ -7,7 +7,7 @@
7
7
  *
8
8
  * Features:
9
9
  * - Adaptive truncation (PERF-001): calculates fraction from actual overage
10
- * instead of fixed 50%, with iterative refinement up to 3 passes.
10
+ * instead of fixed 50%, with iterative refinement up to 6 passes.
11
11
  * - Small conversation handling (BUG-005): for <= 4 messages, truncates
12
12
  * message content proportionally instead of returning no-op.
13
13
  */
@@ -7,7 +7,7 @@
7
7
  *
8
8
  * Features:
9
9
  * - Adaptive truncation (PERF-001): calculates fraction from actual overage
10
- * instead of fixed 50%, with iterative refinement up to 3 passes.
10
+ * instead of fixed 50%, with iterative refinement up to 6 passes.
11
11
  * - Small conversation handling (BUG-005): for <= 4 messages, truncates
12
12
  * message content proportionally instead of returning no-op.
13
13
  */
@@ -153,8 +153,8 @@ export function truncateWithSlidingWindow(messages, config) {
153
153
  messagesRemoved: evenRemoveCount,
154
154
  };
155
155
  }
156
- // Not enough -- increase fraction by 25% for next iteration
157
- currentFraction = Math.min(0.95, currentFraction + 0.25);
156
+ // Not enough -- increase fraction by 10% for finer-grained escalation
157
+ currentFraction = Math.min(0.95, currentFraction + 0.1);
158
158
  continue;
159
159
  }
160
160
  // No token targets -- single-pass with calculated fraction
@@ -57,6 +57,11 @@ export declare class Utilities {
57
57
  * Supports number or string formats (e.g., '30s', '2m', '1h')
58
58
  */
59
59
  getTimeout(options: TextGenerationOptions | StreamOptions): number;
60
+ /**
61
+ * Get timeout scaled by estimated input token count.
62
+ * For large contexts (>100K tokens), increase timeout proportionally.
63
+ */
64
+ getContextAwareTimeout(options: TextGenerationOptions | StreamOptions, estimatedTokens?: number): number;
60
65
  /**
61
66
  * Check if a schema is a Zod schema
62
67
  */
@@ -19,7 +19,7 @@
19
19
  import { z } from "zod";
20
20
  import { logger } from "../../utils/logger.js";
21
21
  import { getSafeMaxTokens } from "../../utils/tokenLimits.js";
22
- import { TimeoutError } from "../../utils/timeout.js";
22
+ import { TimeoutError, getDefaultTimeout, parseTimeout, } from "../../utils/timeout.js";
23
23
  import { validateStreamOptions as validateStreamOpts, validateTextGenerationOptions, ValidationError, createValidationSummary, } from "../../utils/parameterValidation.js";
24
24
  import { STEP_LIMITS } from "../constants.js";
25
25
  /**
@@ -157,25 +157,36 @@ export class Utilities {
157
157
  * Supports number or string formats (e.g., '30s', '2m', '1h')
158
158
  */
159
159
  getTimeout(options) {
160
- if (!options.timeout) {
161
- return this.defaultTimeout;
162
- }
163
- if (typeof options.timeout === "number") {
164
- return options.timeout;
165
- }
166
- // Parse string timeout (e.g., '30s', '2m', '1h')
167
- const timeoutStr = options.timeout.toLowerCase();
168
- const value = parseInt(timeoutStr);
169
- if (timeoutStr.includes("h")) {
170
- return value * 60 * 60 * 1000;
171
- }
172
- else if (timeoutStr.includes("m")) {
173
- return value * 60 * 1000;
160
+ // If caller specified a timeout, use it (supports number ms and string formats)
161
+ if (options.timeout !== undefined && options.timeout !== null) {
162
+ const parsed = parseTimeout(options.timeout);
163
+ if (parsed !== undefined) {
164
+ return parsed;
165
+ }
174
166
  }
175
- else if (timeoutStr.includes("s")) {
176
- return value * 1000;
167
+ // Use per-provider default (e.g., vertex=60s, ollama=5m) instead of global 30s.
168
+ // Always use "generate" operation here — streaming operations have their own
169
+ // longer timeout (DEFAULT_TIMEOUTS.streaming = 2m) applied by the streaming
170
+ // infrastructure in BaseProvider.stream(). Both TextGenerationOptions and
171
+ // StreamOptions share the same `input` property, so there is no reliable
172
+ // discriminator to detect streaming at this level.
173
+ const providerDefault = parseTimeout(getDefaultTimeout(this.providerName, "generate"));
174
+ return providerDefault ?? this.defaultTimeout;
175
+ }
176
+ /**
177
+ * Get timeout scaled by estimated input token count.
178
+ * For large contexts (>100K tokens), increase timeout proportionally.
179
+ */
180
+ getContextAwareTimeout(options, estimatedTokens) {
181
+ const baseTimeout = this.getTimeout(options);
182
+ if (!estimatedTokens || estimatedTokens <= 100_000) {
183
+ return baseTimeout;
177
184
  }
178
- return this.defaultTimeout;
185
+ // Scale: >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x)
186
+ // Use (estimatedTokens - 1) so exact multiples stay in the lower tier
187
+ // (e.g., 100_000 → 1x, 100_001 → 1.5x)
188
+ const scale = 1 + Math.floor((estimatedTokens - 1) / 100_000) * 0.5;
189
+ return Math.round(baseTimeout * Math.min(scale, 4));
179
190
  }
180
191
  /**
181
192
  * Check if a schema is a Zod schema
@@ -36,6 +36,11 @@ export declare class ExternalServerManager extends EventEmitter {
36
36
  * Get current HITL manager
37
37
  */
38
38
  getHITLManager(): HITLManager | undefined;
39
+ /**
40
+ * Resolve the human-readable server name for an event payload.
41
+ * Falls back to serverId if the instance or config.name isn't available.
42
+ */
43
+ getServerName(serverId: string): string;
39
44
  /**
40
45
  * Load MCP server configurations from .mcp-config.json file with parallel loading support
41
46
  * Automatically registers servers found in the configuration
@@ -194,10 +194,16 @@ export class ExternalServerManager extends EventEmitter {
194
194
  this.toolDiscovery = new ToolDiscoveryService();
195
195
  // Forward tool discovery events
196
196
  this.toolDiscovery.on("toolRegistered", (event) => {
197
- this.emit("toolDiscovered", event);
197
+ this.emit("toolDiscovered", {
198
+ ...event,
199
+ serverName: this.getServerName(event.serverId),
200
+ });
198
201
  });
199
202
  this.toolDiscovery.on("toolUnregistered", (event) => {
200
- this.emit("toolRemoved", event);
203
+ this.emit("toolRemoved", {
204
+ ...event,
205
+ serverName: this.getServerName(event.serverId),
206
+ });
201
207
  });
202
208
  // Handle process cleanup
203
209
  process.on("SIGINT", () => this.shutdown());
@@ -223,6 +229,14 @@ export class ExternalServerManager extends EventEmitter {
223
229
  getHITLManager() {
224
230
  return this.hitlManager;
225
231
  }
232
+ /**
233
+ * Resolve the human-readable server name for an event payload.
234
+ * Falls back to serverId if the instance or config.name isn't available.
235
+ */
236
+ getServerName(serverId) {
237
+ const instance = this.servers.get(serverId);
238
+ return instance?.config?.name || serverId;
239
+ }
226
240
  /**
227
241
  * Load MCP server configurations from .mcp-config.json file with parallel loading support
228
242
  * Automatically registers servers found in the configuration
@@ -712,6 +726,8 @@ export class ExternalServerManager extends EventEmitter {
712
726
  };
713
727
  }
714
728
  mcpLogger.info(`[ExternalServerManager] Removing server: ${serverId}`);
729
+ // Capture name before deletion removes the instance
730
+ const serverName = this.getServerName(serverId);
715
731
  // Stop the server
716
732
  await this.stopServer(serverId);
717
733
  // Remove from registry
@@ -719,6 +735,7 @@ export class ExternalServerManager extends EventEmitter {
719
735
  // Emit event
720
736
  this.emit("disconnected", {
721
737
  serverId,
738
+ serverName,
722
739
  reason: "Manually removed",
723
740
  timestamp: new Date(),
724
741
  });
@@ -816,6 +833,7 @@ export class ExternalServerManager extends EventEmitter {
816
833
  // Emit connected event
817
834
  this.emit("connected", {
818
835
  serverId,
836
+ serverName: this.getServerName(serverId),
819
837
  toolCount: instance.toolsMap.size,
820
838
  timestamp: new Date(),
821
839
  });
@@ -921,6 +939,7 @@ export class ExternalServerManager extends EventEmitter {
921
939
  // Emit status change event
922
940
  this.emit("statusChanged", {
923
941
  serverId,
942
+ serverName: this.getServerName(serverId),
924
943
  oldStatus,
925
944
  newStatus,
926
945
  timestamp: new Date(),
@@ -941,6 +960,7 @@ export class ExternalServerManager extends EventEmitter {
941
960
  // Emit failed event
942
961
  this.emit("failed", {
943
962
  serverId,
963
+ serverName: this.getServerName(serverId),
944
964
  error: error.message,
945
965
  timestamp: new Date(),
946
966
  });
@@ -965,6 +985,7 @@ export class ExternalServerManager extends EventEmitter {
965
985
  // Emit disconnected event
966
986
  this.emit("disconnected", {
967
987
  serverId,
988
+ serverName: this.getServerName(serverId),
968
989
  reason,
969
990
  timestamp: new Date(),
970
991
  });
@@ -1078,6 +1099,7 @@ export class ExternalServerManager extends EventEmitter {
1078
1099
  // Emit health check event
1079
1100
  this.emit("healthCheck", {
1080
1101
  serverId,
1102
+ serverName: this.getServerName(serverId),
1081
1103
  health,
1082
1104
  timestamp: new Date(),
1083
1105
  });
@@ -50,7 +50,7 @@ import { createMemoryRetrievalTools } from "./memory/memoryRetrievalTools.js";
50
50
  import { getMetricsAggregator, MetricsAggregator, } from "./observability/metricsAggregator.js";
51
51
  import { SpanStatus, SpanType } from "./observability/types/spanTypes.js";
52
52
  import { SpanSerializer } from "./observability/utils/spanSerializer.js";
53
- import { flushOpenTelemetry, getLangfuseHealthStatus, initializeOpenTelemetry, isOpenTelemetryInitialized, setLangfuseContext, shutdownOpenTelemetry, } from "./services/server/ai/observability/instrumentation.js";
53
+ import { flushOpenTelemetry, getLangfuseHealthStatus, initializeOpenTelemetry, isOpenTelemetryInitialized, runWithCurrentLangfuseContext, setLangfuseContext, shutdownOpenTelemetry, } from "./services/server/ai/observability/instrumentation.js";
54
54
  import { TaskManager } from "./tasks/taskManager.js";
55
55
  import { createTaskTools } from "./tasks/tools/taskTools.js";
56
56
  import { ATTR } from "./telemetry/attributes.js";
@@ -1129,7 +1129,10 @@ Current user's request: ${currentInput}`;
1129
1129
  * Supports additional users with per-user prompt and maxWords overrides.
1130
1130
  */
1131
1131
  storeMemoryInBackground(originalPrompt, responseContent, userId, additionalUsers) {
1132
- setImmediate(async () => {
1132
+ // Preserve AsyncLocalStorage context across setImmediate boundary so that
1133
+ // memory writes appear under the originating Langfuse trace instead of
1134
+ // becoming orphan spans.
1135
+ const wrappedMemoryWrite = runWithCurrentLangfuseContext(async () => {
1133
1136
  try {
1134
1137
  const client = this.ensureMemoryReady();
1135
1138
  if (!client) {
@@ -1145,12 +1148,18 @@ Current user's request: ${currentInput}`;
1145
1148
  : undefined;
1146
1149
  writeOps.push(client.add(user.userId, content, addOptions));
1147
1150
  }
1148
- await Promise.all(writeOps);
1151
+ // withTimeout races against Promise.all — if the timeout fires, the
1152
+ // await resolves with an error but the underlying client.add() calls
1153
+ // may still complete in the background. This is acceptable: the memory
1154
+ // client API (Mem0) doesn't support AbortSignal, and these are
1155
+ // fire-and-forget background writes where a stale completion is harmless.
1156
+ await withTimeout(Promise.all(writeOps), 30_000, new Error("Background memory write timed out after 30s"));
1149
1157
  }
1150
1158
  catch (error) {
1151
1159
  logger.warn("Memory storage failed:", error);
1152
1160
  }
1153
1161
  });
1162
+ setImmediate(wrappedMemoryWrite);
1154
1163
  }
1155
1164
  /**
1156
1165
  * Set up HITL event forwarding to main emitter
@@ -3723,6 +3732,21 @@ Current user's request: ${currentInput}`;
3723
3732
  conversationMessageCount: conversationMessages.length,
3724
3733
  shouldCompact: budgetResult.shouldCompact,
3725
3734
  });
3735
+ // Scale timeout for large contexts if caller didn't set one explicitly.
3736
+ // Providers read options.timeout via getTimeout(), so setting it here
3737
+ // propagates to any downstream provider call.
3738
+ if (options.timeout === undefined &&
3739
+ budgetResult.estimatedInputTokens > 100_000) {
3740
+ // >100K → 1.5x, >200K → 2x, >300K → 2.5x (capped at 4x) of 60s base
3741
+ const scale = 1 + Math.floor((budgetResult.estimatedInputTokens - 1) / 100_000) * 0.5;
3742
+ const scaledMs = Math.round(60_000 * Math.min(scale, 4));
3743
+ options.timeout = scaledMs;
3744
+ logger.info("[TokenBudget] Scaled timeout for large context", {
3745
+ requestId,
3746
+ estimatedTokens: budgetResult.estimatedInputTokens,
3747
+ scaledTimeoutMs: scaledMs,
3748
+ });
3749
+ }
3726
3750
  const compactionSessionId = this.getCompactionSessionId(options);
3727
3751
  const lastCompactionCount = this.lastCompactionMessageCount.get(compactionSessionId) ?? 0;
3728
3752
  if (!budgetResult.shouldCompact ||
@@ -3798,6 +3822,8 @@ Current user's request: ${currentInput}`;
3798
3822
  toolDefinitions: availableTools,
3799
3823
  });
3800
3824
  if (!finalBudget.withinBudget) {
3825
+ // Clear watermark so handleContextOverflow recovery can re-compact
3826
+ this.lastCompactionMessageCount.delete(compactionSessionId);
3801
3827
  throw new ContextBudgetExceededError(`Context exceeds model budget after all compaction stages. ` +
3802
3828
  `Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
3803
3829
  `Budget: ${finalBudget.availableInputTokens} tokens. ` +
@@ -3993,6 +4019,8 @@ Current user's request: ${currentInput}`;
3993
4019
  : undefined,
3994
4020
  });
3995
4021
  if (!finalBudget.withinBudget) {
4022
+ // Clear watermark so handleContextOverflow recovery can re-compact
4023
+ this.lastCompactionMessageCount.delete(dpgCompactionSessionId);
3996
4024
  throw new ContextBudgetExceededError(`Context exceeds model budget after all compaction stages. ` +
3997
4025
  `Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
3998
4026
  `Budget: ${finalBudget.availableInputTokens} tokens.`, {
@@ -5016,6 +5044,8 @@ Current user's request: ${currentInput}`;
5016
5044
  toolDefinitions: availableTools,
5017
5045
  });
5018
5046
  if (!finalBudget.withinBudget) {
5047
+ // Clear watermark so handleContextOverflow recovery can re-compact
5048
+ this.lastCompactionMessageCount.delete(streamCompactionSessionId);
5019
5049
  throw new ContextBudgetExceededError(`Stream context exceeds model budget after all compaction stages. ` +
5020
5050
  `Estimated: ${finalBudget.estimatedInputTokens} tokens, ` +
5021
5051
  `Budget: ${finalBudget.availableInputTokens} tokens.`, {
@@ -7508,6 +7538,7 @@ Current user's request: ${currentInput}`;
7508
7538
  // Emit server added event
7509
7539
  this.emitter.emit("externalMCP:serverAdded", {
7510
7540
  serverId,
7541
+ serverName: config.name || serverId,
7511
7542
  config,
7512
7543
  toolCount: result.metadata?.toolsDiscovered || 0,
7513
7544
  timestamp: Date.now(),
@@ -7535,12 +7566,15 @@ Current user's request: ${currentInput}`;
7535
7566
  this.invalidateToolCache(); // Invalidate cache when an external server is removed
7536
7567
  try {
7537
7568
  mcpLogger.info(`[NeuroLink] Removing external MCP server: ${serverId}`);
7569
+ // Capture the configured name before removal destroys the instance
7570
+ const serverName = this.externalServerManager.getServerName(serverId);
7538
7571
  const result = await this.externalServerManager.removeServer(serverId);
7539
7572
  if (result.success) {
7540
7573
  mcpLogger.info(`[NeuroLink] External MCP server removed successfully: ${serverId}`);
7541
7574
  // Emit server removed event
7542
7575
  this.emitter.emit("externalMCP:serverRemoved", {
7543
7576
  serverId,
7577
+ serverName,
7544
7578
  timestamp: Date.now(),
7545
7579
  });
7546
7580
  }
@@ -445,7 +445,45 @@ function initializeExternalOpenTelemetryMode(config, resource, otlpEndpoint, ser
445
445
  const provider = globalProvider;
446
446
  if (globalProvider && typeof provider.addSpanProcessor === "function") {
447
447
  provider.addSpanProcessor(new ContextEnricher());
448
- const skipLangfuse = config.skipLangfuseSpanProcessor === true || !langfuseProcessor;
448
+ // Auto-detect: skip if consumer already registered a LangfuseSpanProcessor.
449
+ //
450
+ // Detection strategy (ordered by robustness):
451
+ // 1. `instanceof LangfuseSpanProcessor` — reliable when both sides use
452
+ // the same @langfuse/otel package instance (same module identity).
453
+ // 2. Duck-type check for Langfuse-specific public member
454
+ // (`langfuseClient` property) — survives minification.
455
+ // 3. `constructor.name === "LangfuseSpanProcessor"` — last resort,
456
+ // brittle under minification or bundler renaming.
457
+ //
458
+ // NOTE: `_registeredSpanProcessors` is an internal OpenTelemetry field.
459
+ // If the OTel SDK removes or renames it, the array defaults to [] and
460
+ // `hasExistingLangfuse` is false — NeuroLink registers its own processor
461
+ // (same behavior as before this check). Consumers can always force skip
462
+ // via `skipLangfuseSpanProcessor: true`.
463
+ const existingProcessors = provider
464
+ ._registeredSpanProcessors ?? [];
465
+ const hasExistingLangfuse = existingProcessors.some((p) => {
466
+ if (p === null || p === undefined || typeof p !== "object") {
467
+ return false;
468
+ }
469
+ // Prefer instanceof — works when same @langfuse/otel package is shared
470
+ if (p instanceof LangfuseSpanProcessor) {
471
+ return true;
472
+ }
473
+ // Duck-type: Langfuse processor exposes a langfuseClient property
474
+ if ("langfuseClient" in p) {
475
+ return true;
476
+ }
477
+ // Fallback: constructor name (brittle under minification)
478
+ return (p.constructor?.name ===
479
+ "LangfuseSpanProcessor");
480
+ });
481
+ const skipLangfuse = config.skipLangfuseSpanProcessor === true ||
482
+ !langfuseProcessor ||
483
+ hasExistingLangfuse;
484
+ if (hasExistingLangfuse && !config.skipLangfuseSpanProcessor) {
485
+ logger.info(`${LOG_PREFIX} Auto-detected existing LangfuseSpanProcessor — skipping SDK registration to avoid duplicates`);
486
+ }
449
487
  if (!skipLangfuse && langfuseProcessor) {
450
488
  provider.addSpanProcessor(langfuseProcessor);
451
489
  }
@@ -228,6 +228,7 @@ export type ExternalMCPServerEvents = {
228
228
  /** Server status changed */
229
229
  statusChanged: {
230
230
  serverId: string;
231
+ serverName: string;
231
232
  oldStatus: ExternalMCPServerStatus;
232
233
  newStatus: ExternalMCPServerStatus;
233
234
  timestamp: Date;
@@ -235,24 +236,28 @@ export type ExternalMCPServerEvents = {
235
236
  /** Server connected successfully */
236
237
  connected: {
237
238
  serverId: string;
239
+ serverName: string;
238
240
  toolCount: number;
239
241
  timestamp: Date;
240
242
  };
241
243
  /** Server disconnected */
242
244
  disconnected: {
243
245
  serverId: string;
246
+ serverName: string;
244
247
  reason?: string;
245
248
  timestamp: Date;
246
249
  };
247
250
  /** Server failed */
248
251
  failed: {
249
252
  serverId: string;
253
+ serverName: string;
250
254
  error: string;
251
255
  timestamp: Date;
252
256
  };
253
257
  /** Tool discovered */
254
258
  toolDiscovered: {
255
259
  serverId: string;
260
+ serverName: string;
256
261
  toolName: string;
257
262
  toolInfo: ExternalMCPToolInfo;
258
263
  timestamp: Date;
@@ -260,12 +265,14 @@ export type ExternalMCPServerEvents = {
260
265
  /** Tool removed */
261
266
  toolRemoved: {
262
267
  serverId: string;
268
+ serverName: string;
263
269
  toolName: string;
264
270
  timestamp: Date;
265
271
  };
266
272
  /** Health check completed */
267
273
  healthCheck: {
268
274
  serverId: string;
275
+ serverName: string;
269
276
  health: ExternalMCPServerHealth;
270
277
  timestamp: Date;
271
278
  };
@@ -36,6 +36,11 @@ export declare class ExternalServerManager extends EventEmitter {
36
36
  * Get current HITL manager
37
37
  */
38
38
  getHITLManager(): HITLManager | undefined;
39
+ /**
40
+ * Resolve the human-readable server name for an event payload.
41
+ * Falls back to serverId if the instance or config.name isn't available.
42
+ */
43
+ getServerName(serverId: string): string;
39
44
  /**
40
45
  * Load MCP server configurations from .mcp-config.json file with parallel loading support
41
46
  * Automatically registers servers found in the configuration