@juspay/neurolink 7.44.0 → 7.46.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (74) hide show
  1. package/CHANGELOG.md +12 -0
  2. package/dist/cli/commands/config.d.ts +2 -2
  3. package/dist/cli/loop/optionsSchema.d.ts +1 -1
  4. package/dist/core/factory.d.ts +3 -1
  5. package/dist/core/factory.js +5 -3
  6. package/dist/evaluation/contextBuilder.d.ts +48 -0
  7. package/dist/evaluation/contextBuilder.js +134 -0
  8. package/dist/evaluation/index.d.ts +36 -0
  9. package/dist/evaluation/index.js +61 -0
  10. package/dist/evaluation/prompts.d.ts +22 -0
  11. package/dist/evaluation/prompts.js +73 -0
  12. package/dist/evaluation/ragasEvaluator.d.ts +28 -0
  13. package/dist/evaluation/ragasEvaluator.js +90 -0
  14. package/dist/evaluation/retryManager.d.ts +40 -0
  15. package/dist/evaluation/retryManager.js +78 -0
  16. package/dist/evaluation/scoring.d.ts +16 -0
  17. package/dist/evaluation/scoring.js +35 -0
  18. package/dist/factories/providerFactory.d.ts +3 -3
  19. package/dist/factories/providerFactory.js +3 -3
  20. package/dist/factories/providerRegistry.js +6 -6
  21. package/dist/lib/core/factory.d.ts +3 -1
  22. package/dist/lib/core/factory.js +5 -3
  23. package/dist/lib/evaluation/contextBuilder.d.ts +48 -0
  24. package/dist/lib/evaluation/contextBuilder.js +134 -0
  25. package/dist/lib/evaluation/index.d.ts +36 -0
  26. package/dist/lib/evaluation/index.js +61 -0
  27. package/dist/lib/evaluation/prompts.d.ts +22 -0
  28. package/dist/lib/evaluation/prompts.js +73 -0
  29. package/dist/lib/evaluation/ragasEvaluator.d.ts +28 -0
  30. package/dist/lib/evaluation/ragasEvaluator.js +90 -0
  31. package/dist/lib/evaluation/retryManager.d.ts +40 -0
  32. package/dist/lib/evaluation/retryManager.js +78 -0
  33. package/dist/lib/evaluation/scoring.d.ts +16 -0
  34. package/dist/lib/evaluation/scoring.js +35 -0
  35. package/dist/lib/factories/providerFactory.d.ts +3 -3
  36. package/dist/lib/factories/providerFactory.js +3 -3
  37. package/dist/lib/factories/providerRegistry.js +6 -6
  38. package/dist/lib/middleware/builtin/autoEvaluation.d.ts +14 -0
  39. package/dist/lib/middleware/builtin/autoEvaluation.js +181 -0
  40. package/dist/lib/middleware/factory.js +6 -0
  41. package/dist/lib/neurolink.js +7 -3
  42. package/dist/lib/providers/amazonBedrock.d.ts +2 -1
  43. package/dist/lib/providers/amazonBedrock.js +6 -4
  44. package/dist/lib/providers/amazonSagemaker.d.ts +1 -1
  45. package/dist/lib/providers/amazonSagemaker.js +2 -2
  46. package/dist/lib/providers/googleVertex.d.ts +1 -1
  47. package/dist/lib/providers/googleVertex.js +9 -10
  48. package/dist/lib/providers/sagemaker/config.d.ts +7 -5
  49. package/dist/lib/providers/sagemaker/config.js +11 -6
  50. package/dist/lib/types/evaluation.d.ts +2 -0
  51. package/dist/lib/types/evaluationTypes.d.ts +142 -0
  52. package/dist/lib/types/evaluationTypes.js +1 -0
  53. package/dist/lib/types/generateTypes.d.ts +2 -0
  54. package/dist/lib/types/middlewareTypes.d.ts +28 -2
  55. package/dist/lib/types/streamTypes.d.ts +1 -0
  56. package/dist/middleware/builtin/autoEvaluation.d.ts +14 -0
  57. package/dist/middleware/builtin/autoEvaluation.js +181 -0
  58. package/dist/middleware/factory.js +6 -0
  59. package/dist/neurolink.js +7 -3
  60. package/dist/providers/amazonBedrock.d.ts +2 -1
  61. package/dist/providers/amazonBedrock.js +6 -4
  62. package/dist/providers/amazonSagemaker.d.ts +1 -1
  63. package/dist/providers/amazonSagemaker.js +2 -2
  64. package/dist/providers/googleVertex.d.ts +1 -1
  65. package/dist/providers/googleVertex.js +9 -10
  66. package/dist/providers/sagemaker/config.d.ts +7 -5
  67. package/dist/providers/sagemaker/config.js +11 -6
  68. package/dist/types/evaluation.d.ts +2 -0
  69. package/dist/types/evaluationTypes.d.ts +142 -0
  70. package/dist/types/evaluationTypes.js +1 -0
  71. package/dist/types/generateTypes.d.ts +2 -0
  72. package/dist/types/middlewareTypes.d.ts +28 -2
  73. package/dist/types/streamTypes.d.ts +1 -0
  74. package/package.json +1 -1
@@ -2,6 +2,7 @@ import { wrapLanguageModel } from "ai";
2
2
  import { MiddlewareRegistry } from "./registry.js";
3
3
  import { createAnalyticsMiddleware } from "./builtin/analytics.js";
4
4
  import { createGuardrailsMiddleware } from "./builtin/guardrails.js";
5
+ import { createAutoEvaluationMiddleware } from "./builtin/autoEvaluation.js";
5
6
  import { logger } from "../utils/logger.js";
6
7
  /**
7
8
  * Middleware factory for creating and applying middleware chains.
@@ -24,6 +25,7 @@ export class MiddlewareFactory {
24
25
  const builtInMiddlewareCreators = {
25
26
  analytics: createAnalyticsMiddleware,
26
27
  guardrails: createGuardrailsMiddleware,
28
+ autoEvaluation: createAutoEvaluationMiddleware,
27
29
  };
28
30
  // Register built-in presets
29
31
  this.registerPreset({
@@ -54,6 +56,7 @@ export class MiddlewareFactory {
54
56
  if (!this.registry.has(middlewareId)) {
55
57
  const creator = builtInMiddlewareCreators[middlewareId];
56
58
  const config = options.middlewareConfig?.[middlewareId]?.config;
59
+ logger.debug(`Registering built-in middleware '${middlewareId}'`, config);
57
60
  this.registry.register(creator(config));
58
61
  }
59
62
  }
@@ -92,6 +95,7 @@ export class MiddlewareFactory {
92
95
  const middlewareConfig = this.buildMiddlewareConfig(mergedOptions);
93
96
  // Re-register middleware with the correct configuration for this call
94
97
  for (const [id, config] of Object.entries(middlewareConfig)) {
98
+ logger.debug(`Configuring middleware '${id}'`, { config });
95
99
  if (config.enabled && this.registry.has(id)) {
96
100
  const creator = this.getCreator(id);
97
101
  if (creator) {
@@ -137,7 +141,9 @@ export class MiddlewareFactory {
137
141
  const builtInMiddlewareCreators = {
138
142
  analytics: createAnalyticsMiddleware,
139
143
  guardrails: createGuardrailsMiddleware,
144
+ autoEvaluation: createAutoEvaluationMiddleware,
140
145
  };
146
+ logger.debug("Getting creator for middleware ID:", id);
141
147
  return builtInMiddlewareCreators[id];
142
148
  }
143
149
  /**
package/dist/neurolink.js CHANGED
@@ -1082,6 +1082,7 @@ export class NeuroLink {
1082
1082
  evaluationDomain: options.evaluationDomain,
1083
1083
  toolUsageContext: options.toolUsageContext,
1084
1084
  input: options.input, // This includes text, images, and content arrays
1085
+ region: options.region,
1085
1086
  };
1086
1087
  // Apply factory enhancement using centralized utilities
1087
1088
  const textOptions = enhanceTextGenerationOptions(baseOptions, factoryResult);
@@ -1410,7 +1411,8 @@ export class NeuroLink {
1410
1411
  const conversationMessages = await getConversationMessages(this.conversationMemory, options);
1411
1412
  // Create provider and generate
1412
1413
  const provider = await AIProviderFactory.createProvider(providerName, options.model, !options.disableTools, // Pass disableTools as inverse of enableMCP
1413
- this);
1414
+ this, // Pass SDK instance
1415
+ options.region);
1414
1416
  // ADD: Emit connection events for all providers (Bedrock-compatible)
1415
1417
  this.emitter.emit("connected");
1416
1418
  this.emitter.emit("message", `${providerName} provider initialized successfully`);
@@ -1520,7 +1522,8 @@ export class NeuroLink {
1520
1522
  // Get conversation messages for context
1521
1523
  const conversationMessages = await getConversationMessages(this.conversationMemory, options);
1522
1524
  const provider = await AIProviderFactory.createProvider(providerName, options.model, !options.disableTools, // Pass disableTools as inverse of enableMCP
1523
- this);
1525
+ this, // Pass SDK instance
1526
+ options.region);
1524
1527
  // ADD: Emit connection events for successful provider creation (Bedrock-compatible)
1525
1528
  this.emitter.emit("connected");
1526
1529
  this.emitter.emit("message", `${providerName} provider initialized successfully`);
@@ -1897,7 +1900,8 @@ export class NeuroLink {
1897
1900
  // Simplified placeholder - in the actual implementation this would contain the complex MCP stream logic
1898
1901
  const providerName = await getBestProvider(options.provider);
1899
1902
  const provider = await AIProviderFactory.createProvider(providerName, options.model, !options.disableTools, // Pass disableTools as inverse of enableMCP
1900
- this);
1903
+ this, // Pass SDK instance
1904
+ options.region);
1901
1905
  // Enable tool execution for the provider using BaseProvider method
1902
1906
  provider.setupToolExecutor({
1903
1907
  customTools: this.getCustomTools(),
@@ -6,7 +6,8 @@ import type { NeuroLink } from "../neurolink.js";
6
6
  export declare class AmazonBedrockProvider extends BaseProvider {
7
7
  private bedrockClient;
8
8
  private conversationHistory;
9
- constructor(modelName?: string, neurolink?: NeuroLink);
9
+ private region;
10
+ constructor(modelName?: string, neurolink?: NeuroLink, region?: string);
10
11
  /**
11
12
  * Perform initial health check to catch credential/connectivity issues early
12
13
  * This prevents the health check failure we saw in production logs
@@ -7,8 +7,10 @@ import { convertZodToJsonSchema } from "../utils/schemaConversion.js";
7
7
  export class AmazonBedrockProvider extends BaseProvider {
8
8
  bedrockClient;
9
9
  conversationHistory = [];
10
- constructor(modelName, neurolink) {
10
+ region;
11
+ constructor(modelName, neurolink, region) {
11
12
  super(modelName, "bedrock", neurolink);
13
+ this.region = region || process.env.AWS_REGION || "us-east-1";
12
14
  logger.debug("[AmazonBedrockProvider] Starting constructor with extensive logging for debugging");
13
15
  // Log environment variables for debugging
14
16
  logger.debug(`[AmazonBedrockProvider] Environment check: AWS_REGION=${process.env.AWS_REGION || "undefined"}, AWS_ACCESS_KEY_ID=${process.env.AWS_ACCESS_KEY_ID ? "SET" : "undefined"}, AWS_SECRET_ACCESS_KEY=${process.env.AWS_SECRET_ACCESS_KEY ? "SET" : "undefined"}`);
@@ -17,14 +19,14 @@ export class AmazonBedrockProvider extends BaseProvider {
17
19
  // Absolutely no proxy interference - let AWS SDK handle everything natively
18
20
  logger.debug("[AmazonBedrockProvider] Creating BedrockRuntimeClient with clean configuration");
19
21
  this.bedrockClient = new BedrockRuntimeClient({
20
- region: process.env.AWS_REGION || "us-east-1",
22
+ region: this.region,
21
23
  // Clean configuration - AWS SDK will handle credentials via:
22
24
  // 1. IAM roles (preferred in production)
23
25
  // 2. Environment variables
24
26
  // 3. AWS config files
25
27
  // 4. Instance metadata
26
28
  });
27
- logger.debug(`[AmazonBedrockProvider] Successfully created BedrockRuntimeClient with model: ${this.modelName}, region: ${process.env.AWS_REGION || "us-east-1"}`);
29
+ logger.debug(`[AmazonBedrockProvider] Successfully created BedrockRuntimeClient with model: ${this.modelName}, region: ${this.region}`);
28
30
  // Immediate health check to catch credential issues early
29
31
  this.performInitialHealthCheck();
30
32
  }
@@ -39,7 +41,7 @@ export class AmazonBedrockProvider extends BaseProvider {
39
41
  */
40
42
  async performInitialHealthCheck() {
41
43
  const bedrockClient = new BedrockClient({
42
- region: process.env.AWS_REGION || "us-east-1",
44
+ region: this.region,
43
45
  });
44
46
  try {
45
47
  logger.debug("[AmazonBedrockProvider] Starting initial health check to validate credentials and connectivity");
@@ -16,7 +16,7 @@ export declare class AmazonSageMakerProvider extends BaseProvider {
16
16
  private sagemakerModel;
17
17
  private sagemakerConfig;
18
18
  private modelConfig;
19
- constructor(modelName?: string, endpointName?: string);
19
+ constructor(modelName?: string, endpointName?: string, region?: string);
20
20
  protected getProviderName(): AIProviderName;
21
21
  protected getDefaultModel(): string;
22
22
  protected getAISDKModel(): LanguageModelV1;
@@ -17,11 +17,11 @@ export class AmazonSageMakerProvider extends BaseProvider {
17
17
  sagemakerModel;
18
18
  sagemakerConfig;
19
19
  modelConfig;
20
- constructor(modelName, endpointName) {
20
+ constructor(modelName, endpointName, region) {
21
21
  super(modelName, "sagemaker");
22
22
  try {
23
23
  // Load and validate configuration
24
- this.sagemakerConfig = getSageMakerConfig();
24
+ this.sagemakerConfig = getSageMakerConfig(region);
25
25
  this.modelConfig = getSageMakerModelConfig(endpointName || getDefaultSageMakerEndpoint());
26
26
  // Create the proper LanguageModel (v2) implementation
27
27
  this.sagemakerModel = new SageMakerLanguageModel(this.modelName, this.sagemakerConfig, this.modelConfig);
@@ -25,7 +25,7 @@ export declare class GoogleVertexProvider extends BaseProvider {
25
25
  private static readonly MAX_CACHE_SIZE;
26
26
  private static maxTokensCache;
27
27
  private static maxTokensCacheTime;
28
- constructor(modelName?: string, _providerName?: string, sdk?: unknown);
28
+ constructor(modelName?: string, _providerName?: string, sdk?: unknown, region?: string);
29
29
  protected getProviderName(): AIProviderName;
30
30
  protected getDefaultModel(): string;
31
31
  /**
@@ -46,10 +46,10 @@ const hasGoogleCredentials = () => {
46
46
  process.env.GOOGLE_AUTH_PRIVATE_KEY));
47
47
  };
48
48
  // Enhanced Vertex settings creation with authentication fallback and proxy support
49
- const createVertexSettings = async () => {
49
+ const createVertexSettings = async (region) => {
50
50
  const baseSettings = {
51
51
  project: getVertexProjectId(),
52
- location: getVertexLocation(),
52
+ location: region || getVertexLocation(),
53
53
  fetch: createProxyFetch(),
54
54
  };
55
55
  // 🎯 OPTION 2: Create credentials file from environment variables at runtime
@@ -157,8 +157,7 @@ const createVertexSettings = async () => {
157
157
  private_key: requiredEnvVars.private_key.replace(/\\n/g, "\n"),
158
158
  client_email: requiredEnvVars.client_email,
159
159
  client_id: requiredEnvVars.client_id || "",
160
- auth_uri: requiredEnvVars.auth_uri ||
161
- "https://accounts.google.com/o/oauth2/auth",
160
+ auth_uri: requiredEnvVars.auth_uri || "https://accounts.google.com/o/oauth2/auth",
162
161
  token_uri: requiredEnvVars.token_uri || "https://oauth2.googleapis.com/token",
163
162
  auth_provider_x509_cert_url: requiredEnvVars.auth_provider_x509_cert_url ||
164
163
  "https://www.googleapis.com/oauth2/v1/certs",
@@ -199,8 +198,8 @@ const createVertexSettings = async () => {
199
198
  return baseSettings;
200
199
  };
201
200
  // Create Anthropic-specific Vertex settings with the same authentication and proxy support
202
- const createVertexAnthropicSettings = async () => {
203
- const baseVertexSettings = await createVertexSettings();
201
+ const createVertexAnthropicSettings = async (region) => {
202
+ const baseVertexSettings = await createVertexSettings(region);
204
203
  // GoogleVertexAnthropicProviderSettings extends GoogleVertexProviderSettings
205
204
  // so we can use the same settings with proper typing
206
205
  return {
@@ -241,7 +240,7 @@ export class GoogleVertexProvider extends BaseProvider {
241
240
  // Memory-managed cache for maxTokens handling decisions to optimize streaming performance
242
241
  static maxTokensCache = new Map();
243
242
  static maxTokensCacheTime = 0;
244
- constructor(modelName, _providerName, sdk) {
243
+ constructor(modelName, _providerName, sdk, region) {
245
244
  super(modelName, "vertex", sdk);
246
245
  // Validate Google Cloud credentials - now using consolidated utility
247
246
  if (!hasGoogleCredentials()) {
@@ -249,7 +248,7 @@ export class GoogleVertexProvider extends BaseProvider {
249
248
  }
250
249
  // Initialize Google Cloud configuration
251
250
  this.projectId = getVertexProjectId();
252
- this.location = getVertexLocation();
251
+ this.location = region || getVertexLocation();
253
252
  logger.debug("Google Vertex AI BaseProvider v2 initialized", {
254
253
  modelName: this.modelName,
255
254
  projectId: this.projectId,
@@ -360,7 +359,7 @@ export class GoogleVertexProvider extends BaseProvider {
360
359
  message: "Starting Vertex settings creation with network configuration analysis",
361
360
  });
362
361
  try {
363
- const vertexSettings = await createVertexSettings();
362
+ const vertexSettings = await createVertexSettings(this.location);
364
363
  const vertexSettingsEndTime = process.hrtime.bigint();
365
364
  const vertexSettingsDurationNs = vertexSettingsEndTime - vertexSettingsStartTime;
366
365
  logger.debug(`[GoogleVertexProvider] ✅ LOG_POINT_V009_VERTEX_SETTINGS_SUCCESS`, {
@@ -920,7 +919,7 @@ export class GoogleVertexProvider extends BaseProvider {
920
919
  projectId: projectValidation.projectId,
921
920
  region: projectValidation.region,
922
921
  });
923
- const vertexAnthropicSettings = await createVertexAnthropicSettings();
922
+ const vertexAnthropicSettings = await createVertexAnthropicSettings(this.location);
924
923
  // 7. Settings Validation
925
924
  if (!vertexAnthropicSettings.project ||
926
925
  !vertexAnthropicSettings.location) {
@@ -8,15 +8,17 @@ import type { SageMakerConfig, SageMakerModelConfig } from "./types.js";
8
8
  /**
9
9
  * Load and validate SageMaker configuration from environment variables
10
10
  *
11
- * Environment variable priority:
12
- * 1. SAGEMAKER_* variables (highest priority)
13
- * 2. AWS_* variables (standard AWS SDK variables)
14
- * 3. Default values (lowest priority)
11
+ * Region priority:
12
+ * 1. region parameter (highest priority)
13
+ * 2. SAGEMAKER_REGION environment variable
14
+ * 3. AWS_REGION environment variable
15
+ * 4. Default value "us-east-1" (lowest priority)
15
16
  *
17
+ * @param region - Optional region parameter override
16
18
  * @returns Validated SageMaker configuration
17
19
  * @throws {Error} When required configuration is missing or invalid
18
20
  */
19
- export declare function getSageMakerConfig(): SageMakerConfig;
21
+ export declare function getSageMakerConfig(region?: string): SageMakerConfig;
20
22
  /**
21
23
  * Load and validate SageMaker model configuration
22
24
  *
@@ -45,21 +45,26 @@ const modelConfigCache = new Map();
45
45
  /**
46
46
  * Load and validate SageMaker configuration from environment variables
47
47
  *
48
- * Environment variable priority:
49
- * 1. SAGEMAKER_* variables (highest priority)
50
- * 2. AWS_* variables (standard AWS SDK variables)
51
- * 3. Default values (lowest priority)
48
+ * Region priority:
49
+ * 1. region parameter (highest priority)
50
+ * 2. SAGEMAKER_REGION environment variable
51
+ * 3. AWS_REGION environment variable
52
+ * 4. Default value "us-east-1" (lowest priority)
52
53
  *
54
+ * @param region - Optional region parameter override
53
55
  * @returns Validated SageMaker configuration
54
56
  * @throws {Error} When required configuration is missing or invalid
55
57
  */
56
- export function getSageMakerConfig() {
58
+ export function getSageMakerConfig(region) {
57
59
  // Return cached config if available
58
60
  if (configCache) {
59
61
  return configCache;
60
62
  }
61
63
  const config = {
62
- region: process.env.SAGEMAKER_REGION || process.env.AWS_REGION || "us-east-1",
64
+ region: region ||
65
+ process.env.SAGEMAKER_REGION ||
66
+ process.env.AWS_REGION ||
67
+ "us-east-1",
63
68
  accessKeyId: process.env.AWS_ACCESS_KEY_ID || "",
64
69
  secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY || "",
65
70
  sessionToken: process.env.AWS_SESSION_TOKEN,
@@ -25,6 +25,8 @@ export type EvaluationData = {
25
25
  domainAlignment?: number;
26
26
  terminologyAccuracy?: number;
27
27
  toolEffectiveness?: number;
28
+ responseContent?: string;
29
+ queryContent?: string;
28
30
  isOffTopic: boolean;
29
31
  alertSeverity: AlertSeverity;
30
32
  reasoning: string;
@@ -0,0 +1,142 @@
1
+ import type { LanguageModelV1CallOptions } from "ai";
2
+ import type { TokenUsage } from "./analytics.js";
3
+ import type { GenerateResult } from "./generateTypes.js";
4
+ import type { ToolExecution } from "./tools.js";
5
+ /**
6
+ * Represents the analysis of the user's query intent.
7
+ * This provides a basic understanding of what the user is trying to achieve.
8
+ */
9
+ export interface QueryIntentAnalysis {
10
+ /** The type of query, e.g., asking a question or giving a command. */
11
+ type: "question" | "command" | "greeting" | "unknown";
12
+ /** The estimated complexity of the query. */
13
+ complexity: "low" | "medium" | "high";
14
+ /** Whether the query likely required the use of tools to be answered correctly. */
15
+ shouldHaveUsedTools: boolean;
16
+ }
17
+ /**
18
+ * Represents a single turn in an enhanced conversation history,
19
+ * including tool executions and evaluations for richer context.
20
+ */
21
+ export interface EnhancedConversationTurn {
22
+ /** The role of the speaker, either 'user' or 'assistant'. */
23
+ role: "user" | "assistant";
24
+ /** The content of the message. */
25
+ content: string;
26
+ /** The timestamp of the message. */
27
+ timestamp: string;
28
+ /** Any tools that were executed as part of this turn. */
29
+ toolExecutions?: ToolExecution[];
30
+ /** The evaluation result for this turn, if applicable. */
31
+ evaluation?: EvaluationResult;
32
+ }
33
+ /**
34
+ * Contains all the rich context needed for a thorough, RAGAS-style evaluation.
35
+ * This object is constructed by the `ContextBuilder` and used by the `RAGASEvaluator`.
36
+ */
37
+ export interface EnhancedEvaluationContext {
38
+ /** The original user query. */
39
+ userQuery: string;
40
+ /** An analysis of the user's query intent. */
41
+ queryAnalysis: QueryIntentAnalysis;
42
+ /** The AI's response that is being evaluated. */
43
+ aiResponse: string;
44
+ /** The AI provider that generated the response. */
45
+ provider: string;
46
+ /** The specific model that generated the response. */
47
+ model: string;
48
+ /** The parameters used for the generation call. */
49
+ generationParams: {
50
+ temperature?: number;
51
+ maxTokens?: number;
52
+ systemPrompt?: string;
53
+ };
54
+ /** A list of tools that were executed. */
55
+ toolExecutions: ToolExecution[];
56
+ /** The history of the conversation leading up to this turn. */
57
+ conversationHistory: EnhancedConversationTurn[];
58
+ /** The response time of the AI in milliseconds. */
59
+ responseTime: number;
60
+ /** The token usage for the generation. */
61
+ tokenUsage: TokenUsage;
62
+ /** The results of any previous evaluation attempts for this response. */
63
+ previousEvaluations?: EvaluationResult[];
64
+ /** The current attempt number for this evaluation (1-based). */
65
+ attemptNumber: number;
66
+ }
67
+ /**
68
+ * Represents the result of a single evaluation attempt, based on RAGAS principles.
69
+ */
70
+ export interface EvaluationResult {
71
+ /** The final, overall score for the response, typically from 1 to 10. */
72
+ finalScore: number;
73
+ /** How well the response addresses the user's query. */
74
+ relevanceScore: number;
75
+ /** The factual accuracy of the information in the response. */
76
+ accuracyScore: number;
77
+ /** How completely the response answers the user's query. */
78
+ completenessScore: number;
79
+ /** Whether the final score meets the passing threshold. */
80
+ isPassing: boolean;
81
+ /** Constructive response from the judge LLM on how to improve the response. */
82
+ reasoning: string;
83
+ /** Specific suggestions for improving the response. */
84
+ suggestedImprovements: string;
85
+ /** The raw, unparsed response from the judge LLM. */
86
+ rawEvaluationResponse: string;
87
+ /** The model used to perform the evaluation. */
88
+ evaluationModel: string;
89
+ /** The time taken for the evaluation in milliseconds. */
90
+ evaluationTime: number;
91
+ /** The attempt number for this evaluation. */
92
+ attemptNumber: number;
93
+ }
94
+ /**
95
+ * Provides detailed information when a response fails quality assurance checks.
96
+ */
97
+ export interface QualityErrorDetails {
98
+ /** The history of all evaluation attempts for this response. */
99
+ evaluationHistory: EvaluationResult[];
100
+ /** The final score of the last attempt. */
101
+ finalScore: number;
102
+ /** The total number of evaluation attempts made. */
103
+ attempts: number;
104
+ /** A summary message of the failure. */
105
+ message: string;
106
+ }
107
+ /**
108
+ * Configuration for the main `Evaluator` class.
109
+ */
110
+ export interface EvaluationConfig {
111
+ /** The minimum score (1-10) for a response to be considered passing. */
112
+ threshold?: number;
113
+ /** The evaluation strategy to use. Currently only 'ragas' is supported. */
114
+ evaluationStrategy?: "ragas" | "custom";
115
+ /** The model to use for the LLM-as-judge evaluation. */
116
+ evaluationModel?: string;
117
+ /** The maximum number of evaluation attempts before failing. */
118
+ maxAttempts?: number;
119
+ /** The provider to use for the evaluation model. */
120
+ provider?: string;
121
+ /** A custom evaluator function to override the default behavior. */
122
+ customEvaluator?: (options: LanguageModelV1CallOptions, result: GenerateResult) => Promise<{
123
+ evaluationResult: EvaluationResult;
124
+ evalContext: EnhancedEvaluationContext;
125
+ }>;
126
+ /** The score below which a response is considered off-topic. */
127
+ offTopicThreshold?: number;
128
+ /** The score below which a failing response is considered a high severity alert. */
129
+ highSeverityThreshold?: number;
130
+ /** An optional function to generate custom evaluation prompts. */
131
+ promptGenerator?: GetPromptFunction;
132
+ }
133
+ /**
134
+ * A function that generates the main body of an evaluation prompt.
135
+ */
136
+ export type GetPromptFunction = (context: {
137
+ userQuery: string;
138
+ history: string;
139
+ tools: string;
140
+ retryInfo: string;
141
+ aiResponse: string;
142
+ }) => string;
@@ -0,0 +1 @@
1
+ export {};
@@ -22,6 +22,7 @@ export type GenerateOptions = {
22
22
  };
23
23
  provider?: AIProviderName | string;
24
24
  model?: string;
25
+ region?: string;
25
26
  temperature?: number;
26
27
  maxTokens?: number;
27
28
  systemPrompt?: string;
@@ -143,6 +144,7 @@ export type TextGenerationOptions = {
143
144
  };
144
145
  provider?: AIProviderName;
145
146
  model?: string;
147
+ region?: string;
146
148
  temperature?: number;
147
149
  maxTokens?: number;
148
150
  systemPrompt?: string;
@@ -1,5 +1,7 @@
1
1
  import type { LanguageModelV1Middleware } from "ai";
2
2
  import type { JsonValue } from "../types/common.js";
3
+ import type { EvaluationData } from "./evaluation.js";
4
+ import type { GetPromptFunction } from "./evaluationTypes.js";
3
5
  /**
4
6
  * Metadata interface for NeuroLink middleware
5
7
  * Provides additional information about middleware without affecting execution
@@ -33,7 +35,7 @@ export interface MiddlewareConfig {
33
35
  /** Whether the middleware is enabled */
34
36
  enabled?: boolean;
35
37
  /** Middleware-specific configuration */
36
- config?: Record<string, JsonValue>;
38
+ config?: Record<string, unknown>;
37
39
  /** Conditions under which to apply this middleware */
38
40
  conditions?: MiddlewareConditions;
39
41
  }
@@ -108,7 +110,7 @@ export interface MiddlewareChainStats {
108
110
  /**
109
111
  * Built-in middleware types
110
112
  */
111
- export type BuiltInMiddlewareType = "analytics" | "guardrails" | "logging" | "caching" | "rateLimit" | "retry" | "timeout";
113
+ export type BuiltInMiddlewareType = "analytics" | "guardrails" | "logging" | "caching" | "rateLimit" | "retry" | "timeout" | "autoEvaluation";
112
114
  /**
113
115
  * Middleware preset configurations
114
116
  */
@@ -144,3 +146,27 @@ export interface MiddlewareFactoryOptions {
144
146
  collectStats?: boolean;
145
147
  };
146
148
  }
149
+ /**
150
+ * Configuration for the Auto-Evaluation Middleware.
151
+ */
152
+ export interface AutoEvaluationConfig {
153
+ /** The minimum score (1-10) for a response to be considered passing. */
154
+ threshold?: number;
155
+ /** The maximum number of retry attempts before failing. */
156
+ maxRetries?: number;
157
+ /** The model to use for the LLM-as-judge evaluation. */
158
+ evaluationModel?: string;
159
+ /**
160
+ * If true, the middleware will wait for the evaluation to complete before returning.
161
+ * If the evaluation fails, it will throw an error. Defaults to true.
162
+ */
163
+ blocking?: boolean;
164
+ /** A callback function to be invoked with the evaluation result. */
165
+ onEvaluationComplete?: (evaluation: EvaluationData) => void | Promise<void>;
166
+ /** The score below which a response is considered off-topic. */
167
+ offTopicThreshold?: number;
168
+ /** The score below which a failing response is considered a high severity alert. */
169
+ highSeverityThreshold?: number;
170
+ promptGenerator?: GetPromptFunction;
171
+ provider?: string;
172
+ }
@@ -137,6 +137,7 @@ export interface StreamOptions {
137
137
  };
138
138
  provider?: AIProviderName | string;
139
139
  model?: string;
140
+ region?: string;
140
141
  temperature?: number;
141
142
  maxTokens?: number;
142
143
  systemPrompt?: string;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@juspay/neurolink",
3
- "version": "7.44.0",
3
+ "version": "7.46.0",
4
4
  "description": "Universal AI Development Platform with working MCP integration, multi-provider support, and professional CLI. Built-in tools operational, 58+ external MCP servers discoverable. Connect to filesystem, GitHub, database operations, and more. Build, test, and deploy AI applications with 9 major providers: OpenAI, Anthropic, Google AI, AWS Bedrock, Azure, Hugging Face, Ollama, and Mistral AI.",
5
5
  "author": {
6
6
  "name": "Juspay Technologies",