peerbench 0.0.2-alpha.0 → 0.0.2-alpha.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +123 -99
- package/dist/aggregators/index.d.ts +67 -0
- package/dist/aggregators/index.js +46 -0
- package/dist/aggregators/index.js.map +1 -0
- package/dist/benchmarks/index.d.ts +615 -1271
- package/dist/benchmarks/index.js +358 -805
- package/dist/benchmarks/index.js.map +1 -1
- package/dist/{chunk-DUBKY73H.js → chunk-4UBK6452.js} +13 -13
- package/dist/chunk-4UBK6452.js.map +1 -0
- package/dist/chunk-ERALDEZY.js +112 -0
- package/dist/chunk-ERALDEZY.js.map +1 -0
- package/dist/{chunk-ZJWSK4VO.js → chunk-HMQYGCKI.js} +1 -1
- package/dist/chunk-HMQYGCKI.js.map +1 -0
- package/dist/chunk-NUEOE3K5.js +8 -0
- package/dist/chunk-NUEOE3K5.js.map +1 -0
- package/dist/chunk-OQE6TQXZ.js +42 -0
- package/dist/chunk-OQE6TQXZ.js.map +1 -0
- package/dist/chunk-QY5MPNNB.js +28 -0
- package/dist/chunk-QY5MPNNB.js.map +1 -0
- package/dist/chunk-R76XA2K6.js +229 -0
- package/dist/chunk-R76XA2K6.js.map +1 -0
- package/dist/chunk-TRNCF2BG.js +35 -0
- package/dist/chunk-TRNCF2BG.js.map +1 -0
- package/dist/chunk-UHHHSYVE.js +11 -0
- package/dist/chunk-UHHHSYVE.js.map +1 -0
- package/dist/{chunk-232PY7K3.js → chunk-YY33MNMV.js} +29 -14
- package/dist/chunk-YY33MNMV.js.map +1 -0
- package/dist/chunk-ZEWI24CV.js +365 -0
- package/dist/chunk-ZEWI24CV.js.map +1 -0
- package/dist/chunk-ZXTQJFGL.js +44 -0
- package/dist/chunk-ZXTQJFGL.js.map +1 -0
- package/dist/index-BAioQhp2.d.ts +27 -0
- package/dist/index.d.ts +51 -26
- package/dist/index.js +28 -25
- package/dist/index.js.map +1 -1
- package/dist/json-file-ZwzLUbje.d.ts +73 -0
- package/dist/llm-judge-QThCZ9TQ.d.ts +67 -0
- package/dist/providers/index.d.ts +16 -19
- package/dist/providers/index.js +8 -253
- package/dist/providers/index.js.map +1 -1
- package/dist/schemas/extensions/index.d.ts +16 -2
- package/dist/schemas/extensions/index.js +9 -3
- package/dist/schemas/extensions/index.js.map +1 -1
- package/dist/schemas/index.d.ts +108 -141
- package/dist/schemas/index.js +7 -10
- package/dist/schemas/llm/index.d.ts +100 -82
- package/dist/schemas/llm/index.js +7 -29
- package/dist/schemas/llm/index.js.map +1 -1
- package/dist/scorers/index.d.ts +3 -2
- package/dist/scorers/index.js +8 -486
- package/dist/scorers/index.js.map +1 -1
- package/dist/storages/index.d.ts +69 -0
- package/dist/storages/index.js +98 -0
- package/dist/storages/index.js.map +1 -0
- package/package.json +12 -6
- package/dist/catalogs/index.d.ts +0 -75
- package/dist/catalogs/index.js +0 -88
- package/dist/catalogs/index.js.map +0 -1
- package/dist/chunk-22HU24QF.js +0 -8
- package/dist/chunk-22HU24QF.js.map +0 -1
- package/dist/chunk-232PY7K3.js.map +0 -1
- package/dist/chunk-7TREBPSJ.js +0 -26
- package/dist/chunk-7TREBPSJ.js.map +0 -1
- package/dist/chunk-DUBKY73H.js.map +0 -1
- package/dist/chunk-GVF4YZF3.js +0 -15
- package/dist/chunk-GVF4YZF3.js.map +0 -1
- package/dist/chunk-HJH3SW3L.js +0 -103
- package/dist/chunk-HJH3SW3L.js.map +0 -1
- package/dist/chunk-IUN2IUCS.js +0 -58
- package/dist/chunk-IUN2IUCS.js.map +0 -1
- package/dist/chunk-VBOM2YEG.js +0 -47
- package/dist/chunk-VBOM2YEG.js.map +0 -1
- package/dist/chunk-ZJWSK4VO.js.map +0 -1
- package/dist/data-BmN5WjZ4.d.ts +0 -57
- package/dist/generic-array-DLHWSvf1.d.ts +0 -22
- package/dist/index-WiPjF2AL.d.ts +0 -15
- package/dist/llm-judge-DIG1f1Az.d.ts +0 -67
- package/dist/simple-system-prompt-CzPYuvo0.d.ts +0 -49
- package/dist/system-prompt--0FdPWqK.d.ts +0 -58
- package/dist/utilities-BrRH32rD.d.ts +0 -30
|
@@ -0,0 +1,229 @@
|
|
|
1
|
+
import {
|
|
2
|
+
PEERBENCH_NAMESPACE
|
|
3
|
+
} from "./chunk-UHHHSYVE.js";
|
|
4
|
+
import {
|
|
5
|
+
RateLimiter
|
|
6
|
+
} from "./chunk-4UBK6452.js";
|
|
7
|
+
|
|
8
|
+
// src/providers/abstract/provider.ts
|
|
9
|
+
var AbstractProvider = class {
|
|
10
|
+
};
|
|
11
|
+
|
|
12
|
+
// src/providers/abstract/llm.ts
|
|
13
|
+
var AbstractLLMProvider = class extends AbstractProvider {
|
|
14
|
+
};
|
|
15
|
+
|
|
16
|
+
// src/providers/mastra.ts
|
|
17
|
+
import { MastraClient } from "@mastra/client-js";
|
|
18
|
+
var MastraProvider = class extends AbstractLLMProvider {
|
|
19
|
+
kind = "mastra";
|
|
20
|
+
endpoint;
|
|
21
|
+
authToken;
|
|
22
|
+
client;
|
|
23
|
+
warnedAboutSystemMessages = false;
|
|
24
|
+
warnedAboutResponseFormat = false;
|
|
25
|
+
constructor(params) {
|
|
26
|
+
super();
|
|
27
|
+
this.endpoint = params.endpoint;
|
|
28
|
+
this.authToken = params.authToken;
|
|
29
|
+
this.client = new MastraClient({
|
|
30
|
+
baseUrl: this.endpoint,
|
|
31
|
+
headers: {
|
|
32
|
+
Authorization: `Bearer ${this.authToken}`
|
|
33
|
+
}
|
|
34
|
+
});
|
|
35
|
+
}
|
|
36
|
+
async forward(args) {
|
|
37
|
+
const apiMessages = args.messages.filter((m) => m.role === "user" || m.role === "assistant").map((m) => ({
|
|
38
|
+
role: m.role,
|
|
39
|
+
content: String(m.content ?? "")
|
|
40
|
+
}));
|
|
41
|
+
const agent = this.client.getAgent(args.model);
|
|
42
|
+
const startedAt = Date.now();
|
|
43
|
+
const response = await agent.generate(
|
|
44
|
+
{
|
|
45
|
+
messages: apiMessages,
|
|
46
|
+
runtimeContext: {
|
|
47
|
+
"model-id": args.model
|
|
48
|
+
}
|
|
49
|
+
},
|
|
50
|
+
{ memory: args.memory }
|
|
51
|
+
);
|
|
52
|
+
return {
|
|
53
|
+
data: response.text,
|
|
54
|
+
startedAt,
|
|
55
|
+
completedAt: Date.now()
|
|
56
|
+
};
|
|
57
|
+
}
|
|
58
|
+
};
|
|
59
|
+
|
|
60
|
+
// src/providers/openai.ts
|
|
61
|
+
import OpenAI, { APIError } from "openai";
|
|
62
|
+
var OpenAIProvider = class extends AbstractLLMProvider {
|
|
63
|
+
kind = `${PEERBENCH_NAMESPACE}/llm/openai`;
|
|
64
|
+
client;
|
|
65
|
+
rateLimiter;
|
|
66
|
+
maxRetries;
|
|
67
|
+
constructor(config) {
|
|
68
|
+
super();
|
|
69
|
+
this.maxRetries = config.maxRetries ?? 3;
|
|
70
|
+
this.rateLimiter = config.rateLimiter ?? new RateLimiter({
|
|
71
|
+
maxWeight: 20,
|
|
72
|
+
timeWindow: 3e3
|
|
73
|
+
});
|
|
74
|
+
this.client = new OpenAI({
|
|
75
|
+
baseURL: config.baseURL,
|
|
76
|
+
apiKey: config.apiKey,
|
|
77
|
+
timeout: config.timeout,
|
|
78
|
+
dangerouslyAllowBrowser: true
|
|
79
|
+
});
|
|
80
|
+
}
|
|
81
|
+
async forward(args) {
|
|
82
|
+
let retryCount = this.maxRetries;
|
|
83
|
+
while (retryCount > 0) {
|
|
84
|
+
let startedAt = /* @__PURE__ */ new Date();
|
|
85
|
+
try {
|
|
86
|
+
const response = await this.rateLimiter.execute(
|
|
87
|
+
async () => {
|
|
88
|
+
startedAt = /* @__PURE__ */ new Date();
|
|
89
|
+
return await this.client.chat.completions.create(
|
|
90
|
+
{
|
|
91
|
+
model: args.model,
|
|
92
|
+
messages: args.messages,
|
|
93
|
+
temperature: args.temperature,
|
|
94
|
+
response_format: args.responseFormat
|
|
95
|
+
},
|
|
96
|
+
// Signal for request
|
|
97
|
+
{ signal: args.abortSignal }
|
|
98
|
+
);
|
|
99
|
+
},
|
|
100
|
+
// Signal for rate limiting
|
|
101
|
+
{ signal: args.abortSignal }
|
|
102
|
+
);
|
|
103
|
+
if ("error" in response) {
|
|
104
|
+
const err = response.error;
|
|
105
|
+
throw new Error(
|
|
106
|
+
`${err.message} - Code ${err.code} - ${JSON.stringify(err)}`
|
|
107
|
+
);
|
|
108
|
+
}
|
|
109
|
+
if (!response?.choices?.[0]?.message?.content) {
|
|
110
|
+
throw new Error("No content returned from the model");
|
|
111
|
+
}
|
|
112
|
+
return {
|
|
113
|
+
data: response.choices[0].message.content,
|
|
114
|
+
inputTokensUsed: response?.usage?.prompt_tokens,
|
|
115
|
+
outputTokensUsed: response?.usage?.completion_tokens,
|
|
116
|
+
startedAt: startedAt.getTime(),
|
|
117
|
+
completedAt: Date.now()
|
|
118
|
+
};
|
|
119
|
+
} catch (err) {
|
|
120
|
+
if (err instanceof APIError && err.status === 401) {
|
|
121
|
+
throw new Error(`Invalid credentials provided`, { cause: err });
|
|
122
|
+
}
|
|
123
|
+
retryCount--;
|
|
124
|
+
if (err instanceof SyntaxError) {
|
|
125
|
+
console.debug(err);
|
|
126
|
+
continue;
|
|
127
|
+
}
|
|
128
|
+
if (retryCount !== 0) {
|
|
129
|
+
continue;
|
|
130
|
+
}
|
|
131
|
+
throw new Error(
|
|
132
|
+
`Failed to forward prompt to the model: ${err instanceof Error ? err.message : err}`,
|
|
133
|
+
{ cause: err }
|
|
134
|
+
);
|
|
135
|
+
}
|
|
136
|
+
}
|
|
137
|
+
throw new Error(
|
|
138
|
+
`Failed to forward prompt to the model: Max retries reached`,
|
|
139
|
+
{ cause: new Error("Max retries reached") }
|
|
140
|
+
);
|
|
141
|
+
}
|
|
142
|
+
};
|
|
143
|
+
|
|
144
|
+
// src/providers/openrouter.ts
|
|
145
|
+
import Decimal from "decimal.js";
|
|
146
|
+
import axios from "axios";
|
|
147
|
+
var baseURL = "https://openrouter.ai/api/v1";
|
|
148
|
+
var MODELS_CACHE_TTL = 1e3 * 60 * 60 * 24;
|
|
149
|
+
var OpenRouterProvider = class extends AbstractLLMProvider {
|
|
150
|
+
kind = `${PEERBENCH_NAMESPACE}/llm/openrouter.ai`;
|
|
151
|
+
models = void 0;
|
|
152
|
+
modelsCachePromise = Promise.resolve(void 0);
|
|
153
|
+
modelsUpdatedAt = 0;
|
|
154
|
+
openAIProvider;
|
|
155
|
+
constructor(config) {
|
|
156
|
+
super();
|
|
157
|
+
this.openAIProvider = new OpenAIProvider({
|
|
158
|
+
baseURL,
|
|
159
|
+
apiKey: config.apiKey,
|
|
160
|
+
maxRetries: config.maxRetries,
|
|
161
|
+
timeout: config.timeout,
|
|
162
|
+
rateLimiter: config.rateLimiter
|
|
163
|
+
});
|
|
164
|
+
}
|
|
165
|
+
async forward(args) {
|
|
166
|
+
const [response] = await Promise.all([
|
|
167
|
+
this.openAIProvider.forward(args),
|
|
168
|
+
this.updateModelsCache().catch(() => {
|
|
169
|
+
})
|
|
170
|
+
]);
|
|
171
|
+
const modelInfo = this.models?.data.find((m) => m.id === args.model);
|
|
172
|
+
let inputCost = void 0;
|
|
173
|
+
let outputCost = void 0;
|
|
174
|
+
if (modelInfo !== void 0) {
|
|
175
|
+
if (response.inputTokensUsed !== void 0) {
|
|
176
|
+
inputCost = new Decimal(modelInfo.pricing.prompt).mul(response.inputTokensUsed).toFixed(10);
|
|
177
|
+
}
|
|
178
|
+
if (response.outputTokensUsed !== void 0) {
|
|
179
|
+
outputCost = new Decimal(modelInfo.pricing.completion).mul(response.outputTokensUsed).toFixed(10);
|
|
180
|
+
}
|
|
181
|
+
}
|
|
182
|
+
return {
|
|
183
|
+
...response,
|
|
184
|
+
inputCost,
|
|
185
|
+
outputCost
|
|
186
|
+
};
|
|
187
|
+
}
|
|
188
|
+
/**
|
|
189
|
+
* Updates the cache that holds information about OpenRouter models
|
|
190
|
+
* including pricing information. It will be valid for 24 hours as
|
|
191
|
+
* long as the instance of this Provider object is alive.
|
|
192
|
+
*/
|
|
193
|
+
async updateModelsCache() {
|
|
194
|
+
this.modelsCachePromise = this.modelsCachePromise.then(async () => {
|
|
195
|
+
if (
|
|
196
|
+
// The data presented in the cache
|
|
197
|
+
this.models !== void 0 && // The cache is still valid
|
|
198
|
+
Date.now() - this.modelsUpdatedAt < MODELS_CACHE_TTL
|
|
199
|
+
) {
|
|
200
|
+
return this.models;
|
|
201
|
+
}
|
|
202
|
+
return axios.get(`${baseURL}/models`).then((res) => res.data).then((data) => {
|
|
203
|
+
data = {
|
|
204
|
+
data: data.data.filter(
|
|
205
|
+
(m) => m.architecture.input_modalities.includes("text") && m.architecture.output_modalities.includes("text") && // These models are "fast apply model" and don't support multi turn conversations so don't include them
|
|
206
|
+
![
|
|
207
|
+
"morph/morph-v3-large",
|
|
208
|
+
"morph/morph-v3-fast",
|
|
209
|
+
"relace/relace-apply-3"
|
|
210
|
+
].includes(m.id)
|
|
211
|
+
)
|
|
212
|
+
};
|
|
213
|
+
this.models = data;
|
|
214
|
+
this.modelsUpdatedAt = Date.now();
|
|
215
|
+
return data;
|
|
216
|
+
});
|
|
217
|
+
}).catch(() => void 0);
|
|
218
|
+
await this.modelsCachePromise;
|
|
219
|
+
}
|
|
220
|
+
};
|
|
221
|
+
|
|
222
|
+
export {
|
|
223
|
+
AbstractProvider,
|
|
224
|
+
AbstractLLMProvider,
|
|
225
|
+
MastraProvider,
|
|
226
|
+
OpenAIProvider,
|
|
227
|
+
OpenRouterProvider
|
|
228
|
+
};
|
|
229
|
+
//# sourceMappingURL=chunk-R76XA2K6.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/providers/abstract/provider.ts","../src/providers/abstract/llm.ts","../src/providers/mastra.ts","../src/providers/openai.ts","../src/providers/openrouter.ts"],"sourcesContent":["export abstract class AbstractProvider {\n abstract readonly kind: string;\n}\n\nexport type ProviderResponse<TData = unknown> = {\n startedAt: number;\n completedAt: number;\n data: TData;\n};\n","import { AbstractProvider, ProviderResponse } from \"./provider\";\nimport {\n ResponseFormatJSONObject,\n ResponseFormatJSONSchema,\n ResponseFormatText,\n} from \"openai/resources/shared\";\nimport { ChatCompletionMessageParam } from \"openai/resources/chat/completions\";\n\nexport abstract class AbstractLLMProvider extends AbstractProvider {\n abstract forward(args: LLMProviderForwardArgs): Promise<ChatResponse>;\n}\n\nexport type LLMProviderForwardArgs = {\n messages: ChatCompletionMessageParam[];\n model: string;\n abortSignal?: AbortSignal;\n temperature?: number;\n responseFormat?:\n | ResponseFormatText\n | ResponseFormatJSONSchema\n | ResponseFormatJSONObject;\n};\n\nexport type ChatResponse = ProviderResponse<string> & {\n inputTokensUsed?: number;\n outputTokensUsed?: number;\n inputCost?: string;\n outputCost?: string;\n};\n","import {\n AbstractLLMProvider,\n type ChatResponse,\n type LLMProviderForwardArgs,\n} from \"./abstract/llm\";\nimport { MastraClient } from \"@mastra/client-js\";\n\nexport class MastraProvider extends AbstractLLMProvider {\n override readonly kind = \"mastra\";\n\n private readonly endpoint: string;\n private readonly authToken?: string;\n private client: MastraClient;\n private warnedAboutSystemMessages = false;\n private warnedAboutResponseFormat = false;\n\n constructor(params: { endpoint: string; authToken?: string }) {\n super();\n this.endpoint = params.endpoint;\n this.authToken = params.authToken;\n this.client = new MastraClient({\n baseUrl: this.endpoint,\n headers: {\n Authorization: `Bearer ${this.authToken}`,\n },\n });\n }\n\n override async forward(\n args: LLMProviderForwardArgs & {\n memory?: AgentMemoryOption;\n }\n ): Promise<ChatResponse> {\n const apiMessages = args.messages\n .filter((m) => m.role === \"user\" || m.role === \"assistant\")\n .map((m) => ({\n role: m.role as \"user\" | \"assistant\",\n content: String((m as any).content ?? \"\"),\n }));\n\n const agent = this.client.getAgent(args.model);\n\n const startedAt = Date.now();\n const response = await agent.generate(\n {\n messages: apiMessages,\n runtimeContext: {\n \"model-id\": args.model,\n },\n },\n { memory: args.memory }\n );\n\n return {\n data: response.text,\n startedAt,\n completedAt: Date.now(),\n };\n }\n}\n\n// NOTE: Mastra client does not export this type\nexport type AgentMemoryOption = Parameters<\n Parameters<MastraClient[\"getAgent\"]>[\"0\"] extends string\n ? ReturnType<MastraClient[\"getAgent\"]>[\"generate\"]\n : never\n>[0] extends { memory?: infer M }\n ? M\n : never;\n","import { RateLimiter } from \"@/utils\";\nimport { ChatCompletionMessageParam } from \"openai/resources/chat/completions\";\nimport {\n ResponseFormatJSONObject,\n ResponseFormatJSONSchema,\n ResponseFormatText,\n} from \"openai/resources/shared\";\nimport OpenAI, { APIError } from \"openai\";\nimport { AbstractLLMProvider, ChatResponse } from \"./abstract/llm\";\nimport { PEERBENCH_NAMESPACE } from \"@/constants\";\n\nexport class OpenAIProvider extends AbstractLLMProvider {\n override readonly kind = `${PEERBENCH_NAMESPACE}/llm/openai` as const;\n\n private client: OpenAI;\n private rateLimiter: RateLimiter;\n private maxRetries: number;\n\n constructor(config: {\n apiKey: string;\n baseURL: string;\n maxRetries?: number;\n timeout?: number;\n rateLimiter?: RateLimiter;\n }) {\n super();\n this.maxRetries = config.maxRetries ?? 3;\n this.rateLimiter =\n config.rateLimiter ??\n new RateLimiter({\n maxWeight: 20,\n timeWindow: 3_000,\n });\n\n this.client = new OpenAI({\n baseURL: config.baseURL,\n apiKey: config.apiKey,\n timeout: config.timeout,\n dangerouslyAllowBrowser: true,\n });\n }\n\n async forward(args: {\n messages: ChatCompletionMessageParam[];\n model: string;\n abortSignal?: AbortSignal;\n temperature?: number;\n responseFormat?:\n | ResponseFormatText\n | ResponseFormatJSONSchema\n | ResponseFormatJSONObject;\n }): Promise<ChatResponse> {\n let retryCount = this.maxRetries;\n while (retryCount > 0) {\n let startedAt: Date = new Date();\n\n try {\n const response = await this.rateLimiter.execute(\n async () => {\n // Capture the start time of the request\n startedAt = new Date();\n return await this.client.chat.completions.create(\n {\n model: args.model,\n messages: args.messages,\n temperature: args.temperature,\n response_format: args.responseFormat,\n },\n // Signal for request\n { signal: args.abortSignal }\n );\n },\n // Signal for rate limiting\n { signal: args.abortSignal }\n );\n\n if (\"error\" in response) {\n const err = response.error as any;\n throw new Error(\n `${err.message} - Code ${err.code} - ${JSON.stringify(err)}`\n );\n }\n\n if (!response?.choices?.[0]?.message?.content) {\n throw new Error(\"No content returned from the model\");\n }\n\n return {\n data: response.choices[0].message.content,\n\n inputTokensUsed: response?.usage?.prompt_tokens,\n outputTokensUsed: response?.usage?.completion_tokens,\n\n startedAt: startedAt.getTime(),\n completedAt: Date.now(),\n };\n } catch (err) {\n if (err instanceof APIError && err.status === 401) {\n throw new Error(`Invalid credentials provided`, { cause: err });\n }\n\n retryCount--;\n\n // More likely an empty HTTP response returned by the Provider\n // and it couldn't be parsed as JSON by the OpenAI SDK. We need to retry the request\n // More info can be found in the following links:\n // https://www.reddit.com/r/SillyTavernAI/comments/1ik95vr/deepseek_r1_on_openrouter_returning_blank_messages/\n // https://github.com/cline/cline/issues/60\n if (err instanceof SyntaxError) {\n console.debug(err);\n continue;\n }\n\n // If it was another error, just continue until we run out of retries\n if (retryCount !== 0) {\n continue;\n }\n\n throw new Error(\n `Failed to forward prompt to the model: ${err instanceof Error ? err.message : err}`,\n { cause: err }\n );\n }\n }\n\n throw new Error(\n `Failed to forward prompt to the model: Max retries reached`,\n { cause: new Error(\"Max retries reached\") }\n );\n }\n}\n","import {\n AbstractLLMProvider,\n ChatResponse,\n LLMProviderForwardArgs,\n} from \"./abstract/llm\";\nimport { RateLimiter } from \"@/utils\";\nimport { OpenAIProvider } from \"./openai\";\nimport { PEERBENCH_NAMESPACE } from \"@/constants\";\nimport Decimal from \"decimal.js\";\nimport axios from \"axios\";\n\nconst baseURL = \"https://openrouter.ai/api/v1\";\nconst MODELS_CACHE_TTL = 1000 * 60 * 60 * 24; // 24 hours\n\nexport class OpenRouterProvider extends AbstractLLMProvider {\n override readonly kind = `${PEERBENCH_NAMESPACE}/llm/openrouter.ai` as const;\n\n private models: ModelsResponse | undefined = undefined;\n private modelsCachePromise: Promise<ModelsResponse | undefined> =\n Promise.resolve(undefined);\n private modelsUpdatedAt = 0;\n private openAIProvider: OpenAIProvider;\n\n constructor(config: {\n apiKey: string;\n maxRetries?: number;\n timeout?: number;\n rateLimiter?: RateLimiter;\n }) {\n super();\n this.openAIProvider = new OpenAIProvider({\n baseURL,\n apiKey: config.apiKey,\n maxRetries: config.maxRetries,\n timeout: config.timeout,\n rateLimiter: config.rateLimiter,\n });\n }\n\n override async forward(args: LLMProviderForwardArgs): Promise<ChatResponse> {\n // Update models cache concurrently (non-blocking)\n const [response] = await Promise.all([\n this.openAIProvider.forward(args),\n this.updateModelsCache().catch(() => {\n // Silently fail if cache update fails so we won't have cost info in the result\n }),\n ]);\n\n // Get the model info from the cache\n const modelInfo = this.models?.data.find((m) => m.id === args.model);\n let inputCost: string | undefined = undefined;\n let outputCost: string | undefined = undefined;\n\n if (modelInfo !== undefined) {\n // Use Decimal.js for more accurate calculation\n if (response.inputTokensUsed !== undefined) {\n inputCost = new Decimal(modelInfo.pricing.prompt)\n .mul(response.inputTokensUsed)\n .toFixed(10);\n }\n if (response.outputTokensUsed !== undefined) {\n outputCost = new Decimal(modelInfo.pricing.completion)\n .mul(response.outputTokensUsed)\n .toFixed(10);\n }\n }\n\n return {\n ...response,\n inputCost,\n outputCost,\n };\n }\n\n /**\n * Updates the cache that holds information about OpenRouter models\n * including pricing information. It will be valid for 24 hours as\n * long as the instance of this Provider object is alive.\n */\n private async updateModelsCache() {\n // Chain each update method call to the promise.\n // This approach prevents race conditions between multiple calls.\n // Since each call is chained to the end of the previous one,\n // each promise makes a request only if the models cache is not updated\n // in the last call. Otherwise it simply resolves to the cached value.\n this.modelsCachePromise = this.modelsCachePromise\n .then(async () => {\n if (\n // The data presented in the cache\n this.models !== undefined &&\n // The cache is still valid\n Date.now() - this.modelsUpdatedAt < MODELS_CACHE_TTL\n ) {\n return this.models;\n }\n\n // If the cache is not valid, update it\n return axios\n .get<ModelsResponse>(`${baseURL}/models`)\n .then((res) => res.data)\n .then((data) => {\n // Only get the models that supports text input and output\n data = {\n data: data.data.filter(\n (m) =>\n m.architecture.input_modalities.includes(\"text\") &&\n m.architecture.output_modalities.includes(\"text\") &&\n // These models are \"fast apply model\" and don't support multi turn conversations so don't include them\n ![\n \"morph/morph-v3-large\",\n \"morph/morph-v3-fast\",\n \"relace/relace-apply-3\",\n ].includes(m.id)\n ),\n };\n\n this.models = data;\n this.modelsUpdatedAt = Date.now();\n\n return data;\n });\n })\n .catch(() => undefined);\n\n // Wait for the promise chain to resolve\n await this.modelsCachePromise;\n }\n}\n\ntype PutModality = \"text\" | \"image\" | \"file\" | \"audio\";\ntype Modality = \"text->text\" | \"text+image->text\" | \"text+image->text+image\";\ntype ModelsResponse = {\n data: {\n readonly id: string;\n readonly canonical_slug: string;\n readonly hugging_face_id: null | string;\n readonly name: string;\n readonly created: number;\n readonly description: string;\n readonly context_length: number;\n readonly architecture: {\n readonly modality: Modality;\n readonly input_modalities: PutModality[];\n readonly output_modalities: PutModality[];\n readonly instruct_type: null | string;\n };\n readonly pricing: {\n readonly prompt: string;\n readonly completion: string;\n readonly request?: string;\n readonly image?: string;\n readonly web_search?: string;\n readonly internal_reasoning?: string;\n readonly input_cache_read?: string;\n readonly input_cache_write?: string;\n readonly audio?: string;\n };\n }[];\n};\n"],"mappings":";;;;;;;;AAAO,IAAe,mBAAf,MAAgC;AAEvC;;;ACMO,IAAe,sBAAf,cAA2C,iBAAiB;AAEnE;;;ACLA,SAAS,oBAAoB;AAEtB,IAAM,iBAAN,cAA6B,oBAAoB;AAAA,EACpC,OAAO;AAAA,EAER;AAAA,EACA;AAAA,EACT;AAAA,EACA,4BAA4B;AAAA,EAC5B,4BAA4B;AAAA,EAEpC,YAAY,QAAkD;AAC5D,UAAM;AACN,SAAK,WAAW,OAAO;AACvB,SAAK,YAAY,OAAO;AACxB,SAAK,SAAS,IAAI,aAAa;AAAA,MAC7B,SAAS,KAAK;AAAA,MACd,SAAS;AAAA,QACP,eAAe,UAAU,KAAK,SAAS;AAAA,MACzC;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAe,QACb,MAGuB;AACvB,UAAM,cAAc,KAAK,SACtB,OAAO,CAAC,MAAM,EAAE,SAAS,UAAU,EAAE,SAAS,WAAW,EACzD,IAAI,CAAC,OAAO;AAAA,MACX,MAAM,EAAE;AAAA,MACR,SAAS,OAAQ,EAAU,WAAW,EAAE;AAAA,IAC1C,EAAE;AAEJ,UAAM,QAAQ,KAAK,OAAO,SAAS,KAAK,KAAK;AAE7C,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM,WAAW,MAAM,MAAM;AAAA,MAC3B;AAAA,QACE,UAAU;AAAA,QACV,gBAAgB;AAAA,UACd,YAAY,KAAK;AAAA,QACnB;AAAA,MACF;AAAA,MACA,EAAE,QAAQ,KAAK,OAAO;AAAA,IACxB;AAEA,WAAO;AAAA,MACL,MAAM,SAAS;AAAA,MACf;AAAA,MACA,aAAa,KAAK,IAAI;AAAA,IACxB;AAAA,EACF;AACF;;;ACpDA,OAAO,UAAU,gBAAgB;AAI1B,IAAM,iBAAN,cAA6B,oBAAoB;AAAA,EACpC,OAAO,GAAG,mBAAmB;AAAA,EAEvC;AAAA,EACA;AAAA,EACA;AAAA,EAER,YAAY,QAMT;AACD,UAAM;AACN,SAAK,aAAa,OAAO,cAAc;AACvC,SAAK,cACH,OAAO,eACP,IAAI,YAAY;AAAA,MACd,WAAW;AAAA,MACX,YAAY;AAAA,IACd,CAAC;AAEH,SAAK,SAAS,IAAI,OAAO;AAAA,MACvB,SAAS,OAAO;AAAA,MAChB,QAAQ,OAAO;AAAA,MACf,SAAS,OAAO;AAAA,MAChB,yBAAyB;AAAA,IAC3B,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,QAAQ,MASY;AACxB,QAAI,aAAa,KAAK;AACtB,WAAO,aAAa,GAAG;AACrB,UAAI,YAAkB,oBAAI,KAAK;AAE/B,UAAI;AACF,cAAM,WAAW,MAAM,KAAK,YAAY;AAAA,UACtC,YAAY;AAEV,wBAAY,oBAAI,KAAK;AACrB,mBAAO,MAAM,KAAK,OAAO,KAAK,YAAY;AAAA,cACxC;AAAA,gBACE,OAAO,KAAK;AAAA,gBACZ,UAAU,KAAK;AAAA,gBACf,aAAa,KAAK;AAAA,gBAClB,iBAAiB,KAAK;AAAA,cACxB;AAAA;AAAA,cAEA,EAAE,QAAQ,KAAK,YAAY;AAAA,YAC7B;AAAA,UACF;AAAA;AAAA,UAEA,EAAE,QAAQ,KAAK,YAAY;AAAA,QAC7B;AAEA,YAAI,WAAW,UAAU;AACvB,gBAAM,MAAM,SAAS;AACrB,gBAAM,IAAI;AAAA,YACR,GAAG,IAAI,OAAO,WAAW,IAAI,IAAI,MAAM,KAAK,UAAU,GAAG,CAAC;AAAA,UAC5D;AAAA,QACF;AAEA,YAAI,CAAC,UAAU,UAAU,CAAC,GAAG,SAAS,SAAS;AAC7C,gBAAM,IAAI,MAAM,oCAAoC;AAAA,QACtD;AAEA,eAAO;AAAA,UACL,MAAM,SAAS,QAAQ,CAAC,EAAE,QAAQ;AAAA,UAElC,iBAAiB,UAAU,OAAO;AAAA,UAClC,kBAAkB,UAAU,OAAO;AAAA,UAEnC,WAAW,UAAU,QAAQ;AAAA,UAC7B,aAAa,KAAK,IAAI;AAAA,QACxB;AAAA,MACF,SAAS,KAAK;AACZ,YAAI,eAAe,YAAY,IAAI,WAAW,KAAK;AACjD,gBAAM,IAAI,MAAM,gCAAgC,EAAE,OAAO,IAAI,CAAC;AAAA,QAChE;AAEA;AAOA,YAAI,eAAe,aAAa;AAC9B,kBAAQ,MAAM,GAAG;AACjB;AAAA,QACF;AAGA,YAAI,eAAe,GAAG;AACpB;AAAA,QACF;AAEA,cAAM,IAAI;AAAA,UACR,0CAA0C,eAAe,QAAQ,IAAI,UAAU,GAAG;AAAA,UAClF,EAAE,OAAO,IAAI;AAAA,QACf;AAAA,MACF;AAAA,IACF;AAEA,UAAM,IAAI;AAAA,MACR;AAAA,MACA,EAAE,OAAO,IAAI,MAAM,qBAAqB,EAAE;AAAA,IAC5C;AAAA,EACF;AACF;;;AC1HA,OAAO,aAAa;AACpB,OAAO,WAAW;AAElB,IAAM,UAAU;AAChB,IAAM,mBAAmB,MAAO,KAAK,KAAK;AAEnC,IAAM,qBAAN,cAAiC,oBAAoB;AAAA,EACxC,OAAO,GAAG,mBAAmB;AAAA,EAEvC,SAAqC;AAAA,EACrC,qBACN,QAAQ,QAAQ,MAAS;AAAA,EACnB,kBAAkB;AAAA,EAClB;AAAA,EAER,YAAY,QAKT;AACD,UAAM;AACN,SAAK,iBAAiB,IAAI,eAAe;AAAA,MACvC;AAAA,MACA,QAAQ,OAAO;AAAA,MACf,YAAY,OAAO;AAAA,MACnB,SAAS,OAAO;AAAA,MAChB,aAAa,OAAO;AAAA,IACtB,CAAC;AAAA,EACH;AAAA,EAEA,MAAe,QAAQ,MAAqD;AAE1E,UAAM,CAAC,QAAQ,IAAI,MAAM,QAAQ,IAAI;AAAA,MACnC,KAAK,eAAe,QAAQ,IAAI;AAAA,MAChC,KAAK,kBAAkB,EAAE,MAAM,MAAM;AAAA,MAErC,CAAC;AAAA,IACH,CAAC;AAGD,UAAM,YAAY,KAAK,QAAQ,KAAK,KAAK,CAAC,MAAM,EAAE,OAAO,KAAK,KAAK;AACnE,QAAI,YAAgC;AACpC,QAAI,aAAiC;AAErC,QAAI,cAAc,QAAW;AAE3B,UAAI,SAAS,oBAAoB,QAAW;AAC1C,oBAAY,IAAI,QAAQ,UAAU,QAAQ,MAAM,EAC7C,IAAI,SAAS,eAAe,EAC5B,QAAQ,EAAE;AAAA,MACf;AACA,UAAI,SAAS,qBAAqB,QAAW;AAC3C,qBAAa,IAAI,QAAQ,UAAU,QAAQ,UAAU,EAClD,IAAI,SAAS,gBAAgB,EAC7B,QAAQ,EAAE;AAAA,MACf;AAAA,IACF;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOA,MAAc,oBAAoB;AAMhC,SAAK,qBAAqB,KAAK,mBAC5B,KAAK,YAAY;AAChB;AAAA;AAAA,QAEE,KAAK,WAAW;AAAA,QAEhB,KAAK,IAAI,IAAI,KAAK,kBAAkB;AAAA,QACpC;AACA,eAAO,KAAK;AAAA,MACd;AAGA,aAAO,MACJ,IAAoB,GAAG,OAAO,SAAS,EACvC,KAAK,CAAC,QAAQ,IAAI,IAAI,EACtB,KAAK,CAAC,SAAS;AAEd,eAAO;AAAA,UACL,MAAM,KAAK,KAAK;AAAA,YACd,CAAC,MACC,EAAE,aAAa,iBAAiB,SAAS,MAAM,KAC/C,EAAE,aAAa,kBAAkB,SAAS,MAAM;AAAA,YAEhD,CAAC;AAAA,cACC;AAAA,cACA;AAAA,cACA;AAAA,YACF,EAAE,SAAS,EAAE,EAAE;AAAA,UACnB;AAAA,QACF;AAEA,aAAK,SAAS;AACd,aAAK,kBAAkB,KAAK,IAAI;AAEhC,eAAO;AAAA,MACT,CAAC;AAAA,IACL,CAAC,EACA,MAAM,MAAM,MAAS;AAGxB,UAAM,KAAK;AAAA,EACb;AACF;","names":[]}
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
import {
|
|
2
|
+
IdSchema
|
|
3
|
+
} from "./chunk-NUEOE3K5.js";
|
|
4
|
+
|
|
5
|
+
// src/schemas/extensions/response/llm.ts
|
|
6
|
+
import z from "zod";
|
|
7
|
+
var ExtensionLLMResponseFieldsV1 = {
|
|
8
|
+
data: z.string(),
|
|
9
|
+
modelSlug: z.string(),
|
|
10
|
+
provider: z.string(),
|
|
11
|
+
systemPromptId: IdSchema.optional(),
|
|
12
|
+
inputTokensUsed: z.number().optional(),
|
|
13
|
+
outputTokensUsed: z.number().optional(),
|
|
14
|
+
inputCost: z.string().optional(),
|
|
15
|
+
outputCost: z.string().optional()
|
|
16
|
+
};
|
|
17
|
+
|
|
18
|
+
// src/schemas/extensions/score/llm-as-a-judge-scorer.ts
|
|
19
|
+
import z2 from "zod";
|
|
20
|
+
var ExtensionLLMAsAJudgeScoreFieldsV1 = {
|
|
21
|
+
scorerAISystemPrompt: z2.string().optional(),
|
|
22
|
+
scorerAISystemPromptId: z2.string().optional(),
|
|
23
|
+
scorerAIProvider: z2.string().optional(),
|
|
24
|
+
scorerAIModelSlug: z2.string().optional(),
|
|
25
|
+
scorerAIInputTokensUsed: z2.number().optional(),
|
|
26
|
+
scorerAIOutputTokensUsed: z2.number().optional(),
|
|
27
|
+
scorerAIInputCost: z2.string().optional(),
|
|
28
|
+
scorerAIOutputCost: z2.string().optional()
|
|
29
|
+
};
|
|
30
|
+
|
|
31
|
+
export {
|
|
32
|
+
ExtensionLLMResponseFieldsV1,
|
|
33
|
+
ExtensionLLMAsAJudgeScoreFieldsV1
|
|
34
|
+
};
|
|
35
|
+
//# sourceMappingURL=chunk-TRNCF2BG.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/schemas/extensions/response/llm.ts","../src/schemas/extensions/score/llm-as-a-judge-scorer.ts"],"sourcesContent":["import { IdSchema } from \"@/schemas/id\";\nimport z from \"zod\";\n\n/**\n * Provides a set of fields that holds information about the LLM and its response.\n */\nexport const ExtensionLLMResponseFieldsV1 = {\n data: z.string(),\n modelSlug: z.string(),\n provider: z.string(),\n systemPromptId: IdSchema.optional(),\n\n inputTokensUsed: z.number().optional(),\n outputTokensUsed: z.number().optional(),\n inputCost: z.string().optional(),\n outputCost: z.string().optional(),\n};\n","import z from \"zod\";\n\n/**\n * Provides a set of fields that holds information about the LLM model\n * that was used to judge the response.\n */\nexport const ExtensionLLMAsAJudgeScoreFieldsV1 = {\n scorerAISystemPrompt: z.string().optional(),\n scorerAISystemPromptId: z.string().optional(),\n scorerAIProvider: z.string().optional(),\n scorerAIModelSlug: z.string().optional(),\n scorerAIInputTokensUsed: z.number().optional(),\n scorerAIOutputTokensUsed: z.number().optional(),\n scorerAIInputCost: z.string().optional(),\n scorerAIOutputCost: z.string().optional(),\n};\n"],"mappings":";;;;;AACA,OAAO,OAAO;AAKP,IAAM,+BAA+B;AAAA,EAC1C,MAAM,EAAE,OAAO;AAAA,EACf,WAAW,EAAE,OAAO;AAAA,EACpB,UAAU,EAAE,OAAO;AAAA,EACnB,gBAAgB,SAAS,SAAS;AAAA,EAElC,iBAAiB,EAAE,OAAO,EAAE,SAAS;AAAA,EACrC,kBAAkB,EAAE,OAAO,EAAE,SAAS;AAAA,EACtC,WAAW,EAAE,OAAO,EAAE,SAAS;AAAA,EAC/B,YAAY,EAAE,OAAO,EAAE,SAAS;AAClC;;;AChBA,OAAOA,QAAO;AAMP,IAAM,oCAAoC;AAAA,EAC/C,sBAAsBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC1C,wBAAwBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC5C,kBAAkBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACtC,mBAAmBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACvC,yBAAyBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC7C,0BAA0BA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC9C,mBAAmBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACvC,oBAAoBA,GAAE,OAAO,EAAE,SAAS;AAC1C;","names":["z"]}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/constants.ts"],"sourcesContent":["export const PEERBENCH_NAMESPACE = \"peerbench.ai\" as const;\nexport const CATEGORIES = {\n LLM: \"llm\",\n};\n"],"mappings":";AAAO,IAAM,sBAAsB;AAC5B,IAAM,aAAa;AAAA,EACxB,KAAK;AACP;","names":[]}
|
|
@@ -1,34 +1,49 @@
|
|
|
1
|
+
import {
|
|
2
|
+
buildSchemaDefiner
|
|
3
|
+
} from "./chunk-OQE6TQXZ.js";
|
|
1
4
|
import {
|
|
2
5
|
ScoringMethod
|
|
3
|
-
} from "./chunk-
|
|
6
|
+
} from "./chunk-HMQYGCKI.js";
|
|
4
7
|
import {
|
|
5
|
-
IdSchema
|
|
6
|
-
|
|
7
|
-
} from "./chunk-IUN2IUCS.js";
|
|
8
|
+
IdSchema
|
|
9
|
+
} from "./chunk-NUEOE3K5.js";
|
|
8
10
|
|
|
9
|
-
// src/schemas/
|
|
11
|
+
// src/schemas/response.ts
|
|
10
12
|
import z from "zod";
|
|
11
|
-
var
|
|
12
|
-
|
|
13
|
+
var BaseResponseSchemaV1 = z.object({
|
|
14
|
+
id: IdSchema,
|
|
15
|
+
namespace: z.string(),
|
|
13
16
|
schemaVersion: z.number(),
|
|
17
|
+
kind: z.string(),
|
|
18
|
+
startedAt: z.number(),
|
|
19
|
+
completedAt: z.number(),
|
|
20
|
+
testCaseId: IdSchema,
|
|
14
21
|
metadata: z.record(z.string(), z.unknown()).optional()
|
|
15
22
|
});
|
|
16
|
-
var
|
|
23
|
+
var defineResponseSchema = buildSchemaDefiner(
|
|
24
|
+
BaseResponseSchemaV1,
|
|
25
|
+
"rs"
|
|
26
|
+
);
|
|
17
27
|
|
|
18
28
|
// src/schemas/test-case.ts
|
|
19
29
|
import { z as z2 } from "zod";
|
|
20
30
|
var BaseTestCaseSchemaV1 = z2.object({
|
|
21
31
|
id: IdSchema,
|
|
22
|
-
|
|
32
|
+
namespace: z2.string(),
|
|
23
33
|
schemaVersion: z2.number(),
|
|
34
|
+
kind: z2.string(),
|
|
24
35
|
metadata: z2.record(z2.string(), z2.unknown()).optional()
|
|
25
36
|
});
|
|
26
|
-
var defineTestCaseSchema = buildSchemaDefiner(
|
|
37
|
+
var defineTestCaseSchema = buildSchemaDefiner(
|
|
38
|
+
BaseTestCaseSchemaV1,
|
|
39
|
+
"tc"
|
|
40
|
+
);
|
|
27
41
|
|
|
28
42
|
// src/schemas/score.ts
|
|
29
43
|
import z3 from "zod";
|
|
30
44
|
var BaseScoreSchemaV1 = z3.object({
|
|
31
45
|
id: IdSchema,
|
|
46
|
+
namespace: z3.string(),
|
|
32
47
|
kind: z3.string(),
|
|
33
48
|
schemaVersion: z3.number(),
|
|
34
49
|
value: z3.number(),
|
|
@@ -37,14 +52,14 @@ var BaseScoreSchemaV1 = z3.object({
|
|
|
37
52
|
metadata: z3.record(z3.string(), z3.unknown()).optional(),
|
|
38
53
|
scoringMethod: z3.enum(ScoringMethod)
|
|
39
54
|
});
|
|
40
|
-
var defineScoreSchema = buildSchemaDefiner();
|
|
55
|
+
var defineScoreSchema = buildSchemaDefiner(BaseScoreSchemaV1, "sc");
|
|
41
56
|
|
|
42
57
|
export {
|
|
43
|
-
|
|
44
|
-
|
|
58
|
+
BaseResponseSchemaV1,
|
|
59
|
+
defineResponseSchema,
|
|
45
60
|
BaseTestCaseSchemaV1,
|
|
46
61
|
defineTestCaseSchema,
|
|
47
62
|
BaseScoreSchemaV1,
|
|
48
63
|
defineScoreSchema
|
|
49
64
|
};
|
|
50
|
-
//# sourceMappingURL=chunk-
|
|
65
|
+
//# sourceMappingURL=chunk-YY33MNMV.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/schemas/response.ts","../src/schemas/test-case.ts","../src/schemas/score.ts"],"sourcesContent":["import { IdSchema } from \"./id\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\nimport z from \"zod\";\n\nexport const BaseResponseSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n schemaVersion: z.number(),\n kind: z.string(),\n\n startedAt: z.number(),\n completedAt: z.number(),\n testCaseId: IdSchema,\n metadata: z.record(z.string(), z.unknown()).optional(),\n});\nexport type BaseResponseV1 = z.infer<typeof BaseResponseSchemaV1>;\n\nexport const defineResponseSchema = buildSchemaDefiner(\n BaseResponseSchemaV1,\n \"rs\"\n);\n","import { IdSchema } from \"./id\";\nimport { z } from \"zod\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\n\nexport const BaseTestCaseSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n schemaVersion: z.number(),\n kind: z.string(),\n metadata: z.record(z.string(), z.unknown()).optional(),\n});\nexport type BaseTestCaseV1 = z.infer<typeof BaseTestCaseSchemaV1>;\n\nexport const defineTestCaseSchema = buildSchemaDefiner(\n BaseTestCaseSchemaV1,\n \"tc\"\n);\n","import { IdSchema } from \"./id\";\nimport { ScoringMethod } from \"@/types\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\nimport z from \"zod\";\n\nexport const BaseScoreSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n kind: z.string(),\n schemaVersion: z.number(),\n\n value: z.number(),\n responseId: IdSchema,\n explanation: z.string().optional(),\n metadata: z.record(z.string(), z.unknown()).optional(),\n scoringMethod: z.enum(ScoringMethod),\n});\nexport type BaseScoreV1 = z.infer<typeof BaseScoreSchemaV1>;\n\nexport const defineScoreSchema = buildSchemaDefiner(BaseScoreSchemaV1, \"sc\");\n"],"mappings":";;;;;;;;;;;AAEA,OAAO,OAAO;AAEP,IAAM,uBAAuB,EAAE,OAAO;AAAA,EAC3C,IAAI;AAAA,EACJ,WAAW,EAAE,OAAO;AAAA,EACpB,eAAe,EAAE,OAAO;AAAA,EACxB,MAAM,EAAE,OAAO;AAAA,EAEf,WAAW,EAAE,OAAO;AAAA,EACpB,aAAa,EAAE,OAAO;AAAA,EACtB,YAAY;AAAA,EACZ,UAAU,EAAE,OAAO,EAAE,OAAO,GAAG,EAAE,QAAQ,CAAC,EAAE,SAAS;AACvD,CAAC;AAGM,IAAM,uBAAuB;AAAA,EAClC;AAAA,EACA;AACF;;;ACnBA,SAAS,KAAAA,UAAS;AAGX,IAAM,uBAAuBC,GAAE,OAAO;AAAA,EAC3C,IAAI;AAAA,EACJ,WAAWA,GAAE,OAAO;AAAA,EACpB,eAAeA,GAAE,OAAO;AAAA,EACxB,MAAMA,GAAE,OAAO;AAAA,EACf,UAAUA,GAAE,OAAOA,GAAE,OAAO,GAAGA,GAAE,QAAQ,CAAC,EAAE,SAAS;AACvD,CAAC;AAGM,IAAM,uBAAuB;AAAA,EAClC;AAAA,EACA;AACF;;;ACbA,OAAOC,QAAO;AAEP,IAAM,oBAAoBA,GAAE,OAAO;AAAA,EACxC,IAAI;AAAA,EACJ,WAAWA,GAAE,OAAO;AAAA,EACpB,MAAMA,GAAE,OAAO;AAAA,EACf,eAAeA,GAAE,OAAO;AAAA,EAExB,OAAOA,GAAE,OAAO;AAAA,EAChB,YAAY;AAAA,EACZ,aAAaA,GAAE,OAAO,EAAE,SAAS;AAAA,EACjC,UAAUA,GAAE,OAAOA,GAAE,OAAO,GAAGA,GAAE,QAAQ,CAAC,EAAE,SAAS;AAAA,EACrD,eAAeA,GAAE,KAAK,aAAa;AACrC,CAAC;AAGM,IAAM,oBAAoB,mBAAmB,mBAAmB,IAAI;","names":["z","z","z"]}
|