peerbench 0.0.2-alpha.0 → 0.0.2-alpha.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (80) hide show
  1. package/README.md +123 -99
  2. package/dist/aggregators/index.d.ts +67 -0
  3. package/dist/aggregators/index.js +46 -0
  4. package/dist/aggregators/index.js.map +1 -0
  5. package/dist/benchmarks/index.d.ts +615 -1271
  6. package/dist/benchmarks/index.js +358 -805
  7. package/dist/benchmarks/index.js.map +1 -1
  8. package/dist/{chunk-DUBKY73H.js → chunk-4UBK6452.js} +13 -13
  9. package/dist/chunk-4UBK6452.js.map +1 -0
  10. package/dist/chunk-ERALDEZY.js +112 -0
  11. package/dist/chunk-ERALDEZY.js.map +1 -0
  12. package/dist/{chunk-ZJWSK4VO.js → chunk-HMQYGCKI.js} +1 -1
  13. package/dist/chunk-HMQYGCKI.js.map +1 -0
  14. package/dist/chunk-NUEOE3K5.js +8 -0
  15. package/dist/chunk-NUEOE3K5.js.map +1 -0
  16. package/dist/chunk-OQE6TQXZ.js +42 -0
  17. package/dist/chunk-OQE6TQXZ.js.map +1 -0
  18. package/dist/chunk-QY5MPNNB.js +28 -0
  19. package/dist/chunk-QY5MPNNB.js.map +1 -0
  20. package/dist/chunk-R76XA2K6.js +229 -0
  21. package/dist/chunk-R76XA2K6.js.map +1 -0
  22. package/dist/chunk-TRNCF2BG.js +35 -0
  23. package/dist/chunk-TRNCF2BG.js.map +1 -0
  24. package/dist/chunk-UHHHSYVE.js +11 -0
  25. package/dist/chunk-UHHHSYVE.js.map +1 -0
  26. package/dist/{chunk-232PY7K3.js → chunk-YY33MNMV.js} +29 -14
  27. package/dist/chunk-YY33MNMV.js.map +1 -0
  28. package/dist/chunk-ZEWI24CV.js +365 -0
  29. package/dist/chunk-ZEWI24CV.js.map +1 -0
  30. package/dist/chunk-ZXTQJFGL.js +44 -0
  31. package/dist/chunk-ZXTQJFGL.js.map +1 -0
  32. package/dist/index-BAioQhp2.d.ts +27 -0
  33. package/dist/index.d.ts +51 -26
  34. package/dist/index.js +28 -25
  35. package/dist/index.js.map +1 -1
  36. package/dist/json-file-ZwzLUbje.d.ts +73 -0
  37. package/dist/llm-judge-QThCZ9TQ.d.ts +67 -0
  38. package/dist/providers/index.d.ts +16 -19
  39. package/dist/providers/index.js +8 -253
  40. package/dist/providers/index.js.map +1 -1
  41. package/dist/schemas/extensions/index.d.ts +16 -2
  42. package/dist/schemas/extensions/index.js +9 -3
  43. package/dist/schemas/extensions/index.js.map +1 -1
  44. package/dist/schemas/index.d.ts +108 -141
  45. package/dist/schemas/index.js +7 -10
  46. package/dist/schemas/llm/index.d.ts +100 -82
  47. package/dist/schemas/llm/index.js +7 -29
  48. package/dist/schemas/llm/index.js.map +1 -1
  49. package/dist/scorers/index.d.ts +3 -2
  50. package/dist/scorers/index.js +8 -486
  51. package/dist/scorers/index.js.map +1 -1
  52. package/dist/storages/index.d.ts +69 -0
  53. package/dist/storages/index.js +98 -0
  54. package/dist/storages/index.js.map +1 -0
  55. package/package.json +12 -6
  56. package/dist/catalogs/index.d.ts +0 -75
  57. package/dist/catalogs/index.js +0 -88
  58. package/dist/catalogs/index.js.map +0 -1
  59. package/dist/chunk-22HU24QF.js +0 -8
  60. package/dist/chunk-22HU24QF.js.map +0 -1
  61. package/dist/chunk-232PY7K3.js.map +0 -1
  62. package/dist/chunk-7TREBPSJ.js +0 -26
  63. package/dist/chunk-7TREBPSJ.js.map +0 -1
  64. package/dist/chunk-DUBKY73H.js.map +0 -1
  65. package/dist/chunk-GVF4YZF3.js +0 -15
  66. package/dist/chunk-GVF4YZF3.js.map +0 -1
  67. package/dist/chunk-HJH3SW3L.js +0 -103
  68. package/dist/chunk-HJH3SW3L.js.map +0 -1
  69. package/dist/chunk-IUN2IUCS.js +0 -58
  70. package/dist/chunk-IUN2IUCS.js.map +0 -1
  71. package/dist/chunk-VBOM2YEG.js +0 -47
  72. package/dist/chunk-VBOM2YEG.js.map +0 -1
  73. package/dist/chunk-ZJWSK4VO.js.map +0 -1
  74. package/dist/data-BmN5WjZ4.d.ts +0 -57
  75. package/dist/generic-array-DLHWSvf1.d.ts +0 -22
  76. package/dist/index-WiPjF2AL.d.ts +0 -15
  77. package/dist/llm-judge-DIG1f1Az.d.ts +0 -67
  78. package/dist/simple-system-prompt-CzPYuvo0.d.ts +0 -49
  79. package/dist/system-prompt--0FdPWqK.d.ts +0 -58
  80. package/dist/utilities-BrRH32rD.d.ts +0 -30
@@ -0,0 +1,229 @@
1
+ import {
2
+ PEERBENCH_NAMESPACE
3
+ } from "./chunk-UHHHSYVE.js";
4
+ import {
5
+ RateLimiter
6
+ } from "./chunk-4UBK6452.js";
7
+
8
+ // src/providers/abstract/provider.ts
9
+ var AbstractProvider = class {
10
+ };
11
+
12
+ // src/providers/abstract/llm.ts
13
+ var AbstractLLMProvider = class extends AbstractProvider {
14
+ };
15
+
16
+ // src/providers/mastra.ts
17
+ import { MastraClient } from "@mastra/client-js";
18
+ var MastraProvider = class extends AbstractLLMProvider {
19
+ kind = "mastra";
20
+ endpoint;
21
+ authToken;
22
+ client;
23
+ warnedAboutSystemMessages = false;
24
+ warnedAboutResponseFormat = false;
25
+ constructor(params) {
26
+ super();
27
+ this.endpoint = params.endpoint;
28
+ this.authToken = params.authToken;
29
+ this.client = new MastraClient({
30
+ baseUrl: this.endpoint,
31
+ headers: {
32
+ Authorization: `Bearer ${this.authToken}`
33
+ }
34
+ });
35
+ }
36
+ async forward(args) {
37
+ const apiMessages = args.messages.filter((m) => m.role === "user" || m.role === "assistant").map((m) => ({
38
+ role: m.role,
39
+ content: String(m.content ?? "")
40
+ }));
41
+ const agent = this.client.getAgent(args.model);
42
+ const startedAt = Date.now();
43
+ const response = await agent.generate(
44
+ {
45
+ messages: apiMessages,
46
+ runtimeContext: {
47
+ "model-id": args.model
48
+ }
49
+ },
50
+ { memory: args.memory }
51
+ );
52
+ return {
53
+ data: response.text,
54
+ startedAt,
55
+ completedAt: Date.now()
56
+ };
57
+ }
58
+ };
59
+
60
+ // src/providers/openai.ts
61
+ import OpenAI, { APIError } from "openai";
62
+ var OpenAIProvider = class extends AbstractLLMProvider {
63
+ kind = `${PEERBENCH_NAMESPACE}/llm/openai`;
64
+ client;
65
+ rateLimiter;
66
+ maxRetries;
67
+ constructor(config) {
68
+ super();
69
+ this.maxRetries = config.maxRetries ?? 3;
70
+ this.rateLimiter = config.rateLimiter ?? new RateLimiter({
71
+ maxWeight: 20,
72
+ timeWindow: 3e3
73
+ });
74
+ this.client = new OpenAI({
75
+ baseURL: config.baseURL,
76
+ apiKey: config.apiKey,
77
+ timeout: config.timeout,
78
+ dangerouslyAllowBrowser: true
79
+ });
80
+ }
81
+ async forward(args) {
82
+ let retryCount = this.maxRetries;
83
+ while (retryCount > 0) {
84
+ let startedAt = /* @__PURE__ */ new Date();
85
+ try {
86
+ const response = await this.rateLimiter.execute(
87
+ async () => {
88
+ startedAt = /* @__PURE__ */ new Date();
89
+ return await this.client.chat.completions.create(
90
+ {
91
+ model: args.model,
92
+ messages: args.messages,
93
+ temperature: args.temperature,
94
+ response_format: args.responseFormat
95
+ },
96
+ // Signal for request
97
+ { signal: args.abortSignal }
98
+ );
99
+ },
100
+ // Signal for rate limiting
101
+ { signal: args.abortSignal }
102
+ );
103
+ if ("error" in response) {
104
+ const err = response.error;
105
+ throw new Error(
106
+ `${err.message} - Code ${err.code} - ${JSON.stringify(err)}`
107
+ );
108
+ }
109
+ if (!response?.choices?.[0]?.message?.content) {
110
+ throw new Error("No content returned from the model");
111
+ }
112
+ return {
113
+ data: response.choices[0].message.content,
114
+ inputTokensUsed: response?.usage?.prompt_tokens,
115
+ outputTokensUsed: response?.usage?.completion_tokens,
116
+ startedAt: startedAt.getTime(),
117
+ completedAt: Date.now()
118
+ };
119
+ } catch (err) {
120
+ if (err instanceof APIError && err.status === 401) {
121
+ throw new Error(`Invalid credentials provided`, { cause: err });
122
+ }
123
+ retryCount--;
124
+ if (err instanceof SyntaxError) {
125
+ console.debug(err);
126
+ continue;
127
+ }
128
+ if (retryCount !== 0) {
129
+ continue;
130
+ }
131
+ throw new Error(
132
+ `Failed to forward prompt to the model: ${err instanceof Error ? err.message : err}`,
133
+ { cause: err }
134
+ );
135
+ }
136
+ }
137
+ throw new Error(
138
+ `Failed to forward prompt to the model: Max retries reached`,
139
+ { cause: new Error("Max retries reached") }
140
+ );
141
+ }
142
+ };
143
+
144
+ // src/providers/openrouter.ts
145
+ import Decimal from "decimal.js";
146
+ import axios from "axios";
147
+ var baseURL = "https://openrouter.ai/api/v1";
148
+ var MODELS_CACHE_TTL = 1e3 * 60 * 60 * 24;
149
+ var OpenRouterProvider = class extends AbstractLLMProvider {
150
+ kind = `${PEERBENCH_NAMESPACE}/llm/openrouter.ai`;
151
+ models = void 0;
152
+ modelsCachePromise = Promise.resolve(void 0);
153
+ modelsUpdatedAt = 0;
154
+ openAIProvider;
155
+ constructor(config) {
156
+ super();
157
+ this.openAIProvider = new OpenAIProvider({
158
+ baseURL,
159
+ apiKey: config.apiKey,
160
+ maxRetries: config.maxRetries,
161
+ timeout: config.timeout,
162
+ rateLimiter: config.rateLimiter
163
+ });
164
+ }
165
+ async forward(args) {
166
+ const [response] = await Promise.all([
167
+ this.openAIProvider.forward(args),
168
+ this.updateModelsCache().catch(() => {
169
+ })
170
+ ]);
171
+ const modelInfo = this.models?.data.find((m) => m.id === args.model);
172
+ let inputCost = void 0;
173
+ let outputCost = void 0;
174
+ if (modelInfo !== void 0) {
175
+ if (response.inputTokensUsed !== void 0) {
176
+ inputCost = new Decimal(modelInfo.pricing.prompt).mul(response.inputTokensUsed).toFixed(10);
177
+ }
178
+ if (response.outputTokensUsed !== void 0) {
179
+ outputCost = new Decimal(modelInfo.pricing.completion).mul(response.outputTokensUsed).toFixed(10);
180
+ }
181
+ }
182
+ return {
183
+ ...response,
184
+ inputCost,
185
+ outputCost
186
+ };
187
+ }
188
+ /**
189
+ * Updates the cache that holds information about OpenRouter models
190
+ * including pricing information. It will be valid for 24 hours as
191
+ * long as the instance of this Provider object is alive.
192
+ */
193
+ async updateModelsCache() {
194
+ this.modelsCachePromise = this.modelsCachePromise.then(async () => {
195
+ if (
196
+ // The data presented in the cache
197
+ this.models !== void 0 && // The cache is still valid
198
+ Date.now() - this.modelsUpdatedAt < MODELS_CACHE_TTL
199
+ ) {
200
+ return this.models;
201
+ }
202
+ return axios.get(`${baseURL}/models`).then((res) => res.data).then((data) => {
203
+ data = {
204
+ data: data.data.filter(
205
+ (m) => m.architecture.input_modalities.includes("text") && m.architecture.output_modalities.includes("text") && // These models are "fast apply model" and don't support multi turn conversations so don't include them
206
+ ![
207
+ "morph/morph-v3-large",
208
+ "morph/morph-v3-fast",
209
+ "relace/relace-apply-3"
210
+ ].includes(m.id)
211
+ )
212
+ };
213
+ this.models = data;
214
+ this.modelsUpdatedAt = Date.now();
215
+ return data;
216
+ });
217
+ }).catch(() => void 0);
218
+ await this.modelsCachePromise;
219
+ }
220
+ };
221
+
222
+ export {
223
+ AbstractProvider,
224
+ AbstractLLMProvider,
225
+ MastraProvider,
226
+ OpenAIProvider,
227
+ OpenRouterProvider
228
+ };
229
+ //# sourceMappingURL=chunk-R76XA2K6.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/providers/abstract/provider.ts","../src/providers/abstract/llm.ts","../src/providers/mastra.ts","../src/providers/openai.ts","../src/providers/openrouter.ts"],"sourcesContent":["export abstract class AbstractProvider {\n abstract readonly kind: string;\n}\n\nexport type ProviderResponse<TData = unknown> = {\n startedAt: number;\n completedAt: number;\n data: TData;\n};\n","import { AbstractProvider, ProviderResponse } from \"./provider\";\nimport {\n ResponseFormatJSONObject,\n ResponseFormatJSONSchema,\n ResponseFormatText,\n} from \"openai/resources/shared\";\nimport { ChatCompletionMessageParam } from \"openai/resources/chat/completions\";\n\nexport abstract class AbstractLLMProvider extends AbstractProvider {\n abstract forward(args: LLMProviderForwardArgs): Promise<ChatResponse>;\n}\n\nexport type LLMProviderForwardArgs = {\n messages: ChatCompletionMessageParam[];\n model: string;\n abortSignal?: AbortSignal;\n temperature?: number;\n responseFormat?:\n | ResponseFormatText\n | ResponseFormatJSONSchema\n | ResponseFormatJSONObject;\n};\n\nexport type ChatResponse = ProviderResponse<string> & {\n inputTokensUsed?: number;\n outputTokensUsed?: number;\n inputCost?: string;\n outputCost?: string;\n};\n","import {\n AbstractLLMProvider,\n type ChatResponse,\n type LLMProviderForwardArgs,\n} from \"./abstract/llm\";\nimport { MastraClient } from \"@mastra/client-js\";\n\nexport class MastraProvider extends AbstractLLMProvider {\n override readonly kind = \"mastra\";\n\n private readonly endpoint: string;\n private readonly authToken?: string;\n private client: MastraClient;\n private warnedAboutSystemMessages = false;\n private warnedAboutResponseFormat = false;\n\n constructor(params: { endpoint: string; authToken?: string }) {\n super();\n this.endpoint = params.endpoint;\n this.authToken = params.authToken;\n this.client = new MastraClient({\n baseUrl: this.endpoint,\n headers: {\n Authorization: `Bearer ${this.authToken}`,\n },\n });\n }\n\n override async forward(\n args: LLMProviderForwardArgs & {\n memory?: AgentMemoryOption;\n }\n ): Promise<ChatResponse> {\n const apiMessages = args.messages\n .filter((m) => m.role === \"user\" || m.role === \"assistant\")\n .map((m) => ({\n role: m.role as \"user\" | \"assistant\",\n content: String((m as any).content ?? \"\"),\n }));\n\n const agent = this.client.getAgent(args.model);\n\n const startedAt = Date.now();\n const response = await agent.generate(\n {\n messages: apiMessages,\n runtimeContext: {\n \"model-id\": args.model,\n },\n },\n { memory: args.memory }\n );\n\n return {\n data: response.text,\n startedAt,\n completedAt: Date.now(),\n };\n }\n}\n\n// NOTE: Mastra client does not export this type\nexport type AgentMemoryOption = Parameters<\n Parameters<MastraClient[\"getAgent\"]>[\"0\"] extends string\n ? ReturnType<MastraClient[\"getAgent\"]>[\"generate\"]\n : never\n>[0] extends { memory?: infer M }\n ? M\n : never;\n","import { RateLimiter } from \"@/utils\";\nimport { ChatCompletionMessageParam } from \"openai/resources/chat/completions\";\nimport {\n ResponseFormatJSONObject,\n ResponseFormatJSONSchema,\n ResponseFormatText,\n} from \"openai/resources/shared\";\nimport OpenAI, { APIError } from \"openai\";\nimport { AbstractLLMProvider, ChatResponse } from \"./abstract/llm\";\nimport { PEERBENCH_NAMESPACE } from \"@/constants\";\n\nexport class OpenAIProvider extends AbstractLLMProvider {\n override readonly kind = `${PEERBENCH_NAMESPACE}/llm/openai` as const;\n\n private client: OpenAI;\n private rateLimiter: RateLimiter;\n private maxRetries: number;\n\n constructor(config: {\n apiKey: string;\n baseURL: string;\n maxRetries?: number;\n timeout?: number;\n rateLimiter?: RateLimiter;\n }) {\n super();\n this.maxRetries = config.maxRetries ?? 3;\n this.rateLimiter =\n config.rateLimiter ??\n new RateLimiter({\n maxWeight: 20,\n timeWindow: 3_000,\n });\n\n this.client = new OpenAI({\n baseURL: config.baseURL,\n apiKey: config.apiKey,\n timeout: config.timeout,\n dangerouslyAllowBrowser: true,\n });\n }\n\n async forward(args: {\n messages: ChatCompletionMessageParam[];\n model: string;\n abortSignal?: AbortSignal;\n temperature?: number;\n responseFormat?:\n | ResponseFormatText\n | ResponseFormatJSONSchema\n | ResponseFormatJSONObject;\n }): Promise<ChatResponse> {\n let retryCount = this.maxRetries;\n while (retryCount > 0) {\n let startedAt: Date = new Date();\n\n try {\n const response = await this.rateLimiter.execute(\n async () => {\n // Capture the start time of the request\n startedAt = new Date();\n return await this.client.chat.completions.create(\n {\n model: args.model,\n messages: args.messages,\n temperature: args.temperature,\n response_format: args.responseFormat,\n },\n // Signal for request\n { signal: args.abortSignal }\n );\n },\n // Signal for rate limiting\n { signal: args.abortSignal }\n );\n\n if (\"error\" in response) {\n const err = response.error as any;\n throw new Error(\n `${err.message} - Code ${err.code} - ${JSON.stringify(err)}`\n );\n }\n\n if (!response?.choices?.[0]?.message?.content) {\n throw new Error(\"No content returned from the model\");\n }\n\n return {\n data: response.choices[0].message.content,\n\n inputTokensUsed: response?.usage?.prompt_tokens,\n outputTokensUsed: response?.usage?.completion_tokens,\n\n startedAt: startedAt.getTime(),\n completedAt: Date.now(),\n };\n } catch (err) {\n if (err instanceof APIError && err.status === 401) {\n throw new Error(`Invalid credentials provided`, { cause: err });\n }\n\n retryCount--;\n\n // More likely an empty HTTP response returned by the Provider\n // and it couldn't be parsed as JSON by the OpenAI SDK. We need to retry the request\n // More info can be found in the following links:\n // https://www.reddit.com/r/SillyTavernAI/comments/1ik95vr/deepseek_r1_on_openrouter_returning_blank_messages/\n // https://github.com/cline/cline/issues/60\n if (err instanceof SyntaxError) {\n console.debug(err);\n continue;\n }\n\n // If it was another error, just continue until we run out of retries\n if (retryCount !== 0) {\n continue;\n }\n\n throw new Error(\n `Failed to forward prompt to the model: ${err instanceof Error ? err.message : err}`,\n { cause: err }\n );\n }\n }\n\n throw new Error(\n `Failed to forward prompt to the model: Max retries reached`,\n { cause: new Error(\"Max retries reached\") }\n );\n }\n}\n","import {\n AbstractLLMProvider,\n ChatResponse,\n LLMProviderForwardArgs,\n} from \"./abstract/llm\";\nimport { RateLimiter } from \"@/utils\";\nimport { OpenAIProvider } from \"./openai\";\nimport { PEERBENCH_NAMESPACE } from \"@/constants\";\nimport Decimal from \"decimal.js\";\nimport axios from \"axios\";\n\nconst baseURL = \"https://openrouter.ai/api/v1\";\nconst MODELS_CACHE_TTL = 1000 * 60 * 60 * 24; // 24 hours\n\nexport class OpenRouterProvider extends AbstractLLMProvider {\n override readonly kind = `${PEERBENCH_NAMESPACE}/llm/openrouter.ai` as const;\n\n private models: ModelsResponse | undefined = undefined;\n private modelsCachePromise: Promise<ModelsResponse | undefined> =\n Promise.resolve(undefined);\n private modelsUpdatedAt = 0;\n private openAIProvider: OpenAIProvider;\n\n constructor(config: {\n apiKey: string;\n maxRetries?: number;\n timeout?: number;\n rateLimiter?: RateLimiter;\n }) {\n super();\n this.openAIProvider = new OpenAIProvider({\n baseURL,\n apiKey: config.apiKey,\n maxRetries: config.maxRetries,\n timeout: config.timeout,\n rateLimiter: config.rateLimiter,\n });\n }\n\n override async forward(args: LLMProviderForwardArgs): Promise<ChatResponse> {\n // Update models cache concurrently (non-blocking)\n const [response] = await Promise.all([\n this.openAIProvider.forward(args),\n this.updateModelsCache().catch(() => {\n // Silently fail if cache update fails so we won't have cost info in the result\n }),\n ]);\n\n // Get the model info from the cache\n const modelInfo = this.models?.data.find((m) => m.id === args.model);\n let inputCost: string | undefined = undefined;\n let outputCost: string | undefined = undefined;\n\n if (modelInfo !== undefined) {\n // Use Decimal.js for more accurate calculation\n if (response.inputTokensUsed !== undefined) {\n inputCost = new Decimal(modelInfo.pricing.prompt)\n .mul(response.inputTokensUsed)\n .toFixed(10);\n }\n if (response.outputTokensUsed !== undefined) {\n outputCost = new Decimal(modelInfo.pricing.completion)\n .mul(response.outputTokensUsed)\n .toFixed(10);\n }\n }\n\n return {\n ...response,\n inputCost,\n outputCost,\n };\n }\n\n /**\n * Updates the cache that holds information about OpenRouter models\n * including pricing information. It will be valid for 24 hours as\n * long as the instance of this Provider object is alive.\n */\n private async updateModelsCache() {\n // Chain each update method call to the promise.\n // This approach prevents race conditions between multiple calls.\n // Since each call is chained to the end of the previous one,\n // each promise makes a request only if the models cache is not updated\n // in the last call. Otherwise it simply resolves to the cached value.\n this.modelsCachePromise = this.modelsCachePromise\n .then(async () => {\n if (\n // The data presented in the cache\n this.models !== undefined &&\n // The cache is still valid\n Date.now() - this.modelsUpdatedAt < MODELS_CACHE_TTL\n ) {\n return this.models;\n }\n\n // If the cache is not valid, update it\n return axios\n .get<ModelsResponse>(`${baseURL}/models`)\n .then((res) => res.data)\n .then((data) => {\n // Only get the models that supports text input and output\n data = {\n data: data.data.filter(\n (m) =>\n m.architecture.input_modalities.includes(\"text\") &&\n m.architecture.output_modalities.includes(\"text\") &&\n // These models are \"fast apply model\" and don't support multi turn conversations so don't include them\n ![\n \"morph/morph-v3-large\",\n \"morph/morph-v3-fast\",\n \"relace/relace-apply-3\",\n ].includes(m.id)\n ),\n };\n\n this.models = data;\n this.modelsUpdatedAt = Date.now();\n\n return data;\n });\n })\n .catch(() => undefined);\n\n // Wait for the promise chain to resolve\n await this.modelsCachePromise;\n }\n}\n\ntype PutModality = \"text\" | \"image\" | \"file\" | \"audio\";\ntype Modality = \"text->text\" | \"text+image->text\" | \"text+image->text+image\";\ntype ModelsResponse = {\n data: {\n readonly id: string;\n readonly canonical_slug: string;\n readonly hugging_face_id: null | string;\n readonly name: string;\n readonly created: number;\n readonly description: string;\n readonly context_length: number;\n readonly architecture: {\n readonly modality: Modality;\n readonly input_modalities: PutModality[];\n readonly output_modalities: PutModality[];\n readonly instruct_type: null | string;\n };\n readonly pricing: {\n readonly prompt: string;\n readonly completion: string;\n readonly request?: string;\n readonly image?: string;\n readonly web_search?: string;\n readonly internal_reasoning?: string;\n readonly input_cache_read?: string;\n readonly input_cache_write?: string;\n readonly audio?: string;\n };\n }[];\n};\n"],"mappings":";;;;;;;;AAAO,IAAe,mBAAf,MAAgC;AAEvC;;;ACMO,IAAe,sBAAf,cAA2C,iBAAiB;AAEnE;;;ACLA,SAAS,oBAAoB;AAEtB,IAAM,iBAAN,cAA6B,oBAAoB;AAAA,EACpC,OAAO;AAAA,EAER;AAAA,EACA;AAAA,EACT;AAAA,EACA,4BAA4B;AAAA,EAC5B,4BAA4B;AAAA,EAEpC,YAAY,QAAkD;AAC5D,UAAM;AACN,SAAK,WAAW,OAAO;AACvB,SAAK,YAAY,OAAO;AACxB,SAAK,SAAS,IAAI,aAAa;AAAA,MAC7B,SAAS,KAAK;AAAA,MACd,SAAS;AAAA,QACP,eAAe,UAAU,KAAK,SAAS;AAAA,MACzC;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,MAAe,QACb,MAGuB;AACvB,UAAM,cAAc,KAAK,SACtB,OAAO,CAAC,MAAM,EAAE,SAAS,UAAU,EAAE,SAAS,WAAW,EACzD,IAAI,CAAC,OAAO;AAAA,MACX,MAAM,EAAE;AAAA,MACR,SAAS,OAAQ,EAAU,WAAW,EAAE;AAAA,IAC1C,EAAE;AAEJ,UAAM,QAAQ,KAAK,OAAO,SAAS,KAAK,KAAK;AAE7C,UAAM,YAAY,KAAK,IAAI;AAC3B,UAAM,WAAW,MAAM,MAAM;AAAA,MAC3B;AAAA,QACE,UAAU;AAAA,QACV,gBAAgB;AAAA,UACd,YAAY,KAAK;AAAA,QACnB;AAAA,MACF;AAAA,MACA,EAAE,QAAQ,KAAK,OAAO;AAAA,IACxB;AAEA,WAAO;AAAA,MACL,MAAM,SAAS;AAAA,MACf;AAAA,MACA,aAAa,KAAK,IAAI;AAAA,IACxB;AAAA,EACF;AACF;;;ACpDA,OAAO,UAAU,gBAAgB;AAI1B,IAAM,iBAAN,cAA6B,oBAAoB;AAAA,EACpC,OAAO,GAAG,mBAAmB;AAAA,EAEvC;AAAA,EACA;AAAA,EACA;AAAA,EAER,YAAY,QAMT;AACD,UAAM;AACN,SAAK,aAAa,OAAO,cAAc;AACvC,SAAK,cACH,OAAO,eACP,IAAI,YAAY;AAAA,MACd,WAAW;AAAA,MACX,YAAY;AAAA,IACd,CAAC;AAEH,SAAK,SAAS,IAAI,OAAO;AAAA,MACvB,SAAS,OAAO;AAAA,MAChB,QAAQ,OAAO;AAAA,MACf,SAAS,OAAO;AAAA,MAChB,yBAAyB;AAAA,IAC3B,CAAC;AAAA,EACH;AAAA,EAEA,MAAM,QAAQ,MASY;AACxB,QAAI,aAAa,KAAK;AACtB,WAAO,aAAa,GAAG;AACrB,UAAI,YAAkB,oBAAI,KAAK;AAE/B,UAAI;AACF,cAAM,WAAW,MAAM,KAAK,YAAY;AAAA,UACtC,YAAY;AAEV,wBAAY,oBAAI,KAAK;AACrB,mBAAO,MAAM,KAAK,OAAO,KAAK,YAAY;AAAA,cACxC;AAAA,gBACE,OAAO,KAAK;AAAA,gBACZ,UAAU,KAAK;AAAA,gBACf,aAAa,KAAK;AAAA,gBAClB,iBAAiB,KAAK;AAAA,cACxB;AAAA;AAAA,cAEA,EAAE,QAAQ,KAAK,YAAY;AAAA,YAC7B;AAAA,UACF;AAAA;AAAA,UAEA,EAAE,QAAQ,KAAK,YAAY;AAAA,QAC7B;AAEA,YAAI,WAAW,UAAU;AACvB,gBAAM,MAAM,SAAS;AACrB,gBAAM,IAAI;AAAA,YACR,GAAG,IAAI,OAAO,WAAW,IAAI,IAAI,MAAM,KAAK,UAAU,GAAG,CAAC;AAAA,UAC5D;AAAA,QACF;AAEA,YAAI,CAAC,UAAU,UAAU,CAAC,GAAG,SAAS,SAAS;AAC7C,gBAAM,IAAI,MAAM,oCAAoC;AAAA,QACtD;AAEA,eAAO;AAAA,UACL,MAAM,SAAS,QAAQ,CAAC,EAAE,QAAQ;AAAA,UAElC,iBAAiB,UAAU,OAAO;AAAA,UAClC,kBAAkB,UAAU,OAAO;AAAA,UAEnC,WAAW,UAAU,QAAQ;AAAA,UAC7B,aAAa,KAAK,IAAI;AAAA,QACxB;AAAA,MACF,SAAS,KAAK;AACZ,YAAI,eAAe,YAAY,IAAI,WAAW,KAAK;AACjD,gBAAM,IAAI,MAAM,gCAAgC,EAAE,OAAO,IAAI,CAAC;AAAA,QAChE;AAEA;AAOA,YAAI,eAAe,aAAa;AAC9B,kBAAQ,MAAM,GAAG;AACjB;AAAA,QACF;AAGA,YAAI,eAAe,GAAG;AACpB;AAAA,QACF;AAEA,cAAM,IAAI;AAAA,UACR,0CAA0C,eAAe,QAAQ,IAAI,UAAU,GAAG;AAAA,UAClF,EAAE,OAAO,IAAI;AAAA,QACf;AAAA,MACF;AAAA,IACF;AAEA,UAAM,IAAI;AAAA,MACR;AAAA,MACA,EAAE,OAAO,IAAI,MAAM,qBAAqB,EAAE;AAAA,IAC5C;AAAA,EACF;AACF;;;AC1HA,OAAO,aAAa;AACpB,OAAO,WAAW;AAElB,IAAM,UAAU;AAChB,IAAM,mBAAmB,MAAO,KAAK,KAAK;AAEnC,IAAM,qBAAN,cAAiC,oBAAoB;AAAA,EACxC,OAAO,GAAG,mBAAmB;AAAA,EAEvC,SAAqC;AAAA,EACrC,qBACN,QAAQ,QAAQ,MAAS;AAAA,EACnB,kBAAkB;AAAA,EAClB;AAAA,EAER,YAAY,QAKT;AACD,UAAM;AACN,SAAK,iBAAiB,IAAI,eAAe;AAAA,MACvC;AAAA,MACA,QAAQ,OAAO;AAAA,MACf,YAAY,OAAO;AAAA,MACnB,SAAS,OAAO;AAAA,MAChB,aAAa,OAAO;AAAA,IACtB,CAAC;AAAA,EACH;AAAA,EAEA,MAAe,QAAQ,MAAqD;AAE1E,UAAM,CAAC,QAAQ,IAAI,MAAM,QAAQ,IAAI;AAAA,MACnC,KAAK,eAAe,QAAQ,IAAI;AAAA,MAChC,KAAK,kBAAkB,EAAE,MAAM,MAAM;AAAA,MAErC,CAAC;AAAA,IACH,CAAC;AAGD,UAAM,YAAY,KAAK,QAAQ,KAAK,KAAK,CAAC,MAAM,EAAE,OAAO,KAAK,KAAK;AACnE,QAAI,YAAgC;AACpC,QAAI,aAAiC;AAErC,QAAI,cAAc,QAAW;AAE3B,UAAI,SAAS,oBAAoB,QAAW;AAC1C,oBAAY,IAAI,QAAQ,UAAU,QAAQ,MAAM,EAC7C,IAAI,SAAS,eAAe,EAC5B,QAAQ,EAAE;AAAA,MACf;AACA,UAAI,SAAS,qBAAqB,QAAW;AAC3C,qBAAa,IAAI,QAAQ,UAAU,QAAQ,UAAU,EAClD,IAAI,SAAS,gBAAgB,EAC7B,QAAQ,EAAE;AAAA,MACf;AAAA,IACF;AAEA,WAAO;AAAA,MACL,GAAG;AAAA,MACH;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOA,MAAc,oBAAoB;AAMhC,SAAK,qBAAqB,KAAK,mBAC5B,KAAK,YAAY;AAChB;AAAA;AAAA,QAEE,KAAK,WAAW;AAAA,QAEhB,KAAK,IAAI,IAAI,KAAK,kBAAkB;AAAA,QACpC;AACA,eAAO,KAAK;AAAA,MACd;AAGA,aAAO,MACJ,IAAoB,GAAG,OAAO,SAAS,EACvC,KAAK,CAAC,QAAQ,IAAI,IAAI,EACtB,KAAK,CAAC,SAAS;AAEd,eAAO;AAAA,UACL,MAAM,KAAK,KAAK;AAAA,YACd,CAAC,MACC,EAAE,aAAa,iBAAiB,SAAS,MAAM,KAC/C,EAAE,aAAa,kBAAkB,SAAS,MAAM;AAAA,YAEhD,CAAC;AAAA,cACC;AAAA,cACA;AAAA,cACA;AAAA,YACF,EAAE,SAAS,EAAE,EAAE;AAAA,UACnB;AAAA,QACF;AAEA,aAAK,SAAS;AACd,aAAK,kBAAkB,KAAK,IAAI;AAEhC,eAAO;AAAA,MACT,CAAC;AAAA,IACL,CAAC,EACA,MAAM,MAAM,MAAS;AAGxB,UAAM,KAAK;AAAA,EACb;AACF;","names":[]}
@@ -0,0 +1,35 @@
1
+ import {
2
+ IdSchema
3
+ } from "./chunk-NUEOE3K5.js";
4
+
5
+ // src/schemas/extensions/response/llm.ts
6
+ import z from "zod";
7
+ var ExtensionLLMResponseFieldsV1 = {
8
+ data: z.string(),
9
+ modelSlug: z.string(),
10
+ provider: z.string(),
11
+ systemPromptId: IdSchema.optional(),
12
+ inputTokensUsed: z.number().optional(),
13
+ outputTokensUsed: z.number().optional(),
14
+ inputCost: z.string().optional(),
15
+ outputCost: z.string().optional()
16
+ };
17
+
18
+ // src/schemas/extensions/score/llm-as-a-judge-scorer.ts
19
+ import z2 from "zod";
20
+ var ExtensionLLMAsAJudgeScoreFieldsV1 = {
21
+ scorerAISystemPrompt: z2.string().optional(),
22
+ scorerAISystemPromptId: z2.string().optional(),
23
+ scorerAIProvider: z2.string().optional(),
24
+ scorerAIModelSlug: z2.string().optional(),
25
+ scorerAIInputTokensUsed: z2.number().optional(),
26
+ scorerAIOutputTokensUsed: z2.number().optional(),
27
+ scorerAIInputCost: z2.string().optional(),
28
+ scorerAIOutputCost: z2.string().optional()
29
+ };
30
+
31
+ export {
32
+ ExtensionLLMResponseFieldsV1,
33
+ ExtensionLLMAsAJudgeScoreFieldsV1
34
+ };
35
+ //# sourceMappingURL=chunk-TRNCF2BG.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/schemas/extensions/response/llm.ts","../src/schemas/extensions/score/llm-as-a-judge-scorer.ts"],"sourcesContent":["import { IdSchema } from \"@/schemas/id\";\nimport z from \"zod\";\n\n/**\n * Provides a set of fields that holds information about the LLM and its response.\n */\nexport const ExtensionLLMResponseFieldsV1 = {\n data: z.string(),\n modelSlug: z.string(),\n provider: z.string(),\n systemPromptId: IdSchema.optional(),\n\n inputTokensUsed: z.number().optional(),\n outputTokensUsed: z.number().optional(),\n inputCost: z.string().optional(),\n outputCost: z.string().optional(),\n};\n","import z from \"zod\";\n\n/**\n * Provides a set of fields that holds information about the LLM model\n * that was used to judge the response.\n */\nexport const ExtensionLLMAsAJudgeScoreFieldsV1 = {\n scorerAISystemPrompt: z.string().optional(),\n scorerAISystemPromptId: z.string().optional(),\n scorerAIProvider: z.string().optional(),\n scorerAIModelSlug: z.string().optional(),\n scorerAIInputTokensUsed: z.number().optional(),\n scorerAIOutputTokensUsed: z.number().optional(),\n scorerAIInputCost: z.string().optional(),\n scorerAIOutputCost: z.string().optional(),\n};\n"],"mappings":";;;;;AACA,OAAO,OAAO;AAKP,IAAM,+BAA+B;AAAA,EAC1C,MAAM,EAAE,OAAO;AAAA,EACf,WAAW,EAAE,OAAO;AAAA,EACpB,UAAU,EAAE,OAAO;AAAA,EACnB,gBAAgB,SAAS,SAAS;AAAA,EAElC,iBAAiB,EAAE,OAAO,EAAE,SAAS;AAAA,EACrC,kBAAkB,EAAE,OAAO,EAAE,SAAS;AAAA,EACtC,WAAW,EAAE,OAAO,EAAE,SAAS;AAAA,EAC/B,YAAY,EAAE,OAAO,EAAE,SAAS;AAClC;;;AChBA,OAAOA,QAAO;AAMP,IAAM,oCAAoC;AAAA,EAC/C,sBAAsBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC1C,wBAAwBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC5C,kBAAkBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACtC,mBAAmBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACvC,yBAAyBA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC7C,0BAA0BA,GAAE,OAAO,EAAE,SAAS;AAAA,EAC9C,mBAAmBA,GAAE,OAAO,EAAE,SAAS;AAAA,EACvC,oBAAoBA,GAAE,OAAO,EAAE,SAAS;AAC1C;","names":["z"]}
@@ -0,0 +1,11 @@
1
+ // src/constants.ts
2
+ var PEERBENCH_NAMESPACE = "peerbench.ai";
3
+ var CATEGORIES = {
4
+ LLM: "llm"
5
+ };
6
+
7
+ export {
8
+ PEERBENCH_NAMESPACE,
9
+ CATEGORIES
10
+ };
11
+ //# sourceMappingURL=chunk-UHHHSYVE.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/constants.ts"],"sourcesContent":["export const PEERBENCH_NAMESPACE = \"peerbench.ai\" as const;\nexport const CATEGORIES = {\n LLM: \"llm\",\n};\n"],"mappings":";AAAO,IAAM,sBAAsB;AAC5B,IAAM,aAAa;AAAA,EACxB,KAAK;AACP;","names":[]}
@@ -1,34 +1,49 @@
1
+ import {
2
+ buildSchemaDefiner
3
+ } from "./chunk-OQE6TQXZ.js";
1
4
  import {
2
5
  ScoringMethod
3
- } from "./chunk-ZJWSK4VO.js";
6
+ } from "./chunk-HMQYGCKI.js";
4
7
  import {
5
- IdSchema,
6
- buildSchemaDefiner
7
- } from "./chunk-IUN2IUCS.js";
8
+ IdSchema
9
+ } from "./chunk-NUEOE3K5.js";
8
10
 
9
- // src/schemas/benchmark-spec.ts
11
+ // src/schemas/response.ts
10
12
  import z from "zod";
11
- var BaseBenchmarkSpecSchemaV1 = z.object({
12
- kind: z.string(),
13
+ var BaseResponseSchemaV1 = z.object({
14
+ id: IdSchema,
15
+ namespace: z.string(),
13
16
  schemaVersion: z.number(),
17
+ kind: z.string(),
18
+ startedAt: z.number(),
19
+ completedAt: z.number(),
20
+ testCaseId: IdSchema,
14
21
  metadata: z.record(z.string(), z.unknown()).optional()
15
22
  });
16
- var defineBenchmarkSpecSchema = buildSchemaDefiner();
23
+ var defineResponseSchema = buildSchemaDefiner(
24
+ BaseResponseSchemaV1,
25
+ "rs"
26
+ );
17
27
 
18
28
  // src/schemas/test-case.ts
19
29
  import { z as z2 } from "zod";
20
30
  var BaseTestCaseSchemaV1 = z2.object({
21
31
  id: IdSchema,
22
- kind: z2.string(),
32
+ namespace: z2.string(),
23
33
  schemaVersion: z2.number(),
34
+ kind: z2.string(),
24
35
  metadata: z2.record(z2.string(), z2.unknown()).optional()
25
36
  });
26
- var defineTestCaseSchema = buildSchemaDefiner();
37
+ var defineTestCaseSchema = buildSchemaDefiner(
38
+ BaseTestCaseSchemaV1,
39
+ "tc"
40
+ );
27
41
 
28
42
  // src/schemas/score.ts
29
43
  import z3 from "zod";
30
44
  var BaseScoreSchemaV1 = z3.object({
31
45
  id: IdSchema,
46
+ namespace: z3.string(),
32
47
  kind: z3.string(),
33
48
  schemaVersion: z3.number(),
34
49
  value: z3.number(),
@@ -37,14 +52,14 @@ var BaseScoreSchemaV1 = z3.object({
37
52
  metadata: z3.record(z3.string(), z3.unknown()).optional(),
38
53
  scoringMethod: z3.enum(ScoringMethod)
39
54
  });
40
- var defineScoreSchema = buildSchemaDefiner();
55
+ var defineScoreSchema = buildSchemaDefiner(BaseScoreSchemaV1, "sc");
41
56
 
42
57
  export {
43
- BaseBenchmarkSpecSchemaV1,
44
- defineBenchmarkSpecSchema,
58
+ BaseResponseSchemaV1,
59
+ defineResponseSchema,
45
60
  BaseTestCaseSchemaV1,
46
61
  defineTestCaseSchema,
47
62
  BaseScoreSchemaV1,
48
63
  defineScoreSchema
49
64
  };
50
- //# sourceMappingURL=chunk-232PY7K3.js.map
65
+ //# sourceMappingURL=chunk-YY33MNMV.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/schemas/response.ts","../src/schemas/test-case.ts","../src/schemas/score.ts"],"sourcesContent":["import { IdSchema } from \"./id\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\nimport z from \"zod\";\n\nexport const BaseResponseSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n schemaVersion: z.number(),\n kind: z.string(),\n\n startedAt: z.number(),\n completedAt: z.number(),\n testCaseId: IdSchema,\n metadata: z.record(z.string(), z.unknown()).optional(),\n});\nexport type BaseResponseV1 = z.infer<typeof BaseResponseSchemaV1>;\n\nexport const defineResponseSchema = buildSchemaDefiner(\n BaseResponseSchemaV1,\n \"rs\"\n);\n","import { IdSchema } from \"./id\";\nimport { z } from \"zod\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\n\nexport const BaseTestCaseSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n schemaVersion: z.number(),\n kind: z.string(),\n metadata: z.record(z.string(), z.unknown()).optional(),\n});\nexport type BaseTestCaseV1 = z.infer<typeof BaseTestCaseSchemaV1>;\n\nexport const defineTestCaseSchema = buildSchemaDefiner(\n BaseTestCaseSchemaV1,\n \"tc\"\n);\n","import { IdSchema } from \"./id\";\nimport { ScoringMethod } from \"@/types\";\nimport { buildSchemaDefiner } from \"./schema-definer\";\nimport z from \"zod\";\n\nexport const BaseScoreSchemaV1 = z.object({\n id: IdSchema,\n namespace: z.string(),\n kind: z.string(),\n schemaVersion: z.number(),\n\n value: z.number(),\n responseId: IdSchema,\n explanation: z.string().optional(),\n metadata: z.record(z.string(), z.unknown()).optional(),\n scoringMethod: z.enum(ScoringMethod),\n});\nexport type BaseScoreV1 = z.infer<typeof BaseScoreSchemaV1>;\n\nexport const defineScoreSchema = buildSchemaDefiner(BaseScoreSchemaV1, \"sc\");\n"],"mappings":";;;;;;;;;;;AAEA,OAAO,OAAO;AAEP,IAAM,uBAAuB,EAAE,OAAO;AAAA,EAC3C,IAAI;AAAA,EACJ,WAAW,EAAE,OAAO;AAAA,EACpB,eAAe,EAAE,OAAO;AAAA,EACxB,MAAM,EAAE,OAAO;AAAA,EAEf,WAAW,EAAE,OAAO;AAAA,EACpB,aAAa,EAAE,OAAO;AAAA,EACtB,YAAY;AAAA,EACZ,UAAU,EAAE,OAAO,EAAE,OAAO,GAAG,EAAE,QAAQ,CAAC,EAAE,SAAS;AACvD,CAAC;AAGM,IAAM,uBAAuB;AAAA,EAClC;AAAA,EACA;AACF;;;ACnBA,SAAS,KAAAA,UAAS;AAGX,IAAM,uBAAuBC,GAAE,OAAO;AAAA,EAC3C,IAAI;AAAA,EACJ,WAAWA,GAAE,OAAO;AAAA,EACpB,eAAeA,GAAE,OAAO;AAAA,EACxB,MAAMA,GAAE,OAAO;AAAA,EACf,UAAUA,GAAE,OAAOA,GAAE,OAAO,GAAGA,GAAE,QAAQ,CAAC,EAAE,SAAS;AACvD,CAAC;AAGM,IAAM,uBAAuB;AAAA,EAClC;AAAA,EACA;AACF;;;ACbA,OAAOC,QAAO;AAEP,IAAM,oBAAoBA,GAAE,OAAO;AAAA,EACxC,IAAI;AAAA,EACJ,WAAWA,GAAE,OAAO;AAAA,EACpB,MAAMA,GAAE,OAAO;AAAA,EACf,eAAeA,GAAE,OAAO;AAAA,EAExB,OAAOA,GAAE,OAAO;AAAA,EAChB,YAAY;AAAA,EACZ,aAAaA,GAAE,OAAO,EAAE,SAAS;AAAA,EACjC,UAAUA,GAAE,OAAOA,GAAE,OAAO,GAAGA,GAAE,QAAQ,CAAC,EAAE,SAAS;AAAA,EACrD,eAAeA,GAAE,KAAK,aAAa;AACrC,CAAC;AAGM,IAAM,oBAAoB,mBAAmB,mBAAmB,IAAI;","names":["z","z","z"]}