web-agent-bridge 3.0.0 → 3.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (202) hide show
  1. package/LICENSE +72 -21
  2. package/README.ar.md +1286 -1073
  3. package/README.md +1764 -1535
  4. package/bin/agent-runner.js +474 -474
  5. package/bin/cli.js +237 -138
  6. package/bin/wab.js +80 -80
  7. package/examples/bidi-agent.js +119 -119
  8. package/examples/cross-site-agent.js +91 -91
  9. package/examples/mcp-agent.js +94 -94
  10. package/examples/next-app-router/README.md +44 -44
  11. package/examples/puppeteer-agent.js +108 -108
  12. package/examples/saas-dashboard/README.md +55 -55
  13. package/examples/shopify-hydrogen/README.md +74 -74
  14. package/examples/vision-agent.js +171 -171
  15. package/examples/wordpress-elementor/README.md +77 -77
  16. package/package.json +17 -3
  17. package/public/.well-known/agent-tools.json +180 -180
  18. package/public/.well-known/ai-assets.json +59 -59
  19. package/public/.well-known/ai-plugin.json +28 -0
  20. package/public/.well-known/security.txt +8 -0
  21. package/public/agent-workspace.html +349 -347
  22. package/public/ai.html +198 -196
  23. package/public/api.html +413 -0
  24. package/public/browser.html +486 -484
  25. package/public/commander-dashboard.html +243 -243
  26. package/public/cookies.html +210 -208
  27. package/public/css/agent-workspace.css +1713 -1713
  28. package/public/css/premium.css +317 -317
  29. package/public/css/styles.css +1235 -1235
  30. package/public/dashboard.html +706 -704
  31. package/public/demo.html +1770 -1
  32. package/public/dns.html +507 -0
  33. package/public/docs.html +587 -585
  34. package/public/feed.xml +89 -89
  35. package/public/growth.html +463 -0
  36. package/public/index.html +341 -9
  37. package/public/integrations.html +556 -0
  38. package/public/js/agent-workspace.js +1740 -1740
  39. package/public/js/auth-nav.js +31 -31
  40. package/public/js/auth-redirect.js +12 -12
  41. package/public/js/cookie-consent.js +56 -56
  42. package/public/js/wab-demo-page.js +721 -721
  43. package/public/js/ws-client.js +74 -74
  44. package/public/llms-full.txt +360 -309
  45. package/public/llms.txt +125 -86
  46. package/public/login.html +85 -83
  47. package/public/mesh-dashboard.html +328 -328
  48. package/public/openapi.json +580 -580
  49. package/public/phone-shield.html +281 -0
  50. package/public/premium-dashboard.html +2489 -2487
  51. package/public/premium.html +793 -791
  52. package/public/privacy.html +297 -295
  53. package/public/register.html +105 -103
  54. package/public/robots.txt +87 -87
  55. package/public/script/wab-consent.d.ts +36 -36
  56. package/public/script/wab-consent.js +104 -104
  57. package/public/script/wab-schema.js +131 -131
  58. package/public/script/wab.d.ts +108 -108
  59. package/public/script/wab.min.js +580 -580
  60. package/public/security.txt +8 -0
  61. package/public/terms.html +256 -254
  62. package/script/ai-agent-bridge.js +1754 -1754
  63. package/sdk/README.md +99 -99
  64. package/sdk/agent-mesh.js +449 -449
  65. package/sdk/commander.js +262 -262
  66. package/sdk/index.d.ts +464 -464
  67. package/sdk/index.js +18 -1
  68. package/sdk/multi-agent.js +318 -318
  69. package/sdk/package.json +12 -1
  70. package/sdk/safety-shield.js +219 -0
  71. package/sdk/schema-discovery.js +83 -83
  72. package/server/adapters/index.js +520 -520
  73. package/server/config/plans.js +367 -367
  74. package/server/config/secrets.js +102 -102
  75. package/server/control-plane/index.js +301 -301
  76. package/server/data-plane/index.js +354 -354
  77. package/server/index.js +175 -19
  78. package/server/llm/index.js +404 -404
  79. package/server/middleware/adminAuth.js +35 -35
  80. package/server/middleware/auth.js +50 -50
  81. package/server/middleware/featureGate.js +88 -88
  82. package/server/middleware/rateLimits.js +100 -100
  83. package/server/middleware/sensitiveAction.js +157 -0
  84. package/server/migrations/001_add_analytics_indexes.sql +7 -7
  85. package/server/migrations/002_premium_features.sql +418 -418
  86. package/server/migrations/003_ads_integer_cents.sql +33 -33
  87. package/server/migrations/004_agent_os.sql +158 -158
  88. package/server/migrations/005_marketplace_metering.sql +126 -126
  89. package/server/models/adapters/index.js +33 -33
  90. package/server/models/adapters/mysql.js +183 -183
  91. package/server/models/adapters/postgresql.js +172 -172
  92. package/server/models/adapters/sqlite.js +7 -7
  93. package/server/models/db.js +681 -681
  94. package/server/observability/failure-analysis.js +337 -337
  95. package/server/observability/index.js +394 -394
  96. package/server/protocol/capabilities.js +223 -223
  97. package/server/protocol/index.js +243 -243
  98. package/server/protocol/schema.js +584 -584
  99. package/server/registry/certification.js +271 -271
  100. package/server/registry/index.js +326 -326
  101. package/server/routes/admin-premium.js +671 -671
  102. package/server/routes/admin.js +261 -261
  103. package/server/routes/ads.js +130 -130
  104. package/server/routes/agent-workspace.js +540 -378
  105. package/server/routes/api.js +150 -150
  106. package/server/routes/auth.js +71 -71
  107. package/server/routes/billing.js +45 -45
  108. package/server/routes/commander.js +316 -316
  109. package/server/routes/demo-showcase.js +332 -0
  110. package/server/routes/demo-store.js +154 -0
  111. package/server/routes/discovery.js +417 -406
  112. package/server/routes/gateway.js +173 -0
  113. package/server/routes/license.js +251 -240
  114. package/server/routes/mesh.js +469 -469
  115. package/server/routes/noscript.js +543 -543
  116. package/server/routes/premium-v2.js +686 -686
  117. package/server/routes/premium.js +724 -724
  118. package/server/routes/runtime.js +2148 -2147
  119. package/server/routes/sovereign.js +465 -385
  120. package/server/routes/universal.js +200 -177
  121. package/server/routes/wab-api.js +850 -491
  122. package/server/runtime/container-worker.js +111 -111
  123. package/server/runtime/container.js +448 -448
  124. package/server/runtime/distributed-worker.js +362 -362
  125. package/server/runtime/event-bus.js +210 -210
  126. package/server/runtime/index.js +253 -253
  127. package/server/runtime/queue.js +599 -599
  128. package/server/runtime/replay.js +666 -666
  129. package/server/runtime/sandbox.js +266 -266
  130. package/server/runtime/scheduler.js +534 -534
  131. package/server/runtime/session-engine.js +293 -293
  132. package/server/runtime/state-manager.js +188 -188
  133. package/server/security/cross-site-redactor.js +196 -0
  134. package/server/security/dry-run.js +180 -0
  135. package/server/security/human-gate-rate-limit.js +147 -0
  136. package/server/security/human-gate-transports.js +178 -0
  137. package/server/security/human-gate.js +281 -0
  138. package/server/security/index.js +368 -368
  139. package/server/security/intent-engine.js +245 -0
  140. package/server/security/reward-guard.js +171 -0
  141. package/server/security/rollback-store.js +239 -0
  142. package/server/security/token-scope.js +404 -0
  143. package/server/security/url-policy.js +139 -0
  144. package/server/services/agent-chat.js +506 -506
  145. package/server/services/agent-learning.js +601 -575
  146. package/server/services/agent-memory.js +625 -625
  147. package/server/services/agent-mesh.js +555 -539
  148. package/server/services/agent-symphony.js +717 -717
  149. package/server/services/agent-tasks.js +1807 -1807
  150. package/server/services/api-key-engine.js +292 -0
  151. package/server/services/cluster.js +894 -894
  152. package/server/services/commander.js +738 -738
  153. package/server/services/edge-compute.js +440 -440
  154. package/server/services/email.js +204 -204
  155. package/server/services/hosted-runtime.js +205 -205
  156. package/server/services/lfd.js +635 -616
  157. package/server/services/local-ai.js +389 -389
  158. package/server/services/marketplace.js +270 -270
  159. package/server/services/metering.js +182 -182
  160. package/server/services/modules/affiliate-intelligence.js +93 -0
  161. package/server/services/modules/agent-firewall.js +90 -0
  162. package/server/services/modules/bounty.js +89 -0
  163. package/server/services/modules/collective-bargaining.js +92 -0
  164. package/server/services/modules/dark-pattern.js +66 -0
  165. package/server/services/modules/gov-intelligence.js +45 -0
  166. package/server/services/modules/neural.js +55 -0
  167. package/server/services/modules/notary.js +49 -0
  168. package/server/services/modules/price-time-machine.js +86 -0
  169. package/server/services/modules/protocol.js +104 -0
  170. package/server/services/negotiation.js +439 -439
  171. package/server/services/plugins.js +771 -771
  172. package/server/services/premium.js +1 -1
  173. package/server/services/price-intelligence.js +566 -565
  174. package/server/services/price-shield.js +1137 -1137
  175. package/server/services/reputation.js +465 -465
  176. package/server/services/search-engine.js +357 -357
  177. package/server/services/security.js +513 -513
  178. package/server/services/self-healing.js +843 -843
  179. package/server/services/sovereign-shield.js +542 -0
  180. package/server/services/stripe.js +192 -192
  181. package/server/services/swarm.js +788 -788
  182. package/server/services/universal-scraper.js +662 -661
  183. package/server/services/verification.js +481 -481
  184. package/server/services/vision.js +1163 -1163
  185. package/server/utils/cache.js +125 -125
  186. package/server/utils/migrate.js +81 -81
  187. package/server/utils/safe-fetch.js +228 -0
  188. package/server/utils/secureFields.js +50 -50
  189. package/server/ws.js +161 -161
  190. package/templates/artisan-marketplace.yaml +104 -104
  191. package/templates/book-price-scout.yaml +98 -98
  192. package/templates/electronics-price-tracker.yaml +108 -108
  193. package/templates/flight-deal-hunter.yaml +113 -113
  194. package/templates/freelancer-direct.yaml +116 -116
  195. package/templates/grocery-price-compare.yaml +93 -93
  196. package/templates/hotel-direct-booking.yaml +113 -113
  197. package/templates/local-services.yaml +98 -98
  198. package/templates/olive-oil-tunisia.yaml +88 -88
  199. package/templates/organic-farm-fresh.yaml +101 -101
  200. package/templates/restaurant-direct.yaml +97 -97
  201. package/server/services/fairness-engine.js +0 -409
  202. package/server/services/fairness.js +0 -420
@@ -1,404 +1,404 @@
1
- 'use strict';
2
-
3
- /**
4
- * WAB LLM Abstraction Layer
5
- *
6
- * Model-agnostic LLM interface. Supports:
7
- * - OpenAI (GPT-4, GPT-3.5)
8
- * - Anthropic (Claude)
9
- * - Ollama (local models)
10
- * - Custom providers
11
- *
12
- * Provides a unified API with automatic fallback,
13
- * cost tracking, and response caching.
14
- */
15
-
16
- const { metrics, logger } = require('../observability');
17
-
18
- // ─── Provider Interface ─────────────────────────────────────────────────────
19
-
20
- class LLMProvider {
21
- constructor(name, config = {}) {
22
- this.name = name;
23
- this.config = config;
24
- this.available = false;
25
- this.models = [];
26
- }
27
-
28
- async initialize() { throw new Error('Not implemented'); }
29
- async complete(prompt, options) { throw new Error('Not implemented'); }
30
- async embed(text) { throw new Error('Not implemented'); }
31
- async listModels() { return this.models; }
32
- }
33
-
34
- // ─── OpenAI Provider ────────────────────────────────────────────────────────
35
-
36
- class OpenAIProvider extends LLMProvider {
37
- constructor(config) {
38
- super('openai', config);
39
- this.apiKey = config.apiKey || process.env.OPENAI_API_KEY;
40
- this.baseUrl = config.baseUrl || 'https://api.openai.com/v1';
41
- this.models = ['gpt-4o', 'gpt-4o-mini', 'gpt-4-turbo', 'gpt-3.5-turbo'];
42
- }
43
-
44
- async initialize() {
45
- this.available = !!this.apiKey;
46
- return this.available;
47
- }
48
-
49
- async complete(prompt, options = {}) {
50
- if (!this.available) throw new Error('OpenAI provider not initialized');
51
-
52
- const model = options.model || 'gpt-4o-mini';
53
- const messages = [];
54
- if (options.systemPrompt) messages.push({ role: 'system', content: options.systemPrompt });
55
- messages.push({ role: 'user', content: prompt });
56
-
57
- const body = {
58
- model,
59
- messages,
60
- temperature: options.temperature ?? 0.7,
61
- max_tokens: options.maxTokens || 2048,
62
- };
63
-
64
- const res = await fetch(`${this.baseUrl}/chat/completions`, {
65
- method: 'POST',
66
- headers: {
67
- 'Content-Type': 'application/json',
68
- 'Authorization': `Bearer ${this.apiKey}`,
69
- },
70
- body: JSON.stringify(body),
71
- });
72
-
73
- if (!res.ok) {
74
- const err = await res.text();
75
- throw new Error(`OpenAI error ${res.status}: ${err}`);
76
- }
77
-
78
- const data = await res.json();
79
- return {
80
- text: data.choices[0]?.message?.content || '',
81
- model,
82
- provider: 'openai',
83
- usage: {
84
- promptTokens: data.usage?.prompt_tokens || 0,
85
- completionTokens: data.usage?.completion_tokens || 0,
86
- totalTokens: data.usage?.total_tokens || 0,
87
- },
88
- finishReason: data.choices[0]?.finish_reason,
89
- };
90
- }
91
-
92
- async embed(text) {
93
- if (!this.available) throw new Error('OpenAI provider not initialized');
94
-
95
- const res = await fetch(`${this.baseUrl}/embeddings`, {
96
- method: 'POST',
97
- headers: {
98
- 'Content-Type': 'application/json',
99
- 'Authorization': `Bearer ${this.apiKey}`,
100
- },
101
- body: JSON.stringify({ model: 'text-embedding-3-small', input: text }),
102
- });
103
-
104
- if (!res.ok) throw new Error(`OpenAI embed error ${res.status}`);
105
- const data = await res.json();
106
- return { embedding: data.data[0]?.embedding || [], model: 'text-embedding-3-small', provider: 'openai' };
107
- }
108
- }
109
-
110
- // ─── Anthropic Provider ─────────────────────────────────────────────────────
111
-
112
- class AnthropicProvider extends LLMProvider {
113
- constructor(config) {
114
- super('anthropic', config);
115
- this.apiKey = config.apiKey || process.env.ANTHROPIC_API_KEY;
116
- this.baseUrl = config.baseUrl || 'https://api.anthropic.com/v1';
117
- this.models = ['claude-sonnet-4-20250514', 'claude-3-5-haiku-20241022', 'claude-3-5-sonnet-20241022'];
118
- }
119
-
120
- async initialize() {
121
- this.available = !!this.apiKey;
122
- return this.available;
123
- }
124
-
125
- async complete(prompt, options = {}) {
126
- if (!this.available) throw new Error('Anthropic provider not initialized');
127
-
128
- const model = options.model || 'claude-3-5-haiku-20241022';
129
- const body = {
130
- model,
131
- max_tokens: options.maxTokens || 2048,
132
- messages: [{ role: 'user', content: prompt }],
133
- };
134
- if (options.systemPrompt) body.system = options.systemPrompt;
135
- if (options.temperature !== undefined) body.temperature = options.temperature;
136
-
137
- const res = await fetch(`${this.baseUrl}/messages`, {
138
- method: 'POST',
139
- headers: {
140
- 'Content-Type': 'application/json',
141
- 'x-api-key': this.apiKey,
142
- 'anthropic-version': '2023-06-01',
143
- },
144
- body: JSON.stringify(body),
145
- });
146
-
147
- if (!res.ok) {
148
- const err = await res.text();
149
- throw new Error(`Anthropic error ${res.status}: ${err}`);
150
- }
151
-
152
- const data = await res.json();
153
- return {
154
- text: data.content?.[0]?.text || '',
155
- model,
156
- provider: 'anthropic',
157
- usage: {
158
- promptTokens: data.usage?.input_tokens || 0,
159
- completionTokens: data.usage?.output_tokens || 0,
160
- totalTokens: (data.usage?.input_tokens || 0) + (data.usage?.output_tokens || 0),
161
- },
162
- finishReason: data.stop_reason,
163
- };
164
- }
165
- }
166
-
167
- // ─── Ollama Provider (Local) ────────────────────────────────────────────────
168
-
169
- class OllamaProvider extends LLMProvider {
170
- constructor(config) {
171
- super('ollama', config);
172
- this.baseUrl = config.baseUrl || process.env.OLLAMA_URL || 'http://localhost:11434';
173
- }
174
-
175
- async initialize() {
176
- try {
177
- const res = await fetch(`${this.baseUrl}/api/tags`, { signal: AbortSignal.timeout(3000) });
178
- if (res.ok) {
179
- const data = await res.json();
180
- this.models = (data.models || []).map(m => m.name);
181
- this.available = true;
182
- }
183
- } catch (_) {
184
- this.available = false;
185
- }
186
- return this.available;
187
- }
188
-
189
- async complete(prompt, options = {}) {
190
- if (!this.available) throw new Error('Ollama not available');
191
-
192
- const model = options.model || this.models[0] || 'llama3.2';
193
- const body = {
194
- model,
195
- prompt: options.systemPrompt ? `${options.systemPrompt}\n\n${prompt}` : prompt,
196
- stream: false,
197
- options: {},
198
- };
199
- if (options.temperature !== undefined) body.options.temperature = options.temperature;
200
-
201
- const res = await fetch(`${this.baseUrl}/api/generate`, {
202
- method: 'POST',
203
- headers: { 'Content-Type': 'application/json' },
204
- body: JSON.stringify(body),
205
- });
206
-
207
- if (!res.ok) throw new Error(`Ollama error ${res.status}`);
208
- const data = await res.json();
209
-
210
- return {
211
- text: data.response || '',
212
- model,
213
- provider: 'ollama',
214
- usage: {
215
- promptTokens: data.prompt_eval_count || 0,
216
- completionTokens: data.eval_count || 0,
217
- totalTokens: (data.prompt_eval_count || 0) + (data.eval_count || 0),
218
- },
219
- finishReason: data.done ? 'stop' : 'length',
220
- };
221
- }
222
-
223
- async embed(text) {
224
- if (!this.available) throw new Error('Ollama not available');
225
-
226
- const model = this.models.find(m => m.includes('embed')) || 'nomic-embed-text';
227
- const res = await fetch(`${this.baseUrl}/api/embeddings`, {
228
- method: 'POST',
229
- headers: { 'Content-Type': 'application/json' },
230
- body: JSON.stringify({ model, prompt: text }),
231
- });
232
-
233
- if (!res.ok) throw new Error(`Ollama embed error ${res.status}`);
234
- const data = await res.json();
235
- return { embedding: data.embedding || [], model, provider: 'ollama' };
236
- }
237
- }
238
-
239
- // ─── LLM Manager (Unified Interface) ───────────────────────────────────────
240
-
241
- class LLMManager {
242
- constructor() {
243
- this._providers = new Map();
244
- this._defaultProvider = null;
245
- this._fallbackOrder = [];
246
- this._cache = new Map();
247
- this._maxCache = 500;
248
- this._stats = { requests: 0, cacheHits: 0, failures: 0, totalTokens: 0 };
249
- }
250
-
251
- /**
252
- * Register a provider
253
- */
254
- registerProvider(provider) {
255
- this._providers.set(provider.name, provider);
256
- if (!this._defaultProvider) this._defaultProvider = provider.name;
257
- this._fallbackOrder.push(provider.name);
258
- }
259
-
260
- /**
261
- * Initialize all providers
262
- */
263
- async initialize() {
264
- const results = {};
265
- for (const [name, provider] of this._providers) {
266
- try {
267
- results[name] = await provider.initialize();
268
- } catch (_) {
269
- results[name] = false;
270
- }
271
- }
272
-
273
- // Set default to first available
274
- for (const name of this._fallbackOrder) {
275
- if (this._providers.get(name)?.available) {
276
- this._defaultProvider = name;
277
- break;
278
- }
279
- }
280
-
281
- return results;
282
- }
283
-
284
- /**
285
- * Complete a prompt (with automatic fallback)
286
- */
287
- async complete(prompt, options = {}) {
288
- this._stats.requests++;
289
-
290
- // Check cache
291
- if (options.cache !== false) {
292
- const cacheKey = this._cacheKey(prompt, options);
293
- const cached = this._cache.get(cacheKey);
294
- if (cached && (Date.now() - cached.timestamp < 300_000)) {
295
- this._stats.cacheHits++;
296
- return { ...cached.result, cached: true };
297
- }
298
- }
299
-
300
- const providerName = options.provider || this._defaultProvider;
301
- const providers = [providerName, ...this._fallbackOrder.filter(p => p !== providerName)];
302
-
303
- const endTimer = metrics.startTimer('llm.request.duration');
304
-
305
- for (const name of providers) {
306
- const provider = this._providers.get(name);
307
- if (!provider?.available) continue;
308
-
309
- try {
310
- const result = await provider.complete(prompt, options);
311
-
312
- endTimer();
313
- metrics.increment('llm.requests.success', 1, { provider: name });
314
- this._stats.totalTokens += result.usage?.totalTokens || 0;
315
-
316
- // Cache result
317
- if (options.cache !== false) {
318
- const cacheKey = this._cacheKey(prompt, options);
319
- this._cache.set(cacheKey, { result, timestamp: Date.now() });
320
- if (this._cache.size > this._maxCache) {
321
- const oldest = this._cache.keys().next().value;
322
- this._cache.delete(oldest);
323
- }
324
- }
325
-
326
- return { ...result, duration: endTimer() };
327
- } catch (err) {
328
- metrics.increment('llm.requests.failure', 1, { provider: name });
329
- this._stats.failures++;
330
- // Try next provider
331
- continue;
332
- }
333
- }
334
-
335
- endTimer();
336
- throw new Error('All LLM providers failed');
337
- }
338
-
339
- /**
340
- * Generate embeddings
341
- */
342
- async embed(text, options = {}) {
343
- const providerName = options.provider || this._defaultProvider;
344
- const provider = this._providers.get(providerName);
345
- if (!provider?.available) throw new Error(`Provider ${providerName} not available`);
346
- if (!provider.embed) throw new Error(`Provider ${providerName} does not support embeddings`);
347
- return provider.embed(text);
348
- }
349
-
350
- /**
351
- * List available models across all providers
352
- */
353
- listModels() {
354
- const models = [];
355
- for (const [name, provider] of this._providers) {
356
- if (!provider.available) continue;
357
- for (const model of provider.models) {
358
- models.push({ model, provider: name });
359
- }
360
- }
361
- return models;
362
- }
363
-
364
- /**
365
- * Get provider status
366
- */
367
- getStatus() {
368
- const providers = {};
369
- for (const [name, provider] of this._providers) {
370
- providers[name] = {
371
- available: provider.available,
372
- models: provider.models,
373
- };
374
- }
375
- return {
376
- defaultProvider: this._defaultProvider,
377
- providers,
378
- stats: { ...this._stats },
379
- };
380
- }
381
-
382
- _cacheKey(prompt, options) {
383
- const key = `${options.provider || ''}:${options.model || ''}:${prompt.slice(0, 200)}`;
384
- return require('crypto').createHash('md5').update(key).digest('hex');
385
- }
386
- }
387
-
388
- // ─── Singleton ──────────────────────────────────────────────────────────────
389
-
390
- const llm = new LLMManager();
391
-
392
- // Register default providers
393
- llm.registerProvider(new OpenAIProvider({}));
394
- llm.registerProvider(new AnthropicProvider({}));
395
- llm.registerProvider(new OllamaProvider({}));
396
-
397
- module.exports = {
398
- LLMProvider,
399
- OpenAIProvider,
400
- AnthropicProvider,
401
- OllamaProvider,
402
- LLMManager,
403
- llm,
404
- };
1
+ 'use strict';
2
+
3
+ /**
4
+ * WAB LLM Abstraction Layer
5
+ *
6
+ * Model-agnostic LLM interface. Supports:
7
+ * - OpenAI (GPT-4, GPT-3.5)
8
+ * - Anthropic (Claude)
9
+ * - Ollama (local models)
10
+ * - Custom providers
11
+ *
12
+ * Provides a unified API with automatic fallback,
13
+ * cost tracking, and response caching.
14
+ */
15
+
16
+ const { metrics, logger } = require('../observability');
17
+
18
+ // ─── Provider Interface ─────────────────────────────────────────────────────
19
+
20
+ class LLMProvider {
21
+ constructor(name, config = {}) {
22
+ this.name = name;
23
+ this.config = config;
24
+ this.available = false;
25
+ this.models = [];
26
+ }
27
+
28
+ async initialize() { throw new Error('Not implemented'); }
29
+ async complete(prompt, options) { throw new Error('Not implemented'); }
30
+ async embed(text) { throw new Error('Not implemented'); }
31
+ async listModels() { return this.models; }
32
+ }
33
+
34
+ // ─── OpenAI Provider ────────────────────────────────────────────────────────
35
+
36
+ class OpenAIProvider extends LLMProvider {
37
+ constructor(config) {
38
+ super('openai', config);
39
+ this.apiKey = config.apiKey || process.env.OPENAI_API_KEY;
40
+ this.baseUrl = config.baseUrl || 'https://api.openai.com/v1';
41
+ this.models = ['gpt-4o', 'gpt-4o-mini', 'gpt-4-turbo', 'gpt-3.5-turbo'];
42
+ }
43
+
44
+ async initialize() {
45
+ this.available = !!this.apiKey;
46
+ return this.available;
47
+ }
48
+
49
+ async complete(prompt, options = {}) {
50
+ if (!this.available) throw new Error('OpenAI provider not initialized');
51
+
52
+ const model = options.model || 'gpt-4o-mini';
53
+ const messages = [];
54
+ if (options.systemPrompt) messages.push({ role: 'system', content: options.systemPrompt });
55
+ messages.push({ role: 'user', content: prompt });
56
+
57
+ const body = {
58
+ model,
59
+ messages,
60
+ temperature: options.temperature ?? 0.7,
61
+ max_tokens: options.maxTokens || 2048,
62
+ };
63
+
64
+ const res = await fetch(`${this.baseUrl}/chat/completions`, {
65
+ method: 'POST',
66
+ headers: {
67
+ 'Content-Type': 'application/json',
68
+ 'Authorization': `Bearer ${this.apiKey}`,
69
+ },
70
+ body: JSON.stringify(body),
71
+ });
72
+
73
+ if (!res.ok) {
74
+ const err = await res.text();
75
+ throw new Error(`OpenAI error ${res.status}: ${err}`);
76
+ }
77
+
78
+ const data = await res.json();
79
+ return {
80
+ text: data.choices[0]?.message?.content || '',
81
+ model,
82
+ provider: 'openai',
83
+ usage: {
84
+ promptTokens: data.usage?.prompt_tokens || 0,
85
+ completionTokens: data.usage?.completion_tokens || 0,
86
+ totalTokens: data.usage?.total_tokens || 0,
87
+ },
88
+ finishReason: data.choices[0]?.finish_reason,
89
+ };
90
+ }
91
+
92
+ async embed(text) {
93
+ if (!this.available) throw new Error('OpenAI provider not initialized');
94
+
95
+ const res = await fetch(`${this.baseUrl}/embeddings`, {
96
+ method: 'POST',
97
+ headers: {
98
+ 'Content-Type': 'application/json',
99
+ 'Authorization': `Bearer ${this.apiKey}`,
100
+ },
101
+ body: JSON.stringify({ model: 'text-embedding-3-small', input: text }),
102
+ });
103
+
104
+ if (!res.ok) throw new Error(`OpenAI embed error ${res.status}`);
105
+ const data = await res.json();
106
+ return { embedding: data.data[0]?.embedding || [], model: 'text-embedding-3-small', provider: 'openai' };
107
+ }
108
+ }
109
+
110
+ // ─── Anthropic Provider ─────────────────────────────────────────────────────
111
+
112
+ class AnthropicProvider extends LLMProvider {
113
+ constructor(config) {
114
+ super('anthropic', config);
115
+ this.apiKey = config.apiKey || process.env.ANTHROPIC_API_KEY;
116
+ this.baseUrl = config.baseUrl || 'https://api.anthropic.com/v1';
117
+ this.models = ['claude-sonnet-4-20250514', 'claude-3-5-haiku-20241022', 'claude-3-5-sonnet-20241022'];
118
+ }
119
+
120
+ async initialize() {
121
+ this.available = !!this.apiKey;
122
+ return this.available;
123
+ }
124
+
125
+ async complete(prompt, options = {}) {
126
+ if (!this.available) throw new Error('Anthropic provider not initialized');
127
+
128
+ const model = options.model || 'claude-3-5-haiku-20241022';
129
+ const body = {
130
+ model,
131
+ max_tokens: options.maxTokens || 2048,
132
+ messages: [{ role: 'user', content: prompt }],
133
+ };
134
+ if (options.systemPrompt) body.system = options.systemPrompt;
135
+ if (options.temperature !== undefined) body.temperature = options.temperature;
136
+
137
+ const res = await fetch(`${this.baseUrl}/messages`, {
138
+ method: 'POST',
139
+ headers: {
140
+ 'Content-Type': 'application/json',
141
+ 'x-api-key': this.apiKey,
142
+ 'anthropic-version': '2023-06-01',
143
+ },
144
+ body: JSON.stringify(body),
145
+ });
146
+
147
+ if (!res.ok) {
148
+ const err = await res.text();
149
+ throw new Error(`Anthropic error ${res.status}: ${err}`);
150
+ }
151
+
152
+ const data = await res.json();
153
+ return {
154
+ text: data.content?.[0]?.text || '',
155
+ model,
156
+ provider: 'anthropic',
157
+ usage: {
158
+ promptTokens: data.usage?.input_tokens || 0,
159
+ completionTokens: data.usage?.output_tokens || 0,
160
+ totalTokens: (data.usage?.input_tokens || 0) + (data.usage?.output_tokens || 0),
161
+ },
162
+ finishReason: data.stop_reason,
163
+ };
164
+ }
165
+ }
166
+
167
+ // ─── Ollama Provider (Local) ────────────────────────────────────────────────
168
+
169
+ class OllamaProvider extends LLMProvider {
170
+ constructor(config) {
171
+ super('ollama', config);
172
+ this.baseUrl = config.baseUrl || process.env.OLLAMA_URL || 'http://localhost:11434';
173
+ }
174
+
175
+ async initialize() {
176
+ try {
177
+ const res = await fetch(`${this.baseUrl}/api/tags`, { signal: AbortSignal.timeout(3000) });
178
+ if (res.ok) {
179
+ const data = await res.json();
180
+ this.models = (data.models || []).map(m => m.name);
181
+ this.available = true;
182
+ }
183
+ } catch (_) {
184
+ this.available = false;
185
+ }
186
+ return this.available;
187
+ }
188
+
189
+ async complete(prompt, options = {}) {
190
+ if (!this.available) throw new Error('Ollama not available');
191
+
192
+ const model = options.model || this.models[0] || 'llama3.2';
193
+ const body = {
194
+ model,
195
+ prompt: options.systemPrompt ? `${options.systemPrompt}\n\n${prompt}` : prompt,
196
+ stream: false,
197
+ options: {},
198
+ };
199
+ if (options.temperature !== undefined) body.options.temperature = options.temperature;
200
+
201
+ const res = await fetch(`${this.baseUrl}/api/generate`, {
202
+ method: 'POST',
203
+ headers: { 'Content-Type': 'application/json' },
204
+ body: JSON.stringify(body),
205
+ });
206
+
207
+ if (!res.ok) throw new Error(`Ollama error ${res.status}`);
208
+ const data = await res.json();
209
+
210
+ return {
211
+ text: data.response || '',
212
+ model,
213
+ provider: 'ollama',
214
+ usage: {
215
+ promptTokens: data.prompt_eval_count || 0,
216
+ completionTokens: data.eval_count || 0,
217
+ totalTokens: (data.prompt_eval_count || 0) + (data.eval_count || 0),
218
+ },
219
+ finishReason: data.done ? 'stop' : 'length',
220
+ };
221
+ }
222
+
223
+ async embed(text) {
224
+ if (!this.available) throw new Error('Ollama not available');
225
+
226
+ const model = this.models.find(m => m.includes('embed')) || 'nomic-embed-text';
227
+ const res = await fetch(`${this.baseUrl}/api/embeddings`, {
228
+ method: 'POST',
229
+ headers: { 'Content-Type': 'application/json' },
230
+ body: JSON.stringify({ model, prompt: text }),
231
+ });
232
+
233
+ if (!res.ok) throw new Error(`Ollama embed error ${res.status}`);
234
+ const data = await res.json();
235
+ return { embedding: data.embedding || [], model, provider: 'ollama' };
236
+ }
237
+ }
238
+
239
+ // ─── LLM Manager (Unified Interface) ───────────────────────────────────────
240
+
241
+ class LLMManager {
242
+ constructor() {
243
+ this._providers = new Map();
244
+ this._defaultProvider = null;
245
+ this._fallbackOrder = [];
246
+ this._cache = new Map();
247
+ this._maxCache = 500;
248
+ this._stats = { requests: 0, cacheHits: 0, failures: 0, totalTokens: 0 };
249
+ }
250
+
251
+ /**
252
+ * Register a provider
253
+ */
254
+ registerProvider(provider) {
255
+ this._providers.set(provider.name, provider);
256
+ if (!this._defaultProvider) this._defaultProvider = provider.name;
257
+ this._fallbackOrder.push(provider.name);
258
+ }
259
+
260
+ /**
261
+ * Initialize all providers
262
+ */
263
+ async initialize() {
264
+ const results = {};
265
+ for (const [name, provider] of this._providers) {
266
+ try {
267
+ results[name] = await provider.initialize();
268
+ } catch (_) {
269
+ results[name] = false;
270
+ }
271
+ }
272
+
273
+ // Set default to first available
274
+ for (const name of this._fallbackOrder) {
275
+ if (this._providers.get(name)?.available) {
276
+ this._defaultProvider = name;
277
+ break;
278
+ }
279
+ }
280
+
281
+ return results;
282
+ }
283
+
284
+ /**
285
+ * Complete a prompt (with automatic fallback)
286
+ */
287
+ async complete(prompt, options = {}) {
288
+ this._stats.requests++;
289
+
290
+ // Check cache
291
+ if (options.cache !== false) {
292
+ const cacheKey = this._cacheKey(prompt, options);
293
+ const cached = this._cache.get(cacheKey);
294
+ if (cached && (Date.now() - cached.timestamp < 300_000)) {
295
+ this._stats.cacheHits++;
296
+ return { ...cached.result, cached: true };
297
+ }
298
+ }
299
+
300
+ const providerName = options.provider || this._defaultProvider;
301
+ const providers = [providerName, ...this._fallbackOrder.filter(p => p !== providerName)];
302
+
303
+ const endTimer = metrics.startTimer('llm.request.duration');
304
+
305
+ for (const name of providers) {
306
+ const provider = this._providers.get(name);
307
+ if (!provider?.available) continue;
308
+
309
+ try {
310
+ const result = await provider.complete(prompt, options);
311
+
312
+ endTimer();
313
+ metrics.increment('llm.requests.success', 1, { provider: name });
314
+ this._stats.totalTokens += result.usage?.totalTokens || 0;
315
+
316
+ // Cache result
317
+ if (options.cache !== false) {
318
+ const cacheKey = this._cacheKey(prompt, options);
319
+ this._cache.set(cacheKey, { result, timestamp: Date.now() });
320
+ if (this._cache.size > this._maxCache) {
321
+ const oldest = this._cache.keys().next().value;
322
+ this._cache.delete(oldest);
323
+ }
324
+ }
325
+
326
+ return { ...result, duration: endTimer() };
327
+ } catch (err) {
328
+ metrics.increment('llm.requests.failure', 1, { provider: name });
329
+ this._stats.failures++;
330
+ // Try next provider
331
+ continue;
332
+ }
333
+ }
334
+
335
+ endTimer();
336
+ throw new Error('All LLM providers failed');
337
+ }
338
+
339
+ /**
340
+ * Generate embeddings
341
+ */
342
+ async embed(text, options = {}) {
343
+ const providerName = options.provider || this._defaultProvider;
344
+ const provider = this._providers.get(providerName);
345
+ if (!provider?.available) throw new Error(`Provider ${providerName} not available`);
346
+ if (!provider.embed) throw new Error(`Provider ${providerName} does not support embeddings`);
347
+ return provider.embed(text);
348
+ }
349
+
350
+ /**
351
+ * List available models across all providers
352
+ */
353
+ listModels() {
354
+ const models = [];
355
+ for (const [name, provider] of this._providers) {
356
+ if (!provider.available) continue;
357
+ for (const model of provider.models) {
358
+ models.push({ model, provider: name });
359
+ }
360
+ }
361
+ return models;
362
+ }
363
+
364
+ /**
365
+ * Get provider status
366
+ */
367
+ getStatus() {
368
+ const providers = {};
369
+ for (const [name, provider] of this._providers) {
370
+ providers[name] = {
371
+ available: provider.available,
372
+ models: provider.models,
373
+ };
374
+ }
375
+ return {
376
+ defaultProvider: this._defaultProvider,
377
+ providers,
378
+ stats: { ...this._stats },
379
+ };
380
+ }
381
+
382
+ _cacheKey(prompt, options) {
383
+ const key = `${options.provider || ''}:${options.model || ''}:${prompt.slice(0, 200)}`;
384
+ return require('crypto').createHash('md5').update(key).digest('hex');
385
+ }
386
+ }
387
+
388
+ // ─── Singleton ──────────────────────────────────────────────────────────────
389
+
390
+ const llm = new LLMManager();
391
+
392
+ // Register default providers
393
+ llm.registerProvider(new OpenAIProvider({}));
394
+ llm.registerProvider(new AnthropicProvider({}));
395
+ llm.registerProvider(new OllamaProvider({}));
396
+
397
+ module.exports = {
398
+ LLMProvider,
399
+ OpenAIProvider,
400
+ AnthropicProvider,
401
+ OllamaProvider,
402
+ LLMManager,
403
+ llm,
404
+ };