@yigitahmetsahin/captcha-solver 2.0.1 → 3.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -30,8 +30,10 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
30
30
  // src/index.ts
31
31
  var index_exports = {};
32
32
  __export(index_exports, {
33
+ LEGACY_CONFUSION_GROUPS: () => LEGACY_CONFUSION_GROUPS,
33
34
  Solver: () => Solver,
34
35
  imageToBase64: () => imageToBase64,
36
+ majorityVote: () => majorityVote,
35
37
  preprocessCaptcha: () => preprocessCaptcha,
36
38
  preprocessCaptchaToBuffer: () => preprocessCaptchaToBuffer
37
39
  });
@@ -44,47 +46,98 @@ var import_ai = require("ai");
44
46
  var import_fs = __toESM(require("fs"), 1);
45
47
  var import_path = __toESM(require("path"), 1);
46
48
  var import_sharp = __toESM(require("sharp"), 1);
47
- async function preprocessCaptcha(input) {
48
- const buf = await preprocessCaptchaToBuffer(input);
49
+ var LEGACY_CROP = { left: 0.1, top: 0.02, right: 0.9, bottom: 0.6 };
50
+ async function preprocessCaptcha(input, options) {
51
+ const buf = await preprocessCaptchaToBuffer(input, options);
49
52
  return buf.toString("base64");
50
53
  }
51
- async function preprocessCaptchaToBuffer(input) {
54
+ async function preprocessCaptchaToBuffer(input, options) {
55
+ const {
56
+ blur = 1.5,
57
+ scale = 4,
58
+ contrast = 3,
59
+ sharpen = true,
60
+ crop = "auto",
61
+ padding = true,
62
+ negate = false,
63
+ greyscale = true
64
+ } = options ?? {};
52
65
  const source = typeof input === "string" ? import_path.default.resolve(input) : input;
53
66
  const metadata = await (0, import_sharp.default)(source).metadata();
54
67
  const origW = metadata.width;
55
68
  const origH = metadata.height;
56
- const smoothed = await (0, import_sharp.default)(source).blur(1.5).greyscale().toBuffer();
57
- const upscaled = await (0, import_sharp.default)(smoothed).resize(origW * 4, origH * 4, { kernel: "lanczos3" }).toBuffer();
58
- const stats = await (0, import_sharp.default)(upscaled).stats();
59
- const mean = stats.channels[0].mean;
60
- const enhanced = await (0, import_sharp.default)(upscaled).linear(3, mean * (1 - 3)).sharpen({ sigma: 1, m1: 2, m2: 1 }).toBuffer();
61
- const scaledW = origW * 4;
62
- const scaledH = origH * 4;
63
- const cropLeft = Math.floor(scaledW * 0.1);
64
- const cropTop = Math.floor(scaledH * 0.02);
65
- const cropRight = Math.floor(scaledW * 0.9);
66
- const cropBottom = Math.floor(scaledH * 0.6);
67
- const cropW = cropRight - cropLeft;
68
- const cropH = cropBottom - cropTop;
69
- return (0, import_sharp.default)(enhanced).extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH }).extend({
70
- top: 20,
71
- bottom: 20,
72
- left: 30,
73
- right: 30,
69
+ let pipeline = (0, import_sharp.default)(source);
70
+ if (blur > 0) pipeline = pipeline.blur(blur);
71
+ if (greyscale) pipeline = pipeline.greyscale();
72
+ const smoothed = await pipeline.toBuffer();
73
+ const upscaled = await (0, import_sharp.default)(smoothed).resize(origW * scale, origH * scale, { kernel: "lanczos3" }).toBuffer();
74
+ let enhanced;
75
+ if (contrast !== 1) {
76
+ const stats = await (0, import_sharp.default)(upscaled).stats();
77
+ const mean = stats.channels[0].mean;
78
+ let pipe = (0, import_sharp.default)(upscaled).linear(contrast, mean * (1 - contrast));
79
+ if (sharpen) pipe = pipe.sharpen({ sigma: 1, m1: 2, m2: 1 });
80
+ enhanced = await pipe.toBuffer();
81
+ } else {
82
+ enhanced = sharpen ? await (0, import_sharp.default)(upscaled).sharpen({ sigma: 1, m1: 2, m2: 1 }).toBuffer() : upscaled;
83
+ }
84
+ let cropped;
85
+ if (crop === "none") {
86
+ cropped = enhanced;
87
+ } else if (crop === "auto") {
88
+ cropped = await autoCrop(enhanced);
89
+ } else {
90
+ const fractions = crop === "legacy" ? LEGACY_CROP : crop;
91
+ const scaledW = origW * scale;
92
+ const scaledH = origH * scale;
93
+ const cropLeft = Math.floor(scaledW * fractions.left);
94
+ const cropTop = Math.floor(scaledH * fractions.top);
95
+ const cropRight = Math.floor(scaledW * fractions.right);
96
+ const cropBottom = Math.floor(scaledH * fractions.bottom);
97
+ const cropW = cropRight - cropLeft;
98
+ const cropH = cropBottom - cropTop;
99
+ cropped = await (0, import_sharp.default)(enhanced).extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH }).toBuffer();
100
+ }
101
+ const final = negate ? await (0, import_sharp.default)(cropped).negate().toBuffer() : cropped;
102
+ if (padding === false) {
103
+ return (0, import_sharp.default)(final).png().toBuffer();
104
+ }
105
+ const pad = typeof padding === "number" ? padding : void 0;
106
+ const vPad = pad ?? 20;
107
+ const hPad = pad ?? 30;
108
+ return (0, import_sharp.default)(final).extend({
109
+ top: vPad,
110
+ bottom: vPad,
111
+ left: hPad,
112
+ right: hPad,
74
113
  background: { r: 255, g: 255, b: 255 }
75
114
  }).png().toBuffer();
76
115
  }
116
+ async function autoCrop(enhanced) {
117
+ try {
118
+ const trimmed = (0, import_sharp.default)(enhanced).trim({ threshold: 30 });
119
+ const trimmedBuf = await trimmed.toBuffer({ resolveWithObject: true });
120
+ const { width, height } = trimmedBuf.info;
121
+ if (width > 2 && height > 2) {
122
+ return trimmedBuf.data;
123
+ }
124
+ } catch {
125
+ }
126
+ return enhanced;
127
+ }
77
128
  function imageToBase64(imagePath) {
78
129
  const buffer = import_fs.default.readFileSync(imagePath);
79
130
  return buffer.toString("base64");
80
131
  }
81
132
 
82
133
  // src/solver.ts
83
- var PROMPT = `You are an assistant helping a visually impaired person read distorted text from an image.
84
- The text contains uppercase letters A-Z and/or digits 0-9.
85
- A thin vertical stroke is the digit 1. Never read it as the letter I or L.
86
- A round closed shape is the letter O, not the letter D.
87
- Output ONLY the exact characters you read, nothing else.`;
134
+ var PROMPT = `You are an expert OCR assistant reading distorted text from a CAPTCHA image.
135
+ Two versions of the same captcha are provided. Cross-reference both to determine the correct text.
136
+ The text may contain uppercase letters (A-Z), lowercase letters (a-z), and/or digits (0-9).
137
+ Pay close attention to:
138
+ - Letter case: lowercase "e" has a horizontal bar inside, digit "0" does not. Lowercase "r" has a short descender, uppercase "T" has a flat top.
139
+ - Similar shapes: "5" has a flat top + curved bottom, "S" is fully curved. "4" has an angled stroke, "A" has a pointed top. "6" has a closed bottom loop, "8" has two loops. "2" has a curved top + flat bottom, "Z" has all straight lines.
140
+ Output ONLY the exact characters you read, preserving case. Nothing else.`;
88
141
  var DEFAULT_MODELS = {
89
142
  openai: "gpt-4o",
90
143
  anthropic: "claude-sonnet-4-20250514",
@@ -110,7 +163,7 @@ async function resolveModel(apiKey, provider, modelId) {
110
163
  );
111
164
  }
112
165
  }
113
- var CONFUSION_GROUPS = {
166
+ var LEGACY_CONFUSION_GROUPS = {
114
167
  "1": "1",
115
168
  I: "1",
116
169
  L: "1",
@@ -122,7 +175,7 @@ var CONFUSION_GROUPS = {
122
175
  Z: "Z",
123
176
  "2": "Z"
124
177
  };
125
- function majorityVote(attempts, expectedLength) {
178
+ function majorityVote(attempts, expectedLength, groups) {
126
179
  let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;
127
180
  if (filtered.length === 0) {
128
181
  filtered = attempts;
@@ -142,6 +195,7 @@ function majorityVote(attempts, expectedLength) {
142
195
  }
143
196
  const sameLenAttempts = filtered.filter((a) => a.length === bestLen);
144
197
  if (sameLenAttempts.length === 0) return filtered[0];
198
+ const useGroups = groups && typeof groups === "object" ? groups : void 0;
145
199
  const result = [];
146
200
  for (let pos = 0; pos < bestLen; pos++) {
147
201
  const charCounts = /* @__PURE__ */ new Map();
@@ -149,20 +203,32 @@ function majorityVote(attempts, expectedLength) {
149
203
  const ch = a[pos];
150
204
  charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);
151
205
  }
152
- const groupCounts = /* @__PURE__ */ new Map();
153
- for (const [ch, count] of charCounts) {
154
- const canonical = CONFUSION_GROUPS[ch] ?? ch;
155
- groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);
156
- }
157
- let bestGroup = "";
158
- let bestGroupCount = 0;
159
- for (const [canonical, count] of groupCounts) {
160
- if (count > bestGroupCount) {
161
- bestGroup = canonical;
162
- bestGroupCount = count;
206
+ if (useGroups) {
207
+ const groupCounts = /* @__PURE__ */ new Map();
208
+ for (const [ch, count] of charCounts) {
209
+ const canonical = useGroups[ch] ?? ch;
210
+ groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);
211
+ }
212
+ let bestGroup = "";
213
+ let bestGroupCount = 0;
214
+ for (const [canonical, count] of groupCounts) {
215
+ if (count > bestGroupCount) {
216
+ bestGroup = canonical;
217
+ bestGroupCount = count;
218
+ }
219
+ }
220
+ result.push(bestGroup);
221
+ } else {
222
+ let bestChar = "";
223
+ let bestCharCount = 0;
224
+ for (const [ch, count] of charCounts) {
225
+ if (count > bestCharCount) {
226
+ bestChar = ch;
227
+ bestCharCount = count;
228
+ }
163
229
  }
230
+ result.push(bestChar);
164
231
  }
165
- result.push(bestGroup);
166
232
  }
167
233
  return result.join("");
168
234
  }
@@ -258,11 +324,32 @@ var Solver = class {
258
324
  * @returns Solved text, per-attempt answers, and token usage
259
325
  */
260
326
  async solve(input, options = {}) {
261
- const { numAttempts = 5, expectedLength, maxRetries = 2, verbose = true } = options;
327
+ const {
328
+ numAttempts = 7,
329
+ expectedLength,
330
+ maxRetries = 2,
331
+ verbose = true,
332
+ confusionGroups = false,
333
+ preprocess
334
+ } = options;
262
335
  const model = await this.getModel();
263
- const imageBuffer = await preprocessCaptchaToBuffer(input);
336
+ const [enhancedBuffer, colorBuffer] = await Promise.all([
337
+ preprocessCaptchaToBuffer(input, preprocess),
338
+ preprocessCaptchaToBuffer(input, {
339
+ blur: 0,
340
+ scale: 4,
341
+ contrast: 1,
342
+ sharpen: false,
343
+ crop: "none",
344
+ padding: 40,
345
+ greyscale: false
346
+ })
347
+ ]);
264
348
  const results = await Promise.all(
265
- Array.from({ length: numAttempts }, () => this.singleAttempt(model, imageBuffer, maxRetries))
349
+ Array.from(
350
+ { length: numAttempts },
351
+ () => this.singleAttempt(model, enhancedBuffer, colorBuffer, maxRetries)
352
+ )
266
353
  );
267
354
  const valid = results.filter((r) => r !== null);
268
355
  if (verbose) {
@@ -275,13 +362,18 @@ var Solver = class {
275
362
  if (verbose) console.log(" All attempts failed!");
276
363
  return { text: "", attempts, usage, attemptUsages };
277
364
  }
278
- return { text: majorityVote(attempts, expectedLength), attempts, usage, attemptUsages };
365
+ return {
366
+ text: majorityVote(attempts, expectedLength, confusionGroups),
367
+ attempts,
368
+ usage,
369
+ attemptUsages
370
+ };
279
371
  }
280
372
  /**
281
373
  * Make a single API call to read the captcha.
282
374
  * Retries up to `maxRetries` times on failure.
283
375
  */
284
- async singleAttempt(model, imageBuffer, maxRetries) {
376
+ async singleAttempt(model, primaryBuffer, secondaryBuffer, maxRetries) {
285
377
  for (let retry = 0; retry <= maxRetries; retry++) {
286
378
  try {
287
379
  const { text, usage } = await (0, import_ai.generateText)({
@@ -291,7 +383,8 @@ var Solver = class {
291
383
  role: "user",
292
384
  content: [
293
385
  { type: "text", text: PROMPT },
294
- { type: "image", image: imageBuffer }
386
+ { type: "image", image: primaryBuffer },
387
+ { type: "image", image: secondaryBuffer }
295
388
  ]
296
389
  }
297
390
  ],
@@ -303,7 +396,7 @@ var Solver = class {
303
396
  if (lower.includes("sorry") || lower.includes("can't help") || lower.includes("cannot help") || lower.includes("unable to") || lower.includes("i can't") || raw.length > 20) {
304
397
  return null;
305
398
  }
306
- const cleaned = raw.toUpperCase().replace(/[^A-Z0-9]/g, "");
399
+ const cleaned = raw.replace(/[^A-Za-z0-9]/g, "");
307
400
  return cleaned ? { text: cleaned, usage } : null;
308
401
  } catch (_err) {
309
402
  if (retry < maxRetries) {
@@ -318,8 +411,10 @@ var Solver = class {
318
411
  };
319
412
  // Annotate the CommonJS export names for ESM import in node:
320
413
  0 && (module.exports = {
414
+ LEGACY_CONFUSION_GROUPS,
321
415
  Solver,
322
416
  imageToBase64,
417
+ majorityVote,
323
418
  preprocessCaptcha,
324
419
  preprocessCaptchaToBuffer
325
420
  });
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/index.ts","../src/solver.ts","../src/preprocess.ts"],"sourcesContent":["export { Solver } from './solver.js';\nexport type { SolverOptions, SolveOptions, SolveResult, Provider } from './solver.js';\nexport type { LanguageModelUsage } from 'ai';\nexport { preprocessCaptcha, preprocessCaptchaToBuffer, imageToBase64 } from './preprocess.js';\n","import type { LanguageModel, LanguageModelUsage } from 'ai';\nimport { generateText } from 'ai';\nimport { preprocessCaptchaToBuffer } from './preprocess.js';\n\nconst PROMPT = `You are an assistant helping a visually impaired person read distorted text from an image.\nThe text contains uppercase letters A-Z and/or digits 0-9.\nA thin vertical stroke is the digit 1. Never read it as the letter I or L.\nA round closed shape is the letter O, not the letter D.\nOutput ONLY the exact characters you read, nothing else.`;\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport type Provider = 'openai' | 'anthropic' | 'google';\n\nexport interface SolverOptions {\n /** AI provider to use when constructing the model from an API key (default: \"openai\") */\n provider?: Provider;\n /** Model ID passed to the provider (default: \"gpt-4o\") */\n model?: string;\n}\n\nexport interface SolveOptions {\n /** Number of voting attempts (default: 5) */\n numAttempts?: number;\n /** Expected captcha length — results of other lengths are discarded */\n expectedLength?: number;\n /** Max retries per attempt on API failure (default: 2) */\n maxRetries?: number;\n /** Whether to log attempt details (default: true) */\n verbose?: boolean;\n}\n\nexport interface SolveResult {\n /** The solved captcha text (majority-voted) */\n text: string;\n /** Per-attempt raw answers (before voting) */\n attempts: string[];\n /** Aggregated token usage across all parallel attempts */\n usage: LanguageModelUsage;\n /** Per-attempt usage breakdown */\n attemptUsages: LanguageModelUsage[];\n}\n\ninterface AttemptResult {\n text: string;\n usage: LanguageModelUsage;\n}\n\n// ── Provider resolution ──────────────────────────────────────────────\n\nconst DEFAULT_MODELS: Record<Provider, string> = {\n openai: 'gpt-4o',\n anthropic: 'claude-sonnet-4-20250514',\n google: 'gemini-2.0-flash',\n};\n\nasync function resolveModel(\n apiKey: string,\n provider: Provider,\n modelId: string\n): Promise<LanguageModel> {\n switch (provider) {\n case 'openai': {\n const { createOpenAI } = await import('@ai-sdk/openai');\n return createOpenAI({ apiKey })(modelId);\n }\n case 'anthropic': {\n // @ts-expect-error — optional peer dependency\n const { createAnthropic } = await import('@ai-sdk/anthropic');\n return createAnthropic({ apiKey })(modelId);\n }\n case 'google': {\n // @ts-expect-error — optional peer dependency\n const { createGoogleGenerativeAI } = await import('@ai-sdk/google');\n return createGoogleGenerativeAI({ apiKey })(modelId);\n }\n default:\n throw new Error(\n `Unknown provider \"${provider}\". Install the matching @ai-sdk/* package and pass the model directly.`\n );\n }\n}\n\n// ── Confusion groups ─────────────────────────────────────────────────\n\n/**\n * Characters the model commonly misreads as each other.\n * Each group maps to its canonical (most likely correct) character.\n */\nconst CONFUSION_GROUPS: Record<string, string> = {\n '1': '1',\n I: '1',\n L: '1',\n O: 'O',\n D: 'O',\n '0': 'O',\n S: 'S',\n '5': 'S',\n Z: 'Z',\n '2': 'Z',\n};\n\n// ── Majority voting ──────────────────────────────────────────────────\n\n/**\n * Character-level majority vote across multiple attempts.\n * Uses confusion-aware voting: characters that the model commonly\n * confuses (e.g. 1/I/L, O/D/0) are grouped together during counting.\n */\nfunction majorityVote(attempts: string[], expectedLength?: number): string {\n let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;\n\n if (filtered.length === 0) {\n filtered = attempts;\n }\n if (filtered.length === 0) return '';\n\n // Find most common length\n const lenCounts = new Map<number, number>();\n for (const a of filtered) {\n lenCounts.set(a.length, (lenCounts.get(a.length) ?? 0) + 1);\n }\n let bestLen = 0;\n let bestCount = 0;\n for (const [len, count] of lenCounts) {\n if (count > bestCount) {\n bestLen = len;\n bestCount = count;\n }\n }\n\n const sameLenAttempts = filtered.filter((a) => a.length === bestLen);\n if (sameLenAttempts.length === 0) return filtered[0];\n\n // Vote per character position with confusion-aware grouping\n const result: string[] = [];\n for (let pos = 0; pos < bestLen; pos++) {\n const charCounts = new Map<string, number>();\n for (const a of sameLenAttempts) {\n const ch = a[pos];\n charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);\n }\n\n const groupCounts = new Map<string, number>();\n for (const [ch, count] of charCounts) {\n const canonical = CONFUSION_GROUPS[ch] ?? ch;\n groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);\n }\n\n let bestGroup = '';\n let bestGroupCount = 0;\n for (const [canonical, count] of groupCounts) {\n if (count > bestGroupCount) {\n bestGroup = canonical;\n bestGroupCount = count;\n }\n }\n\n result.push(bestGroup);\n }\n\n return result.join('');\n}\n\n// ── Usage aggregation ────────────────────────────────────────────────\n\nfunction sumOptional(a: number | undefined, b: number | undefined): number | undefined {\n if (a === undefined && b === undefined) return undefined;\n return (a ?? 0) + (b ?? 0);\n}\n\nfunction aggregateUsage(usages: LanguageModelUsage[]): LanguageModelUsage {\n const zero: LanguageModelUsage = {\n inputTokens: undefined,\n inputTokenDetails: {\n noCacheTokens: undefined,\n cacheReadTokens: undefined,\n cacheWriteTokens: undefined,\n },\n outputTokens: undefined,\n outputTokenDetails: {\n textTokens: undefined,\n reasoningTokens: undefined,\n },\n totalTokens: undefined,\n };\n return usages.reduce<LanguageModelUsage>(\n (acc, u) => ({\n inputTokens: sumOptional(acc.inputTokens, u.inputTokens),\n inputTokenDetails: {\n noCacheTokens: sumOptional(\n acc.inputTokenDetails.noCacheTokens,\n u.inputTokenDetails.noCacheTokens\n ),\n cacheReadTokens: sumOptional(\n acc.inputTokenDetails.cacheReadTokens,\n u.inputTokenDetails.cacheReadTokens\n ),\n cacheWriteTokens: sumOptional(\n acc.inputTokenDetails.cacheWriteTokens,\n u.inputTokenDetails.cacheWriteTokens\n ),\n },\n outputTokens: sumOptional(acc.outputTokens, u.outputTokens),\n outputTokenDetails: {\n textTokens: sumOptional(acc.outputTokenDetails.textTokens, u.outputTokenDetails.textTokens),\n reasoningTokens: sumOptional(\n acc.outputTokenDetails.reasoningTokens,\n u.outputTokenDetails.reasoningTokens\n ),\n },\n totalTokens: sumOptional(acc.totalTokens, u.totalTokens),\n }),\n zero\n );\n}\n\n// ── Solver class ─────────────────────────────────────────────────────\n\nexport class Solver {\n private _model: LanguageModel | null = null;\n private _pendingModel: Promise<LanguageModel> | null = null;\n\n /**\n * Create a captcha solver.\n *\n * @example\n * // Simple — defaults to OpenAI gpt-4o\n * const solver = new Solver('sk-...');\n *\n * @example\n * // Specify provider and model\n * const solver = new Solver('sk-ant-...', { provider: 'anthropic', model: 'claude-sonnet-4-20250514' });\n *\n * @example\n * // Pass an AI SDK model directly\n * import { createOpenAI } from '@ai-sdk/openai';\n * const openai = createOpenAI({ apiKey: 'sk-...' });\n * const solver = new Solver(openai('gpt-4o'));\n */\n constructor(keyOrModel: string | LanguageModel, options?: SolverOptions) {\n if (typeof keyOrModel === 'string') {\n const provider = options?.provider ?? 'openai';\n const modelId = options?.model ?? DEFAULT_MODELS[provider];\n // Lazily resolve the model on first use\n this._pendingModel = resolveModel(keyOrModel, provider, modelId);\n } else {\n this._model = keyOrModel;\n }\n }\n\n private async getModel(): Promise<LanguageModel> {\n if (this._model) return this._model;\n this._model = await this._pendingModel!;\n this._pendingModel = null;\n return this._model;\n }\n\n /**\n * Solve a captcha image.\n *\n * @param input - File path (string) or raw image Buffer\n * @param options - Solve options (attempts, expected length, etc.)\n * @returns Solved text, per-attempt answers, and token usage\n */\n async solve(input: string | Buffer, options: SolveOptions = {}): Promise<SolveResult> {\n const { numAttempts = 5, expectedLength, maxRetries = 2, verbose = true } = options;\n\n const model = await this.getModel();\n const imageBuffer = await preprocessCaptchaToBuffer(input);\n\n // Fire all attempts in parallel for speed\n const results = await Promise.all(\n Array.from({ length: numAttempts }, () => this.singleAttempt(model, imageBuffer, maxRetries))\n );\n const valid = results.filter((r): r is AttemptResult => r !== null);\n if (verbose) {\n valid.forEach((r, i) => console.log(` Attempt ${i + 1}: ${r.text}`));\n }\n\n const attempts = valid.map((r) => r.text);\n const attemptUsages = valid.map((r) => r.usage);\n const usage = aggregateUsage(attemptUsages);\n\n if (attempts.length === 0) {\n if (verbose) console.log(' All attempts failed!');\n return { text: '', attempts, usage, attemptUsages };\n }\n\n return { text: majorityVote(attempts, expectedLength), attempts, usage, attemptUsages };\n }\n\n /**\n * Make a single API call to read the captcha.\n * Retries up to `maxRetries` times on failure.\n */\n private async singleAttempt(\n model: LanguageModel,\n imageBuffer: Buffer,\n maxRetries: number\n ): Promise<AttemptResult | null> {\n for (let retry = 0; retry <= maxRetries; retry++) {\n try {\n const { text, usage } = await generateText({\n model,\n messages: [\n {\n role: 'user',\n content: [\n { type: 'text', text: PROMPT },\n { type: 'image', image: imageBuffer },\n ],\n },\n ],\n temperature: 1,\n maxOutputTokens: 256,\n });\n\n const raw = text.trim();\n\n // Detect refusals\n const lower = raw.toLowerCase();\n if (\n lower.includes('sorry') ||\n lower.includes(\"can't help\") ||\n lower.includes('cannot help') ||\n lower.includes('unable to') ||\n lower.includes(\"i can't\") ||\n raw.length > 20\n ) {\n return null;\n }\n\n // Clean: keep only uppercase letters and digits\n const cleaned = raw.toUpperCase().replace(/[^A-Z0-9]/g, '');\n return cleaned ? { text: cleaned, usage } : null;\n } catch (_err) {\n if (retry < maxRetries) {\n await new Promise((r) => setTimeout(r, 1000 * (retry + 1)));\n continue;\n }\n return null;\n }\n }\n return null;\n }\n}\n","import fs from 'fs';\nimport path from 'path';\nimport sharp from 'sharp';\n\n/**\n * Preprocess a captcha image using sharp (libvips).\n *\n * Pipeline:\n * 1. Gaussian blur in color space (smooths dither pattern)\n * 2. Grayscale conversion\n * 3. Upscale 4× with Lanczos\n * 4. Contrast boost (3× around image mean) + sharpen\n * 5. Crop decorative borders\n * 6. Add white padding\n *\n * Accepts a file path or a raw image Buffer.\n * Returns a base64-encoded PNG string.\n */\nexport async function preprocessCaptcha(input: string | Buffer): Promise<string> {\n const buf = await preprocessCaptchaToBuffer(input);\n return buf.toString('base64');\n}\n\n/**\n * Same preprocessing pipeline as `preprocessCaptcha`, but returns the\n * resulting PNG as a raw Buffer (useful for AI SDK image content parts).\n */\nexport async function preprocessCaptchaToBuffer(input: string | Buffer): Promise<Buffer> {\n const source = typeof input === 'string' ? path.resolve(input) : input;\n\n // Read original dimensions for crop/resize calculations\n const metadata = await sharp(source).metadata();\n const origW = metadata.width!;\n const origH = metadata.height!;\n\n // Step 1-2: Blur in color space (smooths dither pattern) → greyscale\n // Separate from resize to prevent pipeline reordering\n const smoothed = await sharp(source).blur(1.5).greyscale().toBuffer();\n\n // Step 3: Upscale 4× with Lanczos\n const upscaled = await sharp(smoothed)\n .resize(origW * 4, origH * 4, { kernel: 'lanczos3' })\n .toBuffer();\n\n // Step 4: Contrast 3× around actual image mean + sharpen\n // Matches PIL's ImageEnhance.Contrast: output = factor*input + mean*(1-factor)\n const stats = await sharp(upscaled).stats();\n const mean = stats.channels[0].mean;\n const enhanced = await sharp(upscaled)\n .linear(3.0, mean * (1 - 3.0))\n .sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 })\n .toBuffer();\n\n // Step 5: Crop decorative borders\n // Remove 10% left/right, 2% top, 40% bottom (keep top 60%)\n // Math.floor matches Python's int() truncation\n const scaledW = origW * 4;\n const scaledH = origH * 4;\n const cropLeft = Math.floor(scaledW * 0.1);\n const cropTop = Math.floor(scaledH * 0.02);\n const cropRight = Math.floor(scaledW * 0.9);\n const cropBottom = Math.floor(scaledH * 0.6);\n const cropW = cropRight - cropLeft;\n const cropH = cropBottom - cropTop;\n\n // Step 5-6: Crop → add white padding → output PNG\n return sharp(enhanced)\n .extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH })\n .extend({\n top: 20,\n bottom: 20,\n left: 30,\n right: 30,\n background: { r: 255, g: 255, b: 255 },\n })\n .png()\n .toBuffer();\n}\n\n/**\n * Read an image file and return its base64-encoded content.\n */\nexport function imageToBase64(imagePath: string): string {\n const buffer = fs.readFileSync(imagePath);\n return buffer.toString('base64');\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACCA,gBAA6B;;;ACD7B,gBAAe;AACf,kBAAiB;AACjB,mBAAkB;AAgBlB,eAAsB,kBAAkB,OAAyC;AAC/E,QAAM,MAAM,MAAM,0BAA0B,KAAK;AACjD,SAAO,IAAI,SAAS,QAAQ;AAC9B;AAMA,eAAsB,0BAA0B,OAAyC;AACvF,QAAM,SAAS,OAAO,UAAU,WAAW,YAAAA,QAAK,QAAQ,KAAK,IAAI;AAGjE,QAAM,WAAW,UAAM,aAAAC,SAAM,MAAM,EAAE,SAAS;AAC9C,QAAM,QAAQ,SAAS;AACvB,QAAM,QAAQ,SAAS;AAIvB,QAAM,WAAW,UAAM,aAAAA,SAAM,MAAM,EAAE,KAAK,GAAG,EAAE,UAAU,EAAE,SAAS;AAGpE,QAAM,WAAW,UAAM,aAAAA,SAAM,QAAQ,EAClC,OAAO,QAAQ,GAAG,QAAQ,GAAG,EAAE,QAAQ,WAAW,CAAC,EACnD,SAAS;AAIZ,QAAM,QAAQ,UAAM,aAAAA,SAAM,QAAQ,EAAE,MAAM;AAC1C,QAAM,OAAO,MAAM,SAAS,CAAC,EAAE;AAC/B,QAAM,WAAW,UAAM,aAAAA,SAAM,QAAQ,EAClC,OAAO,GAAK,QAAQ,IAAI,EAAI,EAC5B,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC,EACxC,SAAS;AAKZ,QAAM,UAAU,QAAQ;AACxB,QAAM,UAAU,QAAQ;AACxB,QAAM,WAAW,KAAK,MAAM,UAAU,GAAG;AACzC,QAAM,UAAU,KAAK,MAAM,UAAU,IAAI;AACzC,QAAM,YAAY,KAAK,MAAM,UAAU,GAAG;AAC1C,QAAM,aAAa,KAAK,MAAM,UAAU,GAAG;AAC3C,QAAM,QAAQ,YAAY;AAC1B,QAAM,QAAQ,aAAa;AAG3B,aAAO,aAAAA,SAAM,QAAQ,EAClB,QAAQ,EAAE,MAAM,UAAU,KAAK,SAAS,OAAO,OAAO,QAAQ,MAAM,CAAC,EACrE,OAAO;AAAA,IACN,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,MAAM;AAAA,IACN,OAAO;AAAA,IACP,YAAY,EAAE,GAAG,KAAK,GAAG,KAAK,GAAG,IAAI;AAAA,EACvC,CAAC,EACA,IAAI,EACJ,SAAS;AACd;AAKO,SAAS,cAAc,WAA2B;AACvD,QAAM,SAAS,UAAAC,QAAG,aAAa,SAAS;AACxC,SAAO,OAAO,SAAS,QAAQ;AACjC;;;ADjFA,IAAM,SAAS;AAAA;AAAA;AAAA;AAAA;AA8Cf,IAAM,iBAA2C;AAAA,EAC/C,QAAQ;AAAA,EACR,WAAW;AAAA,EACX,QAAQ;AACV;AAEA,eAAe,aACb,QACA,UACA,SACwB;AACxB,UAAQ,UAAU;AAAA,IAChB,KAAK,UAAU;AACb,YAAM,EAAE,aAAa,IAAI,MAAM,OAAO,gBAAgB;AACtD,aAAO,aAAa,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACzC;AAAA,IACA,KAAK,aAAa;AAEhB,YAAM,EAAE,gBAAgB,IAAI,MAAM,OAAO,mBAAmB;AAC5D,aAAO,gBAAgB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IAC5C;AAAA,IACA,KAAK,UAAU;AAEb,YAAM,EAAE,yBAAyB,IAAI,MAAM,OAAO,gBAAgB;AAClE,aAAO,yBAAyB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACrD;AAAA,IACA;AACE,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ;AAAA,MAC/B;AAAA,EACJ;AACF;AAQA,IAAM,mBAA2C;AAAA,EAC/C,KAAK;AAAA,EACL,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AACP;AASA,SAAS,aAAa,UAAoB,gBAAiC;AACzE,MAAI,WAAW,iBAAiB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,cAAc,IAAI;AAEtF,MAAI,SAAS,WAAW,GAAG;AACzB,eAAW;AAAA,EACb;AACA,MAAI,SAAS,WAAW,EAAG,QAAO;AAGlC,QAAM,YAAY,oBAAI,IAAoB;AAC1C,aAAW,KAAK,UAAU;AACxB,cAAU,IAAI,EAAE,SAAS,UAAU,IAAI,EAAE,MAAM,KAAK,KAAK,CAAC;AAAA,EAC5D;AACA,MAAI,UAAU;AACd,MAAI,YAAY;AAChB,aAAW,CAAC,KAAK,KAAK,KAAK,WAAW;AACpC,QAAI,QAAQ,WAAW;AACrB,gBAAU;AACV,kBAAY;AAAA,IACd;AAAA,EACF;AAEA,QAAM,kBAAkB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,OAAO;AACnE,MAAI,gBAAgB,WAAW,EAAG,QAAO,SAAS,CAAC;AAGnD,QAAM,SAAmB,CAAC;AAC1B,WAAS,MAAM,GAAG,MAAM,SAAS,OAAO;AACtC,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,KAAK,iBAAiB;AAC/B,YAAM,KAAK,EAAE,GAAG;AAChB,iBAAW,IAAI,KAAK,WAAW,IAAI,EAAE,KAAK,KAAK,CAAC;AAAA,IAClD;AAEA,UAAM,cAAc,oBAAI,IAAoB;AAC5C,eAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,YAAM,YAAY,iBAAiB,EAAE,KAAK;AAC1C,kBAAY,IAAI,YAAY,YAAY,IAAI,SAAS,KAAK,KAAK,KAAK;AAAA,IACtE;AAEA,QAAI,YAAY;AAChB,QAAI,iBAAiB;AACrB,eAAW,CAAC,WAAW,KAAK,KAAK,aAAa;AAC5C,UAAI,QAAQ,gBAAgB;AAC1B,oBAAY;AACZ,yBAAiB;AAAA,MACnB;AAAA,IACF;AAEA,WAAO,KAAK,SAAS;AAAA,EACvB;AAEA,SAAO,OAAO,KAAK,EAAE;AACvB;AAIA,SAAS,YAAY,GAAuB,GAA2C;AACrF,MAAI,MAAM,UAAa,MAAM,OAAW,QAAO;AAC/C,UAAQ,KAAK,MAAM,KAAK;AAC1B;AAEA,SAAS,eAAe,QAAkD;AACxE,QAAM,OAA2B;AAAA,IAC/B,aAAa;AAAA,IACb,mBAAmB;AAAA,MACjB,eAAe;AAAA,MACf,iBAAiB;AAAA,MACjB,kBAAkB;AAAA,IACpB;AAAA,IACA,cAAc;AAAA,IACd,oBAAoB;AAAA,MAClB,YAAY;AAAA,MACZ,iBAAiB;AAAA,IACnB;AAAA,IACA,aAAa;AAAA,EACf;AACA,SAAO,OAAO;AAAA,IACZ,CAAC,KAAK,OAAO;AAAA,MACX,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,MACvD,mBAAmB;AAAA,QACjB,eAAe;AAAA,UACb,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,iBAAiB;AAAA,UACf,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,kBAAkB;AAAA,UAChB,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,MACF;AAAA,MACA,cAAc,YAAY,IAAI,cAAc,EAAE,YAAY;AAAA,MAC1D,oBAAoB;AAAA,QAClB,YAAY,YAAY,IAAI,mBAAmB,YAAY,EAAE,mBAAmB,UAAU;AAAA,QAC1F,iBAAiB;AAAA,UACf,IAAI,mBAAmB;AAAA,UACvB,EAAE,mBAAmB;AAAA,QACvB;AAAA,MACF;AAAA,MACA,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,IACzD;AAAA,IACA;AAAA,EACF;AACF;AAIO,IAAM,SAAN,MAAa;AAAA,EACV,SAA+B;AAAA,EAC/B,gBAA+C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAmBvD,YAAY,YAAoC,SAAyB;AACvE,QAAI,OAAO,eAAe,UAAU;AAClC,YAAM,WAAW,SAAS,YAAY;AACtC,YAAM,UAAU,SAAS,SAAS,eAAe,QAAQ;AAEzD,WAAK,gBAAgB,aAAa,YAAY,UAAU,OAAO;AAAA,IACjE,OAAO;AACL,WAAK,SAAS;AAAA,IAChB;AAAA,EACF;AAAA,EAEA,MAAc,WAAmC;AAC/C,QAAI,KAAK,OAAQ,QAAO,KAAK;AAC7B,SAAK,SAAS,MAAM,KAAK;AACzB,SAAK,gBAAgB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EASA,MAAM,MAAM,OAAwB,UAAwB,CAAC,GAAyB;AACpF,UAAM,EAAE,cAAc,GAAG,gBAAgB,aAAa,GAAG,UAAU,KAAK,IAAI;AAE5E,UAAM,QAAQ,MAAM,KAAK,SAAS;AAClC,UAAM,cAAc,MAAM,0BAA0B,KAAK;AAGzD,UAAM,UAAU,MAAM,QAAQ;AAAA,MAC5B,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,MAAM,KAAK,cAAc,OAAO,aAAa,UAAU,CAAC;AAAA,IAC9F;AACA,UAAM,QAAQ,QAAQ,OAAO,CAAC,MAA0B,MAAM,IAAI;AAClE,QAAI,SAAS;AACX,YAAM,QAAQ,CAAC,GAAG,MAAM,QAAQ,IAAI,aAAa,IAAI,CAAC,KAAK,EAAE,IAAI,EAAE,CAAC;AAAA,IACtE;AAEA,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,EAAE,IAAI;AACxC,UAAM,gBAAgB,MAAM,IAAI,CAAC,MAAM,EAAE,KAAK;AAC9C,UAAM,QAAQ,eAAe,aAAa;AAE1C,QAAI,SAAS,WAAW,GAAG;AACzB,UAAI,QAAS,SAAQ,IAAI,wBAAwB;AACjD,aAAO,EAAE,MAAM,IAAI,UAAU,OAAO,cAAc;AAAA,IACpD;AAEA,WAAO,EAAE,MAAM,aAAa,UAAU,cAAc,GAAG,UAAU,OAAO,cAAc;AAAA,EACxF;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,cACZ,OACA,aACA,YAC+B;AAC/B,aAAS,QAAQ,GAAG,SAAS,YAAY,SAAS;AAChD,UAAI;AACF,cAAM,EAAE,MAAM,MAAM,IAAI,UAAM,wBAAa;AAAA,UACzC;AAAA,UACA,UAAU;AAAA,YACR;AAAA,cACE,MAAM;AAAA,cACN,SAAS;AAAA,gBACP,EAAE,MAAM,QAAQ,MAAM,OAAO;AAAA,gBAC7B,EAAE,MAAM,SAAS,OAAO,YAAY;AAAA,cACtC;AAAA,YACF;AAAA,UACF;AAAA,UACA,aAAa;AAAA,UACb,iBAAiB;AAAA,QACnB,CAAC;AAED,cAAM,MAAM,KAAK,KAAK;AAGtB,cAAM,QAAQ,IAAI,YAAY;AAC9B,YACE,MAAM,SAAS,OAAO,KACtB,MAAM,SAAS,YAAY,KAC3B,MAAM,SAAS,aAAa,KAC5B,MAAM,SAAS,WAAW,KAC1B,MAAM,SAAS,SAAS,KACxB,IAAI,SAAS,IACb;AACA,iBAAO;AAAA,QACT;AAGA,cAAM,UAAU,IAAI,YAAY,EAAE,QAAQ,cAAc,EAAE;AAC1D,eAAO,UAAU,EAAE,MAAM,SAAS,MAAM,IAAI;AAAA,MAC9C,SAAS,MAAM;AACb,YAAI,QAAQ,YAAY;AACtB,gBAAM,IAAI,QAAQ,CAAC,MAAM,WAAW,GAAG,OAAQ,QAAQ,EAAE,CAAC;AAC1D;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACF;","names":["path","sharp","fs"]}
1
+ {"version":3,"sources":["../src/index.ts","../src/solver.ts","../src/preprocess.ts"],"sourcesContent":["export { Solver } from './solver.js';\nexport { majorityVote, LEGACY_CONFUSION_GROUPS } from './solver.js';\nexport type { SolverOptions, SolveOptions, SolveResult, Provider } from './solver.js';\nexport type { LanguageModelUsage } from 'ai';\nexport { preprocessCaptcha, preprocessCaptchaToBuffer, imageToBase64 } from './preprocess.js';\nexport type { PreprocessOptions, CropFractions } from './preprocess.js';\n","import type { LanguageModel, LanguageModelUsage } from 'ai';\nimport { generateText } from 'ai';\nimport type { PreprocessOptions } from './preprocess.js';\nimport { preprocessCaptchaToBuffer } from './preprocess.js';\n\nconst PROMPT = `You are an expert OCR assistant reading distorted text from a CAPTCHA image.\nTwo versions of the same captcha are provided. Cross-reference both to determine the correct text.\nThe text may contain uppercase letters (A-Z), lowercase letters (a-z), and/or digits (0-9).\nPay close attention to:\n- Letter case: lowercase \"e\" has a horizontal bar inside, digit \"0\" does not. Lowercase \"r\" has a short descender, uppercase \"T\" has a flat top.\n- Similar shapes: \"5\" has a flat top + curved bottom, \"S\" is fully curved. \"4\" has an angled stroke, \"A\" has a pointed top. \"6\" has a closed bottom loop, \"8\" has two loops. \"2\" has a curved top + flat bottom, \"Z\" has all straight lines.\nOutput ONLY the exact characters you read, preserving case. Nothing else.`;\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport type Provider = 'openai' | 'anthropic' | 'google';\n\nexport interface SolverOptions {\n /** AI provider to use when constructing the model from an API key (default: \"openai\") */\n provider?: Provider;\n /** Model ID passed to the provider (default: \"gpt-4o\") */\n model?: string;\n}\n\nexport interface SolveOptions {\n /** Number of voting attempts (default: 7) */\n numAttempts?: number;\n /** Expected captcha length — results of other lengths are discarded */\n expectedLength?: number;\n /** Max retries per attempt on API failure (default: 2) */\n maxRetries?: number;\n /** Whether to log attempt details (default: true) */\n verbose?: boolean;\n /**\n * Confusion groups for majority voting.\n * Pass a Record<string, string> to merge visually similar characters,\n * or `false` to disable (default: false).\n * Use LEGACY_CONFUSION_GROUPS to restore pre-3.0 behavior.\n */\n confusionGroups?: Record<string, string> | false;\n /** Preprocessing options passed to the image pipeline */\n preprocess?: PreprocessOptions;\n}\n\nexport interface SolveResult {\n /** The solved captcha text (majority-voted) */\n text: string;\n /** Per-attempt raw answers (before voting) */\n attempts: string[];\n /** Aggregated token usage across all parallel attempts */\n usage: LanguageModelUsage;\n /** Per-attempt usage breakdown */\n attemptUsages: LanguageModelUsage[];\n}\n\ninterface AttemptResult {\n text: string;\n usage: LanguageModelUsage;\n}\n\n// ── Provider resolution ──────────────────────────────────────────────\n\nconst DEFAULT_MODELS: Record<Provider, string> = {\n openai: 'gpt-4o',\n anthropic: 'claude-sonnet-4-20250514',\n google: 'gemini-2.0-flash',\n};\n\nasync function resolveModel(\n apiKey: string,\n provider: Provider,\n modelId: string\n): Promise<LanguageModel> {\n switch (provider) {\n case 'openai': {\n const { createOpenAI } = await import('@ai-sdk/openai');\n return createOpenAI({ apiKey })(modelId);\n }\n case 'anthropic': {\n // @ts-expect-error — optional peer dependency\n const { createAnthropic } = await import('@ai-sdk/anthropic');\n return createAnthropic({ apiKey })(modelId);\n }\n case 'google': {\n // @ts-expect-error — optional peer dependency\n const { createGoogleGenerativeAI } = await import('@ai-sdk/google');\n return createGoogleGenerativeAI({ apiKey })(modelId);\n }\n default:\n throw new Error(\n `Unknown provider \"${provider}\". Install the matching @ai-sdk/* package and pass the model directly.`\n );\n }\n}\n\n// ── Confusion groups ─────────────────────────────────────────────────\n\n/**\n * Pre-3.0 confusion groups that merge visually similar characters.\n * Opt-in via `{ confusionGroups: LEGACY_CONFUSION_GROUPS }`.\n *\n * Maps: 1/I/L → '1', O/D/0 → 'O', S/5 → 'S', Z/2 → 'Z'\n */\nexport const LEGACY_CONFUSION_GROUPS: Record<string, string> = {\n '1': '1',\n I: '1',\n L: '1',\n O: 'O',\n D: 'O',\n '0': 'O',\n S: 'S',\n '5': 'S',\n Z: 'Z',\n '2': 'Z',\n};\n\n// ── Majority voting ──────────────────────────────────────────────────\n\n/**\n * Character-level majority vote across multiple attempts.\n * When `groups` is provided, visually similar characters are merged\n * during counting (e.g. 1/I/L all count toward '1').\n */\nexport function majorityVote(\n attempts: string[],\n expectedLength?: number,\n groups?: Record<string, string> | false\n): string {\n let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;\n\n if (filtered.length === 0) {\n filtered = attempts;\n }\n if (filtered.length === 0) return '';\n\n // Find most common length\n const lenCounts = new Map<number, number>();\n for (const a of filtered) {\n lenCounts.set(a.length, (lenCounts.get(a.length) ?? 0) + 1);\n }\n let bestLen = 0;\n let bestCount = 0;\n for (const [len, count] of lenCounts) {\n if (count > bestCount) {\n bestLen = len;\n bestCount = count;\n }\n }\n\n const sameLenAttempts = filtered.filter((a) => a.length === bestLen);\n if (sameLenAttempts.length === 0) return filtered[0];\n\n const useGroups = groups && typeof groups === 'object' ? groups : undefined;\n\n // Vote per character position\n const result: string[] = [];\n for (let pos = 0; pos < bestLen; pos++) {\n const charCounts = new Map<string, number>();\n for (const a of sameLenAttempts) {\n const ch = a[pos];\n charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);\n }\n\n if (useGroups) {\n // Confusion-aware voting\n const groupCounts = new Map<string, number>();\n for (const [ch, count] of charCounts) {\n const canonical = useGroups[ch] ?? ch;\n groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);\n }\n\n let bestGroup = '';\n let bestGroupCount = 0;\n for (const [canonical, count] of groupCounts) {\n if (count > bestGroupCount) {\n bestGroup = canonical;\n bestGroupCount = count;\n }\n }\n result.push(bestGroup);\n } else {\n // Simple majority — pick the most frequent raw character\n let bestChar = '';\n let bestCharCount = 0;\n for (const [ch, count] of charCounts) {\n if (count > bestCharCount) {\n bestChar = ch;\n bestCharCount = count;\n }\n }\n result.push(bestChar);\n }\n }\n\n return result.join('');\n}\n\n// ── Usage aggregation ────────────────────────────────────────────────\n\nfunction sumOptional(a: number | undefined, b: number | undefined): number | undefined {\n if (a === undefined && b === undefined) return undefined;\n return (a ?? 0) + (b ?? 0);\n}\n\nfunction aggregateUsage(usages: LanguageModelUsage[]): LanguageModelUsage {\n const zero: LanguageModelUsage = {\n inputTokens: undefined,\n inputTokenDetails: {\n noCacheTokens: undefined,\n cacheReadTokens: undefined,\n cacheWriteTokens: undefined,\n },\n outputTokens: undefined,\n outputTokenDetails: {\n textTokens: undefined,\n reasoningTokens: undefined,\n },\n totalTokens: undefined,\n };\n return usages.reduce<LanguageModelUsage>(\n (acc, u) => ({\n inputTokens: sumOptional(acc.inputTokens, u.inputTokens),\n inputTokenDetails: {\n noCacheTokens: sumOptional(\n acc.inputTokenDetails.noCacheTokens,\n u.inputTokenDetails.noCacheTokens\n ),\n cacheReadTokens: sumOptional(\n acc.inputTokenDetails.cacheReadTokens,\n u.inputTokenDetails.cacheReadTokens\n ),\n cacheWriteTokens: sumOptional(\n acc.inputTokenDetails.cacheWriteTokens,\n u.inputTokenDetails.cacheWriteTokens\n ),\n },\n outputTokens: sumOptional(acc.outputTokens, u.outputTokens),\n outputTokenDetails: {\n textTokens: sumOptional(acc.outputTokenDetails.textTokens, u.outputTokenDetails.textTokens),\n reasoningTokens: sumOptional(\n acc.outputTokenDetails.reasoningTokens,\n u.outputTokenDetails.reasoningTokens\n ),\n },\n totalTokens: sumOptional(acc.totalTokens, u.totalTokens),\n }),\n zero\n );\n}\n\n// ── Solver class ─────────────────────────────────────────────────────\n\nexport class Solver {\n private _model: LanguageModel | null = null;\n private _pendingModel: Promise<LanguageModel> | null = null;\n\n /**\n * Create a captcha solver.\n *\n * @example\n * // Simple — defaults to OpenAI gpt-4o\n * const solver = new Solver('sk-...');\n *\n * @example\n * // Specify provider and model\n * const solver = new Solver('sk-ant-...', { provider: 'anthropic', model: 'claude-sonnet-4-20250514' });\n *\n * @example\n * // Pass an AI SDK model directly\n * import { createOpenAI } from '@ai-sdk/openai';\n * const openai = createOpenAI({ apiKey: 'sk-...' });\n * const solver = new Solver(openai('gpt-4o'));\n */\n constructor(keyOrModel: string | LanguageModel, options?: SolverOptions) {\n if (typeof keyOrModel === 'string') {\n const provider = options?.provider ?? 'openai';\n const modelId = options?.model ?? DEFAULT_MODELS[provider];\n // Lazily resolve the model on first use\n this._pendingModel = resolveModel(keyOrModel, provider, modelId);\n } else {\n this._model = keyOrModel;\n }\n }\n\n private async getModel(): Promise<LanguageModel> {\n if (this._model) return this._model;\n this._model = await this._pendingModel!;\n this._pendingModel = null;\n return this._model;\n }\n\n /**\n * Solve a captcha image.\n *\n * @param input - File path (string) or raw image Buffer\n * @param options - Solve options (attempts, expected length, etc.)\n * @returns Solved text, per-attempt answers, and token usage\n */\n async solve(input: string | Buffer, options: SolveOptions = {}): Promise<SolveResult> {\n const {\n numAttempts = 7,\n expectedLength,\n maxRetries = 2,\n verbose = true,\n confusionGroups = false,\n preprocess,\n } = options;\n\n const model = await this.getModel();\n\n // Two complementary views:\n // 1. Enhanced grayscale (high contrast + auto-crop) — great for clear text\n // 2. Color original (upscaled, no greyscale, no contrast) — preserves subtle features\n const [enhancedBuffer, colorBuffer] = await Promise.all([\n preprocessCaptchaToBuffer(input, preprocess),\n preprocessCaptchaToBuffer(input, {\n blur: 0,\n scale: 4,\n contrast: 1.0,\n sharpen: false,\n crop: 'none',\n padding: 40,\n greyscale: false,\n }),\n ]);\n\n // Fire all attempts in parallel for speed\n const results = await Promise.all(\n Array.from({ length: numAttempts }, () =>\n this.singleAttempt(model, enhancedBuffer, colorBuffer, maxRetries)\n )\n );\n const valid = results.filter((r): r is AttemptResult => r !== null);\n if (verbose) {\n valid.forEach((r, i) => console.log(` Attempt ${i + 1}: ${r.text}`));\n }\n\n const attempts = valid.map((r) => r.text);\n const attemptUsages = valid.map((r) => r.usage);\n const usage = aggregateUsage(attemptUsages);\n\n if (attempts.length === 0) {\n if (verbose) console.log(' All attempts failed!');\n return { text: '', attempts, usage, attemptUsages };\n }\n\n return {\n text: majorityVote(attempts, expectedLength, confusionGroups),\n attempts,\n usage,\n attemptUsages,\n };\n }\n\n /**\n * Make a single API call to read the captcha.\n * Retries up to `maxRetries` times on failure.\n */\n private async singleAttempt(\n model: LanguageModel,\n primaryBuffer: Buffer,\n secondaryBuffer: Buffer,\n maxRetries: number\n ): Promise<AttemptResult | null> {\n for (let retry = 0; retry <= maxRetries; retry++) {\n try {\n const { text, usage } = await generateText({\n model,\n messages: [\n {\n role: 'user',\n content: [\n { type: 'text', text: PROMPT },\n { type: 'image', image: primaryBuffer },\n { type: 'image', image: secondaryBuffer },\n ],\n },\n ],\n temperature: 1,\n maxOutputTokens: 256,\n });\n\n const raw = text.trim();\n\n // Detect refusals\n const lower = raw.toLowerCase();\n if (\n lower.includes('sorry') ||\n lower.includes(\"can't help\") ||\n lower.includes('cannot help') ||\n lower.includes('unable to') ||\n lower.includes(\"i can't\") ||\n raw.length > 20\n ) {\n return null;\n }\n\n // Clean: keep only uppercase letters and digits\n const cleaned = raw.replace(/[^A-Za-z0-9]/g, '');\n return cleaned ? { text: cleaned, usage } : null;\n } catch (_err) {\n if (retry < maxRetries) {\n await new Promise((r) => setTimeout(r, 1000 * (retry + 1)));\n continue;\n }\n return null;\n }\n }\n return null;\n }\n}\n","import fs from 'fs';\nimport path from 'path';\nimport sharp from 'sharp';\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport interface CropFractions {\n /** Fraction from left edge (0–1, default: 0.1) */\n left: number;\n /** Fraction from top edge (0–1, default: 0.02) */\n top: number;\n /** Fraction from left to keep (0–1, default: 0.9) */\n right: number;\n /** Fraction from top to keep (0–1, default: 0.6) */\n bottom: number;\n}\n\nexport interface PreprocessOptions {\n /** Gaussian blur radius (default: 1.5). Set to 0 to skip. */\n blur?: number;\n /** Upscale factor (default: 4) */\n scale?: number;\n /** Contrast multiplier around image mean (default: 3.0). Set to 1 to skip. */\n contrast?: number;\n /** Enable unsharp-mask sharpening (default: true) */\n sharpen?: boolean;\n /**\n * Crop mode (default: 'auto'):\n * - 'auto' – trim whitespace after contrast enhancement, with margin\n * - 'legacy' – fixed-percentage crop (original behavior)\n * - 'none' – skip cropping\n * - CropFractions – custom crop percentages\n */\n crop?: 'auto' | 'legacy' | 'none' | CropFractions;\n /** Add white padding around the result (default: true). Pass false to skip, or a number for custom px. */\n padding?: boolean | number;\n /** Invert colors (negate) after processing (default: false) */\n negate?: boolean;\n /** Convert to greyscale (default: true) */\n greyscale?: boolean;\n}\n\nconst LEGACY_CROP: CropFractions = { left: 0.1, top: 0.02, right: 0.9, bottom: 0.6 };\n\n// ── Public API ───────────────────────────────────────────────────────\n\n/**\n * Preprocess a captcha image and return a base64-encoded PNG string.\n */\nexport async function preprocessCaptcha(\n input: string | Buffer,\n options?: PreprocessOptions\n): Promise<string> {\n const buf = await preprocessCaptchaToBuffer(input, options);\n return buf.toString('base64');\n}\n\n/**\n * Preprocess a captcha image and return the resulting PNG as a raw Buffer.\n *\n * Pipeline:\n * 1. Gaussian blur in color space (smooths dither pattern)\n * 2. Grayscale conversion\n * 3. Upscale with Lanczos\n * 4. Contrast boost around image mean + sharpen\n * 5. Crop (auto-detect, legacy fixed, none, or custom)\n * 6. Add white padding\n */\nexport async function preprocessCaptchaToBuffer(\n input: string | Buffer,\n options?: PreprocessOptions\n): Promise<Buffer> {\n const {\n blur = 1.5,\n scale = 4,\n contrast = 3.0,\n sharpen = true,\n crop = 'auto',\n padding = true,\n negate = false,\n greyscale = true,\n } = options ?? {};\n\n const source = typeof input === 'string' ? path.resolve(input) : input;\n\n // Read original dimensions\n const metadata = await sharp(source).metadata();\n const origW = metadata.width!;\n const origH = metadata.height!;\n\n // Step 1-2: Blur (optional) + greyscale (optional)\n let pipeline = sharp(source);\n if (blur > 0) pipeline = pipeline.blur(blur);\n if (greyscale) pipeline = pipeline.greyscale();\n const smoothed = await pipeline.toBuffer();\n\n // Step 3: Upscale with Lanczos\n const upscaled = await sharp(smoothed)\n .resize(origW * scale, origH * scale, { kernel: 'lanczos3' })\n .toBuffer();\n\n // Step 4: Contrast boost + sharpen\n let enhanced: Buffer;\n if (contrast !== 1.0) {\n const stats = await sharp(upscaled).stats();\n const mean = stats.channels[0].mean;\n let pipe = sharp(upscaled).linear(contrast, mean * (1 - contrast));\n if (sharpen) pipe = pipe.sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 });\n enhanced = await pipe.toBuffer();\n } else {\n enhanced = sharpen\n ? await sharp(upscaled).sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 }).toBuffer()\n : upscaled;\n }\n\n // Step 5: Crop\n let cropped: Buffer;\n if (crop === 'none') {\n cropped = enhanced;\n } else if (crop === 'auto') {\n cropped = await autoCrop(enhanced);\n } else {\n const fractions = crop === 'legacy' ? LEGACY_CROP : crop;\n const scaledW = origW * scale;\n const scaledH = origH * scale;\n const cropLeft = Math.floor(scaledW * fractions.left);\n const cropTop = Math.floor(scaledH * fractions.top);\n const cropRight = Math.floor(scaledW * fractions.right);\n const cropBottom = Math.floor(scaledH * fractions.bottom);\n const cropW = cropRight - cropLeft;\n const cropH = cropBottom - cropTop;\n cropped = await sharp(enhanced)\n .extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH })\n .toBuffer();\n }\n\n // Step 6: Negate (optional)\n const final = negate ? await sharp(cropped).negate().toBuffer() : cropped;\n\n // Step 7: Padding\n if (padding === false) {\n return sharp(final).png().toBuffer();\n }\n const pad = typeof padding === 'number' ? padding : undefined;\n const vPad = pad ?? 20;\n const hPad = pad ?? 30;\n return sharp(final)\n .extend({\n top: vPad,\n bottom: vPad,\n left: hPad,\n right: hPad,\n background: { r: 255, g: 255, b: 255 },\n })\n .png()\n .toBuffer();\n}\n\n/**\n * Auto-crop: use sharp.trim() to detect the content bounding box after\n * contrast enhancement, then add a small margin. Falls back to the\n * untrimmed image if trim removes everything.\n */\nasync function autoCrop(enhanced: Buffer): Promise<Buffer> {\n try {\n const trimmed = sharp(enhanced).trim({ threshold: 30 });\n const trimmedBuf = await trimmed.toBuffer({ resolveWithObject: true });\n\n // If trim left a reasonable image, add a margin\n const { width, height } = trimmedBuf.info;\n if (width > 2 && height > 2) {\n return trimmedBuf.data;\n }\n } catch {\n // trim() can throw if image is uniform — fall through\n }\n\n // Fallback: return untrimmed\n return enhanced;\n}\n\n/**\n * Read an image file and return its base64-encoded content.\n */\nexport function imageToBase64(imagePath: string): string {\n const buffer = fs.readFileSync(imagePath);\n return buffer.toString('base64');\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACCA,gBAA6B;;;ACD7B,gBAAe;AACf,kBAAiB;AACjB,mBAAkB;AAwClB,IAAM,cAA6B,EAAE,MAAM,KAAK,KAAK,MAAM,OAAO,KAAK,QAAQ,IAAI;AAOnF,eAAsB,kBACpB,OACA,SACiB;AACjB,QAAM,MAAM,MAAM,0BAA0B,OAAO,OAAO;AAC1D,SAAO,IAAI,SAAS,QAAQ;AAC9B;AAaA,eAAsB,0BACpB,OACA,SACiB;AACjB,QAAM;AAAA,IACJ,OAAO;AAAA,IACP,QAAQ;AAAA,IACR,WAAW;AAAA,IACX,UAAU;AAAA,IACV,OAAO;AAAA,IACP,UAAU;AAAA,IACV,SAAS;AAAA,IACT,YAAY;AAAA,EACd,IAAI,WAAW,CAAC;AAEhB,QAAM,SAAS,OAAO,UAAU,WAAW,YAAAA,QAAK,QAAQ,KAAK,IAAI;AAGjE,QAAM,WAAW,UAAM,aAAAC,SAAM,MAAM,EAAE,SAAS;AAC9C,QAAM,QAAQ,SAAS;AACvB,QAAM,QAAQ,SAAS;AAGvB,MAAI,eAAW,aAAAA,SAAM,MAAM;AAC3B,MAAI,OAAO,EAAG,YAAW,SAAS,KAAK,IAAI;AAC3C,MAAI,UAAW,YAAW,SAAS,UAAU;AAC7C,QAAM,WAAW,MAAM,SAAS,SAAS;AAGzC,QAAM,WAAW,UAAM,aAAAA,SAAM,QAAQ,EAClC,OAAO,QAAQ,OAAO,QAAQ,OAAO,EAAE,QAAQ,WAAW,CAAC,EAC3D,SAAS;AAGZ,MAAI;AACJ,MAAI,aAAa,GAAK;AACpB,UAAM,QAAQ,UAAM,aAAAA,SAAM,QAAQ,EAAE,MAAM;AAC1C,UAAM,OAAO,MAAM,SAAS,CAAC,EAAE;AAC/B,QAAI,WAAO,aAAAA,SAAM,QAAQ,EAAE,OAAO,UAAU,QAAQ,IAAI,SAAS;AACjE,QAAI,QAAS,QAAO,KAAK,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC;AACjE,eAAW,MAAM,KAAK,SAAS;AAAA,EACjC,OAAO;AACL,eAAW,UACP,UAAM,aAAAA,SAAM,QAAQ,EAAE,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC,EAAE,SAAS,IACzE;AAAA,EACN;AAGA,MAAI;AACJ,MAAI,SAAS,QAAQ;AACnB,cAAU;AAAA,EACZ,WAAW,SAAS,QAAQ;AAC1B,cAAU,MAAM,SAAS,QAAQ;AAAA,EACnC,OAAO;AACL,UAAM,YAAY,SAAS,WAAW,cAAc;AACpD,UAAM,UAAU,QAAQ;AACxB,UAAM,UAAU,QAAQ;AACxB,UAAM,WAAW,KAAK,MAAM,UAAU,UAAU,IAAI;AACpD,UAAM,UAAU,KAAK,MAAM,UAAU,UAAU,GAAG;AAClD,UAAM,YAAY,KAAK,MAAM,UAAU,UAAU,KAAK;AACtD,UAAM,aAAa,KAAK,MAAM,UAAU,UAAU,MAAM;AACxD,UAAM,QAAQ,YAAY;AAC1B,UAAM,QAAQ,aAAa;AAC3B,cAAU,UAAM,aAAAA,SAAM,QAAQ,EAC3B,QAAQ,EAAE,MAAM,UAAU,KAAK,SAAS,OAAO,OAAO,QAAQ,MAAM,CAAC,EACrE,SAAS;AAAA,EACd;AAGA,QAAM,QAAQ,SAAS,UAAM,aAAAA,SAAM,OAAO,EAAE,OAAO,EAAE,SAAS,IAAI;AAGlE,MAAI,YAAY,OAAO;AACrB,eAAO,aAAAA,SAAM,KAAK,EAAE,IAAI,EAAE,SAAS;AAAA,EACrC;AACA,QAAM,MAAM,OAAO,YAAY,WAAW,UAAU;AACpD,QAAM,OAAO,OAAO;AACpB,QAAM,OAAO,OAAO;AACpB,aAAO,aAAAA,SAAM,KAAK,EACf,OAAO;AAAA,IACN,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,MAAM;AAAA,IACN,OAAO;AAAA,IACP,YAAY,EAAE,GAAG,KAAK,GAAG,KAAK,GAAG,IAAI;AAAA,EACvC,CAAC,EACA,IAAI,EACJ,SAAS;AACd;AAOA,eAAe,SAAS,UAAmC;AACzD,MAAI;AACF,UAAM,cAAU,aAAAA,SAAM,QAAQ,EAAE,KAAK,EAAE,WAAW,GAAG,CAAC;AACtD,UAAM,aAAa,MAAM,QAAQ,SAAS,EAAE,mBAAmB,KAAK,CAAC;AAGrE,UAAM,EAAE,OAAO,OAAO,IAAI,WAAW;AACrC,QAAI,QAAQ,KAAK,SAAS,GAAG;AAC3B,aAAO,WAAW;AAAA,IACpB;AAAA,EACF,QAAQ;AAAA,EAER;AAGA,SAAO;AACT;AAKO,SAAS,cAAc,WAA2B;AACvD,QAAM,SAAS,UAAAC,QAAG,aAAa,SAAS;AACxC,SAAO,OAAO,SAAS,QAAQ;AACjC;;;ADtLA,IAAM,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAyDf,IAAM,iBAA2C;AAAA,EAC/C,QAAQ;AAAA,EACR,WAAW;AAAA,EACX,QAAQ;AACV;AAEA,eAAe,aACb,QACA,UACA,SACwB;AACxB,UAAQ,UAAU;AAAA,IAChB,KAAK,UAAU;AACb,YAAM,EAAE,aAAa,IAAI,MAAM,OAAO,gBAAgB;AACtD,aAAO,aAAa,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACzC;AAAA,IACA,KAAK,aAAa;AAEhB,YAAM,EAAE,gBAAgB,IAAI,MAAM,OAAO,mBAAmB;AAC5D,aAAO,gBAAgB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IAC5C;AAAA,IACA,KAAK,UAAU;AAEb,YAAM,EAAE,yBAAyB,IAAI,MAAM,OAAO,gBAAgB;AAClE,aAAO,yBAAyB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACrD;AAAA,IACA;AACE,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ;AAAA,MAC/B;AAAA,EACJ;AACF;AAUO,IAAM,0BAAkD;AAAA,EAC7D,KAAK;AAAA,EACL,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AACP;AASO,SAAS,aACd,UACA,gBACA,QACQ;AACR,MAAI,WAAW,iBAAiB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,cAAc,IAAI;AAEtF,MAAI,SAAS,WAAW,GAAG;AACzB,eAAW;AAAA,EACb;AACA,MAAI,SAAS,WAAW,EAAG,QAAO;AAGlC,QAAM,YAAY,oBAAI,IAAoB;AAC1C,aAAW,KAAK,UAAU;AACxB,cAAU,IAAI,EAAE,SAAS,UAAU,IAAI,EAAE,MAAM,KAAK,KAAK,CAAC;AAAA,EAC5D;AACA,MAAI,UAAU;AACd,MAAI,YAAY;AAChB,aAAW,CAAC,KAAK,KAAK,KAAK,WAAW;AACpC,QAAI,QAAQ,WAAW;AACrB,gBAAU;AACV,kBAAY;AAAA,IACd;AAAA,EACF;AAEA,QAAM,kBAAkB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,OAAO;AACnE,MAAI,gBAAgB,WAAW,EAAG,QAAO,SAAS,CAAC;AAEnD,QAAM,YAAY,UAAU,OAAO,WAAW,WAAW,SAAS;AAGlE,QAAM,SAAmB,CAAC;AAC1B,WAAS,MAAM,GAAG,MAAM,SAAS,OAAO;AACtC,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,KAAK,iBAAiB;AAC/B,YAAM,KAAK,EAAE,GAAG;AAChB,iBAAW,IAAI,KAAK,WAAW,IAAI,EAAE,KAAK,KAAK,CAAC;AAAA,IAClD;AAEA,QAAI,WAAW;AAEb,YAAM,cAAc,oBAAI,IAAoB;AAC5C,iBAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,cAAM,YAAY,UAAU,EAAE,KAAK;AACnC,oBAAY,IAAI,YAAY,YAAY,IAAI,SAAS,KAAK,KAAK,KAAK;AAAA,MACtE;AAEA,UAAI,YAAY;AAChB,UAAI,iBAAiB;AACrB,iBAAW,CAAC,WAAW,KAAK,KAAK,aAAa;AAC5C,YAAI,QAAQ,gBAAgB;AAC1B,sBAAY;AACZ,2BAAiB;AAAA,QACnB;AAAA,MACF;AACA,aAAO,KAAK,SAAS;AAAA,IACvB,OAAO;AAEL,UAAI,WAAW;AACf,UAAI,gBAAgB;AACpB,iBAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,YAAI,QAAQ,eAAe;AACzB,qBAAW;AACX,0BAAgB;AAAA,QAClB;AAAA,MACF;AACA,aAAO,KAAK,QAAQ;AAAA,IACtB;AAAA,EACF;AAEA,SAAO,OAAO,KAAK,EAAE;AACvB;AAIA,SAAS,YAAY,GAAuB,GAA2C;AACrF,MAAI,MAAM,UAAa,MAAM,OAAW,QAAO;AAC/C,UAAQ,KAAK,MAAM,KAAK;AAC1B;AAEA,SAAS,eAAe,QAAkD;AACxE,QAAM,OAA2B;AAAA,IAC/B,aAAa;AAAA,IACb,mBAAmB;AAAA,MACjB,eAAe;AAAA,MACf,iBAAiB;AAAA,MACjB,kBAAkB;AAAA,IACpB;AAAA,IACA,cAAc;AAAA,IACd,oBAAoB;AAAA,MAClB,YAAY;AAAA,MACZ,iBAAiB;AAAA,IACnB;AAAA,IACA,aAAa;AAAA,EACf;AACA,SAAO,OAAO;AAAA,IACZ,CAAC,KAAK,OAAO;AAAA,MACX,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,MACvD,mBAAmB;AAAA,QACjB,eAAe;AAAA,UACb,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,iBAAiB;AAAA,UACf,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,kBAAkB;AAAA,UAChB,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,MACF;AAAA,MACA,cAAc,YAAY,IAAI,cAAc,EAAE,YAAY;AAAA,MAC1D,oBAAoB;AAAA,QAClB,YAAY,YAAY,IAAI,mBAAmB,YAAY,EAAE,mBAAmB,UAAU;AAAA,QAC1F,iBAAiB;AAAA,UACf,IAAI,mBAAmB;AAAA,UACvB,EAAE,mBAAmB;AAAA,QACvB;AAAA,MACF;AAAA,MACA,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,IACzD;AAAA,IACA;AAAA,EACF;AACF;AAIO,IAAM,SAAN,MAAa;AAAA,EACV,SAA+B;AAAA,EAC/B,gBAA+C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAmBvD,YAAY,YAAoC,SAAyB;AACvE,QAAI,OAAO,eAAe,UAAU;AAClC,YAAM,WAAW,SAAS,YAAY;AACtC,YAAM,UAAU,SAAS,SAAS,eAAe,QAAQ;AAEzD,WAAK,gBAAgB,aAAa,YAAY,UAAU,OAAO;AAAA,IACjE,OAAO;AACL,WAAK,SAAS;AAAA,IAChB;AAAA,EACF;AAAA,EAEA,MAAc,WAAmC;AAC/C,QAAI,KAAK,OAAQ,QAAO,KAAK;AAC7B,SAAK,SAAS,MAAM,KAAK;AACzB,SAAK,gBAAgB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EASA,MAAM,MAAM,OAAwB,UAAwB,CAAC,GAAyB;AACpF,UAAM;AAAA,MACJ,cAAc;AAAA,MACd;AAAA,MACA,aAAa;AAAA,MACb,UAAU;AAAA,MACV,kBAAkB;AAAA,MAClB;AAAA,IACF,IAAI;AAEJ,UAAM,QAAQ,MAAM,KAAK,SAAS;AAKlC,UAAM,CAAC,gBAAgB,WAAW,IAAI,MAAM,QAAQ,IAAI;AAAA,MACtD,0BAA0B,OAAO,UAAU;AAAA,MAC3C,0BAA0B,OAAO;AAAA,QAC/B,MAAM;AAAA,QACN,OAAO;AAAA,QACP,UAAU;AAAA,QACV,SAAS;AAAA,QACT,MAAM;AAAA,QACN,SAAS;AAAA,QACT,WAAW;AAAA,MACb,CAAC;AAAA,IACH,CAAC;AAGD,UAAM,UAAU,MAAM,QAAQ;AAAA,MAC5B,MAAM;AAAA,QAAK,EAAE,QAAQ,YAAY;AAAA,QAAG,MAClC,KAAK,cAAc,OAAO,gBAAgB,aAAa,UAAU;AAAA,MACnE;AAAA,IACF;AACA,UAAM,QAAQ,QAAQ,OAAO,CAAC,MAA0B,MAAM,IAAI;AAClE,QAAI,SAAS;AACX,YAAM,QAAQ,CAAC,GAAG,MAAM,QAAQ,IAAI,aAAa,IAAI,CAAC,KAAK,EAAE,IAAI,EAAE,CAAC;AAAA,IACtE;AAEA,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,EAAE,IAAI;AACxC,UAAM,gBAAgB,MAAM,IAAI,CAAC,MAAM,EAAE,KAAK;AAC9C,UAAM,QAAQ,eAAe,aAAa;AAE1C,QAAI,SAAS,WAAW,GAAG;AACzB,UAAI,QAAS,SAAQ,IAAI,wBAAwB;AACjD,aAAO,EAAE,MAAM,IAAI,UAAU,OAAO,cAAc;AAAA,IACpD;AAEA,WAAO;AAAA,MACL,MAAM,aAAa,UAAU,gBAAgB,eAAe;AAAA,MAC5D;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,cACZ,OACA,eACA,iBACA,YAC+B;AAC/B,aAAS,QAAQ,GAAG,SAAS,YAAY,SAAS;AAChD,UAAI;AACF,cAAM,EAAE,MAAM,MAAM,IAAI,UAAM,wBAAa;AAAA,UACzC;AAAA,UACA,UAAU;AAAA,YACR;AAAA,cACE,MAAM;AAAA,cACN,SAAS;AAAA,gBACP,EAAE,MAAM,QAAQ,MAAM,OAAO;AAAA,gBAC7B,EAAE,MAAM,SAAS,OAAO,cAAc;AAAA,gBACtC,EAAE,MAAM,SAAS,OAAO,gBAAgB;AAAA,cAC1C;AAAA,YACF;AAAA,UACF;AAAA,UACA,aAAa;AAAA,UACb,iBAAiB;AAAA,QACnB,CAAC;AAED,cAAM,MAAM,KAAK,KAAK;AAGtB,cAAM,QAAQ,IAAI,YAAY;AAC9B,YACE,MAAM,SAAS,OAAO,KACtB,MAAM,SAAS,YAAY,KAC3B,MAAM,SAAS,aAAa,KAC5B,MAAM,SAAS,WAAW,KAC1B,MAAM,SAAS,SAAS,KACxB,IAAI,SAAS,IACb;AACA,iBAAO;AAAA,QACT;AAGA,cAAM,UAAU,IAAI,QAAQ,iBAAiB,EAAE;AAC/C,eAAO,UAAU,EAAE,MAAM,SAAS,MAAM,IAAI;AAAA,MAC9C,SAAS,MAAM;AACb,YAAI,QAAQ,YAAY;AACtB,gBAAM,IAAI,QAAQ,CAAC,MAAM,WAAW,GAAG,OAAQ,QAAQ,EAAE,CAAC;AAC1D;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACF;","names":["path","sharp","fs"]}
package/dist/index.d.cts CHANGED
@@ -1,6 +1,61 @@
1
1
  import { LanguageModelUsage, LanguageModel } from 'ai';
2
2
  export { LanguageModelUsage } from 'ai';
3
3
 
4
+ interface CropFractions {
5
+ /** Fraction from left edge (0–1, default: 0.1) */
6
+ left: number;
7
+ /** Fraction from top edge (0–1, default: 0.02) */
8
+ top: number;
9
+ /** Fraction from left to keep (0–1, default: 0.9) */
10
+ right: number;
11
+ /** Fraction from top to keep (0–1, default: 0.6) */
12
+ bottom: number;
13
+ }
14
+ interface PreprocessOptions {
15
+ /** Gaussian blur radius (default: 1.5). Set to 0 to skip. */
16
+ blur?: number;
17
+ /** Upscale factor (default: 4) */
18
+ scale?: number;
19
+ /** Contrast multiplier around image mean (default: 3.0). Set to 1 to skip. */
20
+ contrast?: number;
21
+ /** Enable unsharp-mask sharpening (default: true) */
22
+ sharpen?: boolean;
23
+ /**
24
+ * Crop mode (default: 'auto'):
25
+ * - 'auto' – trim whitespace after contrast enhancement, with margin
26
+ * - 'legacy' – fixed-percentage crop (original behavior)
27
+ * - 'none' – skip cropping
28
+ * - CropFractions – custom crop percentages
29
+ */
30
+ crop?: 'auto' | 'legacy' | 'none' | CropFractions;
31
+ /** Add white padding around the result (default: true). Pass false to skip, or a number for custom px. */
32
+ padding?: boolean | number;
33
+ /** Invert colors (negate) after processing (default: false) */
34
+ negate?: boolean;
35
+ /** Convert to greyscale (default: true) */
36
+ greyscale?: boolean;
37
+ }
38
+ /**
39
+ * Preprocess a captcha image and return a base64-encoded PNG string.
40
+ */
41
+ declare function preprocessCaptcha(input: string | Buffer, options?: PreprocessOptions): Promise<string>;
42
+ /**
43
+ * Preprocess a captcha image and return the resulting PNG as a raw Buffer.
44
+ *
45
+ * Pipeline:
46
+ * 1. Gaussian blur in color space (smooths dither pattern)
47
+ * 2. Grayscale conversion
48
+ * 3. Upscale with Lanczos
49
+ * 4. Contrast boost around image mean + sharpen
50
+ * 5. Crop (auto-detect, legacy fixed, none, or custom)
51
+ * 6. Add white padding
52
+ */
53
+ declare function preprocessCaptchaToBuffer(input: string | Buffer, options?: PreprocessOptions): Promise<Buffer>;
54
+ /**
55
+ * Read an image file and return its base64-encoded content.
56
+ */
57
+ declare function imageToBase64(imagePath: string): string;
58
+
4
59
  type Provider = 'openai' | 'anthropic' | 'google';
5
60
  interface SolverOptions {
6
61
  /** AI provider to use when constructing the model from an API key (default: "openai") */
@@ -9,7 +64,7 @@ interface SolverOptions {
9
64
  model?: string;
10
65
  }
11
66
  interface SolveOptions {
12
- /** Number of voting attempts (default: 5) */
67
+ /** Number of voting attempts (default: 7) */
13
68
  numAttempts?: number;
14
69
  /** Expected captcha length — results of other lengths are discarded */
15
70
  expectedLength?: number;
@@ -17,6 +72,15 @@ interface SolveOptions {
17
72
  maxRetries?: number;
18
73
  /** Whether to log attempt details (default: true) */
19
74
  verbose?: boolean;
75
+ /**
76
+ * Confusion groups for majority voting.
77
+ * Pass a Record<string, string> to merge visually similar characters,
78
+ * or `false` to disable (default: false).
79
+ * Use LEGACY_CONFUSION_GROUPS to restore pre-3.0 behavior.
80
+ */
81
+ confusionGroups?: Record<string, string> | false;
82
+ /** Preprocessing options passed to the image pipeline */
83
+ preprocess?: PreprocessOptions;
20
84
  }
21
85
  interface SolveResult {
22
86
  /** The solved captcha text (majority-voted) */
@@ -28,6 +92,19 @@ interface SolveResult {
28
92
  /** Per-attempt usage breakdown */
29
93
  attemptUsages: LanguageModelUsage[];
30
94
  }
95
+ /**
96
+ * Pre-3.0 confusion groups that merge visually similar characters.
97
+ * Opt-in via `{ confusionGroups: LEGACY_CONFUSION_GROUPS }`.
98
+ *
99
+ * Maps: 1/I/L → '1', O/D/0 → 'O', S/5 → 'S', Z/2 → 'Z'
100
+ */
101
+ declare const LEGACY_CONFUSION_GROUPS: Record<string, string>;
102
+ /**
103
+ * Character-level majority vote across multiple attempts.
104
+ * When `groups` is provided, visually similar characters are merged
105
+ * during counting (e.g. 1/I/L all count toward '1').
106
+ */
107
+ declare function majorityVote(attempts: string[], expectedLength?: number, groups?: Record<string, string> | false): string;
31
108
  declare class Solver {
32
109
  private _model;
33
110
  private _pendingModel;
@@ -65,29 +142,4 @@ declare class Solver {
65
142
  private singleAttempt;
66
143
  }
67
144
 
68
- /**
69
- * Preprocess a captcha image using sharp (libvips).
70
- *
71
- * Pipeline:
72
- * 1. Gaussian blur in color space (smooths dither pattern)
73
- * 2. Grayscale conversion
74
- * 3. Upscale 4× with Lanczos
75
- * 4. Contrast boost (3× around image mean) + sharpen
76
- * 5. Crop decorative borders
77
- * 6. Add white padding
78
- *
79
- * Accepts a file path or a raw image Buffer.
80
- * Returns a base64-encoded PNG string.
81
- */
82
- declare function preprocessCaptcha(input: string | Buffer): Promise<string>;
83
- /**
84
- * Same preprocessing pipeline as `preprocessCaptcha`, but returns the
85
- * resulting PNG as a raw Buffer (useful for AI SDK image content parts).
86
- */
87
- declare function preprocessCaptchaToBuffer(input: string | Buffer): Promise<Buffer>;
88
- /**
89
- * Read an image file and return its base64-encoded content.
90
- */
91
- declare function imageToBase64(imagePath: string): string;
92
-
93
- export { type Provider, type SolveOptions, type SolveResult, Solver, type SolverOptions, imageToBase64, preprocessCaptcha, preprocessCaptchaToBuffer };
145
+ export { type CropFractions, LEGACY_CONFUSION_GROUPS, type PreprocessOptions, type Provider, type SolveOptions, type SolveResult, Solver, type SolverOptions, imageToBase64, majorityVote, preprocessCaptcha, preprocessCaptchaToBuffer };
package/dist/index.d.ts CHANGED
@@ -1,6 +1,61 @@
1
1
  import { LanguageModelUsage, LanguageModel } from 'ai';
2
2
  export { LanguageModelUsage } from 'ai';
3
3
 
4
+ interface CropFractions {
5
+ /** Fraction from left edge (0–1, default: 0.1) */
6
+ left: number;
7
+ /** Fraction from top edge (0–1, default: 0.02) */
8
+ top: number;
9
+ /** Fraction from left to keep (0–1, default: 0.9) */
10
+ right: number;
11
+ /** Fraction from top to keep (0–1, default: 0.6) */
12
+ bottom: number;
13
+ }
14
+ interface PreprocessOptions {
15
+ /** Gaussian blur radius (default: 1.5). Set to 0 to skip. */
16
+ blur?: number;
17
+ /** Upscale factor (default: 4) */
18
+ scale?: number;
19
+ /** Contrast multiplier around image mean (default: 3.0). Set to 1 to skip. */
20
+ contrast?: number;
21
+ /** Enable unsharp-mask sharpening (default: true) */
22
+ sharpen?: boolean;
23
+ /**
24
+ * Crop mode (default: 'auto'):
25
+ * - 'auto' – trim whitespace after contrast enhancement, with margin
26
+ * - 'legacy' – fixed-percentage crop (original behavior)
27
+ * - 'none' – skip cropping
28
+ * - CropFractions – custom crop percentages
29
+ */
30
+ crop?: 'auto' | 'legacy' | 'none' | CropFractions;
31
+ /** Add white padding around the result (default: true). Pass false to skip, or a number for custom px. */
32
+ padding?: boolean | number;
33
+ /** Invert colors (negate) after processing (default: false) */
34
+ negate?: boolean;
35
+ /** Convert to greyscale (default: true) */
36
+ greyscale?: boolean;
37
+ }
38
+ /**
39
+ * Preprocess a captcha image and return a base64-encoded PNG string.
40
+ */
41
+ declare function preprocessCaptcha(input: string | Buffer, options?: PreprocessOptions): Promise<string>;
42
+ /**
43
+ * Preprocess a captcha image and return the resulting PNG as a raw Buffer.
44
+ *
45
+ * Pipeline:
46
+ * 1. Gaussian blur in color space (smooths dither pattern)
47
+ * 2. Grayscale conversion
48
+ * 3. Upscale with Lanczos
49
+ * 4. Contrast boost around image mean + sharpen
50
+ * 5. Crop (auto-detect, legacy fixed, none, or custom)
51
+ * 6. Add white padding
52
+ */
53
+ declare function preprocessCaptchaToBuffer(input: string | Buffer, options?: PreprocessOptions): Promise<Buffer>;
54
+ /**
55
+ * Read an image file and return its base64-encoded content.
56
+ */
57
+ declare function imageToBase64(imagePath: string): string;
58
+
4
59
  type Provider = 'openai' | 'anthropic' | 'google';
5
60
  interface SolverOptions {
6
61
  /** AI provider to use when constructing the model from an API key (default: "openai") */
@@ -9,7 +64,7 @@ interface SolverOptions {
9
64
  model?: string;
10
65
  }
11
66
  interface SolveOptions {
12
- /** Number of voting attempts (default: 5) */
67
+ /** Number of voting attempts (default: 7) */
13
68
  numAttempts?: number;
14
69
  /** Expected captcha length — results of other lengths are discarded */
15
70
  expectedLength?: number;
@@ -17,6 +72,15 @@ interface SolveOptions {
17
72
  maxRetries?: number;
18
73
  /** Whether to log attempt details (default: true) */
19
74
  verbose?: boolean;
75
+ /**
76
+ * Confusion groups for majority voting.
77
+ * Pass a Record<string, string> to merge visually similar characters,
78
+ * or `false` to disable (default: false).
79
+ * Use LEGACY_CONFUSION_GROUPS to restore pre-3.0 behavior.
80
+ */
81
+ confusionGroups?: Record<string, string> | false;
82
+ /** Preprocessing options passed to the image pipeline */
83
+ preprocess?: PreprocessOptions;
20
84
  }
21
85
  interface SolveResult {
22
86
  /** The solved captcha text (majority-voted) */
@@ -28,6 +92,19 @@ interface SolveResult {
28
92
  /** Per-attempt usage breakdown */
29
93
  attemptUsages: LanguageModelUsage[];
30
94
  }
95
+ /**
96
+ * Pre-3.0 confusion groups that merge visually similar characters.
97
+ * Opt-in via `{ confusionGroups: LEGACY_CONFUSION_GROUPS }`.
98
+ *
99
+ * Maps: 1/I/L → '1', O/D/0 → 'O', S/5 → 'S', Z/2 → 'Z'
100
+ */
101
+ declare const LEGACY_CONFUSION_GROUPS: Record<string, string>;
102
+ /**
103
+ * Character-level majority vote across multiple attempts.
104
+ * When `groups` is provided, visually similar characters are merged
105
+ * during counting (e.g. 1/I/L all count toward '1').
106
+ */
107
+ declare function majorityVote(attempts: string[], expectedLength?: number, groups?: Record<string, string> | false): string;
31
108
  declare class Solver {
32
109
  private _model;
33
110
  private _pendingModel;
@@ -65,29 +142,4 @@ declare class Solver {
65
142
  private singleAttempt;
66
143
  }
67
144
 
68
- /**
69
- * Preprocess a captcha image using sharp (libvips).
70
- *
71
- * Pipeline:
72
- * 1. Gaussian blur in color space (smooths dither pattern)
73
- * 2. Grayscale conversion
74
- * 3. Upscale 4× with Lanczos
75
- * 4. Contrast boost (3× around image mean) + sharpen
76
- * 5. Crop decorative borders
77
- * 6. Add white padding
78
- *
79
- * Accepts a file path or a raw image Buffer.
80
- * Returns a base64-encoded PNG string.
81
- */
82
- declare function preprocessCaptcha(input: string | Buffer): Promise<string>;
83
- /**
84
- * Same preprocessing pipeline as `preprocessCaptcha`, but returns the
85
- * resulting PNG as a raw Buffer (useful for AI SDK image content parts).
86
- */
87
- declare function preprocessCaptchaToBuffer(input: string | Buffer): Promise<Buffer>;
88
- /**
89
- * Read an image file and return its base64-encoded content.
90
- */
91
- declare function imageToBase64(imagePath: string): string;
92
-
93
- export { type Provider, type SolveOptions, type SolveResult, Solver, type SolverOptions, imageToBase64, preprocessCaptcha, preprocessCaptchaToBuffer };
145
+ export { type CropFractions, LEGACY_CONFUSION_GROUPS, type PreprocessOptions, type Provider, type SolveOptions, type SolveResult, Solver, type SolverOptions, imageToBase64, majorityVote, preprocessCaptcha, preprocessCaptchaToBuffer };
package/dist/index.js CHANGED
@@ -5,47 +5,98 @@ import { generateText } from "ai";
5
5
  import fs from "fs";
6
6
  import path from "path";
7
7
  import sharp from "sharp";
8
- async function preprocessCaptcha(input) {
9
- const buf = await preprocessCaptchaToBuffer(input);
8
+ var LEGACY_CROP = { left: 0.1, top: 0.02, right: 0.9, bottom: 0.6 };
9
+ async function preprocessCaptcha(input, options) {
10
+ const buf = await preprocessCaptchaToBuffer(input, options);
10
11
  return buf.toString("base64");
11
12
  }
12
- async function preprocessCaptchaToBuffer(input) {
13
+ async function preprocessCaptchaToBuffer(input, options) {
14
+ const {
15
+ blur = 1.5,
16
+ scale = 4,
17
+ contrast = 3,
18
+ sharpen = true,
19
+ crop = "auto",
20
+ padding = true,
21
+ negate = false,
22
+ greyscale = true
23
+ } = options ?? {};
13
24
  const source = typeof input === "string" ? path.resolve(input) : input;
14
25
  const metadata = await sharp(source).metadata();
15
26
  const origW = metadata.width;
16
27
  const origH = metadata.height;
17
- const smoothed = await sharp(source).blur(1.5).greyscale().toBuffer();
18
- const upscaled = await sharp(smoothed).resize(origW * 4, origH * 4, { kernel: "lanczos3" }).toBuffer();
19
- const stats = await sharp(upscaled).stats();
20
- const mean = stats.channels[0].mean;
21
- const enhanced = await sharp(upscaled).linear(3, mean * (1 - 3)).sharpen({ sigma: 1, m1: 2, m2: 1 }).toBuffer();
22
- const scaledW = origW * 4;
23
- const scaledH = origH * 4;
24
- const cropLeft = Math.floor(scaledW * 0.1);
25
- const cropTop = Math.floor(scaledH * 0.02);
26
- const cropRight = Math.floor(scaledW * 0.9);
27
- const cropBottom = Math.floor(scaledH * 0.6);
28
- const cropW = cropRight - cropLeft;
29
- const cropH = cropBottom - cropTop;
30
- return sharp(enhanced).extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH }).extend({
31
- top: 20,
32
- bottom: 20,
33
- left: 30,
34
- right: 30,
28
+ let pipeline = sharp(source);
29
+ if (blur > 0) pipeline = pipeline.blur(blur);
30
+ if (greyscale) pipeline = pipeline.greyscale();
31
+ const smoothed = await pipeline.toBuffer();
32
+ const upscaled = await sharp(smoothed).resize(origW * scale, origH * scale, { kernel: "lanczos3" }).toBuffer();
33
+ let enhanced;
34
+ if (contrast !== 1) {
35
+ const stats = await sharp(upscaled).stats();
36
+ const mean = stats.channels[0].mean;
37
+ let pipe = sharp(upscaled).linear(contrast, mean * (1 - contrast));
38
+ if (sharpen) pipe = pipe.sharpen({ sigma: 1, m1: 2, m2: 1 });
39
+ enhanced = await pipe.toBuffer();
40
+ } else {
41
+ enhanced = sharpen ? await sharp(upscaled).sharpen({ sigma: 1, m1: 2, m2: 1 }).toBuffer() : upscaled;
42
+ }
43
+ let cropped;
44
+ if (crop === "none") {
45
+ cropped = enhanced;
46
+ } else if (crop === "auto") {
47
+ cropped = await autoCrop(enhanced);
48
+ } else {
49
+ const fractions = crop === "legacy" ? LEGACY_CROP : crop;
50
+ const scaledW = origW * scale;
51
+ const scaledH = origH * scale;
52
+ const cropLeft = Math.floor(scaledW * fractions.left);
53
+ const cropTop = Math.floor(scaledH * fractions.top);
54
+ const cropRight = Math.floor(scaledW * fractions.right);
55
+ const cropBottom = Math.floor(scaledH * fractions.bottom);
56
+ const cropW = cropRight - cropLeft;
57
+ const cropH = cropBottom - cropTop;
58
+ cropped = await sharp(enhanced).extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH }).toBuffer();
59
+ }
60
+ const final = negate ? await sharp(cropped).negate().toBuffer() : cropped;
61
+ if (padding === false) {
62
+ return sharp(final).png().toBuffer();
63
+ }
64
+ const pad = typeof padding === "number" ? padding : void 0;
65
+ const vPad = pad ?? 20;
66
+ const hPad = pad ?? 30;
67
+ return sharp(final).extend({
68
+ top: vPad,
69
+ bottom: vPad,
70
+ left: hPad,
71
+ right: hPad,
35
72
  background: { r: 255, g: 255, b: 255 }
36
73
  }).png().toBuffer();
37
74
  }
75
+ async function autoCrop(enhanced) {
76
+ try {
77
+ const trimmed = sharp(enhanced).trim({ threshold: 30 });
78
+ const trimmedBuf = await trimmed.toBuffer({ resolveWithObject: true });
79
+ const { width, height } = trimmedBuf.info;
80
+ if (width > 2 && height > 2) {
81
+ return trimmedBuf.data;
82
+ }
83
+ } catch {
84
+ }
85
+ return enhanced;
86
+ }
38
87
  function imageToBase64(imagePath) {
39
88
  const buffer = fs.readFileSync(imagePath);
40
89
  return buffer.toString("base64");
41
90
  }
42
91
 
43
92
  // src/solver.ts
44
- var PROMPT = `You are an assistant helping a visually impaired person read distorted text from an image.
45
- The text contains uppercase letters A-Z and/or digits 0-9.
46
- A thin vertical stroke is the digit 1. Never read it as the letter I or L.
47
- A round closed shape is the letter O, not the letter D.
48
- Output ONLY the exact characters you read, nothing else.`;
93
+ var PROMPT = `You are an expert OCR assistant reading distorted text from a CAPTCHA image.
94
+ Two versions of the same captcha are provided. Cross-reference both to determine the correct text.
95
+ The text may contain uppercase letters (A-Z), lowercase letters (a-z), and/or digits (0-9).
96
+ Pay close attention to:
97
+ - Letter case: lowercase "e" has a horizontal bar inside, digit "0" does not. Lowercase "r" has a short descender, uppercase "T" has a flat top.
98
+ - Similar shapes: "5" has a flat top + curved bottom, "S" is fully curved. "4" has an angled stroke, "A" has a pointed top. "6" has a closed bottom loop, "8" has two loops. "2" has a curved top + flat bottom, "Z" has all straight lines.
99
+ Output ONLY the exact characters you read, preserving case. Nothing else.`;
49
100
  var DEFAULT_MODELS = {
50
101
  openai: "gpt-4o",
51
102
  anthropic: "claude-sonnet-4-20250514",
@@ -71,7 +122,7 @@ async function resolveModel(apiKey, provider, modelId) {
71
122
  );
72
123
  }
73
124
  }
74
- var CONFUSION_GROUPS = {
125
+ var LEGACY_CONFUSION_GROUPS = {
75
126
  "1": "1",
76
127
  I: "1",
77
128
  L: "1",
@@ -83,7 +134,7 @@ var CONFUSION_GROUPS = {
83
134
  Z: "Z",
84
135
  "2": "Z"
85
136
  };
86
- function majorityVote(attempts, expectedLength) {
137
+ function majorityVote(attempts, expectedLength, groups) {
87
138
  let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;
88
139
  if (filtered.length === 0) {
89
140
  filtered = attempts;
@@ -103,6 +154,7 @@ function majorityVote(attempts, expectedLength) {
103
154
  }
104
155
  const sameLenAttempts = filtered.filter((a) => a.length === bestLen);
105
156
  if (sameLenAttempts.length === 0) return filtered[0];
157
+ const useGroups = groups && typeof groups === "object" ? groups : void 0;
106
158
  const result = [];
107
159
  for (let pos = 0; pos < bestLen; pos++) {
108
160
  const charCounts = /* @__PURE__ */ new Map();
@@ -110,20 +162,32 @@ function majorityVote(attempts, expectedLength) {
110
162
  const ch = a[pos];
111
163
  charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);
112
164
  }
113
- const groupCounts = /* @__PURE__ */ new Map();
114
- for (const [ch, count] of charCounts) {
115
- const canonical = CONFUSION_GROUPS[ch] ?? ch;
116
- groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);
117
- }
118
- let bestGroup = "";
119
- let bestGroupCount = 0;
120
- for (const [canonical, count] of groupCounts) {
121
- if (count > bestGroupCount) {
122
- bestGroup = canonical;
123
- bestGroupCount = count;
165
+ if (useGroups) {
166
+ const groupCounts = /* @__PURE__ */ new Map();
167
+ for (const [ch, count] of charCounts) {
168
+ const canonical = useGroups[ch] ?? ch;
169
+ groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);
170
+ }
171
+ let bestGroup = "";
172
+ let bestGroupCount = 0;
173
+ for (const [canonical, count] of groupCounts) {
174
+ if (count > bestGroupCount) {
175
+ bestGroup = canonical;
176
+ bestGroupCount = count;
177
+ }
178
+ }
179
+ result.push(bestGroup);
180
+ } else {
181
+ let bestChar = "";
182
+ let bestCharCount = 0;
183
+ for (const [ch, count] of charCounts) {
184
+ if (count > bestCharCount) {
185
+ bestChar = ch;
186
+ bestCharCount = count;
187
+ }
124
188
  }
189
+ result.push(bestChar);
125
190
  }
126
- result.push(bestGroup);
127
191
  }
128
192
  return result.join("");
129
193
  }
@@ -219,11 +283,32 @@ var Solver = class {
219
283
  * @returns Solved text, per-attempt answers, and token usage
220
284
  */
221
285
  async solve(input, options = {}) {
222
- const { numAttempts = 5, expectedLength, maxRetries = 2, verbose = true } = options;
286
+ const {
287
+ numAttempts = 7,
288
+ expectedLength,
289
+ maxRetries = 2,
290
+ verbose = true,
291
+ confusionGroups = false,
292
+ preprocess
293
+ } = options;
223
294
  const model = await this.getModel();
224
- const imageBuffer = await preprocessCaptchaToBuffer(input);
295
+ const [enhancedBuffer, colorBuffer] = await Promise.all([
296
+ preprocessCaptchaToBuffer(input, preprocess),
297
+ preprocessCaptchaToBuffer(input, {
298
+ blur: 0,
299
+ scale: 4,
300
+ contrast: 1,
301
+ sharpen: false,
302
+ crop: "none",
303
+ padding: 40,
304
+ greyscale: false
305
+ })
306
+ ]);
225
307
  const results = await Promise.all(
226
- Array.from({ length: numAttempts }, () => this.singleAttempt(model, imageBuffer, maxRetries))
308
+ Array.from(
309
+ { length: numAttempts },
310
+ () => this.singleAttempt(model, enhancedBuffer, colorBuffer, maxRetries)
311
+ )
227
312
  );
228
313
  const valid = results.filter((r) => r !== null);
229
314
  if (verbose) {
@@ -236,13 +321,18 @@ var Solver = class {
236
321
  if (verbose) console.log(" All attempts failed!");
237
322
  return { text: "", attempts, usage, attemptUsages };
238
323
  }
239
- return { text: majorityVote(attempts, expectedLength), attempts, usage, attemptUsages };
324
+ return {
325
+ text: majorityVote(attempts, expectedLength, confusionGroups),
326
+ attempts,
327
+ usage,
328
+ attemptUsages
329
+ };
240
330
  }
241
331
  /**
242
332
  * Make a single API call to read the captcha.
243
333
  * Retries up to `maxRetries` times on failure.
244
334
  */
245
- async singleAttempt(model, imageBuffer, maxRetries) {
335
+ async singleAttempt(model, primaryBuffer, secondaryBuffer, maxRetries) {
246
336
  for (let retry = 0; retry <= maxRetries; retry++) {
247
337
  try {
248
338
  const { text, usage } = await generateText({
@@ -252,7 +342,8 @@ var Solver = class {
252
342
  role: "user",
253
343
  content: [
254
344
  { type: "text", text: PROMPT },
255
- { type: "image", image: imageBuffer }
345
+ { type: "image", image: primaryBuffer },
346
+ { type: "image", image: secondaryBuffer }
256
347
  ]
257
348
  }
258
349
  ],
@@ -264,7 +355,7 @@ var Solver = class {
264
355
  if (lower.includes("sorry") || lower.includes("can't help") || lower.includes("cannot help") || lower.includes("unable to") || lower.includes("i can't") || raw.length > 20) {
265
356
  return null;
266
357
  }
267
- const cleaned = raw.toUpperCase().replace(/[^A-Z0-9]/g, "");
358
+ const cleaned = raw.replace(/[^A-Za-z0-9]/g, "");
268
359
  return cleaned ? { text: cleaned, usage } : null;
269
360
  } catch (_err) {
270
361
  if (retry < maxRetries) {
@@ -278,8 +369,10 @@ var Solver = class {
278
369
  }
279
370
  };
280
371
  export {
372
+ LEGACY_CONFUSION_GROUPS,
281
373
  Solver,
282
374
  imageToBase64,
375
+ majorityVote,
283
376
  preprocessCaptcha,
284
377
  preprocessCaptchaToBuffer
285
378
  };
package/dist/index.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/solver.ts","../src/preprocess.ts"],"sourcesContent":["import type { LanguageModel, LanguageModelUsage } from 'ai';\nimport { generateText } from 'ai';\nimport { preprocessCaptchaToBuffer } from './preprocess.js';\n\nconst PROMPT = `You are an assistant helping a visually impaired person read distorted text from an image.\nThe text contains uppercase letters A-Z and/or digits 0-9.\nA thin vertical stroke is the digit 1. Never read it as the letter I or L.\nA round closed shape is the letter O, not the letter D.\nOutput ONLY the exact characters you read, nothing else.`;\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport type Provider = 'openai' | 'anthropic' | 'google';\n\nexport interface SolverOptions {\n /** AI provider to use when constructing the model from an API key (default: \"openai\") */\n provider?: Provider;\n /** Model ID passed to the provider (default: \"gpt-4o\") */\n model?: string;\n}\n\nexport interface SolveOptions {\n /** Number of voting attempts (default: 5) */\n numAttempts?: number;\n /** Expected captcha length — results of other lengths are discarded */\n expectedLength?: number;\n /** Max retries per attempt on API failure (default: 2) */\n maxRetries?: number;\n /** Whether to log attempt details (default: true) */\n verbose?: boolean;\n}\n\nexport interface SolveResult {\n /** The solved captcha text (majority-voted) */\n text: string;\n /** Per-attempt raw answers (before voting) */\n attempts: string[];\n /** Aggregated token usage across all parallel attempts */\n usage: LanguageModelUsage;\n /** Per-attempt usage breakdown */\n attemptUsages: LanguageModelUsage[];\n}\n\ninterface AttemptResult {\n text: string;\n usage: LanguageModelUsage;\n}\n\n// ── Provider resolution ──────────────────────────────────────────────\n\nconst DEFAULT_MODELS: Record<Provider, string> = {\n openai: 'gpt-4o',\n anthropic: 'claude-sonnet-4-20250514',\n google: 'gemini-2.0-flash',\n};\n\nasync function resolveModel(\n apiKey: string,\n provider: Provider,\n modelId: string\n): Promise<LanguageModel> {\n switch (provider) {\n case 'openai': {\n const { createOpenAI } = await import('@ai-sdk/openai');\n return createOpenAI({ apiKey })(modelId);\n }\n case 'anthropic': {\n // @ts-expect-error — optional peer dependency\n const { createAnthropic } = await import('@ai-sdk/anthropic');\n return createAnthropic({ apiKey })(modelId);\n }\n case 'google': {\n // @ts-expect-error — optional peer dependency\n const { createGoogleGenerativeAI } = await import('@ai-sdk/google');\n return createGoogleGenerativeAI({ apiKey })(modelId);\n }\n default:\n throw new Error(\n `Unknown provider \"${provider}\". Install the matching @ai-sdk/* package and pass the model directly.`\n );\n }\n}\n\n// ── Confusion groups ─────────────────────────────────────────────────\n\n/**\n * Characters the model commonly misreads as each other.\n * Each group maps to its canonical (most likely correct) character.\n */\nconst CONFUSION_GROUPS: Record<string, string> = {\n '1': '1',\n I: '1',\n L: '1',\n O: 'O',\n D: 'O',\n '0': 'O',\n S: 'S',\n '5': 'S',\n Z: 'Z',\n '2': 'Z',\n};\n\n// ── Majority voting ──────────────────────────────────────────────────\n\n/**\n * Character-level majority vote across multiple attempts.\n * Uses confusion-aware voting: characters that the model commonly\n * confuses (e.g. 1/I/L, O/D/0) are grouped together during counting.\n */\nfunction majorityVote(attempts: string[], expectedLength?: number): string {\n let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;\n\n if (filtered.length === 0) {\n filtered = attempts;\n }\n if (filtered.length === 0) return '';\n\n // Find most common length\n const lenCounts = new Map<number, number>();\n for (const a of filtered) {\n lenCounts.set(a.length, (lenCounts.get(a.length) ?? 0) + 1);\n }\n let bestLen = 0;\n let bestCount = 0;\n for (const [len, count] of lenCounts) {\n if (count > bestCount) {\n bestLen = len;\n bestCount = count;\n }\n }\n\n const sameLenAttempts = filtered.filter((a) => a.length === bestLen);\n if (sameLenAttempts.length === 0) return filtered[0];\n\n // Vote per character position with confusion-aware grouping\n const result: string[] = [];\n for (let pos = 0; pos < bestLen; pos++) {\n const charCounts = new Map<string, number>();\n for (const a of sameLenAttempts) {\n const ch = a[pos];\n charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);\n }\n\n const groupCounts = new Map<string, number>();\n for (const [ch, count] of charCounts) {\n const canonical = CONFUSION_GROUPS[ch] ?? ch;\n groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);\n }\n\n let bestGroup = '';\n let bestGroupCount = 0;\n for (const [canonical, count] of groupCounts) {\n if (count > bestGroupCount) {\n bestGroup = canonical;\n bestGroupCount = count;\n }\n }\n\n result.push(bestGroup);\n }\n\n return result.join('');\n}\n\n// ── Usage aggregation ────────────────────────────────────────────────\n\nfunction sumOptional(a: number | undefined, b: number | undefined): number | undefined {\n if (a === undefined && b === undefined) return undefined;\n return (a ?? 0) + (b ?? 0);\n}\n\nfunction aggregateUsage(usages: LanguageModelUsage[]): LanguageModelUsage {\n const zero: LanguageModelUsage = {\n inputTokens: undefined,\n inputTokenDetails: {\n noCacheTokens: undefined,\n cacheReadTokens: undefined,\n cacheWriteTokens: undefined,\n },\n outputTokens: undefined,\n outputTokenDetails: {\n textTokens: undefined,\n reasoningTokens: undefined,\n },\n totalTokens: undefined,\n };\n return usages.reduce<LanguageModelUsage>(\n (acc, u) => ({\n inputTokens: sumOptional(acc.inputTokens, u.inputTokens),\n inputTokenDetails: {\n noCacheTokens: sumOptional(\n acc.inputTokenDetails.noCacheTokens,\n u.inputTokenDetails.noCacheTokens\n ),\n cacheReadTokens: sumOptional(\n acc.inputTokenDetails.cacheReadTokens,\n u.inputTokenDetails.cacheReadTokens\n ),\n cacheWriteTokens: sumOptional(\n acc.inputTokenDetails.cacheWriteTokens,\n u.inputTokenDetails.cacheWriteTokens\n ),\n },\n outputTokens: sumOptional(acc.outputTokens, u.outputTokens),\n outputTokenDetails: {\n textTokens: sumOptional(acc.outputTokenDetails.textTokens, u.outputTokenDetails.textTokens),\n reasoningTokens: sumOptional(\n acc.outputTokenDetails.reasoningTokens,\n u.outputTokenDetails.reasoningTokens\n ),\n },\n totalTokens: sumOptional(acc.totalTokens, u.totalTokens),\n }),\n zero\n );\n}\n\n// ── Solver class ─────────────────────────────────────────────────────\n\nexport class Solver {\n private _model: LanguageModel | null = null;\n private _pendingModel: Promise<LanguageModel> | null = null;\n\n /**\n * Create a captcha solver.\n *\n * @example\n * // Simple — defaults to OpenAI gpt-4o\n * const solver = new Solver('sk-...');\n *\n * @example\n * // Specify provider and model\n * const solver = new Solver('sk-ant-...', { provider: 'anthropic', model: 'claude-sonnet-4-20250514' });\n *\n * @example\n * // Pass an AI SDK model directly\n * import { createOpenAI } from '@ai-sdk/openai';\n * const openai = createOpenAI({ apiKey: 'sk-...' });\n * const solver = new Solver(openai('gpt-4o'));\n */\n constructor(keyOrModel: string | LanguageModel, options?: SolverOptions) {\n if (typeof keyOrModel === 'string') {\n const provider = options?.provider ?? 'openai';\n const modelId = options?.model ?? DEFAULT_MODELS[provider];\n // Lazily resolve the model on first use\n this._pendingModel = resolveModel(keyOrModel, provider, modelId);\n } else {\n this._model = keyOrModel;\n }\n }\n\n private async getModel(): Promise<LanguageModel> {\n if (this._model) return this._model;\n this._model = await this._pendingModel!;\n this._pendingModel = null;\n return this._model;\n }\n\n /**\n * Solve a captcha image.\n *\n * @param input - File path (string) or raw image Buffer\n * @param options - Solve options (attempts, expected length, etc.)\n * @returns Solved text, per-attempt answers, and token usage\n */\n async solve(input: string | Buffer, options: SolveOptions = {}): Promise<SolveResult> {\n const { numAttempts = 5, expectedLength, maxRetries = 2, verbose = true } = options;\n\n const model = await this.getModel();\n const imageBuffer = await preprocessCaptchaToBuffer(input);\n\n // Fire all attempts in parallel for speed\n const results = await Promise.all(\n Array.from({ length: numAttempts }, () => this.singleAttempt(model, imageBuffer, maxRetries))\n );\n const valid = results.filter((r): r is AttemptResult => r !== null);\n if (verbose) {\n valid.forEach((r, i) => console.log(` Attempt ${i + 1}: ${r.text}`));\n }\n\n const attempts = valid.map((r) => r.text);\n const attemptUsages = valid.map((r) => r.usage);\n const usage = aggregateUsage(attemptUsages);\n\n if (attempts.length === 0) {\n if (verbose) console.log(' All attempts failed!');\n return { text: '', attempts, usage, attemptUsages };\n }\n\n return { text: majorityVote(attempts, expectedLength), attempts, usage, attemptUsages };\n }\n\n /**\n * Make a single API call to read the captcha.\n * Retries up to `maxRetries` times on failure.\n */\n private async singleAttempt(\n model: LanguageModel,\n imageBuffer: Buffer,\n maxRetries: number\n ): Promise<AttemptResult | null> {\n for (let retry = 0; retry <= maxRetries; retry++) {\n try {\n const { text, usage } = await generateText({\n model,\n messages: [\n {\n role: 'user',\n content: [\n { type: 'text', text: PROMPT },\n { type: 'image', image: imageBuffer },\n ],\n },\n ],\n temperature: 1,\n maxOutputTokens: 256,\n });\n\n const raw = text.trim();\n\n // Detect refusals\n const lower = raw.toLowerCase();\n if (\n lower.includes('sorry') ||\n lower.includes(\"can't help\") ||\n lower.includes('cannot help') ||\n lower.includes('unable to') ||\n lower.includes(\"i can't\") ||\n raw.length > 20\n ) {\n return null;\n }\n\n // Clean: keep only uppercase letters and digits\n const cleaned = raw.toUpperCase().replace(/[^A-Z0-9]/g, '');\n return cleaned ? { text: cleaned, usage } : null;\n } catch (_err) {\n if (retry < maxRetries) {\n await new Promise((r) => setTimeout(r, 1000 * (retry + 1)));\n continue;\n }\n return null;\n }\n }\n return null;\n }\n}\n","import fs from 'fs';\nimport path from 'path';\nimport sharp from 'sharp';\n\n/**\n * Preprocess a captcha image using sharp (libvips).\n *\n * Pipeline:\n * 1. Gaussian blur in color space (smooths dither pattern)\n * 2. Grayscale conversion\n * 3. Upscale 4× with Lanczos\n * 4. Contrast boost (3× around image mean) + sharpen\n * 5. Crop decorative borders\n * 6. Add white padding\n *\n * Accepts a file path or a raw image Buffer.\n * Returns a base64-encoded PNG string.\n */\nexport async function preprocessCaptcha(input: string | Buffer): Promise<string> {\n const buf = await preprocessCaptchaToBuffer(input);\n return buf.toString('base64');\n}\n\n/**\n * Same preprocessing pipeline as `preprocessCaptcha`, but returns the\n * resulting PNG as a raw Buffer (useful for AI SDK image content parts).\n */\nexport async function preprocessCaptchaToBuffer(input: string | Buffer): Promise<Buffer> {\n const source = typeof input === 'string' ? path.resolve(input) : input;\n\n // Read original dimensions for crop/resize calculations\n const metadata = await sharp(source).metadata();\n const origW = metadata.width!;\n const origH = metadata.height!;\n\n // Step 1-2: Blur in color space (smooths dither pattern) → greyscale\n // Separate from resize to prevent pipeline reordering\n const smoothed = await sharp(source).blur(1.5).greyscale().toBuffer();\n\n // Step 3: Upscale 4× with Lanczos\n const upscaled = await sharp(smoothed)\n .resize(origW * 4, origH * 4, { kernel: 'lanczos3' })\n .toBuffer();\n\n // Step 4: Contrast 3× around actual image mean + sharpen\n // Matches PIL's ImageEnhance.Contrast: output = factor*input + mean*(1-factor)\n const stats = await sharp(upscaled).stats();\n const mean = stats.channels[0].mean;\n const enhanced = await sharp(upscaled)\n .linear(3.0, mean * (1 - 3.0))\n .sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 })\n .toBuffer();\n\n // Step 5: Crop decorative borders\n // Remove 10% left/right, 2% top, 40% bottom (keep top 60%)\n // Math.floor matches Python's int() truncation\n const scaledW = origW * 4;\n const scaledH = origH * 4;\n const cropLeft = Math.floor(scaledW * 0.1);\n const cropTop = Math.floor(scaledH * 0.02);\n const cropRight = Math.floor(scaledW * 0.9);\n const cropBottom = Math.floor(scaledH * 0.6);\n const cropW = cropRight - cropLeft;\n const cropH = cropBottom - cropTop;\n\n // Step 5-6: Crop → add white padding → output PNG\n return sharp(enhanced)\n .extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH })\n .extend({\n top: 20,\n bottom: 20,\n left: 30,\n right: 30,\n background: { r: 255, g: 255, b: 255 },\n })\n .png()\n .toBuffer();\n}\n\n/**\n * Read an image file and return its base64-encoded content.\n */\nexport function imageToBase64(imagePath: string): string {\n const buffer = fs.readFileSync(imagePath);\n return buffer.toString('base64');\n}\n"],"mappings":";AACA,SAAS,oBAAoB;;;ACD7B,OAAO,QAAQ;AACf,OAAO,UAAU;AACjB,OAAO,WAAW;AAgBlB,eAAsB,kBAAkB,OAAyC;AAC/E,QAAM,MAAM,MAAM,0BAA0B,KAAK;AACjD,SAAO,IAAI,SAAS,QAAQ;AAC9B;AAMA,eAAsB,0BAA0B,OAAyC;AACvF,QAAM,SAAS,OAAO,UAAU,WAAW,KAAK,QAAQ,KAAK,IAAI;AAGjE,QAAM,WAAW,MAAM,MAAM,MAAM,EAAE,SAAS;AAC9C,QAAM,QAAQ,SAAS;AACvB,QAAM,QAAQ,SAAS;AAIvB,QAAM,WAAW,MAAM,MAAM,MAAM,EAAE,KAAK,GAAG,EAAE,UAAU,EAAE,SAAS;AAGpE,QAAM,WAAW,MAAM,MAAM,QAAQ,EAClC,OAAO,QAAQ,GAAG,QAAQ,GAAG,EAAE,QAAQ,WAAW,CAAC,EACnD,SAAS;AAIZ,QAAM,QAAQ,MAAM,MAAM,QAAQ,EAAE,MAAM;AAC1C,QAAM,OAAO,MAAM,SAAS,CAAC,EAAE;AAC/B,QAAM,WAAW,MAAM,MAAM,QAAQ,EAClC,OAAO,GAAK,QAAQ,IAAI,EAAI,EAC5B,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC,EACxC,SAAS;AAKZ,QAAM,UAAU,QAAQ;AACxB,QAAM,UAAU,QAAQ;AACxB,QAAM,WAAW,KAAK,MAAM,UAAU,GAAG;AACzC,QAAM,UAAU,KAAK,MAAM,UAAU,IAAI;AACzC,QAAM,YAAY,KAAK,MAAM,UAAU,GAAG;AAC1C,QAAM,aAAa,KAAK,MAAM,UAAU,GAAG;AAC3C,QAAM,QAAQ,YAAY;AAC1B,QAAM,QAAQ,aAAa;AAG3B,SAAO,MAAM,QAAQ,EAClB,QAAQ,EAAE,MAAM,UAAU,KAAK,SAAS,OAAO,OAAO,QAAQ,MAAM,CAAC,EACrE,OAAO;AAAA,IACN,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,MAAM;AAAA,IACN,OAAO;AAAA,IACP,YAAY,EAAE,GAAG,KAAK,GAAG,KAAK,GAAG,IAAI;AAAA,EACvC,CAAC,EACA,IAAI,EACJ,SAAS;AACd;AAKO,SAAS,cAAc,WAA2B;AACvD,QAAM,SAAS,GAAG,aAAa,SAAS;AACxC,SAAO,OAAO,SAAS,QAAQ;AACjC;;;ADjFA,IAAM,SAAS;AAAA;AAAA;AAAA;AAAA;AA8Cf,IAAM,iBAA2C;AAAA,EAC/C,QAAQ;AAAA,EACR,WAAW;AAAA,EACX,QAAQ;AACV;AAEA,eAAe,aACb,QACA,UACA,SACwB;AACxB,UAAQ,UAAU;AAAA,IAChB,KAAK,UAAU;AACb,YAAM,EAAE,aAAa,IAAI,MAAM,OAAO,gBAAgB;AACtD,aAAO,aAAa,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACzC;AAAA,IACA,KAAK,aAAa;AAEhB,YAAM,EAAE,gBAAgB,IAAI,MAAM,OAAO,mBAAmB;AAC5D,aAAO,gBAAgB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IAC5C;AAAA,IACA,KAAK,UAAU;AAEb,YAAM,EAAE,yBAAyB,IAAI,MAAM,OAAO,gBAAgB;AAClE,aAAO,yBAAyB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACrD;AAAA,IACA;AACE,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ;AAAA,MAC/B;AAAA,EACJ;AACF;AAQA,IAAM,mBAA2C;AAAA,EAC/C,KAAK;AAAA,EACL,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AACP;AASA,SAAS,aAAa,UAAoB,gBAAiC;AACzE,MAAI,WAAW,iBAAiB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,cAAc,IAAI;AAEtF,MAAI,SAAS,WAAW,GAAG;AACzB,eAAW;AAAA,EACb;AACA,MAAI,SAAS,WAAW,EAAG,QAAO;AAGlC,QAAM,YAAY,oBAAI,IAAoB;AAC1C,aAAW,KAAK,UAAU;AACxB,cAAU,IAAI,EAAE,SAAS,UAAU,IAAI,EAAE,MAAM,KAAK,KAAK,CAAC;AAAA,EAC5D;AACA,MAAI,UAAU;AACd,MAAI,YAAY;AAChB,aAAW,CAAC,KAAK,KAAK,KAAK,WAAW;AACpC,QAAI,QAAQ,WAAW;AACrB,gBAAU;AACV,kBAAY;AAAA,IACd;AAAA,EACF;AAEA,QAAM,kBAAkB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,OAAO;AACnE,MAAI,gBAAgB,WAAW,EAAG,QAAO,SAAS,CAAC;AAGnD,QAAM,SAAmB,CAAC;AAC1B,WAAS,MAAM,GAAG,MAAM,SAAS,OAAO;AACtC,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,KAAK,iBAAiB;AAC/B,YAAM,KAAK,EAAE,GAAG;AAChB,iBAAW,IAAI,KAAK,WAAW,IAAI,EAAE,KAAK,KAAK,CAAC;AAAA,IAClD;AAEA,UAAM,cAAc,oBAAI,IAAoB;AAC5C,eAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,YAAM,YAAY,iBAAiB,EAAE,KAAK;AAC1C,kBAAY,IAAI,YAAY,YAAY,IAAI,SAAS,KAAK,KAAK,KAAK;AAAA,IACtE;AAEA,QAAI,YAAY;AAChB,QAAI,iBAAiB;AACrB,eAAW,CAAC,WAAW,KAAK,KAAK,aAAa;AAC5C,UAAI,QAAQ,gBAAgB;AAC1B,oBAAY;AACZ,yBAAiB;AAAA,MACnB;AAAA,IACF;AAEA,WAAO,KAAK,SAAS;AAAA,EACvB;AAEA,SAAO,OAAO,KAAK,EAAE;AACvB;AAIA,SAAS,YAAY,GAAuB,GAA2C;AACrF,MAAI,MAAM,UAAa,MAAM,OAAW,QAAO;AAC/C,UAAQ,KAAK,MAAM,KAAK;AAC1B;AAEA,SAAS,eAAe,QAAkD;AACxE,QAAM,OAA2B;AAAA,IAC/B,aAAa;AAAA,IACb,mBAAmB;AAAA,MACjB,eAAe;AAAA,MACf,iBAAiB;AAAA,MACjB,kBAAkB;AAAA,IACpB;AAAA,IACA,cAAc;AAAA,IACd,oBAAoB;AAAA,MAClB,YAAY;AAAA,MACZ,iBAAiB;AAAA,IACnB;AAAA,IACA,aAAa;AAAA,EACf;AACA,SAAO,OAAO;AAAA,IACZ,CAAC,KAAK,OAAO;AAAA,MACX,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,MACvD,mBAAmB;AAAA,QACjB,eAAe;AAAA,UACb,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,iBAAiB;AAAA,UACf,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,kBAAkB;AAAA,UAChB,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,MACF;AAAA,MACA,cAAc,YAAY,IAAI,cAAc,EAAE,YAAY;AAAA,MAC1D,oBAAoB;AAAA,QAClB,YAAY,YAAY,IAAI,mBAAmB,YAAY,EAAE,mBAAmB,UAAU;AAAA,QAC1F,iBAAiB;AAAA,UACf,IAAI,mBAAmB;AAAA,UACvB,EAAE,mBAAmB;AAAA,QACvB;AAAA,MACF;AAAA,MACA,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,IACzD;AAAA,IACA;AAAA,EACF;AACF;AAIO,IAAM,SAAN,MAAa;AAAA,EACV,SAA+B;AAAA,EAC/B,gBAA+C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAmBvD,YAAY,YAAoC,SAAyB;AACvE,QAAI,OAAO,eAAe,UAAU;AAClC,YAAM,WAAW,SAAS,YAAY;AACtC,YAAM,UAAU,SAAS,SAAS,eAAe,QAAQ;AAEzD,WAAK,gBAAgB,aAAa,YAAY,UAAU,OAAO;AAAA,IACjE,OAAO;AACL,WAAK,SAAS;AAAA,IAChB;AAAA,EACF;AAAA,EAEA,MAAc,WAAmC;AAC/C,QAAI,KAAK,OAAQ,QAAO,KAAK;AAC7B,SAAK,SAAS,MAAM,KAAK;AACzB,SAAK,gBAAgB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EASA,MAAM,MAAM,OAAwB,UAAwB,CAAC,GAAyB;AACpF,UAAM,EAAE,cAAc,GAAG,gBAAgB,aAAa,GAAG,UAAU,KAAK,IAAI;AAE5E,UAAM,QAAQ,MAAM,KAAK,SAAS;AAClC,UAAM,cAAc,MAAM,0BAA0B,KAAK;AAGzD,UAAM,UAAU,MAAM,QAAQ;AAAA,MAC5B,MAAM,KAAK,EAAE,QAAQ,YAAY,GAAG,MAAM,KAAK,cAAc,OAAO,aAAa,UAAU,CAAC;AAAA,IAC9F;AACA,UAAM,QAAQ,QAAQ,OAAO,CAAC,MAA0B,MAAM,IAAI;AAClE,QAAI,SAAS;AACX,YAAM,QAAQ,CAAC,GAAG,MAAM,QAAQ,IAAI,aAAa,IAAI,CAAC,KAAK,EAAE,IAAI,EAAE,CAAC;AAAA,IACtE;AAEA,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,EAAE,IAAI;AACxC,UAAM,gBAAgB,MAAM,IAAI,CAAC,MAAM,EAAE,KAAK;AAC9C,UAAM,QAAQ,eAAe,aAAa;AAE1C,QAAI,SAAS,WAAW,GAAG;AACzB,UAAI,QAAS,SAAQ,IAAI,wBAAwB;AACjD,aAAO,EAAE,MAAM,IAAI,UAAU,OAAO,cAAc;AAAA,IACpD;AAEA,WAAO,EAAE,MAAM,aAAa,UAAU,cAAc,GAAG,UAAU,OAAO,cAAc;AAAA,EACxF;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,cACZ,OACA,aACA,YAC+B;AAC/B,aAAS,QAAQ,GAAG,SAAS,YAAY,SAAS;AAChD,UAAI;AACF,cAAM,EAAE,MAAM,MAAM,IAAI,MAAM,aAAa;AAAA,UACzC;AAAA,UACA,UAAU;AAAA,YACR;AAAA,cACE,MAAM;AAAA,cACN,SAAS;AAAA,gBACP,EAAE,MAAM,QAAQ,MAAM,OAAO;AAAA,gBAC7B,EAAE,MAAM,SAAS,OAAO,YAAY;AAAA,cACtC;AAAA,YACF;AAAA,UACF;AAAA,UACA,aAAa;AAAA,UACb,iBAAiB;AAAA,QACnB,CAAC;AAED,cAAM,MAAM,KAAK,KAAK;AAGtB,cAAM,QAAQ,IAAI,YAAY;AAC9B,YACE,MAAM,SAAS,OAAO,KACtB,MAAM,SAAS,YAAY,KAC3B,MAAM,SAAS,aAAa,KAC5B,MAAM,SAAS,WAAW,KAC1B,MAAM,SAAS,SAAS,KACxB,IAAI,SAAS,IACb;AACA,iBAAO;AAAA,QACT;AAGA,cAAM,UAAU,IAAI,YAAY,EAAE,QAAQ,cAAc,EAAE;AAC1D,eAAO,UAAU,EAAE,MAAM,SAAS,MAAM,IAAI;AAAA,MAC9C,SAAS,MAAM;AACb,YAAI,QAAQ,YAAY;AACtB,gBAAM,IAAI,QAAQ,CAAC,MAAM,WAAW,GAAG,OAAQ,QAAQ,EAAE,CAAC;AAC1D;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACF;","names":[]}
1
+ {"version":3,"sources":["../src/solver.ts","../src/preprocess.ts"],"sourcesContent":["import type { LanguageModel, LanguageModelUsage } from 'ai';\nimport { generateText } from 'ai';\nimport type { PreprocessOptions } from './preprocess.js';\nimport { preprocessCaptchaToBuffer } from './preprocess.js';\n\nconst PROMPT = `You are an expert OCR assistant reading distorted text from a CAPTCHA image.\nTwo versions of the same captcha are provided. Cross-reference both to determine the correct text.\nThe text may contain uppercase letters (A-Z), lowercase letters (a-z), and/or digits (0-9).\nPay close attention to:\n- Letter case: lowercase \"e\" has a horizontal bar inside, digit \"0\" does not. Lowercase \"r\" has a short descender, uppercase \"T\" has a flat top.\n- Similar shapes: \"5\" has a flat top + curved bottom, \"S\" is fully curved. \"4\" has an angled stroke, \"A\" has a pointed top. \"6\" has a closed bottom loop, \"8\" has two loops. \"2\" has a curved top + flat bottom, \"Z\" has all straight lines.\nOutput ONLY the exact characters you read, preserving case. Nothing else.`;\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport type Provider = 'openai' | 'anthropic' | 'google';\n\nexport interface SolverOptions {\n /** AI provider to use when constructing the model from an API key (default: \"openai\") */\n provider?: Provider;\n /** Model ID passed to the provider (default: \"gpt-4o\") */\n model?: string;\n}\n\nexport interface SolveOptions {\n /** Number of voting attempts (default: 7) */\n numAttempts?: number;\n /** Expected captcha length — results of other lengths are discarded */\n expectedLength?: number;\n /** Max retries per attempt on API failure (default: 2) */\n maxRetries?: number;\n /** Whether to log attempt details (default: true) */\n verbose?: boolean;\n /**\n * Confusion groups for majority voting.\n * Pass a Record<string, string> to merge visually similar characters,\n * or `false` to disable (default: false).\n * Use LEGACY_CONFUSION_GROUPS to restore pre-3.0 behavior.\n */\n confusionGroups?: Record<string, string> | false;\n /** Preprocessing options passed to the image pipeline */\n preprocess?: PreprocessOptions;\n}\n\nexport interface SolveResult {\n /** The solved captcha text (majority-voted) */\n text: string;\n /** Per-attempt raw answers (before voting) */\n attempts: string[];\n /** Aggregated token usage across all parallel attempts */\n usage: LanguageModelUsage;\n /** Per-attempt usage breakdown */\n attemptUsages: LanguageModelUsage[];\n}\n\ninterface AttemptResult {\n text: string;\n usage: LanguageModelUsage;\n}\n\n// ── Provider resolution ──────────────────────────────────────────────\n\nconst DEFAULT_MODELS: Record<Provider, string> = {\n openai: 'gpt-4o',\n anthropic: 'claude-sonnet-4-20250514',\n google: 'gemini-2.0-flash',\n};\n\nasync function resolveModel(\n apiKey: string,\n provider: Provider,\n modelId: string\n): Promise<LanguageModel> {\n switch (provider) {\n case 'openai': {\n const { createOpenAI } = await import('@ai-sdk/openai');\n return createOpenAI({ apiKey })(modelId);\n }\n case 'anthropic': {\n // @ts-expect-error — optional peer dependency\n const { createAnthropic } = await import('@ai-sdk/anthropic');\n return createAnthropic({ apiKey })(modelId);\n }\n case 'google': {\n // @ts-expect-error — optional peer dependency\n const { createGoogleGenerativeAI } = await import('@ai-sdk/google');\n return createGoogleGenerativeAI({ apiKey })(modelId);\n }\n default:\n throw new Error(\n `Unknown provider \"${provider}\". Install the matching @ai-sdk/* package and pass the model directly.`\n );\n }\n}\n\n// ── Confusion groups ─────────────────────────────────────────────────\n\n/**\n * Pre-3.0 confusion groups that merge visually similar characters.\n * Opt-in via `{ confusionGroups: LEGACY_CONFUSION_GROUPS }`.\n *\n * Maps: 1/I/L → '1', O/D/0 → 'O', S/5 → 'S', Z/2 → 'Z'\n */\nexport const LEGACY_CONFUSION_GROUPS: Record<string, string> = {\n '1': '1',\n I: '1',\n L: '1',\n O: 'O',\n D: 'O',\n '0': 'O',\n S: 'S',\n '5': 'S',\n Z: 'Z',\n '2': 'Z',\n};\n\n// ── Majority voting ──────────────────────────────────────────────────\n\n/**\n * Character-level majority vote across multiple attempts.\n * When `groups` is provided, visually similar characters are merged\n * during counting (e.g. 1/I/L all count toward '1').\n */\nexport function majorityVote(\n attempts: string[],\n expectedLength?: number,\n groups?: Record<string, string> | false\n): string {\n let filtered = expectedLength ? attempts.filter((a) => a.length === expectedLength) : attempts;\n\n if (filtered.length === 0) {\n filtered = attempts;\n }\n if (filtered.length === 0) return '';\n\n // Find most common length\n const lenCounts = new Map<number, number>();\n for (const a of filtered) {\n lenCounts.set(a.length, (lenCounts.get(a.length) ?? 0) + 1);\n }\n let bestLen = 0;\n let bestCount = 0;\n for (const [len, count] of lenCounts) {\n if (count > bestCount) {\n bestLen = len;\n bestCount = count;\n }\n }\n\n const sameLenAttempts = filtered.filter((a) => a.length === bestLen);\n if (sameLenAttempts.length === 0) return filtered[0];\n\n const useGroups = groups && typeof groups === 'object' ? groups : undefined;\n\n // Vote per character position\n const result: string[] = [];\n for (let pos = 0; pos < bestLen; pos++) {\n const charCounts = new Map<string, number>();\n for (const a of sameLenAttempts) {\n const ch = a[pos];\n charCounts.set(ch, (charCounts.get(ch) ?? 0) + 1);\n }\n\n if (useGroups) {\n // Confusion-aware voting\n const groupCounts = new Map<string, number>();\n for (const [ch, count] of charCounts) {\n const canonical = useGroups[ch] ?? ch;\n groupCounts.set(canonical, (groupCounts.get(canonical) ?? 0) + count);\n }\n\n let bestGroup = '';\n let bestGroupCount = 0;\n for (const [canonical, count] of groupCounts) {\n if (count > bestGroupCount) {\n bestGroup = canonical;\n bestGroupCount = count;\n }\n }\n result.push(bestGroup);\n } else {\n // Simple majority — pick the most frequent raw character\n let bestChar = '';\n let bestCharCount = 0;\n for (const [ch, count] of charCounts) {\n if (count > bestCharCount) {\n bestChar = ch;\n bestCharCount = count;\n }\n }\n result.push(bestChar);\n }\n }\n\n return result.join('');\n}\n\n// ── Usage aggregation ────────────────────────────────────────────────\n\nfunction sumOptional(a: number | undefined, b: number | undefined): number | undefined {\n if (a === undefined && b === undefined) return undefined;\n return (a ?? 0) + (b ?? 0);\n}\n\nfunction aggregateUsage(usages: LanguageModelUsage[]): LanguageModelUsage {\n const zero: LanguageModelUsage = {\n inputTokens: undefined,\n inputTokenDetails: {\n noCacheTokens: undefined,\n cacheReadTokens: undefined,\n cacheWriteTokens: undefined,\n },\n outputTokens: undefined,\n outputTokenDetails: {\n textTokens: undefined,\n reasoningTokens: undefined,\n },\n totalTokens: undefined,\n };\n return usages.reduce<LanguageModelUsage>(\n (acc, u) => ({\n inputTokens: sumOptional(acc.inputTokens, u.inputTokens),\n inputTokenDetails: {\n noCacheTokens: sumOptional(\n acc.inputTokenDetails.noCacheTokens,\n u.inputTokenDetails.noCacheTokens\n ),\n cacheReadTokens: sumOptional(\n acc.inputTokenDetails.cacheReadTokens,\n u.inputTokenDetails.cacheReadTokens\n ),\n cacheWriteTokens: sumOptional(\n acc.inputTokenDetails.cacheWriteTokens,\n u.inputTokenDetails.cacheWriteTokens\n ),\n },\n outputTokens: sumOptional(acc.outputTokens, u.outputTokens),\n outputTokenDetails: {\n textTokens: sumOptional(acc.outputTokenDetails.textTokens, u.outputTokenDetails.textTokens),\n reasoningTokens: sumOptional(\n acc.outputTokenDetails.reasoningTokens,\n u.outputTokenDetails.reasoningTokens\n ),\n },\n totalTokens: sumOptional(acc.totalTokens, u.totalTokens),\n }),\n zero\n );\n}\n\n// ── Solver class ─────────────────────────────────────────────────────\n\nexport class Solver {\n private _model: LanguageModel | null = null;\n private _pendingModel: Promise<LanguageModel> | null = null;\n\n /**\n * Create a captcha solver.\n *\n * @example\n * // Simple — defaults to OpenAI gpt-4o\n * const solver = new Solver('sk-...');\n *\n * @example\n * // Specify provider and model\n * const solver = new Solver('sk-ant-...', { provider: 'anthropic', model: 'claude-sonnet-4-20250514' });\n *\n * @example\n * // Pass an AI SDK model directly\n * import { createOpenAI } from '@ai-sdk/openai';\n * const openai = createOpenAI({ apiKey: 'sk-...' });\n * const solver = new Solver(openai('gpt-4o'));\n */\n constructor(keyOrModel: string | LanguageModel, options?: SolverOptions) {\n if (typeof keyOrModel === 'string') {\n const provider = options?.provider ?? 'openai';\n const modelId = options?.model ?? DEFAULT_MODELS[provider];\n // Lazily resolve the model on first use\n this._pendingModel = resolveModel(keyOrModel, provider, modelId);\n } else {\n this._model = keyOrModel;\n }\n }\n\n private async getModel(): Promise<LanguageModel> {\n if (this._model) return this._model;\n this._model = await this._pendingModel!;\n this._pendingModel = null;\n return this._model;\n }\n\n /**\n * Solve a captcha image.\n *\n * @param input - File path (string) or raw image Buffer\n * @param options - Solve options (attempts, expected length, etc.)\n * @returns Solved text, per-attempt answers, and token usage\n */\n async solve(input: string | Buffer, options: SolveOptions = {}): Promise<SolveResult> {\n const {\n numAttempts = 7,\n expectedLength,\n maxRetries = 2,\n verbose = true,\n confusionGroups = false,\n preprocess,\n } = options;\n\n const model = await this.getModel();\n\n // Two complementary views:\n // 1. Enhanced grayscale (high contrast + auto-crop) — great for clear text\n // 2. Color original (upscaled, no greyscale, no contrast) — preserves subtle features\n const [enhancedBuffer, colorBuffer] = await Promise.all([\n preprocessCaptchaToBuffer(input, preprocess),\n preprocessCaptchaToBuffer(input, {\n blur: 0,\n scale: 4,\n contrast: 1.0,\n sharpen: false,\n crop: 'none',\n padding: 40,\n greyscale: false,\n }),\n ]);\n\n // Fire all attempts in parallel for speed\n const results = await Promise.all(\n Array.from({ length: numAttempts }, () =>\n this.singleAttempt(model, enhancedBuffer, colorBuffer, maxRetries)\n )\n );\n const valid = results.filter((r): r is AttemptResult => r !== null);\n if (verbose) {\n valid.forEach((r, i) => console.log(` Attempt ${i + 1}: ${r.text}`));\n }\n\n const attempts = valid.map((r) => r.text);\n const attemptUsages = valid.map((r) => r.usage);\n const usage = aggregateUsage(attemptUsages);\n\n if (attempts.length === 0) {\n if (verbose) console.log(' All attempts failed!');\n return { text: '', attempts, usage, attemptUsages };\n }\n\n return {\n text: majorityVote(attempts, expectedLength, confusionGroups),\n attempts,\n usage,\n attemptUsages,\n };\n }\n\n /**\n * Make a single API call to read the captcha.\n * Retries up to `maxRetries` times on failure.\n */\n private async singleAttempt(\n model: LanguageModel,\n primaryBuffer: Buffer,\n secondaryBuffer: Buffer,\n maxRetries: number\n ): Promise<AttemptResult | null> {\n for (let retry = 0; retry <= maxRetries; retry++) {\n try {\n const { text, usage } = await generateText({\n model,\n messages: [\n {\n role: 'user',\n content: [\n { type: 'text', text: PROMPT },\n { type: 'image', image: primaryBuffer },\n { type: 'image', image: secondaryBuffer },\n ],\n },\n ],\n temperature: 1,\n maxOutputTokens: 256,\n });\n\n const raw = text.trim();\n\n // Detect refusals\n const lower = raw.toLowerCase();\n if (\n lower.includes('sorry') ||\n lower.includes(\"can't help\") ||\n lower.includes('cannot help') ||\n lower.includes('unable to') ||\n lower.includes(\"i can't\") ||\n raw.length > 20\n ) {\n return null;\n }\n\n // Clean: keep only uppercase letters and digits\n const cleaned = raw.replace(/[^A-Za-z0-9]/g, '');\n return cleaned ? { text: cleaned, usage } : null;\n } catch (_err) {\n if (retry < maxRetries) {\n await new Promise((r) => setTimeout(r, 1000 * (retry + 1)));\n continue;\n }\n return null;\n }\n }\n return null;\n }\n}\n","import fs from 'fs';\nimport path from 'path';\nimport sharp from 'sharp';\n\n// ── Types ────────────────────────────────────────────────────────────\n\nexport interface CropFractions {\n /** Fraction from left edge (0–1, default: 0.1) */\n left: number;\n /** Fraction from top edge (0–1, default: 0.02) */\n top: number;\n /** Fraction from left to keep (0–1, default: 0.9) */\n right: number;\n /** Fraction from top to keep (0–1, default: 0.6) */\n bottom: number;\n}\n\nexport interface PreprocessOptions {\n /** Gaussian blur radius (default: 1.5). Set to 0 to skip. */\n blur?: number;\n /** Upscale factor (default: 4) */\n scale?: number;\n /** Contrast multiplier around image mean (default: 3.0). Set to 1 to skip. */\n contrast?: number;\n /** Enable unsharp-mask sharpening (default: true) */\n sharpen?: boolean;\n /**\n * Crop mode (default: 'auto'):\n * - 'auto' – trim whitespace after contrast enhancement, with margin\n * - 'legacy' – fixed-percentage crop (original behavior)\n * - 'none' – skip cropping\n * - CropFractions – custom crop percentages\n */\n crop?: 'auto' | 'legacy' | 'none' | CropFractions;\n /** Add white padding around the result (default: true). Pass false to skip, or a number for custom px. */\n padding?: boolean | number;\n /** Invert colors (negate) after processing (default: false) */\n negate?: boolean;\n /** Convert to greyscale (default: true) */\n greyscale?: boolean;\n}\n\nconst LEGACY_CROP: CropFractions = { left: 0.1, top: 0.02, right: 0.9, bottom: 0.6 };\n\n// ── Public API ───────────────────────────────────────────────────────\n\n/**\n * Preprocess a captcha image and return a base64-encoded PNG string.\n */\nexport async function preprocessCaptcha(\n input: string | Buffer,\n options?: PreprocessOptions\n): Promise<string> {\n const buf = await preprocessCaptchaToBuffer(input, options);\n return buf.toString('base64');\n}\n\n/**\n * Preprocess a captcha image and return the resulting PNG as a raw Buffer.\n *\n * Pipeline:\n * 1. Gaussian blur in color space (smooths dither pattern)\n * 2. Grayscale conversion\n * 3. Upscale with Lanczos\n * 4. Contrast boost around image mean + sharpen\n * 5. Crop (auto-detect, legacy fixed, none, or custom)\n * 6. Add white padding\n */\nexport async function preprocessCaptchaToBuffer(\n input: string | Buffer,\n options?: PreprocessOptions\n): Promise<Buffer> {\n const {\n blur = 1.5,\n scale = 4,\n contrast = 3.0,\n sharpen = true,\n crop = 'auto',\n padding = true,\n negate = false,\n greyscale = true,\n } = options ?? {};\n\n const source = typeof input === 'string' ? path.resolve(input) : input;\n\n // Read original dimensions\n const metadata = await sharp(source).metadata();\n const origW = metadata.width!;\n const origH = metadata.height!;\n\n // Step 1-2: Blur (optional) + greyscale (optional)\n let pipeline = sharp(source);\n if (blur > 0) pipeline = pipeline.blur(blur);\n if (greyscale) pipeline = pipeline.greyscale();\n const smoothed = await pipeline.toBuffer();\n\n // Step 3: Upscale with Lanczos\n const upscaled = await sharp(smoothed)\n .resize(origW * scale, origH * scale, { kernel: 'lanczos3' })\n .toBuffer();\n\n // Step 4: Contrast boost + sharpen\n let enhanced: Buffer;\n if (contrast !== 1.0) {\n const stats = await sharp(upscaled).stats();\n const mean = stats.channels[0].mean;\n let pipe = sharp(upscaled).linear(contrast, mean * (1 - contrast));\n if (sharpen) pipe = pipe.sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 });\n enhanced = await pipe.toBuffer();\n } else {\n enhanced = sharpen\n ? await sharp(upscaled).sharpen({ sigma: 1.0, m1: 2.0, m2: 1.0 }).toBuffer()\n : upscaled;\n }\n\n // Step 5: Crop\n let cropped: Buffer;\n if (crop === 'none') {\n cropped = enhanced;\n } else if (crop === 'auto') {\n cropped = await autoCrop(enhanced);\n } else {\n const fractions = crop === 'legacy' ? LEGACY_CROP : crop;\n const scaledW = origW * scale;\n const scaledH = origH * scale;\n const cropLeft = Math.floor(scaledW * fractions.left);\n const cropTop = Math.floor(scaledH * fractions.top);\n const cropRight = Math.floor(scaledW * fractions.right);\n const cropBottom = Math.floor(scaledH * fractions.bottom);\n const cropW = cropRight - cropLeft;\n const cropH = cropBottom - cropTop;\n cropped = await sharp(enhanced)\n .extract({ left: cropLeft, top: cropTop, width: cropW, height: cropH })\n .toBuffer();\n }\n\n // Step 6: Negate (optional)\n const final = negate ? await sharp(cropped).negate().toBuffer() : cropped;\n\n // Step 7: Padding\n if (padding === false) {\n return sharp(final).png().toBuffer();\n }\n const pad = typeof padding === 'number' ? padding : undefined;\n const vPad = pad ?? 20;\n const hPad = pad ?? 30;\n return sharp(final)\n .extend({\n top: vPad,\n bottom: vPad,\n left: hPad,\n right: hPad,\n background: { r: 255, g: 255, b: 255 },\n })\n .png()\n .toBuffer();\n}\n\n/**\n * Auto-crop: use sharp.trim() to detect the content bounding box after\n * contrast enhancement, then add a small margin. Falls back to the\n * untrimmed image if trim removes everything.\n */\nasync function autoCrop(enhanced: Buffer): Promise<Buffer> {\n try {\n const trimmed = sharp(enhanced).trim({ threshold: 30 });\n const trimmedBuf = await trimmed.toBuffer({ resolveWithObject: true });\n\n // If trim left a reasonable image, add a margin\n const { width, height } = trimmedBuf.info;\n if (width > 2 && height > 2) {\n return trimmedBuf.data;\n }\n } catch {\n // trim() can throw if image is uniform — fall through\n }\n\n // Fallback: return untrimmed\n return enhanced;\n}\n\n/**\n * Read an image file and return its base64-encoded content.\n */\nexport function imageToBase64(imagePath: string): string {\n const buffer = fs.readFileSync(imagePath);\n return buffer.toString('base64');\n}\n"],"mappings":";AACA,SAAS,oBAAoB;;;ACD7B,OAAO,QAAQ;AACf,OAAO,UAAU;AACjB,OAAO,WAAW;AAwClB,IAAM,cAA6B,EAAE,MAAM,KAAK,KAAK,MAAM,OAAO,KAAK,QAAQ,IAAI;AAOnF,eAAsB,kBACpB,OACA,SACiB;AACjB,QAAM,MAAM,MAAM,0BAA0B,OAAO,OAAO;AAC1D,SAAO,IAAI,SAAS,QAAQ;AAC9B;AAaA,eAAsB,0BACpB,OACA,SACiB;AACjB,QAAM;AAAA,IACJ,OAAO;AAAA,IACP,QAAQ;AAAA,IACR,WAAW;AAAA,IACX,UAAU;AAAA,IACV,OAAO;AAAA,IACP,UAAU;AAAA,IACV,SAAS;AAAA,IACT,YAAY;AAAA,EACd,IAAI,WAAW,CAAC;AAEhB,QAAM,SAAS,OAAO,UAAU,WAAW,KAAK,QAAQ,KAAK,IAAI;AAGjE,QAAM,WAAW,MAAM,MAAM,MAAM,EAAE,SAAS;AAC9C,QAAM,QAAQ,SAAS;AACvB,QAAM,QAAQ,SAAS;AAGvB,MAAI,WAAW,MAAM,MAAM;AAC3B,MAAI,OAAO,EAAG,YAAW,SAAS,KAAK,IAAI;AAC3C,MAAI,UAAW,YAAW,SAAS,UAAU;AAC7C,QAAM,WAAW,MAAM,SAAS,SAAS;AAGzC,QAAM,WAAW,MAAM,MAAM,QAAQ,EAClC,OAAO,QAAQ,OAAO,QAAQ,OAAO,EAAE,QAAQ,WAAW,CAAC,EAC3D,SAAS;AAGZ,MAAI;AACJ,MAAI,aAAa,GAAK;AACpB,UAAM,QAAQ,MAAM,MAAM,QAAQ,EAAE,MAAM;AAC1C,UAAM,OAAO,MAAM,SAAS,CAAC,EAAE;AAC/B,QAAI,OAAO,MAAM,QAAQ,EAAE,OAAO,UAAU,QAAQ,IAAI,SAAS;AACjE,QAAI,QAAS,QAAO,KAAK,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC;AACjE,eAAW,MAAM,KAAK,SAAS;AAAA,EACjC,OAAO;AACL,eAAW,UACP,MAAM,MAAM,QAAQ,EAAE,QAAQ,EAAE,OAAO,GAAK,IAAI,GAAK,IAAI,EAAI,CAAC,EAAE,SAAS,IACzE;AAAA,EACN;AAGA,MAAI;AACJ,MAAI,SAAS,QAAQ;AACnB,cAAU;AAAA,EACZ,WAAW,SAAS,QAAQ;AAC1B,cAAU,MAAM,SAAS,QAAQ;AAAA,EACnC,OAAO;AACL,UAAM,YAAY,SAAS,WAAW,cAAc;AACpD,UAAM,UAAU,QAAQ;AACxB,UAAM,UAAU,QAAQ;AACxB,UAAM,WAAW,KAAK,MAAM,UAAU,UAAU,IAAI;AACpD,UAAM,UAAU,KAAK,MAAM,UAAU,UAAU,GAAG;AAClD,UAAM,YAAY,KAAK,MAAM,UAAU,UAAU,KAAK;AACtD,UAAM,aAAa,KAAK,MAAM,UAAU,UAAU,MAAM;AACxD,UAAM,QAAQ,YAAY;AAC1B,UAAM,QAAQ,aAAa;AAC3B,cAAU,MAAM,MAAM,QAAQ,EAC3B,QAAQ,EAAE,MAAM,UAAU,KAAK,SAAS,OAAO,OAAO,QAAQ,MAAM,CAAC,EACrE,SAAS;AAAA,EACd;AAGA,QAAM,QAAQ,SAAS,MAAM,MAAM,OAAO,EAAE,OAAO,EAAE,SAAS,IAAI;AAGlE,MAAI,YAAY,OAAO;AACrB,WAAO,MAAM,KAAK,EAAE,IAAI,EAAE,SAAS;AAAA,EACrC;AACA,QAAM,MAAM,OAAO,YAAY,WAAW,UAAU;AACpD,QAAM,OAAO,OAAO;AACpB,QAAM,OAAO,OAAO;AACpB,SAAO,MAAM,KAAK,EACf,OAAO;AAAA,IACN,KAAK;AAAA,IACL,QAAQ;AAAA,IACR,MAAM;AAAA,IACN,OAAO;AAAA,IACP,YAAY,EAAE,GAAG,KAAK,GAAG,KAAK,GAAG,IAAI;AAAA,EACvC,CAAC,EACA,IAAI,EACJ,SAAS;AACd;AAOA,eAAe,SAAS,UAAmC;AACzD,MAAI;AACF,UAAM,UAAU,MAAM,QAAQ,EAAE,KAAK,EAAE,WAAW,GAAG,CAAC;AACtD,UAAM,aAAa,MAAM,QAAQ,SAAS,EAAE,mBAAmB,KAAK,CAAC;AAGrE,UAAM,EAAE,OAAO,OAAO,IAAI,WAAW;AACrC,QAAI,QAAQ,KAAK,SAAS,GAAG;AAC3B,aAAO,WAAW;AAAA,IACpB;AAAA,EACF,QAAQ;AAAA,EAER;AAGA,SAAO;AACT;AAKO,SAAS,cAAc,WAA2B;AACvD,QAAM,SAAS,GAAG,aAAa,SAAS;AACxC,SAAO,OAAO,SAAS,QAAQ;AACjC;;;ADtLA,IAAM,SAAS;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAyDf,IAAM,iBAA2C;AAAA,EAC/C,QAAQ;AAAA,EACR,WAAW;AAAA,EACX,QAAQ;AACV;AAEA,eAAe,aACb,QACA,UACA,SACwB;AACxB,UAAQ,UAAU;AAAA,IAChB,KAAK,UAAU;AACb,YAAM,EAAE,aAAa,IAAI,MAAM,OAAO,gBAAgB;AACtD,aAAO,aAAa,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACzC;AAAA,IACA,KAAK,aAAa;AAEhB,YAAM,EAAE,gBAAgB,IAAI,MAAM,OAAO,mBAAmB;AAC5D,aAAO,gBAAgB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IAC5C;AAAA,IACA,KAAK,UAAU;AAEb,YAAM,EAAE,yBAAyB,IAAI,MAAM,OAAO,gBAAgB;AAClE,aAAO,yBAAyB,EAAE,OAAO,CAAC,EAAE,OAAO;AAAA,IACrD;AAAA,IACA;AACE,YAAM,IAAI;AAAA,QACR,qBAAqB,QAAQ;AAAA,MAC/B;AAAA,EACJ;AACF;AAUO,IAAM,0BAAkD;AAAA,EAC7D,KAAK;AAAA,EACL,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AAAA,EACL,GAAG;AAAA,EACH,KAAK;AACP;AASO,SAAS,aACd,UACA,gBACA,QACQ;AACR,MAAI,WAAW,iBAAiB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,cAAc,IAAI;AAEtF,MAAI,SAAS,WAAW,GAAG;AACzB,eAAW;AAAA,EACb;AACA,MAAI,SAAS,WAAW,EAAG,QAAO;AAGlC,QAAM,YAAY,oBAAI,IAAoB;AAC1C,aAAW,KAAK,UAAU;AACxB,cAAU,IAAI,EAAE,SAAS,UAAU,IAAI,EAAE,MAAM,KAAK,KAAK,CAAC;AAAA,EAC5D;AACA,MAAI,UAAU;AACd,MAAI,YAAY;AAChB,aAAW,CAAC,KAAK,KAAK,KAAK,WAAW;AACpC,QAAI,QAAQ,WAAW;AACrB,gBAAU;AACV,kBAAY;AAAA,IACd;AAAA,EACF;AAEA,QAAM,kBAAkB,SAAS,OAAO,CAAC,MAAM,EAAE,WAAW,OAAO;AACnE,MAAI,gBAAgB,WAAW,EAAG,QAAO,SAAS,CAAC;AAEnD,QAAM,YAAY,UAAU,OAAO,WAAW,WAAW,SAAS;AAGlE,QAAM,SAAmB,CAAC;AAC1B,WAAS,MAAM,GAAG,MAAM,SAAS,OAAO;AACtC,UAAM,aAAa,oBAAI,IAAoB;AAC3C,eAAW,KAAK,iBAAiB;AAC/B,YAAM,KAAK,EAAE,GAAG;AAChB,iBAAW,IAAI,KAAK,WAAW,IAAI,EAAE,KAAK,KAAK,CAAC;AAAA,IAClD;AAEA,QAAI,WAAW;AAEb,YAAM,cAAc,oBAAI,IAAoB;AAC5C,iBAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,cAAM,YAAY,UAAU,EAAE,KAAK;AACnC,oBAAY,IAAI,YAAY,YAAY,IAAI,SAAS,KAAK,KAAK,KAAK;AAAA,MACtE;AAEA,UAAI,YAAY;AAChB,UAAI,iBAAiB;AACrB,iBAAW,CAAC,WAAW,KAAK,KAAK,aAAa;AAC5C,YAAI,QAAQ,gBAAgB;AAC1B,sBAAY;AACZ,2BAAiB;AAAA,QACnB;AAAA,MACF;AACA,aAAO,KAAK,SAAS;AAAA,IACvB,OAAO;AAEL,UAAI,WAAW;AACf,UAAI,gBAAgB;AACpB,iBAAW,CAAC,IAAI,KAAK,KAAK,YAAY;AACpC,YAAI,QAAQ,eAAe;AACzB,qBAAW;AACX,0BAAgB;AAAA,QAClB;AAAA,MACF;AACA,aAAO,KAAK,QAAQ;AAAA,IACtB;AAAA,EACF;AAEA,SAAO,OAAO,KAAK,EAAE;AACvB;AAIA,SAAS,YAAY,GAAuB,GAA2C;AACrF,MAAI,MAAM,UAAa,MAAM,OAAW,QAAO;AAC/C,UAAQ,KAAK,MAAM,KAAK;AAC1B;AAEA,SAAS,eAAe,QAAkD;AACxE,QAAM,OAA2B;AAAA,IAC/B,aAAa;AAAA,IACb,mBAAmB;AAAA,MACjB,eAAe;AAAA,MACf,iBAAiB;AAAA,MACjB,kBAAkB;AAAA,IACpB;AAAA,IACA,cAAc;AAAA,IACd,oBAAoB;AAAA,MAClB,YAAY;AAAA,MACZ,iBAAiB;AAAA,IACnB;AAAA,IACA,aAAa;AAAA,EACf;AACA,SAAO,OAAO;AAAA,IACZ,CAAC,KAAK,OAAO;AAAA,MACX,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,MACvD,mBAAmB;AAAA,QACjB,eAAe;AAAA,UACb,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,iBAAiB;AAAA,UACf,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,QACA,kBAAkB;AAAA,UAChB,IAAI,kBAAkB;AAAA,UACtB,EAAE,kBAAkB;AAAA,QACtB;AAAA,MACF;AAAA,MACA,cAAc,YAAY,IAAI,cAAc,EAAE,YAAY;AAAA,MAC1D,oBAAoB;AAAA,QAClB,YAAY,YAAY,IAAI,mBAAmB,YAAY,EAAE,mBAAmB,UAAU;AAAA,QAC1F,iBAAiB;AAAA,UACf,IAAI,mBAAmB;AAAA,UACvB,EAAE,mBAAmB;AAAA,QACvB;AAAA,MACF;AAAA,MACA,aAAa,YAAY,IAAI,aAAa,EAAE,WAAW;AAAA,IACzD;AAAA,IACA;AAAA,EACF;AACF;AAIO,IAAM,SAAN,MAAa;AAAA,EACV,SAA+B;AAAA,EAC/B,gBAA+C;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAmBvD,YAAY,YAAoC,SAAyB;AACvE,QAAI,OAAO,eAAe,UAAU;AAClC,YAAM,WAAW,SAAS,YAAY;AACtC,YAAM,UAAU,SAAS,SAAS,eAAe,QAAQ;AAEzD,WAAK,gBAAgB,aAAa,YAAY,UAAU,OAAO;AAAA,IACjE,OAAO;AACL,WAAK,SAAS;AAAA,IAChB;AAAA,EACF;AAAA,EAEA,MAAc,WAAmC;AAC/C,QAAI,KAAK,OAAQ,QAAO,KAAK;AAC7B,SAAK,SAAS,MAAM,KAAK;AACzB,SAAK,gBAAgB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EASA,MAAM,MAAM,OAAwB,UAAwB,CAAC,GAAyB;AACpF,UAAM;AAAA,MACJ,cAAc;AAAA,MACd;AAAA,MACA,aAAa;AAAA,MACb,UAAU;AAAA,MACV,kBAAkB;AAAA,MAClB;AAAA,IACF,IAAI;AAEJ,UAAM,QAAQ,MAAM,KAAK,SAAS;AAKlC,UAAM,CAAC,gBAAgB,WAAW,IAAI,MAAM,QAAQ,IAAI;AAAA,MACtD,0BAA0B,OAAO,UAAU;AAAA,MAC3C,0BAA0B,OAAO;AAAA,QAC/B,MAAM;AAAA,QACN,OAAO;AAAA,QACP,UAAU;AAAA,QACV,SAAS;AAAA,QACT,MAAM;AAAA,QACN,SAAS;AAAA,QACT,WAAW;AAAA,MACb,CAAC;AAAA,IACH,CAAC;AAGD,UAAM,UAAU,MAAM,QAAQ;AAAA,MAC5B,MAAM;AAAA,QAAK,EAAE,QAAQ,YAAY;AAAA,QAAG,MAClC,KAAK,cAAc,OAAO,gBAAgB,aAAa,UAAU;AAAA,MACnE;AAAA,IACF;AACA,UAAM,QAAQ,QAAQ,OAAO,CAAC,MAA0B,MAAM,IAAI;AAClE,QAAI,SAAS;AACX,YAAM,QAAQ,CAAC,GAAG,MAAM,QAAQ,IAAI,aAAa,IAAI,CAAC,KAAK,EAAE,IAAI,EAAE,CAAC;AAAA,IACtE;AAEA,UAAM,WAAW,MAAM,IAAI,CAAC,MAAM,EAAE,IAAI;AACxC,UAAM,gBAAgB,MAAM,IAAI,CAAC,MAAM,EAAE,KAAK;AAC9C,UAAM,QAAQ,eAAe,aAAa;AAE1C,QAAI,SAAS,WAAW,GAAG;AACzB,UAAI,QAAS,SAAQ,IAAI,wBAAwB;AACjD,aAAO,EAAE,MAAM,IAAI,UAAU,OAAO,cAAc;AAAA,IACpD;AAEA,WAAO;AAAA,MACL,MAAM,aAAa,UAAU,gBAAgB,eAAe;AAAA,MAC5D;AAAA,MACA;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA;AAAA;AAAA;AAAA;AAAA,EAMA,MAAc,cACZ,OACA,eACA,iBACA,YAC+B;AAC/B,aAAS,QAAQ,GAAG,SAAS,YAAY,SAAS;AAChD,UAAI;AACF,cAAM,EAAE,MAAM,MAAM,IAAI,MAAM,aAAa;AAAA,UACzC;AAAA,UACA,UAAU;AAAA,YACR;AAAA,cACE,MAAM;AAAA,cACN,SAAS;AAAA,gBACP,EAAE,MAAM,QAAQ,MAAM,OAAO;AAAA,gBAC7B,EAAE,MAAM,SAAS,OAAO,cAAc;AAAA,gBACtC,EAAE,MAAM,SAAS,OAAO,gBAAgB;AAAA,cAC1C;AAAA,YACF;AAAA,UACF;AAAA,UACA,aAAa;AAAA,UACb,iBAAiB;AAAA,QACnB,CAAC;AAED,cAAM,MAAM,KAAK,KAAK;AAGtB,cAAM,QAAQ,IAAI,YAAY;AAC9B,YACE,MAAM,SAAS,OAAO,KACtB,MAAM,SAAS,YAAY,KAC3B,MAAM,SAAS,aAAa,KAC5B,MAAM,SAAS,WAAW,KAC1B,MAAM,SAAS,SAAS,KACxB,IAAI,SAAS,IACb;AACA,iBAAO;AAAA,QACT;AAGA,cAAM,UAAU,IAAI,QAAQ,iBAAiB,EAAE;AAC/C,eAAO,UAAU,EAAE,MAAM,SAAS,MAAM,IAAI;AAAA,MAC9C,SAAS,MAAM;AACb,YAAI,QAAQ,YAAY;AACtB,gBAAM,IAAI,QAAQ,CAAC,MAAM,WAAW,GAAG,OAAQ,QAAQ,EAAE,CAAC;AAC1D;AAAA,QACF;AACA,eAAO;AAAA,MACT;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACF;","names":[]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@yigitahmetsahin/captcha-solver",
3
- "version": "2.0.1",
3
+ "version": "3.0.0",
4
4
  "description": "AI-powered captcha solver using image preprocessing and multi-provider vision models (Vercel AI SDK)",
5
5
  "main": "dist/index.js",
6
6
  "module": "dist/index.mjs",
@@ -26,7 +26,8 @@
26
26
  "format": "prettier --write .",
27
27
  "lint": "npm run format && eslint src --fix && tsc --noEmit -p tsconfig.check.json",
28
28
  "lint:check": "prettier --check . && eslint src && tsc --noEmit -p tsconfig.check.json",
29
- "prepublishOnly": "npm run build"
29
+ "prepublishOnly": "npm run build",
30
+ "prepare": "husky"
30
31
  },
31
32
  "keywords": [
32
33
  "captcha",
@@ -62,6 +63,7 @@
62
63
  "@vitest/coverage-v8": "^4.0.18",
63
64
  "eslint": "^9.39.2",
64
65
  "eslint-config-prettier": "^10.1.8",
66
+ "husky": "^9.1.7",
65
67
  "prettier": "^3.8.1",
66
68
  "tsup": "^8.5.1",
67
69
  "tsx": "^4.19.0",
@@ -70,9 +72,9 @@
70
72
  "vitest": "^4.0.17"
71
73
  },
72
74
  "peerDependencies": {
73
- "@ai-sdk/openai": ">=1.0.0",
74
75
  "@ai-sdk/anthropic": ">=1.0.0",
75
- "@ai-sdk/google": ">=1.0.0"
76
+ "@ai-sdk/google": ">=1.0.0",
77
+ "@ai-sdk/openai": ">=1.0.0"
76
78
  },
77
79
  "peerDependenciesMeta": {
78
80
  "@ai-sdk/openai": {