@mhalder/qdrant-mcp-server 1.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (100) hide show
  1. package/.env.example +92 -0
  2. package/.github/workflows/ci.yml +61 -0
  3. package/.github/workflows/claude-code-review.yml +57 -0
  4. package/.github/workflows/claude.yml +50 -0
  5. package/.github/workflows/release.yml +52 -0
  6. package/.husky/commit-msg +1 -0
  7. package/.husky/pre-commit +1 -0
  8. package/.releaserc.json +59 -0
  9. package/.yamlfmt +4 -0
  10. package/CHANGELOG.md +73 -0
  11. package/CONTRIBUTING.md +176 -0
  12. package/LICENSE +21 -0
  13. package/README.md +714 -0
  14. package/build/embeddings/base.d.ts +23 -0
  15. package/build/embeddings/base.d.ts.map +1 -0
  16. package/build/embeddings/base.js +2 -0
  17. package/build/embeddings/base.js.map +1 -0
  18. package/build/embeddings/cohere.d.ts +17 -0
  19. package/build/embeddings/cohere.d.ts.map +1 -0
  20. package/build/embeddings/cohere.js +102 -0
  21. package/build/embeddings/cohere.js.map +1 -0
  22. package/build/embeddings/cohere.test.d.ts +2 -0
  23. package/build/embeddings/cohere.test.d.ts.map +1 -0
  24. package/build/embeddings/cohere.test.js +279 -0
  25. package/build/embeddings/cohere.test.js.map +1 -0
  26. package/build/embeddings/factory.d.ts +10 -0
  27. package/build/embeddings/factory.d.ts.map +1 -0
  28. package/build/embeddings/factory.js +98 -0
  29. package/build/embeddings/factory.js.map +1 -0
  30. package/build/embeddings/factory.test.d.ts +2 -0
  31. package/build/embeddings/factory.test.d.ts.map +1 -0
  32. package/build/embeddings/factory.test.js +329 -0
  33. package/build/embeddings/factory.test.js.map +1 -0
  34. package/build/embeddings/ollama.d.ts +18 -0
  35. package/build/embeddings/ollama.d.ts.map +1 -0
  36. package/build/embeddings/ollama.js +135 -0
  37. package/build/embeddings/ollama.js.map +1 -0
  38. package/build/embeddings/ollama.test.d.ts +2 -0
  39. package/build/embeddings/ollama.test.d.ts.map +1 -0
  40. package/build/embeddings/ollama.test.js +399 -0
  41. package/build/embeddings/ollama.test.js.map +1 -0
  42. package/build/embeddings/openai.d.ts +16 -0
  43. package/build/embeddings/openai.d.ts.map +1 -0
  44. package/build/embeddings/openai.js +108 -0
  45. package/build/embeddings/openai.js.map +1 -0
  46. package/build/embeddings/openai.test.d.ts +2 -0
  47. package/build/embeddings/openai.test.d.ts.map +1 -0
  48. package/build/embeddings/openai.test.js +283 -0
  49. package/build/embeddings/openai.test.js.map +1 -0
  50. package/build/embeddings/voyage.d.ts +19 -0
  51. package/build/embeddings/voyage.d.ts.map +1 -0
  52. package/build/embeddings/voyage.js +113 -0
  53. package/build/embeddings/voyage.js.map +1 -0
  54. package/build/embeddings/voyage.test.d.ts +2 -0
  55. package/build/embeddings/voyage.test.d.ts.map +1 -0
  56. package/build/embeddings/voyage.test.js +371 -0
  57. package/build/embeddings/voyage.test.js.map +1 -0
  58. package/build/index.d.ts +3 -0
  59. package/build/index.d.ts.map +1 -0
  60. package/build/index.js +534 -0
  61. package/build/index.js.map +1 -0
  62. package/build/index.test.d.ts +2 -0
  63. package/build/index.test.d.ts.map +1 -0
  64. package/build/index.test.js +241 -0
  65. package/build/index.test.js.map +1 -0
  66. package/build/qdrant/client.d.ts +37 -0
  67. package/build/qdrant/client.d.ts.map +1 -0
  68. package/build/qdrant/client.js +142 -0
  69. package/build/qdrant/client.js.map +1 -0
  70. package/build/qdrant/client.test.d.ts +2 -0
  71. package/build/qdrant/client.test.d.ts.map +1 -0
  72. package/build/qdrant/client.test.js +340 -0
  73. package/build/qdrant/client.test.js.map +1 -0
  74. package/commitlint.config.js +25 -0
  75. package/docker-compose.yml +22 -0
  76. package/docs/test_report.md +259 -0
  77. package/examples/README.md +315 -0
  78. package/examples/basic/README.md +111 -0
  79. package/examples/filters/README.md +262 -0
  80. package/examples/knowledge-base/README.md +207 -0
  81. package/examples/rate-limiting/README.md +376 -0
  82. package/package.json +59 -0
  83. package/scripts/verify-providers.js +238 -0
  84. package/src/embeddings/base.ts +25 -0
  85. package/src/embeddings/cohere.test.ts +408 -0
  86. package/src/embeddings/cohere.ts +152 -0
  87. package/src/embeddings/factory.test.ts +453 -0
  88. package/src/embeddings/factory.ts +163 -0
  89. package/src/embeddings/ollama.test.ts +543 -0
  90. package/src/embeddings/ollama.ts +196 -0
  91. package/src/embeddings/openai.test.ts +402 -0
  92. package/src/embeddings/openai.ts +158 -0
  93. package/src/embeddings/voyage.test.ts +520 -0
  94. package/src/embeddings/voyage.ts +168 -0
  95. package/src/index.test.ts +304 -0
  96. package/src/index.ts +614 -0
  97. package/src/qdrant/client.test.ts +456 -0
  98. package/src/qdrant/client.ts +195 -0
  99. package/tsconfig.json +19 -0
  100. package/vitest.config.ts +37 -0
@@ -0,0 +1,158 @@
1
+ import OpenAI from "openai";
2
+ import Bottleneck from "bottleneck";
3
+ import {
4
+ EmbeddingProvider,
5
+ EmbeddingResult,
6
+ RateLimitConfig,
7
+ ProviderConfig,
8
+ } from "./base.js";
9
+
10
+ interface OpenAIError {
11
+ status?: number;
12
+ code?: string;
13
+ message?: string;
14
+ headers?: Record<string, string>;
15
+ response?: {
16
+ headers?: Record<string, string>;
17
+ };
18
+ }
19
+
20
+ export class OpenAIEmbeddings implements EmbeddingProvider {
21
+ private client: OpenAI;
22
+ private model: string;
23
+ private dimensions: number;
24
+ private limiter: Bottleneck;
25
+ private retryAttempts: number;
26
+ private retryDelayMs: number;
27
+
28
+ constructor(
29
+ apiKey: string,
30
+ model: string = "text-embedding-3-small",
31
+ dimensions?: number,
32
+ rateLimitConfig?: RateLimitConfig,
33
+ ) {
34
+ this.client = new OpenAI({ apiKey });
35
+ this.model = model;
36
+
37
+ // Default dimensions for different models
38
+ const defaultDimensions: Record<string, number> = {
39
+ "text-embedding-3-small": 1536,
40
+ "text-embedding-3-large": 3072,
41
+ "text-embedding-ada-002": 1536,
42
+ };
43
+
44
+ this.dimensions = dimensions || defaultDimensions[model] || 1536;
45
+
46
+ // Rate limiting configuration
47
+ const maxRequestsPerMinute = rateLimitConfig?.maxRequestsPerMinute || 3500;
48
+ this.retryAttempts = rateLimitConfig?.retryAttempts || 3;
49
+ this.retryDelayMs = rateLimitConfig?.retryDelayMs || 1000;
50
+
51
+ // Initialize bottleneck limiter
52
+ // Uses reservoir (token bucket) pattern for burst handling with per-minute refresh
53
+ // Note: Using both reservoir and minTime provides defense in depth but may be
54
+ // more conservative than necessary. Future optimization could use reservoir-only
55
+ // for better burst handling or minTime-only for simpler even distribution.
56
+ this.limiter = new Bottleneck({
57
+ reservoir: maxRequestsPerMinute,
58
+ reservoirRefreshAmount: maxRequestsPerMinute,
59
+ reservoirRefreshInterval: 60 * 1000, // 1 minute
60
+ maxConcurrent: 10,
61
+ minTime: Math.floor((60 * 1000) / maxRequestsPerMinute),
62
+ });
63
+ }
64
+
65
+ private async retryWithBackoff<T>(
66
+ fn: () => Promise<T>,
67
+ attempt: number = 0,
68
+ ): Promise<T> {
69
+ try {
70
+ return await fn();
71
+ } catch (error: unknown) {
72
+ const apiError = error as OpenAIError;
73
+ const isRateLimitError =
74
+ apiError?.status === 429 ||
75
+ apiError?.code === "rate_limit_exceeded" ||
76
+ apiError?.message?.toLowerCase().includes("rate limit");
77
+
78
+ if (isRateLimitError && attempt < this.retryAttempts) {
79
+ // Check for Retry-After header (different HTTP clients may nest differently)
80
+ const retryAfter =
81
+ apiError?.response?.headers?.["retry-after"] ||
82
+ apiError?.headers?.["retry-after"];
83
+ let delayMs: number;
84
+
85
+ if (retryAfter) {
86
+ // Use Retry-After header if available (in seconds)
87
+ const parsed = parseInt(retryAfter, 10);
88
+ delayMs =
89
+ !isNaN(parsed) && parsed > 0
90
+ ? parsed * 1000
91
+ : this.retryDelayMs * Math.pow(2, attempt);
92
+ } else {
93
+ // Exponential backoff: 1s, 2s, 4s, 8s...
94
+ delayMs = this.retryDelayMs * Math.pow(2, attempt);
95
+ }
96
+
97
+ const waitTimeSeconds = (delayMs / 1000).toFixed(1);
98
+ console.error(
99
+ `Rate limit reached. Retrying in ${waitTimeSeconds}s (attempt ${attempt + 1}/${this.retryAttempts})...`,
100
+ );
101
+
102
+ await new Promise((resolve) => setTimeout(resolve, delayMs));
103
+ return this.retryWithBackoff(fn, attempt + 1);
104
+ }
105
+
106
+ // If not a rate limit error or max retries exceeded, throw
107
+ if (isRateLimitError) {
108
+ throw new Error(
109
+ `OpenAI API rate limit exceeded after ${this.retryAttempts} retry attempts. Please try again later or reduce request frequency.`,
110
+ );
111
+ }
112
+
113
+ throw error;
114
+ }
115
+ }
116
+
117
+ async embed(text: string): Promise<EmbeddingResult> {
118
+ return this.limiter.schedule(() =>
119
+ this.retryWithBackoff(async () => {
120
+ const response = await this.client.embeddings.create({
121
+ model: this.model,
122
+ input: text,
123
+ dimensions: this.dimensions,
124
+ });
125
+
126
+ return {
127
+ embedding: response.data[0].embedding,
128
+ dimensions: this.dimensions,
129
+ };
130
+ }),
131
+ );
132
+ }
133
+
134
+ async embedBatch(texts: string[]): Promise<EmbeddingResult[]> {
135
+ return this.limiter.schedule(() =>
136
+ this.retryWithBackoff(async () => {
137
+ const response = await this.client.embeddings.create({
138
+ model: this.model,
139
+ input: texts,
140
+ dimensions: this.dimensions,
141
+ });
142
+
143
+ return response.data.map((item) => ({
144
+ embedding: item.embedding,
145
+ dimensions: this.dimensions,
146
+ }));
147
+ }),
148
+ );
149
+ }
150
+
151
+ getDimensions(): number {
152
+ return this.dimensions;
153
+ }
154
+
155
+ getModel(): string {
156
+ return this.model;
157
+ }
158
+ }
@@ -0,0 +1,520 @@
1
+ import { describe, it, expect, vi, beforeEach } from "vitest";
2
+ import { VoyageEmbeddings } from "./voyage.js";
3
+
4
+ // Mock fetch globally
5
+ global.fetch = vi.fn();
6
+
7
+ describe("VoyageEmbeddings", () => {
8
+ let embeddings: VoyageEmbeddings;
9
+ let mockFetch: any;
10
+
11
+ beforeEach(() => {
12
+ mockFetch = global.fetch as any;
13
+ mockFetch.mockReset();
14
+
15
+ embeddings = new VoyageEmbeddings("test-api-key");
16
+ });
17
+
18
+ describe("constructor", () => {
19
+ it("should use default model and dimensions", () => {
20
+ expect(embeddings.getModel()).toBe("voyage-2");
21
+ expect(embeddings.getDimensions()).toBe(1024);
22
+ });
23
+
24
+ it("should use custom model", () => {
25
+ const customEmbeddings = new VoyageEmbeddings(
26
+ "test-api-key",
27
+ "voyage-large-2",
28
+ );
29
+ expect(customEmbeddings.getModel()).toBe("voyage-large-2");
30
+ expect(customEmbeddings.getDimensions()).toBe(1536);
31
+ });
32
+
33
+ it("should use custom dimensions", () => {
34
+ const customEmbeddings = new VoyageEmbeddings(
35
+ "test-api-key",
36
+ "voyage-2",
37
+ 512,
38
+ );
39
+ expect(customEmbeddings.getDimensions()).toBe(512);
40
+ });
41
+
42
+ it("should use default base URL", () => {
43
+ const defaultEmbeddings = new VoyageEmbeddings("test-api-key");
44
+ expect(defaultEmbeddings).toBeDefined();
45
+ });
46
+
47
+ it("should use custom base URL", () => {
48
+ const customEmbeddings = new VoyageEmbeddings(
49
+ "test-api-key",
50
+ "voyage-2",
51
+ undefined,
52
+ undefined,
53
+ "https://custom.voyage.com",
54
+ );
55
+ expect(customEmbeddings).toBeDefined();
56
+ });
57
+
58
+ it("should default to 1024 for unknown models", () => {
59
+ const unknownEmbeddings = new VoyageEmbeddings(
60
+ "test-api-key",
61
+ "custom-model",
62
+ );
63
+ expect(unknownEmbeddings.getDimensions()).toBe(1024);
64
+ });
65
+
66
+ it("should accept custom input type", () => {
67
+ const queryEmbeddings = new VoyageEmbeddings(
68
+ "test-api-key",
69
+ "voyage-2",
70
+ undefined,
71
+ undefined,
72
+ undefined,
73
+ "query",
74
+ );
75
+ expect(queryEmbeddings).toBeInstanceOf(VoyageEmbeddings);
76
+ });
77
+ });
78
+
79
+ describe("embed", () => {
80
+ it("should generate embedding for single text", async () => {
81
+ const mockEmbedding = Array(1024)
82
+ .fill(0)
83
+ .map((_, i) => i * 0.001);
84
+ mockFetch.mockResolvedValue({
85
+ ok: true,
86
+ json: async () => ({
87
+ data: [{ embedding: mockEmbedding }],
88
+ model: "voyage-2",
89
+ usage: { total_tokens: 10 },
90
+ }),
91
+ });
92
+
93
+ const result = await embeddings.embed("test text");
94
+
95
+ expect(result).toEqual({
96
+ embedding: mockEmbedding,
97
+ dimensions: 1024,
98
+ });
99
+ expect(mockFetch).toHaveBeenCalledWith(
100
+ "https://api.voyageai.com/v1/embeddings",
101
+ {
102
+ method: "POST",
103
+ headers: {
104
+ "Content-Type": "application/json",
105
+ Authorization: "Bearer test-api-key",
106
+ },
107
+ body: JSON.stringify({
108
+ input: ["test text"],
109
+ model: "voyage-2",
110
+ }),
111
+ },
112
+ );
113
+ });
114
+
115
+ it("should include input_type when specified", async () => {
116
+ const queryEmbeddings = new VoyageEmbeddings(
117
+ "test-api-key",
118
+ "voyage-2",
119
+ undefined,
120
+ undefined,
121
+ undefined,
122
+ "query",
123
+ );
124
+
125
+ const mockEmbedding = Array(1024).fill(0.5);
126
+ mockFetch.mockResolvedValue({
127
+ ok: true,
128
+ json: async () => ({
129
+ data: [{ embedding: mockEmbedding }],
130
+ model: "voyage-2",
131
+ usage: { total_tokens: 10 },
132
+ }),
133
+ });
134
+
135
+ await queryEmbeddings.embed("test text");
136
+
137
+ const callBody = JSON.parse(mockFetch.mock.calls[0][1].body);
138
+ expect(callBody.input_type).toBe("query");
139
+ });
140
+
141
+ it("should handle long text", async () => {
142
+ const longText = "word ".repeat(1000);
143
+ const mockEmbedding = Array(1024).fill(0.5);
144
+ mockFetch.mockResolvedValue({
145
+ ok: true,
146
+ json: async () => ({
147
+ data: [{ embedding: mockEmbedding }],
148
+ model: "voyage-2",
149
+ usage: { total_tokens: 1000 },
150
+ }),
151
+ });
152
+
153
+ const result = await embeddings.embed(longText);
154
+
155
+ expect(result.embedding).toEqual(mockEmbedding);
156
+ });
157
+
158
+ it("should use custom base URL", async () => {
159
+ const customEmbeddings = new VoyageEmbeddings(
160
+ "test-api-key",
161
+ "voyage-2",
162
+ undefined,
163
+ undefined,
164
+ "https://custom.voyage.com/v1",
165
+ );
166
+
167
+ const mockEmbedding = Array(1024).fill(0.1);
168
+ mockFetch.mockResolvedValue({
169
+ ok: true,
170
+ json: async () => ({
171
+ data: [{ embedding: mockEmbedding }],
172
+ model: "voyage-2",
173
+ usage: { total_tokens: 5 },
174
+ }),
175
+ });
176
+
177
+ await customEmbeddings.embed("test");
178
+
179
+ expect(mockFetch).toHaveBeenCalledWith(
180
+ "https://custom.voyage.com/v1/embeddings",
181
+ expect.any(Object),
182
+ );
183
+ });
184
+
185
+ it("should throw error if no embedding returned", async () => {
186
+ mockFetch.mockResolvedValue({
187
+ ok: true,
188
+ json: async () => ({
189
+ data: [],
190
+ model: "voyage-2",
191
+ usage: { total_tokens: 0 },
192
+ }),
193
+ });
194
+
195
+ await expect(embeddings.embed("test")).rejects.toThrow(
196
+ "No embedding returned from Voyage AI API",
197
+ );
198
+ });
199
+
200
+ it("should handle API errors", async () => {
201
+ mockFetch.mockResolvedValue({
202
+ ok: false,
203
+ status: 401,
204
+ text: async () => "Unauthorized: Invalid API key",
205
+ });
206
+
207
+ await expect(embeddings.embed("test")).rejects.toThrow();
208
+ });
209
+
210
+ it("should propagate network errors", async () => {
211
+ mockFetch.mockRejectedValue(new Error("Network Error"));
212
+
213
+ await expect(embeddings.embed("test")).rejects.toThrow("Network Error");
214
+ });
215
+ });
216
+
217
+ describe("embedBatch", () => {
218
+ it("should generate embeddings for multiple texts", async () => {
219
+ const mockEmbeddings = [
220
+ Array(1024).fill(0.1),
221
+ Array(1024).fill(0.2),
222
+ Array(1024).fill(0.3),
223
+ ];
224
+ mockFetch.mockResolvedValue({
225
+ ok: true,
226
+ json: async () => ({
227
+ data: mockEmbeddings.map((embedding) => ({ embedding })),
228
+ model: "voyage-2",
229
+ usage: { total_tokens: 30 },
230
+ }),
231
+ });
232
+
233
+ const texts = ["text1", "text2", "text3"];
234
+ const results = await embeddings.embedBatch(texts);
235
+
236
+ expect(results).toEqual([
237
+ { embedding: mockEmbeddings[0], dimensions: 1024 },
238
+ { embedding: mockEmbeddings[1], dimensions: 1024 },
239
+ { embedding: mockEmbeddings[2], dimensions: 1024 },
240
+ ]);
241
+
242
+ const callBody = JSON.parse(mockFetch.mock.calls[0][1].body);
243
+ expect(callBody.input).toEqual(texts);
244
+ });
245
+
246
+ it("should handle empty batch", async () => {
247
+ mockFetch.mockResolvedValue({
248
+ ok: true,
249
+ json: async () => ({
250
+ data: [],
251
+ model: "voyage-2",
252
+ usage: { total_tokens: 0 },
253
+ }),
254
+ });
255
+
256
+ const results = await embeddings.embedBatch([]);
257
+
258
+ expect(results).toEqual([]);
259
+ });
260
+
261
+ it("should handle single item in batch", async () => {
262
+ const mockEmbedding = Array(1024).fill(0.5);
263
+ mockFetch.mockResolvedValue({
264
+ ok: true,
265
+ json: async () => ({
266
+ data: [{ embedding: mockEmbedding }],
267
+ model: "voyage-2",
268
+ usage: { total_tokens: 10 },
269
+ }),
270
+ });
271
+
272
+ const results = await embeddings.embedBatch(["single text"]);
273
+
274
+ expect(results).toHaveLength(1);
275
+ expect(results[0].embedding).toEqual(mockEmbedding);
276
+ });
277
+
278
+ it("should handle large batches", async () => {
279
+ const batchSize = 100;
280
+ const mockEmbeddings = Array(batchSize)
281
+ .fill(null)
282
+ .map(() => Array(1024).fill(Math.random()));
283
+
284
+ mockFetch.mockResolvedValue({
285
+ ok: true,
286
+ json: async () => ({
287
+ data: mockEmbeddings.map((embedding) => ({ embedding })),
288
+ model: "voyage-2",
289
+ usage: { total_tokens: batchSize * 10 },
290
+ }),
291
+ });
292
+
293
+ const texts = Array(batchSize)
294
+ .fill(null)
295
+ .map((_, i) => `text ${i}`);
296
+ const results = await embeddings.embedBatch(texts);
297
+
298
+ expect(results).toHaveLength(batchSize);
299
+ });
300
+
301
+ it("should throw error if no embeddings returned", async () => {
302
+ mockFetch.mockResolvedValue({
303
+ ok: true,
304
+ json: async () => ({
305
+ model: "voyage-2",
306
+ usage: { total_tokens: 0 },
307
+ }),
308
+ });
309
+
310
+ await expect(embeddings.embedBatch(["text1"])).rejects.toThrow(
311
+ "No embeddings returned from Voyage AI API",
312
+ );
313
+ });
314
+
315
+ it("should propagate errors in batch", async () => {
316
+ mockFetch.mockRejectedValue(new Error("Batch API Error"));
317
+
318
+ await expect(embeddings.embedBatch(["text1", "text2"])).rejects.toThrow(
319
+ "Batch API Error",
320
+ );
321
+ });
322
+ });
323
+
324
+ describe("getDimensions", () => {
325
+ it("should return configured dimensions", () => {
326
+ expect(embeddings.getDimensions()).toBe(1024);
327
+ });
328
+
329
+ it("should return custom dimensions", () => {
330
+ const customEmbeddings = new VoyageEmbeddings(
331
+ "test-api-key",
332
+ "voyage-2",
333
+ 512,
334
+ );
335
+ expect(customEmbeddings.getDimensions()).toBe(512);
336
+ });
337
+ });
338
+
339
+ describe("getModel", () => {
340
+ it("should return configured model", () => {
341
+ expect(embeddings.getModel()).toBe("voyage-2");
342
+ });
343
+
344
+ it("should return custom model", () => {
345
+ const customEmbeddings = new VoyageEmbeddings(
346
+ "test-api-key",
347
+ "voyage-large-2",
348
+ );
349
+ expect(customEmbeddings.getModel()).toBe("voyage-large-2");
350
+ });
351
+ });
352
+
353
+ describe("rate limiting", () => {
354
+ it("should retry on rate limit error (429 status)", async () => {
355
+ const mockEmbedding = Array(1024).fill(0.5);
356
+
357
+ mockFetch
358
+ .mockResolvedValueOnce({
359
+ ok: false,
360
+ status: 429,
361
+ text: async () => "Rate limit exceeded",
362
+ })
363
+ .mockResolvedValueOnce({
364
+ ok: false,
365
+ status: 429,
366
+ text: async () => "Rate limit exceeded",
367
+ })
368
+ .mockResolvedValue({
369
+ ok: true,
370
+ json: async () => ({
371
+ data: [{ embedding: mockEmbedding }],
372
+ model: "voyage-2",
373
+ usage: { total_tokens: 10 },
374
+ }),
375
+ });
376
+
377
+ const result = await embeddings.embed("test text");
378
+
379
+ expect(result.embedding).toEqual(mockEmbedding);
380
+ expect(mockFetch).toHaveBeenCalledTimes(3);
381
+ });
382
+
383
+ it("should retry on rate limit message", async () => {
384
+ const mockEmbedding = Array(1024).fill(0.5);
385
+
386
+ mockFetch
387
+ .mockRejectedValueOnce({
388
+ message: "You have exceeded the rate limit",
389
+ })
390
+ .mockResolvedValue({
391
+ ok: true,
392
+ json: async () => ({
393
+ data: [{ embedding: mockEmbedding }],
394
+ model: "voyage-2",
395
+ usage: { total_tokens: 10 },
396
+ }),
397
+ });
398
+
399
+ const result = await embeddings.embed("test text");
400
+
401
+ expect(result.embedding).toEqual(mockEmbedding);
402
+ expect(mockFetch).toHaveBeenCalledTimes(2);
403
+ });
404
+
405
+ it("should use exponential backoff", async () => {
406
+ const rateLimitEmbeddings = new VoyageEmbeddings(
407
+ "test-api-key",
408
+ "voyage-2",
409
+ undefined,
410
+ {
411
+ retryAttempts: 3,
412
+ retryDelayMs: 100,
413
+ },
414
+ );
415
+
416
+ const mockEmbedding = Array(1024).fill(0.5);
417
+
418
+ mockFetch
419
+ .mockResolvedValueOnce({
420
+ ok: false,
421
+ status: 429,
422
+ text: async () => "Rate limit",
423
+ })
424
+ .mockResolvedValueOnce({
425
+ ok: false,
426
+ status: 429,
427
+ text: async () => "Rate limit",
428
+ })
429
+ .mockResolvedValue({
430
+ ok: true,
431
+ json: async () => ({
432
+ data: [{ embedding: mockEmbedding }],
433
+ model: "voyage-2",
434
+ usage: { total_tokens: 10 },
435
+ }),
436
+ });
437
+
438
+ const startTime = Date.now();
439
+ await rateLimitEmbeddings.embed("test text");
440
+ const duration = Date.now() - startTime;
441
+
442
+ // Should wait: 100ms (first retry) + 200ms (second retry) = 300ms
443
+ expect(duration).toBeGreaterThanOrEqual(250);
444
+ });
445
+
446
+ it("should throw error after max retries exceeded", async () => {
447
+ const rateLimitEmbeddings = new VoyageEmbeddings(
448
+ "test-api-key",
449
+ "voyage-2",
450
+ undefined,
451
+ {
452
+ retryAttempts: 2,
453
+ retryDelayMs: 100,
454
+ },
455
+ );
456
+
457
+ mockFetch.mockResolvedValue({
458
+ ok: false,
459
+ status: 429,
460
+ text: async () => "Rate limit exceeded",
461
+ });
462
+
463
+ await expect(rateLimitEmbeddings.embed("test text")).rejects.toThrow(
464
+ "Voyage AI API rate limit exceeded after 2 retry attempts",
465
+ );
466
+
467
+ expect(mockFetch).toHaveBeenCalledTimes(3);
468
+ });
469
+
470
+ it("should handle rate limit errors in batch operations", async () => {
471
+ const mockEmbeddings = [Array(1024).fill(0.1), Array(1024).fill(0.2)];
472
+
473
+ mockFetch
474
+ .mockResolvedValueOnce({
475
+ ok: false,
476
+ status: 429,
477
+ text: async () => "Rate limit",
478
+ })
479
+ .mockResolvedValue({
480
+ ok: true,
481
+ json: async () => ({
482
+ data: mockEmbeddings.map((embedding) => ({ embedding })),
483
+ model: "voyage-2",
484
+ usage: { total_tokens: 20 },
485
+ }),
486
+ });
487
+
488
+ const results = await embeddings.embedBatch(["text1", "text2"]);
489
+
490
+ expect(results).toHaveLength(2);
491
+ expect(mockFetch).toHaveBeenCalledTimes(2);
492
+ });
493
+
494
+ it("should not retry on non-rate-limit errors", async () => {
495
+ mockFetch.mockResolvedValue({
496
+ ok: false,
497
+ status: 401,
498
+ text: async () => "Unauthorized",
499
+ });
500
+
501
+ await expect(embeddings.embed("test text")).rejects.toThrow();
502
+ expect(mockFetch).toHaveBeenCalledTimes(1);
503
+ });
504
+
505
+ it("should accept custom rate limit configuration", () => {
506
+ const customEmbeddings = new VoyageEmbeddings(
507
+ "test-api-key",
508
+ "voyage-2",
509
+ undefined,
510
+ {
511
+ maxRequestsPerMinute: 500,
512
+ retryAttempts: 5,
513
+ retryDelayMs: 2000,
514
+ },
515
+ );
516
+
517
+ expect(customEmbeddings).toBeDefined();
518
+ });
519
+ });
520
+ });