@mhalder/qdrant-mcp-server 1.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (100) hide show
  1. package/.env.example +92 -0
  2. package/.github/workflows/ci.yml +61 -0
  3. package/.github/workflows/claude-code-review.yml +57 -0
  4. package/.github/workflows/claude.yml +50 -0
  5. package/.github/workflows/release.yml +52 -0
  6. package/.husky/commit-msg +1 -0
  7. package/.husky/pre-commit +1 -0
  8. package/.releaserc.json +59 -0
  9. package/.yamlfmt +4 -0
  10. package/CHANGELOG.md +73 -0
  11. package/CONTRIBUTING.md +176 -0
  12. package/LICENSE +21 -0
  13. package/README.md +714 -0
  14. package/build/embeddings/base.d.ts +23 -0
  15. package/build/embeddings/base.d.ts.map +1 -0
  16. package/build/embeddings/base.js +2 -0
  17. package/build/embeddings/base.js.map +1 -0
  18. package/build/embeddings/cohere.d.ts +17 -0
  19. package/build/embeddings/cohere.d.ts.map +1 -0
  20. package/build/embeddings/cohere.js +102 -0
  21. package/build/embeddings/cohere.js.map +1 -0
  22. package/build/embeddings/cohere.test.d.ts +2 -0
  23. package/build/embeddings/cohere.test.d.ts.map +1 -0
  24. package/build/embeddings/cohere.test.js +279 -0
  25. package/build/embeddings/cohere.test.js.map +1 -0
  26. package/build/embeddings/factory.d.ts +10 -0
  27. package/build/embeddings/factory.d.ts.map +1 -0
  28. package/build/embeddings/factory.js +98 -0
  29. package/build/embeddings/factory.js.map +1 -0
  30. package/build/embeddings/factory.test.d.ts +2 -0
  31. package/build/embeddings/factory.test.d.ts.map +1 -0
  32. package/build/embeddings/factory.test.js +329 -0
  33. package/build/embeddings/factory.test.js.map +1 -0
  34. package/build/embeddings/ollama.d.ts +18 -0
  35. package/build/embeddings/ollama.d.ts.map +1 -0
  36. package/build/embeddings/ollama.js +135 -0
  37. package/build/embeddings/ollama.js.map +1 -0
  38. package/build/embeddings/ollama.test.d.ts +2 -0
  39. package/build/embeddings/ollama.test.d.ts.map +1 -0
  40. package/build/embeddings/ollama.test.js +399 -0
  41. package/build/embeddings/ollama.test.js.map +1 -0
  42. package/build/embeddings/openai.d.ts +16 -0
  43. package/build/embeddings/openai.d.ts.map +1 -0
  44. package/build/embeddings/openai.js +108 -0
  45. package/build/embeddings/openai.js.map +1 -0
  46. package/build/embeddings/openai.test.d.ts +2 -0
  47. package/build/embeddings/openai.test.d.ts.map +1 -0
  48. package/build/embeddings/openai.test.js +283 -0
  49. package/build/embeddings/openai.test.js.map +1 -0
  50. package/build/embeddings/voyage.d.ts +19 -0
  51. package/build/embeddings/voyage.d.ts.map +1 -0
  52. package/build/embeddings/voyage.js +113 -0
  53. package/build/embeddings/voyage.js.map +1 -0
  54. package/build/embeddings/voyage.test.d.ts +2 -0
  55. package/build/embeddings/voyage.test.d.ts.map +1 -0
  56. package/build/embeddings/voyage.test.js +371 -0
  57. package/build/embeddings/voyage.test.js.map +1 -0
  58. package/build/index.d.ts +3 -0
  59. package/build/index.d.ts.map +1 -0
  60. package/build/index.js +534 -0
  61. package/build/index.js.map +1 -0
  62. package/build/index.test.d.ts +2 -0
  63. package/build/index.test.d.ts.map +1 -0
  64. package/build/index.test.js +241 -0
  65. package/build/index.test.js.map +1 -0
  66. package/build/qdrant/client.d.ts +37 -0
  67. package/build/qdrant/client.d.ts.map +1 -0
  68. package/build/qdrant/client.js +142 -0
  69. package/build/qdrant/client.js.map +1 -0
  70. package/build/qdrant/client.test.d.ts +2 -0
  71. package/build/qdrant/client.test.d.ts.map +1 -0
  72. package/build/qdrant/client.test.js +340 -0
  73. package/build/qdrant/client.test.js.map +1 -0
  74. package/commitlint.config.js +25 -0
  75. package/docker-compose.yml +22 -0
  76. package/docs/test_report.md +259 -0
  77. package/examples/README.md +315 -0
  78. package/examples/basic/README.md +111 -0
  79. package/examples/filters/README.md +262 -0
  80. package/examples/knowledge-base/README.md +207 -0
  81. package/examples/rate-limiting/README.md +376 -0
  82. package/package.json +59 -0
  83. package/scripts/verify-providers.js +238 -0
  84. package/src/embeddings/base.ts +25 -0
  85. package/src/embeddings/cohere.test.ts +408 -0
  86. package/src/embeddings/cohere.ts +152 -0
  87. package/src/embeddings/factory.test.ts +453 -0
  88. package/src/embeddings/factory.ts +163 -0
  89. package/src/embeddings/ollama.test.ts +543 -0
  90. package/src/embeddings/ollama.ts +196 -0
  91. package/src/embeddings/openai.test.ts +402 -0
  92. package/src/embeddings/openai.ts +158 -0
  93. package/src/embeddings/voyage.test.ts +520 -0
  94. package/src/embeddings/voyage.ts +168 -0
  95. package/src/index.test.ts +304 -0
  96. package/src/index.ts +614 -0
  97. package/src/qdrant/client.test.ts +456 -0
  98. package/src/qdrant/client.ts +195 -0
  99. package/tsconfig.json +19 -0
  100. package/vitest.config.ts +37 -0
@@ -0,0 +1,163 @@
1
+ import { EmbeddingProvider, ProviderConfig } from "./base.js";
2
+ import { OpenAIEmbeddings } from "./openai.js";
3
+ import { CohereEmbeddings } from "./cohere.js";
4
+ import { VoyageEmbeddings } from "./voyage.js";
5
+ import { OllamaEmbeddings } from "./ollama.js";
6
+
7
+ export type EmbeddingProviderType = "openai" | "cohere" | "voyage" | "ollama";
8
+
9
+ export interface FactoryConfig extends ProviderConfig {
10
+ provider: EmbeddingProviderType;
11
+ }
12
+
13
+ export class EmbeddingProviderFactory {
14
+ static create(config: FactoryConfig): EmbeddingProvider {
15
+ const { provider, model, dimensions, rateLimitConfig, apiKey, baseUrl } =
16
+ config;
17
+
18
+ switch (provider) {
19
+ case "openai":
20
+ if (!apiKey) {
21
+ throw new Error("API key is required for OpenAI provider");
22
+ }
23
+ return new OpenAIEmbeddings(
24
+ apiKey,
25
+ model || "text-embedding-3-small",
26
+ dimensions,
27
+ rateLimitConfig,
28
+ );
29
+
30
+ case "cohere":
31
+ if (!apiKey) {
32
+ throw new Error("API key is required for Cohere provider");
33
+ }
34
+ return new CohereEmbeddings(
35
+ apiKey,
36
+ model || "embed-english-v3.0",
37
+ dimensions,
38
+ rateLimitConfig,
39
+ );
40
+
41
+ case "voyage":
42
+ if (!apiKey) {
43
+ throw new Error("API key is required for Voyage AI provider");
44
+ }
45
+ return new VoyageEmbeddings(
46
+ apiKey,
47
+ model || "voyage-2",
48
+ dimensions,
49
+ rateLimitConfig,
50
+ baseUrl || "https://api.voyageai.com/v1",
51
+ );
52
+
53
+ case "ollama":
54
+ return new OllamaEmbeddings(
55
+ model || "nomic-embed-text",
56
+ dimensions,
57
+ rateLimitConfig,
58
+ baseUrl || "http://localhost:11434",
59
+ );
60
+
61
+ default:
62
+ throw new Error(
63
+ `Unknown embedding provider: ${provider}. Supported providers: openai, cohere, voyage, ollama`,
64
+ );
65
+ }
66
+ }
67
+
68
+ static createFromEnv(): EmbeddingProvider {
69
+ const provider = (
70
+ process.env.EMBEDDING_PROVIDER || "ollama"
71
+ ).toLowerCase() as EmbeddingProviderType;
72
+
73
+ // Select API key based on provider
74
+ let apiKey: string | undefined;
75
+ switch (provider) {
76
+ case "openai":
77
+ apiKey = process.env.OPENAI_API_KEY;
78
+ break;
79
+ case "cohere":
80
+ apiKey = process.env.COHERE_API_KEY;
81
+ break;
82
+ case "voyage":
83
+ apiKey = process.env.VOYAGE_API_KEY;
84
+ break;
85
+ case "ollama":
86
+ // No API key needed for local Ollama
87
+ break;
88
+ }
89
+
90
+ // Common configuration
91
+ const model = process.env.EMBEDDING_MODEL;
92
+ const dimensions = process.env.EMBEDDING_DIMENSIONS
93
+ ? parseInt(process.env.EMBEDDING_DIMENSIONS, 10)
94
+ : undefined;
95
+
96
+ // Validate dimensions
97
+ if (dimensions !== undefined && (isNaN(dimensions) || dimensions <= 0)) {
98
+ throw new Error(
99
+ `Invalid EMBEDDING_DIMENSIONS: must be a positive integer, got "${process.env.EMBEDDING_DIMENSIONS}"`,
100
+ );
101
+ }
102
+
103
+ const baseUrl = process.env.EMBEDDING_BASE_URL;
104
+
105
+ // Rate limiting configuration
106
+ const maxRequestsPerMinute = process.env.EMBEDDING_MAX_REQUESTS_PER_MINUTE
107
+ ? parseInt(process.env.EMBEDDING_MAX_REQUESTS_PER_MINUTE, 10)
108
+ : undefined;
109
+
110
+ // Validate maxRequestsPerMinute
111
+ if (
112
+ maxRequestsPerMinute !== undefined &&
113
+ (isNaN(maxRequestsPerMinute) || maxRequestsPerMinute <= 0)
114
+ ) {
115
+ throw new Error(
116
+ `Invalid EMBEDDING_MAX_REQUESTS_PER_MINUTE: must be a positive integer, got "${process.env.EMBEDDING_MAX_REQUESTS_PER_MINUTE}"`,
117
+ );
118
+ }
119
+
120
+ const retryAttempts = process.env.EMBEDDING_RETRY_ATTEMPTS
121
+ ? parseInt(process.env.EMBEDDING_RETRY_ATTEMPTS, 10)
122
+ : undefined;
123
+
124
+ // Validate retryAttempts
125
+ if (
126
+ retryAttempts !== undefined &&
127
+ (isNaN(retryAttempts) || retryAttempts < 0)
128
+ ) {
129
+ throw new Error(
130
+ `Invalid EMBEDDING_RETRY_ATTEMPTS: must be a non-negative integer, got "${process.env.EMBEDDING_RETRY_ATTEMPTS}"`,
131
+ );
132
+ }
133
+
134
+ const retryDelayMs = process.env.EMBEDDING_RETRY_DELAY
135
+ ? parseInt(process.env.EMBEDDING_RETRY_DELAY, 10)
136
+ : undefined;
137
+
138
+ // Validate retryDelayMs
139
+ if (
140
+ retryDelayMs !== undefined &&
141
+ (isNaN(retryDelayMs) || retryDelayMs < 0)
142
+ ) {
143
+ throw new Error(
144
+ `Invalid EMBEDDING_RETRY_DELAY: must be a non-negative integer, got "${process.env.EMBEDDING_RETRY_DELAY}"`,
145
+ );
146
+ }
147
+
148
+ const rateLimitConfig = {
149
+ maxRequestsPerMinute,
150
+ retryAttempts,
151
+ retryDelayMs,
152
+ };
153
+
154
+ return this.create({
155
+ provider,
156
+ model,
157
+ dimensions,
158
+ rateLimitConfig,
159
+ apiKey,
160
+ baseUrl,
161
+ });
162
+ }
163
+ }
@@ -0,0 +1,543 @@
1
+ import { describe, it, expect, vi, beforeEach } from "vitest";
2
+ import { OllamaEmbeddings } from "./ollama.js";
3
+
4
+ // Mock fetch globally
5
+ global.fetch = vi.fn();
6
+
7
+ describe("OllamaEmbeddings", () => {
8
+ let embeddings: OllamaEmbeddings;
9
+ let mockFetch: any;
10
+
11
+ beforeEach(() => {
12
+ mockFetch = global.fetch as any;
13
+ mockFetch.mockReset();
14
+
15
+ embeddings = new OllamaEmbeddings();
16
+ });
17
+
18
+ describe("constructor", () => {
19
+ it("should use default model and dimensions", () => {
20
+ expect(embeddings.getModel()).toBe("nomic-embed-text");
21
+ expect(embeddings.getDimensions()).toBe(768);
22
+ });
23
+
24
+ it("should use custom model", () => {
25
+ const customEmbeddings = new OllamaEmbeddings("mxbai-embed-large");
26
+ expect(customEmbeddings.getModel()).toBe("mxbai-embed-large");
27
+ expect(customEmbeddings.getDimensions()).toBe(1024);
28
+ });
29
+
30
+ it("should use custom dimensions", () => {
31
+ const customEmbeddings = new OllamaEmbeddings("nomic-embed-text", 512);
32
+ expect(customEmbeddings.getDimensions()).toBe(512);
33
+ });
34
+
35
+ it("should use default base URL", () => {
36
+ const defaultEmbeddings = new OllamaEmbeddings();
37
+ expect(defaultEmbeddings).toBeDefined();
38
+ });
39
+
40
+ it("should use custom base URL", () => {
41
+ const customEmbeddings = new OllamaEmbeddings(
42
+ "nomic-embed-text",
43
+ undefined,
44
+ undefined,
45
+ "http://custom:11434",
46
+ );
47
+ expect(customEmbeddings).toBeDefined();
48
+ });
49
+
50
+ it("should default to 768 for unknown models", () => {
51
+ const unknownEmbeddings = new OllamaEmbeddings("custom-model");
52
+ expect(unknownEmbeddings.getDimensions()).toBe(768);
53
+ });
54
+
55
+ it("should use all-minilm model with 384 dimensions", () => {
56
+ const miniEmbeddings = new OllamaEmbeddings("all-minilm");
57
+ expect(miniEmbeddings.getModel()).toBe("all-minilm");
58
+ expect(miniEmbeddings.getDimensions()).toBe(384);
59
+ });
60
+ });
61
+
62
+ describe("embed", () => {
63
+ it("should generate embedding for single text", async () => {
64
+ const mockEmbedding = Array(768)
65
+ .fill(0)
66
+ .map((_, i) => i * 0.001);
67
+ mockFetch.mockResolvedValue({
68
+ ok: true,
69
+ json: async () => ({
70
+ embedding: mockEmbedding,
71
+ }),
72
+ });
73
+
74
+ const result = await embeddings.embed("test text");
75
+
76
+ expect(result).toEqual({
77
+ embedding: mockEmbedding,
78
+ dimensions: 768,
79
+ });
80
+ expect(mockFetch).toHaveBeenCalledWith(
81
+ "http://localhost:11434/api/embeddings",
82
+ {
83
+ method: "POST",
84
+ headers: {
85
+ "Content-Type": "application/json",
86
+ },
87
+ body: JSON.stringify({
88
+ model: "nomic-embed-text",
89
+ prompt: "test text",
90
+ }),
91
+ },
92
+ );
93
+ });
94
+
95
+ it("should handle long text", async () => {
96
+ const longText = "word ".repeat(1000);
97
+ const mockEmbedding = Array(768).fill(0.5);
98
+ mockFetch.mockResolvedValue({
99
+ ok: true,
100
+ json: async () => ({
101
+ embedding: mockEmbedding,
102
+ }),
103
+ });
104
+
105
+ const result = await embeddings.embed(longText);
106
+
107
+ expect(result.embedding).toEqual(mockEmbedding);
108
+ });
109
+
110
+ it("should use custom base URL", async () => {
111
+ const customEmbeddings = new OllamaEmbeddings(
112
+ "nomic-embed-text",
113
+ undefined,
114
+ undefined,
115
+ "http://custom:11434",
116
+ );
117
+
118
+ const mockEmbedding = Array(768).fill(0.1);
119
+ mockFetch.mockResolvedValue({
120
+ ok: true,
121
+ json: async () => ({
122
+ embedding: mockEmbedding,
123
+ }),
124
+ });
125
+
126
+ await customEmbeddings.embed("test");
127
+
128
+ expect(mockFetch).toHaveBeenCalledWith(
129
+ "http://custom:11434/api/embeddings",
130
+ expect.any(Object),
131
+ );
132
+ });
133
+
134
+ it("should throw error if no embedding returned", async () => {
135
+ mockFetch.mockResolvedValue({
136
+ ok: true,
137
+ json: async () => ({}),
138
+ });
139
+
140
+ await expect(embeddings.embed("test")).rejects.toThrow(
141
+ "No embedding returned from Ollama API",
142
+ );
143
+ });
144
+
145
+ it("should handle API errors", async () => {
146
+ mockFetch.mockResolvedValue({
147
+ ok: false,
148
+ status: 404,
149
+ text: async () => "Model not found",
150
+ });
151
+
152
+ await expect(embeddings.embed("test")).rejects.toThrow();
153
+ });
154
+
155
+ it("should propagate network errors", async () => {
156
+ mockFetch.mockRejectedValue(new Error("Network Error"));
157
+
158
+ await expect(embeddings.embed("test")).rejects.toThrow("Network Error");
159
+ });
160
+
161
+ it("should include text preview in API error for long text", async () => {
162
+ const longText = "a".repeat(150);
163
+ mockFetch.mockResolvedValue({
164
+ ok: false,
165
+ status: 500,
166
+ text: async () => "Server error",
167
+ });
168
+
169
+ await expect(embeddings.embed(longText)).rejects.toThrow(
170
+ expect.objectContaining({
171
+ message: expect.stringContaining("Text preview:"),
172
+ }),
173
+ );
174
+ });
175
+
176
+ it("should include model and URL in network error messages for non-Error objects", async () => {
177
+ mockFetch.mockRejectedValue("Connection refused");
178
+
179
+ await expect(embeddings.embed("test")).rejects.toThrow(
180
+ "Failed to call Ollama API at http://localhost:11434 with model nomic-embed-text",
181
+ );
182
+ });
183
+
184
+ it("should handle errors with message property", async () => {
185
+ mockFetch.mockRejectedValue({
186
+ message: "Custom error message",
187
+ });
188
+
189
+ await expect(embeddings.embed("test")).rejects.toThrow(
190
+ "Custom error message",
191
+ );
192
+ });
193
+
194
+ it("should handle non-Error objects in catch block", async () => {
195
+ mockFetch.mockRejectedValue({ code: "ERR_UNKNOWN", details: "info" });
196
+
197
+ await expect(embeddings.embed("test")).rejects.toThrow(
198
+ "Failed to call Ollama API at http://localhost:11434 with model nomic-embed-text",
199
+ );
200
+ });
201
+ });
202
+
203
+ describe("embedBatch", () => {
204
+ it("should generate embeddings for multiple texts in parallel", async () => {
205
+ const mockEmbeddings = [
206
+ Array(768).fill(0.1),
207
+ Array(768).fill(0.2),
208
+ Array(768).fill(0.3),
209
+ ];
210
+
211
+ // Mock sequential calls for each text
212
+ mockEmbeddings.forEach((embedding) => {
213
+ mockFetch.mockResolvedValueOnce({
214
+ ok: true,
215
+ json: async () => ({ embedding }),
216
+ });
217
+ });
218
+
219
+ const texts = ["text1", "text2", "text3"];
220
+ const results = await embeddings.embedBatch(texts);
221
+
222
+ expect(results).toEqual([
223
+ { embedding: mockEmbeddings[0], dimensions: 768 },
224
+ { embedding: mockEmbeddings[1], dimensions: 768 },
225
+ { embedding: mockEmbeddings[2], dimensions: 768 },
226
+ ]);
227
+
228
+ // Ollama processes each text individually
229
+ expect(mockFetch).toHaveBeenCalledTimes(3);
230
+ });
231
+
232
+ it("should handle empty batch", async () => {
233
+ const results = await embeddings.embedBatch([]);
234
+
235
+ expect(results).toEqual([]);
236
+ expect(mockFetch).not.toHaveBeenCalled();
237
+ });
238
+
239
+ it("should handle single item in batch", async () => {
240
+ const mockEmbedding = Array(768).fill(0.5);
241
+ mockFetch.mockResolvedValue({
242
+ ok: true,
243
+ json: async () => ({ embedding: mockEmbedding }),
244
+ });
245
+
246
+ const results = await embeddings.embedBatch(["single text"]);
247
+
248
+ expect(results).toHaveLength(1);
249
+ expect(results[0].embedding).toEqual(mockEmbedding);
250
+ });
251
+
252
+ it("should handle large batches with parallel processing", async () => {
253
+ const batchSize = 20;
254
+ const mockEmbedding = Array(768).fill(0.5);
255
+
256
+ // Mock all responses
257
+ for (let i = 0; i < batchSize; i++) {
258
+ mockFetch.mockResolvedValueOnce({
259
+ ok: true,
260
+ json: async () => ({ embedding: mockEmbedding }),
261
+ });
262
+ }
263
+
264
+ const texts = Array(batchSize)
265
+ .fill(null)
266
+ .map((_, i) => `text ${i}`);
267
+ const results = await embeddings.embedBatch(texts);
268
+
269
+ expect(results).toHaveLength(batchSize);
270
+ expect(mockFetch).toHaveBeenCalledTimes(batchSize);
271
+ });
272
+
273
+ it("should propagate errors in batch", async () => {
274
+ mockFetch
275
+ .mockResolvedValueOnce({
276
+ ok: true,
277
+ json: async () => ({ embedding: Array(768).fill(0.1) }),
278
+ })
279
+ .mockRejectedValueOnce(new Error("Batch API Error"));
280
+
281
+ await expect(embeddings.embedBatch(["text1", "text2"])).rejects.toThrow(
282
+ "Batch API Error",
283
+ );
284
+ });
285
+
286
+ it("should handle partial failures in batch", async () => {
287
+ mockFetch
288
+ .mockResolvedValueOnce({
289
+ ok: true,
290
+ json: async () => ({ embedding: Array(768).fill(0.1) }),
291
+ })
292
+ .mockResolvedValueOnce({
293
+ ok: false,
294
+ status: 500,
295
+ text: async () => "Internal error",
296
+ });
297
+
298
+ await expect(embeddings.embedBatch(["text1", "text2"])).rejects.toThrow();
299
+ });
300
+ });
301
+
302
+ describe("getDimensions", () => {
303
+ it("should return configured dimensions", () => {
304
+ expect(embeddings.getDimensions()).toBe(768);
305
+ });
306
+
307
+ it("should return custom dimensions", () => {
308
+ const customEmbeddings = new OllamaEmbeddings("nomic-embed-text", 512);
309
+ expect(customEmbeddings.getDimensions()).toBe(512);
310
+ });
311
+ });
312
+
313
+ describe("getModel", () => {
314
+ it("should return configured model", () => {
315
+ expect(embeddings.getModel()).toBe("nomic-embed-text");
316
+ });
317
+
318
+ it("should return custom model", () => {
319
+ const customEmbeddings = new OllamaEmbeddings("mxbai-embed-large");
320
+ expect(customEmbeddings.getModel()).toBe("mxbai-embed-large");
321
+ });
322
+ });
323
+
324
+ describe("rate limiting", () => {
325
+ it("should retry on rate limit error (429 status)", async () => {
326
+ const mockEmbedding = Array(768).fill(0.5);
327
+
328
+ mockFetch
329
+ .mockResolvedValueOnce({
330
+ ok: false,
331
+ status: 429,
332
+ text: async () => "Rate limit exceeded",
333
+ })
334
+ .mockResolvedValueOnce({
335
+ ok: false,
336
+ status: 429,
337
+ text: async () => "Rate limit exceeded",
338
+ })
339
+ .mockResolvedValue({
340
+ ok: true,
341
+ json: async () => ({ embedding: mockEmbedding }),
342
+ });
343
+
344
+ const result = await embeddings.embed("test text");
345
+
346
+ expect(result.embedding).toEqual(mockEmbedding);
347
+ expect(mockFetch).toHaveBeenCalledTimes(3);
348
+ });
349
+
350
+ it("should retry on rate limit message", async () => {
351
+ const mockEmbedding = Array(768).fill(0.5);
352
+
353
+ mockFetch
354
+ .mockRejectedValueOnce({
355
+ message: "You have exceeded the rate limit",
356
+ })
357
+ .mockResolvedValue({
358
+ ok: true,
359
+ json: async () => ({ embedding: mockEmbedding }),
360
+ });
361
+
362
+ const result = await embeddings.embed("test text");
363
+
364
+ expect(result.embedding).toEqual(mockEmbedding);
365
+ expect(mockFetch).toHaveBeenCalledTimes(2);
366
+ });
367
+
368
+ it("should use exponential backoff with faster default delay", async () => {
369
+ const rateLimitEmbeddings = new OllamaEmbeddings(
370
+ "nomic-embed-text",
371
+ undefined,
372
+ {
373
+ retryAttempts: 3,
374
+ retryDelayMs: 100,
375
+ },
376
+ );
377
+
378
+ const mockEmbedding = Array(768).fill(0.5);
379
+
380
+ mockFetch
381
+ .mockResolvedValueOnce({
382
+ ok: false,
383
+ status: 429,
384
+ text: async () => "Rate limit",
385
+ })
386
+ .mockResolvedValueOnce({
387
+ ok: false,
388
+ status: 429,
389
+ text: async () => "Rate limit",
390
+ })
391
+ .mockResolvedValue({
392
+ ok: true,
393
+ json: async () => ({ embedding: mockEmbedding }),
394
+ });
395
+
396
+ const startTime = Date.now();
397
+ await rateLimitEmbeddings.embed("test text");
398
+ const duration = Date.now() - startTime;
399
+
400
+ // Should wait: 100ms (first retry) + 200ms (second retry) = 300ms
401
+ expect(duration).toBeGreaterThanOrEqual(250);
402
+ });
403
+
404
+ it("should throw error after max retries exceeded", async () => {
405
+ const rateLimitEmbeddings = new OllamaEmbeddings(
406
+ "nomic-embed-text",
407
+ undefined,
408
+ {
409
+ retryAttempts: 2,
410
+ retryDelayMs: 100,
411
+ },
412
+ );
413
+
414
+ mockFetch.mockResolvedValue({
415
+ ok: false,
416
+ status: 429,
417
+ text: async () => "Rate limit exceeded",
418
+ });
419
+
420
+ await expect(rateLimitEmbeddings.embed("test text")).rejects.toThrow(
421
+ "Ollama API rate limit exceeded after 2 retry attempts",
422
+ );
423
+
424
+ expect(mockFetch).toHaveBeenCalledTimes(3);
425
+ });
426
+
427
+ it("should handle rate limit errors in batch operations", async () => {
428
+ const mockEmbedding = Array(768).fill(0.5);
429
+
430
+ mockFetch
431
+ .mockResolvedValueOnce({
432
+ ok: false,
433
+ status: 429,
434
+ text: async () => "Rate limit",
435
+ })
436
+ .mockResolvedValueOnce({
437
+ ok: true,
438
+ json: async () => ({ embedding: mockEmbedding }),
439
+ })
440
+ .mockResolvedValueOnce({
441
+ ok: true,
442
+ json: async () => ({ embedding: mockEmbedding }),
443
+ });
444
+
445
+ const results = await embeddings.embedBatch(["text1", "text2"]);
446
+
447
+ expect(results).toHaveLength(2);
448
+ // First call fails and retries, then succeeds. Second call succeeds immediately.
449
+ expect(mockFetch).toHaveBeenCalledTimes(3);
450
+ });
451
+
452
+ it("should not retry on non-rate-limit errors", async () => {
453
+ mockFetch.mockResolvedValue({
454
+ ok: false,
455
+ status: 404,
456
+ text: async () => "Model not found",
457
+ });
458
+
459
+ await expect(embeddings.embed("test text")).rejects.toThrow();
460
+ expect(mockFetch).toHaveBeenCalledTimes(1);
461
+ });
462
+
463
+ it("should accept custom rate limit configuration", () => {
464
+ const customEmbeddings = new OllamaEmbeddings(
465
+ "nomic-embed-text",
466
+ undefined,
467
+ {
468
+ maxRequestsPerMinute: 2000,
469
+ retryAttempts: 5,
470
+ retryDelayMs: 1000,
471
+ },
472
+ );
473
+
474
+ expect(customEmbeddings).toBeDefined();
475
+ });
476
+
477
+ it("should have higher default rate limit for local deployment", () => {
478
+ // Ollama defaults to 1000 requests/minute (more lenient than cloud providers)
479
+ const defaultEmbeddings = new OllamaEmbeddings();
480
+ expect(defaultEmbeddings).toBeDefined();
481
+ });
482
+
483
+ it("should handle primitive error values in retry logic", async () => {
484
+ // This tests line 69: when error is not an OllamaError, convert to { status: 0, message: String(error) }
485
+ mockFetch.mockRejectedValue(null);
486
+
487
+ await expect(embeddings.embed("test")).rejects.toThrow();
488
+ expect(mockFetch).toHaveBeenCalledTimes(1);
489
+ });
490
+
491
+ it("should handle string primitive errors", async () => {
492
+ mockFetch.mockRejectedValue("Network unreachable");
493
+
494
+ await expect(embeddings.embed("test")).rejects.toThrow(
495
+ "Network unreachable",
496
+ );
497
+ });
498
+
499
+ it("should handle error objects with non-string message property", async () => {
500
+ mockFetch.mockRejectedValue({
501
+ message: 404, // Non-string message
502
+ code: "NOT_FOUND",
503
+ });
504
+
505
+ // Should not treat this as a rate limit error even though it has a message property
506
+ await expect(embeddings.embed("test")).rejects.toThrow();
507
+ expect(mockFetch).toHaveBeenCalledTimes(1); // No retries
508
+ });
509
+
510
+ it("should handle Error instance in retry logic", async () => {
511
+ const testError = new Error("Connection timeout");
512
+ mockFetch.mockRejectedValue(testError);
513
+
514
+ await expect(embeddings.embed("test")).rejects.toThrow(
515
+ "Connection timeout",
516
+ );
517
+ });
518
+
519
+ it("should handle Error instance from network error with enhanced message", async () => {
520
+ // This tests error instanceof Error path for network errors
521
+ const networkError = new Error("ECONNREFUSED");
522
+ mockFetch.mockRejectedValue(networkError);
523
+
524
+ await expect(embeddings.embed("test")).rejects.toThrow(
525
+ "Failed to call Ollama API at http://localhost:11434 with model nomic-embed-text: ECONNREFUSED. Text preview:",
526
+ );
527
+ });
528
+
529
+ it("should handle object with string message property", async () => {
530
+ // This tests lines 143-144: object with message property that is a string
531
+ const customError = {
532
+ code: "API_ERROR",
533
+ message: "Custom API failure",
534
+ details: "Something went wrong",
535
+ };
536
+ mockFetch.mockRejectedValue(customError);
537
+
538
+ await expect(embeddings.embed("test")).rejects.toThrow(
539
+ "Custom API failure",
540
+ );
541
+ });
542
+ });
543
+ });