@mhalder/qdrant-mcp-server 1.1.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.env.example +92 -0
- package/.github/workflows/ci.yml +61 -0
- package/.github/workflows/claude-code-review.yml +57 -0
- package/.github/workflows/claude.yml +50 -0
- package/.github/workflows/release.yml +52 -0
- package/.husky/commit-msg +1 -0
- package/.husky/pre-commit +1 -0
- package/.releaserc.json +59 -0
- package/.yamlfmt +4 -0
- package/CHANGELOG.md +73 -0
- package/CONTRIBUTING.md +176 -0
- package/LICENSE +21 -0
- package/README.md +714 -0
- package/build/embeddings/base.d.ts +23 -0
- package/build/embeddings/base.d.ts.map +1 -0
- package/build/embeddings/base.js +2 -0
- package/build/embeddings/base.js.map +1 -0
- package/build/embeddings/cohere.d.ts +17 -0
- package/build/embeddings/cohere.d.ts.map +1 -0
- package/build/embeddings/cohere.js +102 -0
- package/build/embeddings/cohere.js.map +1 -0
- package/build/embeddings/cohere.test.d.ts +2 -0
- package/build/embeddings/cohere.test.d.ts.map +1 -0
- package/build/embeddings/cohere.test.js +279 -0
- package/build/embeddings/cohere.test.js.map +1 -0
- package/build/embeddings/factory.d.ts +10 -0
- package/build/embeddings/factory.d.ts.map +1 -0
- package/build/embeddings/factory.js +98 -0
- package/build/embeddings/factory.js.map +1 -0
- package/build/embeddings/factory.test.d.ts +2 -0
- package/build/embeddings/factory.test.d.ts.map +1 -0
- package/build/embeddings/factory.test.js +329 -0
- package/build/embeddings/factory.test.js.map +1 -0
- package/build/embeddings/ollama.d.ts +18 -0
- package/build/embeddings/ollama.d.ts.map +1 -0
- package/build/embeddings/ollama.js +135 -0
- package/build/embeddings/ollama.js.map +1 -0
- package/build/embeddings/ollama.test.d.ts +2 -0
- package/build/embeddings/ollama.test.d.ts.map +1 -0
- package/build/embeddings/ollama.test.js +399 -0
- package/build/embeddings/ollama.test.js.map +1 -0
- package/build/embeddings/openai.d.ts +16 -0
- package/build/embeddings/openai.d.ts.map +1 -0
- package/build/embeddings/openai.js +108 -0
- package/build/embeddings/openai.js.map +1 -0
- package/build/embeddings/openai.test.d.ts +2 -0
- package/build/embeddings/openai.test.d.ts.map +1 -0
- package/build/embeddings/openai.test.js +283 -0
- package/build/embeddings/openai.test.js.map +1 -0
- package/build/embeddings/voyage.d.ts +19 -0
- package/build/embeddings/voyage.d.ts.map +1 -0
- package/build/embeddings/voyage.js +113 -0
- package/build/embeddings/voyage.js.map +1 -0
- package/build/embeddings/voyage.test.d.ts +2 -0
- package/build/embeddings/voyage.test.d.ts.map +1 -0
- package/build/embeddings/voyage.test.js +371 -0
- package/build/embeddings/voyage.test.js.map +1 -0
- package/build/index.d.ts +3 -0
- package/build/index.d.ts.map +1 -0
- package/build/index.js +534 -0
- package/build/index.js.map +1 -0
- package/build/index.test.d.ts +2 -0
- package/build/index.test.d.ts.map +1 -0
- package/build/index.test.js +241 -0
- package/build/index.test.js.map +1 -0
- package/build/qdrant/client.d.ts +37 -0
- package/build/qdrant/client.d.ts.map +1 -0
- package/build/qdrant/client.js +142 -0
- package/build/qdrant/client.js.map +1 -0
- package/build/qdrant/client.test.d.ts +2 -0
- package/build/qdrant/client.test.d.ts.map +1 -0
- package/build/qdrant/client.test.js +340 -0
- package/build/qdrant/client.test.js.map +1 -0
- package/commitlint.config.js +25 -0
- package/docker-compose.yml +22 -0
- package/docs/test_report.md +259 -0
- package/examples/README.md +315 -0
- package/examples/basic/README.md +111 -0
- package/examples/filters/README.md +262 -0
- package/examples/knowledge-base/README.md +207 -0
- package/examples/rate-limiting/README.md +376 -0
- package/package.json +59 -0
- package/scripts/verify-providers.js +238 -0
- package/src/embeddings/base.ts +25 -0
- package/src/embeddings/cohere.test.ts +408 -0
- package/src/embeddings/cohere.ts +152 -0
- package/src/embeddings/factory.test.ts +453 -0
- package/src/embeddings/factory.ts +163 -0
- package/src/embeddings/ollama.test.ts +543 -0
- package/src/embeddings/ollama.ts +196 -0
- package/src/embeddings/openai.test.ts +402 -0
- package/src/embeddings/openai.ts +158 -0
- package/src/embeddings/voyage.test.ts +520 -0
- package/src/embeddings/voyage.ts +168 -0
- package/src/index.test.ts +304 -0
- package/src/index.ts +614 -0
- package/src/qdrant/client.test.ts +456 -0
- package/src/qdrant/client.ts +195 -0
- package/tsconfig.json +19 -0
- package/vitest.config.ts +37 -0
|
@@ -0,0 +1,158 @@
|
|
|
1
|
+
import OpenAI from "openai";
|
|
2
|
+
import Bottleneck from "bottleneck";
|
|
3
|
+
import {
|
|
4
|
+
EmbeddingProvider,
|
|
5
|
+
EmbeddingResult,
|
|
6
|
+
RateLimitConfig,
|
|
7
|
+
ProviderConfig,
|
|
8
|
+
} from "./base.js";
|
|
9
|
+
|
|
10
|
+
interface OpenAIError {
|
|
11
|
+
status?: number;
|
|
12
|
+
code?: string;
|
|
13
|
+
message?: string;
|
|
14
|
+
headers?: Record<string, string>;
|
|
15
|
+
response?: {
|
|
16
|
+
headers?: Record<string, string>;
|
|
17
|
+
};
|
|
18
|
+
}
|
|
19
|
+
|
|
20
|
+
export class OpenAIEmbeddings implements EmbeddingProvider {
|
|
21
|
+
private client: OpenAI;
|
|
22
|
+
private model: string;
|
|
23
|
+
private dimensions: number;
|
|
24
|
+
private limiter: Bottleneck;
|
|
25
|
+
private retryAttempts: number;
|
|
26
|
+
private retryDelayMs: number;
|
|
27
|
+
|
|
28
|
+
constructor(
|
|
29
|
+
apiKey: string,
|
|
30
|
+
model: string = "text-embedding-3-small",
|
|
31
|
+
dimensions?: number,
|
|
32
|
+
rateLimitConfig?: RateLimitConfig,
|
|
33
|
+
) {
|
|
34
|
+
this.client = new OpenAI({ apiKey });
|
|
35
|
+
this.model = model;
|
|
36
|
+
|
|
37
|
+
// Default dimensions for different models
|
|
38
|
+
const defaultDimensions: Record<string, number> = {
|
|
39
|
+
"text-embedding-3-small": 1536,
|
|
40
|
+
"text-embedding-3-large": 3072,
|
|
41
|
+
"text-embedding-ada-002": 1536,
|
|
42
|
+
};
|
|
43
|
+
|
|
44
|
+
this.dimensions = dimensions || defaultDimensions[model] || 1536;
|
|
45
|
+
|
|
46
|
+
// Rate limiting configuration
|
|
47
|
+
const maxRequestsPerMinute = rateLimitConfig?.maxRequestsPerMinute || 3500;
|
|
48
|
+
this.retryAttempts = rateLimitConfig?.retryAttempts || 3;
|
|
49
|
+
this.retryDelayMs = rateLimitConfig?.retryDelayMs || 1000;
|
|
50
|
+
|
|
51
|
+
// Initialize bottleneck limiter
|
|
52
|
+
// Uses reservoir (token bucket) pattern for burst handling with per-minute refresh
|
|
53
|
+
// Note: Using both reservoir and minTime provides defense in depth but may be
|
|
54
|
+
// more conservative than necessary. Future optimization could use reservoir-only
|
|
55
|
+
// for better burst handling or minTime-only for simpler even distribution.
|
|
56
|
+
this.limiter = new Bottleneck({
|
|
57
|
+
reservoir: maxRequestsPerMinute,
|
|
58
|
+
reservoirRefreshAmount: maxRequestsPerMinute,
|
|
59
|
+
reservoirRefreshInterval: 60 * 1000, // 1 minute
|
|
60
|
+
maxConcurrent: 10,
|
|
61
|
+
minTime: Math.floor((60 * 1000) / maxRequestsPerMinute),
|
|
62
|
+
});
|
|
63
|
+
}
|
|
64
|
+
|
|
65
|
+
private async retryWithBackoff<T>(
|
|
66
|
+
fn: () => Promise<T>,
|
|
67
|
+
attempt: number = 0,
|
|
68
|
+
): Promise<T> {
|
|
69
|
+
try {
|
|
70
|
+
return await fn();
|
|
71
|
+
} catch (error: unknown) {
|
|
72
|
+
const apiError = error as OpenAIError;
|
|
73
|
+
const isRateLimitError =
|
|
74
|
+
apiError?.status === 429 ||
|
|
75
|
+
apiError?.code === "rate_limit_exceeded" ||
|
|
76
|
+
apiError?.message?.toLowerCase().includes("rate limit");
|
|
77
|
+
|
|
78
|
+
if (isRateLimitError && attempt < this.retryAttempts) {
|
|
79
|
+
// Check for Retry-After header (different HTTP clients may nest differently)
|
|
80
|
+
const retryAfter =
|
|
81
|
+
apiError?.response?.headers?.["retry-after"] ||
|
|
82
|
+
apiError?.headers?.["retry-after"];
|
|
83
|
+
let delayMs: number;
|
|
84
|
+
|
|
85
|
+
if (retryAfter) {
|
|
86
|
+
// Use Retry-After header if available (in seconds)
|
|
87
|
+
const parsed = parseInt(retryAfter, 10);
|
|
88
|
+
delayMs =
|
|
89
|
+
!isNaN(parsed) && parsed > 0
|
|
90
|
+
? parsed * 1000
|
|
91
|
+
: this.retryDelayMs * Math.pow(2, attempt);
|
|
92
|
+
} else {
|
|
93
|
+
// Exponential backoff: 1s, 2s, 4s, 8s...
|
|
94
|
+
delayMs = this.retryDelayMs * Math.pow(2, attempt);
|
|
95
|
+
}
|
|
96
|
+
|
|
97
|
+
const waitTimeSeconds = (delayMs / 1000).toFixed(1);
|
|
98
|
+
console.error(
|
|
99
|
+
`Rate limit reached. Retrying in ${waitTimeSeconds}s (attempt ${attempt + 1}/${this.retryAttempts})...`,
|
|
100
|
+
);
|
|
101
|
+
|
|
102
|
+
await new Promise((resolve) => setTimeout(resolve, delayMs));
|
|
103
|
+
return this.retryWithBackoff(fn, attempt + 1);
|
|
104
|
+
}
|
|
105
|
+
|
|
106
|
+
// If not a rate limit error or max retries exceeded, throw
|
|
107
|
+
if (isRateLimitError) {
|
|
108
|
+
throw new Error(
|
|
109
|
+
`OpenAI API rate limit exceeded after ${this.retryAttempts} retry attempts. Please try again later or reduce request frequency.`,
|
|
110
|
+
);
|
|
111
|
+
}
|
|
112
|
+
|
|
113
|
+
throw error;
|
|
114
|
+
}
|
|
115
|
+
}
|
|
116
|
+
|
|
117
|
+
async embed(text: string): Promise<EmbeddingResult> {
|
|
118
|
+
return this.limiter.schedule(() =>
|
|
119
|
+
this.retryWithBackoff(async () => {
|
|
120
|
+
const response = await this.client.embeddings.create({
|
|
121
|
+
model: this.model,
|
|
122
|
+
input: text,
|
|
123
|
+
dimensions: this.dimensions,
|
|
124
|
+
});
|
|
125
|
+
|
|
126
|
+
return {
|
|
127
|
+
embedding: response.data[0].embedding,
|
|
128
|
+
dimensions: this.dimensions,
|
|
129
|
+
};
|
|
130
|
+
}),
|
|
131
|
+
);
|
|
132
|
+
}
|
|
133
|
+
|
|
134
|
+
async embedBatch(texts: string[]): Promise<EmbeddingResult[]> {
|
|
135
|
+
return this.limiter.schedule(() =>
|
|
136
|
+
this.retryWithBackoff(async () => {
|
|
137
|
+
const response = await this.client.embeddings.create({
|
|
138
|
+
model: this.model,
|
|
139
|
+
input: texts,
|
|
140
|
+
dimensions: this.dimensions,
|
|
141
|
+
});
|
|
142
|
+
|
|
143
|
+
return response.data.map((item) => ({
|
|
144
|
+
embedding: item.embedding,
|
|
145
|
+
dimensions: this.dimensions,
|
|
146
|
+
}));
|
|
147
|
+
}),
|
|
148
|
+
);
|
|
149
|
+
}
|
|
150
|
+
|
|
151
|
+
getDimensions(): number {
|
|
152
|
+
return this.dimensions;
|
|
153
|
+
}
|
|
154
|
+
|
|
155
|
+
getModel(): string {
|
|
156
|
+
return this.model;
|
|
157
|
+
}
|
|
158
|
+
}
|
|
@@ -0,0 +1,520 @@
|
|
|
1
|
+
import { describe, it, expect, vi, beforeEach } from "vitest";
|
|
2
|
+
import { VoyageEmbeddings } from "./voyage.js";
|
|
3
|
+
|
|
4
|
+
// Mock fetch globally
|
|
5
|
+
global.fetch = vi.fn();
|
|
6
|
+
|
|
7
|
+
describe("VoyageEmbeddings", () => {
|
|
8
|
+
let embeddings: VoyageEmbeddings;
|
|
9
|
+
let mockFetch: any;
|
|
10
|
+
|
|
11
|
+
beforeEach(() => {
|
|
12
|
+
mockFetch = global.fetch as any;
|
|
13
|
+
mockFetch.mockReset();
|
|
14
|
+
|
|
15
|
+
embeddings = new VoyageEmbeddings("test-api-key");
|
|
16
|
+
});
|
|
17
|
+
|
|
18
|
+
describe("constructor", () => {
|
|
19
|
+
it("should use default model and dimensions", () => {
|
|
20
|
+
expect(embeddings.getModel()).toBe("voyage-2");
|
|
21
|
+
expect(embeddings.getDimensions()).toBe(1024);
|
|
22
|
+
});
|
|
23
|
+
|
|
24
|
+
it("should use custom model", () => {
|
|
25
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
26
|
+
"test-api-key",
|
|
27
|
+
"voyage-large-2",
|
|
28
|
+
);
|
|
29
|
+
expect(customEmbeddings.getModel()).toBe("voyage-large-2");
|
|
30
|
+
expect(customEmbeddings.getDimensions()).toBe(1536);
|
|
31
|
+
});
|
|
32
|
+
|
|
33
|
+
it("should use custom dimensions", () => {
|
|
34
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
35
|
+
"test-api-key",
|
|
36
|
+
"voyage-2",
|
|
37
|
+
512,
|
|
38
|
+
);
|
|
39
|
+
expect(customEmbeddings.getDimensions()).toBe(512);
|
|
40
|
+
});
|
|
41
|
+
|
|
42
|
+
it("should use default base URL", () => {
|
|
43
|
+
const defaultEmbeddings = new VoyageEmbeddings("test-api-key");
|
|
44
|
+
expect(defaultEmbeddings).toBeDefined();
|
|
45
|
+
});
|
|
46
|
+
|
|
47
|
+
it("should use custom base URL", () => {
|
|
48
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
49
|
+
"test-api-key",
|
|
50
|
+
"voyage-2",
|
|
51
|
+
undefined,
|
|
52
|
+
undefined,
|
|
53
|
+
"https://custom.voyage.com",
|
|
54
|
+
);
|
|
55
|
+
expect(customEmbeddings).toBeDefined();
|
|
56
|
+
});
|
|
57
|
+
|
|
58
|
+
it("should default to 1024 for unknown models", () => {
|
|
59
|
+
const unknownEmbeddings = new VoyageEmbeddings(
|
|
60
|
+
"test-api-key",
|
|
61
|
+
"custom-model",
|
|
62
|
+
);
|
|
63
|
+
expect(unknownEmbeddings.getDimensions()).toBe(1024);
|
|
64
|
+
});
|
|
65
|
+
|
|
66
|
+
it("should accept custom input type", () => {
|
|
67
|
+
const queryEmbeddings = new VoyageEmbeddings(
|
|
68
|
+
"test-api-key",
|
|
69
|
+
"voyage-2",
|
|
70
|
+
undefined,
|
|
71
|
+
undefined,
|
|
72
|
+
undefined,
|
|
73
|
+
"query",
|
|
74
|
+
);
|
|
75
|
+
expect(queryEmbeddings).toBeInstanceOf(VoyageEmbeddings);
|
|
76
|
+
});
|
|
77
|
+
});
|
|
78
|
+
|
|
79
|
+
describe("embed", () => {
|
|
80
|
+
it("should generate embedding for single text", async () => {
|
|
81
|
+
const mockEmbedding = Array(1024)
|
|
82
|
+
.fill(0)
|
|
83
|
+
.map((_, i) => i * 0.001);
|
|
84
|
+
mockFetch.mockResolvedValue({
|
|
85
|
+
ok: true,
|
|
86
|
+
json: async () => ({
|
|
87
|
+
data: [{ embedding: mockEmbedding }],
|
|
88
|
+
model: "voyage-2",
|
|
89
|
+
usage: { total_tokens: 10 },
|
|
90
|
+
}),
|
|
91
|
+
});
|
|
92
|
+
|
|
93
|
+
const result = await embeddings.embed("test text");
|
|
94
|
+
|
|
95
|
+
expect(result).toEqual({
|
|
96
|
+
embedding: mockEmbedding,
|
|
97
|
+
dimensions: 1024,
|
|
98
|
+
});
|
|
99
|
+
expect(mockFetch).toHaveBeenCalledWith(
|
|
100
|
+
"https://api.voyageai.com/v1/embeddings",
|
|
101
|
+
{
|
|
102
|
+
method: "POST",
|
|
103
|
+
headers: {
|
|
104
|
+
"Content-Type": "application/json",
|
|
105
|
+
Authorization: "Bearer test-api-key",
|
|
106
|
+
},
|
|
107
|
+
body: JSON.stringify({
|
|
108
|
+
input: ["test text"],
|
|
109
|
+
model: "voyage-2",
|
|
110
|
+
}),
|
|
111
|
+
},
|
|
112
|
+
);
|
|
113
|
+
});
|
|
114
|
+
|
|
115
|
+
it("should include input_type when specified", async () => {
|
|
116
|
+
const queryEmbeddings = new VoyageEmbeddings(
|
|
117
|
+
"test-api-key",
|
|
118
|
+
"voyage-2",
|
|
119
|
+
undefined,
|
|
120
|
+
undefined,
|
|
121
|
+
undefined,
|
|
122
|
+
"query",
|
|
123
|
+
);
|
|
124
|
+
|
|
125
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
126
|
+
mockFetch.mockResolvedValue({
|
|
127
|
+
ok: true,
|
|
128
|
+
json: async () => ({
|
|
129
|
+
data: [{ embedding: mockEmbedding }],
|
|
130
|
+
model: "voyage-2",
|
|
131
|
+
usage: { total_tokens: 10 },
|
|
132
|
+
}),
|
|
133
|
+
});
|
|
134
|
+
|
|
135
|
+
await queryEmbeddings.embed("test text");
|
|
136
|
+
|
|
137
|
+
const callBody = JSON.parse(mockFetch.mock.calls[0][1].body);
|
|
138
|
+
expect(callBody.input_type).toBe("query");
|
|
139
|
+
});
|
|
140
|
+
|
|
141
|
+
it("should handle long text", async () => {
|
|
142
|
+
const longText = "word ".repeat(1000);
|
|
143
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
144
|
+
mockFetch.mockResolvedValue({
|
|
145
|
+
ok: true,
|
|
146
|
+
json: async () => ({
|
|
147
|
+
data: [{ embedding: mockEmbedding }],
|
|
148
|
+
model: "voyage-2",
|
|
149
|
+
usage: { total_tokens: 1000 },
|
|
150
|
+
}),
|
|
151
|
+
});
|
|
152
|
+
|
|
153
|
+
const result = await embeddings.embed(longText);
|
|
154
|
+
|
|
155
|
+
expect(result.embedding).toEqual(mockEmbedding);
|
|
156
|
+
});
|
|
157
|
+
|
|
158
|
+
it("should use custom base URL", async () => {
|
|
159
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
160
|
+
"test-api-key",
|
|
161
|
+
"voyage-2",
|
|
162
|
+
undefined,
|
|
163
|
+
undefined,
|
|
164
|
+
"https://custom.voyage.com/v1",
|
|
165
|
+
);
|
|
166
|
+
|
|
167
|
+
const mockEmbedding = Array(1024).fill(0.1);
|
|
168
|
+
mockFetch.mockResolvedValue({
|
|
169
|
+
ok: true,
|
|
170
|
+
json: async () => ({
|
|
171
|
+
data: [{ embedding: mockEmbedding }],
|
|
172
|
+
model: "voyage-2",
|
|
173
|
+
usage: { total_tokens: 5 },
|
|
174
|
+
}),
|
|
175
|
+
});
|
|
176
|
+
|
|
177
|
+
await customEmbeddings.embed("test");
|
|
178
|
+
|
|
179
|
+
expect(mockFetch).toHaveBeenCalledWith(
|
|
180
|
+
"https://custom.voyage.com/v1/embeddings",
|
|
181
|
+
expect.any(Object),
|
|
182
|
+
);
|
|
183
|
+
});
|
|
184
|
+
|
|
185
|
+
it("should throw error if no embedding returned", async () => {
|
|
186
|
+
mockFetch.mockResolvedValue({
|
|
187
|
+
ok: true,
|
|
188
|
+
json: async () => ({
|
|
189
|
+
data: [],
|
|
190
|
+
model: "voyage-2",
|
|
191
|
+
usage: { total_tokens: 0 },
|
|
192
|
+
}),
|
|
193
|
+
});
|
|
194
|
+
|
|
195
|
+
await expect(embeddings.embed("test")).rejects.toThrow(
|
|
196
|
+
"No embedding returned from Voyage AI API",
|
|
197
|
+
);
|
|
198
|
+
});
|
|
199
|
+
|
|
200
|
+
it("should handle API errors", async () => {
|
|
201
|
+
mockFetch.mockResolvedValue({
|
|
202
|
+
ok: false,
|
|
203
|
+
status: 401,
|
|
204
|
+
text: async () => "Unauthorized: Invalid API key",
|
|
205
|
+
});
|
|
206
|
+
|
|
207
|
+
await expect(embeddings.embed("test")).rejects.toThrow();
|
|
208
|
+
});
|
|
209
|
+
|
|
210
|
+
it("should propagate network errors", async () => {
|
|
211
|
+
mockFetch.mockRejectedValue(new Error("Network Error"));
|
|
212
|
+
|
|
213
|
+
await expect(embeddings.embed("test")).rejects.toThrow("Network Error");
|
|
214
|
+
});
|
|
215
|
+
});
|
|
216
|
+
|
|
217
|
+
describe("embedBatch", () => {
|
|
218
|
+
it("should generate embeddings for multiple texts", async () => {
|
|
219
|
+
const mockEmbeddings = [
|
|
220
|
+
Array(1024).fill(0.1),
|
|
221
|
+
Array(1024).fill(0.2),
|
|
222
|
+
Array(1024).fill(0.3),
|
|
223
|
+
];
|
|
224
|
+
mockFetch.mockResolvedValue({
|
|
225
|
+
ok: true,
|
|
226
|
+
json: async () => ({
|
|
227
|
+
data: mockEmbeddings.map((embedding) => ({ embedding })),
|
|
228
|
+
model: "voyage-2",
|
|
229
|
+
usage: { total_tokens: 30 },
|
|
230
|
+
}),
|
|
231
|
+
});
|
|
232
|
+
|
|
233
|
+
const texts = ["text1", "text2", "text3"];
|
|
234
|
+
const results = await embeddings.embedBatch(texts);
|
|
235
|
+
|
|
236
|
+
expect(results).toEqual([
|
|
237
|
+
{ embedding: mockEmbeddings[0], dimensions: 1024 },
|
|
238
|
+
{ embedding: mockEmbeddings[1], dimensions: 1024 },
|
|
239
|
+
{ embedding: mockEmbeddings[2], dimensions: 1024 },
|
|
240
|
+
]);
|
|
241
|
+
|
|
242
|
+
const callBody = JSON.parse(mockFetch.mock.calls[0][1].body);
|
|
243
|
+
expect(callBody.input).toEqual(texts);
|
|
244
|
+
});
|
|
245
|
+
|
|
246
|
+
it("should handle empty batch", async () => {
|
|
247
|
+
mockFetch.mockResolvedValue({
|
|
248
|
+
ok: true,
|
|
249
|
+
json: async () => ({
|
|
250
|
+
data: [],
|
|
251
|
+
model: "voyage-2",
|
|
252
|
+
usage: { total_tokens: 0 },
|
|
253
|
+
}),
|
|
254
|
+
});
|
|
255
|
+
|
|
256
|
+
const results = await embeddings.embedBatch([]);
|
|
257
|
+
|
|
258
|
+
expect(results).toEqual([]);
|
|
259
|
+
});
|
|
260
|
+
|
|
261
|
+
it("should handle single item in batch", async () => {
|
|
262
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
263
|
+
mockFetch.mockResolvedValue({
|
|
264
|
+
ok: true,
|
|
265
|
+
json: async () => ({
|
|
266
|
+
data: [{ embedding: mockEmbedding }],
|
|
267
|
+
model: "voyage-2",
|
|
268
|
+
usage: { total_tokens: 10 },
|
|
269
|
+
}),
|
|
270
|
+
});
|
|
271
|
+
|
|
272
|
+
const results = await embeddings.embedBatch(["single text"]);
|
|
273
|
+
|
|
274
|
+
expect(results).toHaveLength(1);
|
|
275
|
+
expect(results[0].embedding).toEqual(mockEmbedding);
|
|
276
|
+
});
|
|
277
|
+
|
|
278
|
+
it("should handle large batches", async () => {
|
|
279
|
+
const batchSize = 100;
|
|
280
|
+
const mockEmbeddings = Array(batchSize)
|
|
281
|
+
.fill(null)
|
|
282
|
+
.map(() => Array(1024).fill(Math.random()));
|
|
283
|
+
|
|
284
|
+
mockFetch.mockResolvedValue({
|
|
285
|
+
ok: true,
|
|
286
|
+
json: async () => ({
|
|
287
|
+
data: mockEmbeddings.map((embedding) => ({ embedding })),
|
|
288
|
+
model: "voyage-2",
|
|
289
|
+
usage: { total_tokens: batchSize * 10 },
|
|
290
|
+
}),
|
|
291
|
+
});
|
|
292
|
+
|
|
293
|
+
const texts = Array(batchSize)
|
|
294
|
+
.fill(null)
|
|
295
|
+
.map((_, i) => `text ${i}`);
|
|
296
|
+
const results = await embeddings.embedBatch(texts);
|
|
297
|
+
|
|
298
|
+
expect(results).toHaveLength(batchSize);
|
|
299
|
+
});
|
|
300
|
+
|
|
301
|
+
it("should throw error if no embeddings returned", async () => {
|
|
302
|
+
mockFetch.mockResolvedValue({
|
|
303
|
+
ok: true,
|
|
304
|
+
json: async () => ({
|
|
305
|
+
model: "voyage-2",
|
|
306
|
+
usage: { total_tokens: 0 },
|
|
307
|
+
}),
|
|
308
|
+
});
|
|
309
|
+
|
|
310
|
+
await expect(embeddings.embedBatch(["text1"])).rejects.toThrow(
|
|
311
|
+
"No embeddings returned from Voyage AI API",
|
|
312
|
+
);
|
|
313
|
+
});
|
|
314
|
+
|
|
315
|
+
it("should propagate errors in batch", async () => {
|
|
316
|
+
mockFetch.mockRejectedValue(new Error("Batch API Error"));
|
|
317
|
+
|
|
318
|
+
await expect(embeddings.embedBatch(["text1", "text2"])).rejects.toThrow(
|
|
319
|
+
"Batch API Error",
|
|
320
|
+
);
|
|
321
|
+
});
|
|
322
|
+
});
|
|
323
|
+
|
|
324
|
+
describe("getDimensions", () => {
|
|
325
|
+
it("should return configured dimensions", () => {
|
|
326
|
+
expect(embeddings.getDimensions()).toBe(1024);
|
|
327
|
+
});
|
|
328
|
+
|
|
329
|
+
it("should return custom dimensions", () => {
|
|
330
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
331
|
+
"test-api-key",
|
|
332
|
+
"voyage-2",
|
|
333
|
+
512,
|
|
334
|
+
);
|
|
335
|
+
expect(customEmbeddings.getDimensions()).toBe(512);
|
|
336
|
+
});
|
|
337
|
+
});
|
|
338
|
+
|
|
339
|
+
describe("getModel", () => {
|
|
340
|
+
it("should return configured model", () => {
|
|
341
|
+
expect(embeddings.getModel()).toBe("voyage-2");
|
|
342
|
+
});
|
|
343
|
+
|
|
344
|
+
it("should return custom model", () => {
|
|
345
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
346
|
+
"test-api-key",
|
|
347
|
+
"voyage-large-2",
|
|
348
|
+
);
|
|
349
|
+
expect(customEmbeddings.getModel()).toBe("voyage-large-2");
|
|
350
|
+
});
|
|
351
|
+
});
|
|
352
|
+
|
|
353
|
+
describe("rate limiting", () => {
|
|
354
|
+
it("should retry on rate limit error (429 status)", async () => {
|
|
355
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
356
|
+
|
|
357
|
+
mockFetch
|
|
358
|
+
.mockResolvedValueOnce({
|
|
359
|
+
ok: false,
|
|
360
|
+
status: 429,
|
|
361
|
+
text: async () => "Rate limit exceeded",
|
|
362
|
+
})
|
|
363
|
+
.mockResolvedValueOnce({
|
|
364
|
+
ok: false,
|
|
365
|
+
status: 429,
|
|
366
|
+
text: async () => "Rate limit exceeded",
|
|
367
|
+
})
|
|
368
|
+
.mockResolvedValue({
|
|
369
|
+
ok: true,
|
|
370
|
+
json: async () => ({
|
|
371
|
+
data: [{ embedding: mockEmbedding }],
|
|
372
|
+
model: "voyage-2",
|
|
373
|
+
usage: { total_tokens: 10 },
|
|
374
|
+
}),
|
|
375
|
+
});
|
|
376
|
+
|
|
377
|
+
const result = await embeddings.embed("test text");
|
|
378
|
+
|
|
379
|
+
expect(result.embedding).toEqual(mockEmbedding);
|
|
380
|
+
expect(mockFetch).toHaveBeenCalledTimes(3);
|
|
381
|
+
});
|
|
382
|
+
|
|
383
|
+
it("should retry on rate limit message", async () => {
|
|
384
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
385
|
+
|
|
386
|
+
mockFetch
|
|
387
|
+
.mockRejectedValueOnce({
|
|
388
|
+
message: "You have exceeded the rate limit",
|
|
389
|
+
})
|
|
390
|
+
.mockResolvedValue({
|
|
391
|
+
ok: true,
|
|
392
|
+
json: async () => ({
|
|
393
|
+
data: [{ embedding: mockEmbedding }],
|
|
394
|
+
model: "voyage-2",
|
|
395
|
+
usage: { total_tokens: 10 },
|
|
396
|
+
}),
|
|
397
|
+
});
|
|
398
|
+
|
|
399
|
+
const result = await embeddings.embed("test text");
|
|
400
|
+
|
|
401
|
+
expect(result.embedding).toEqual(mockEmbedding);
|
|
402
|
+
expect(mockFetch).toHaveBeenCalledTimes(2);
|
|
403
|
+
});
|
|
404
|
+
|
|
405
|
+
it("should use exponential backoff", async () => {
|
|
406
|
+
const rateLimitEmbeddings = new VoyageEmbeddings(
|
|
407
|
+
"test-api-key",
|
|
408
|
+
"voyage-2",
|
|
409
|
+
undefined,
|
|
410
|
+
{
|
|
411
|
+
retryAttempts: 3,
|
|
412
|
+
retryDelayMs: 100,
|
|
413
|
+
},
|
|
414
|
+
);
|
|
415
|
+
|
|
416
|
+
const mockEmbedding = Array(1024).fill(0.5);
|
|
417
|
+
|
|
418
|
+
mockFetch
|
|
419
|
+
.mockResolvedValueOnce({
|
|
420
|
+
ok: false,
|
|
421
|
+
status: 429,
|
|
422
|
+
text: async () => "Rate limit",
|
|
423
|
+
})
|
|
424
|
+
.mockResolvedValueOnce({
|
|
425
|
+
ok: false,
|
|
426
|
+
status: 429,
|
|
427
|
+
text: async () => "Rate limit",
|
|
428
|
+
})
|
|
429
|
+
.mockResolvedValue({
|
|
430
|
+
ok: true,
|
|
431
|
+
json: async () => ({
|
|
432
|
+
data: [{ embedding: mockEmbedding }],
|
|
433
|
+
model: "voyage-2",
|
|
434
|
+
usage: { total_tokens: 10 },
|
|
435
|
+
}),
|
|
436
|
+
});
|
|
437
|
+
|
|
438
|
+
const startTime = Date.now();
|
|
439
|
+
await rateLimitEmbeddings.embed("test text");
|
|
440
|
+
const duration = Date.now() - startTime;
|
|
441
|
+
|
|
442
|
+
// Should wait: 100ms (first retry) + 200ms (second retry) = 300ms
|
|
443
|
+
expect(duration).toBeGreaterThanOrEqual(250);
|
|
444
|
+
});
|
|
445
|
+
|
|
446
|
+
it("should throw error after max retries exceeded", async () => {
|
|
447
|
+
const rateLimitEmbeddings = new VoyageEmbeddings(
|
|
448
|
+
"test-api-key",
|
|
449
|
+
"voyage-2",
|
|
450
|
+
undefined,
|
|
451
|
+
{
|
|
452
|
+
retryAttempts: 2,
|
|
453
|
+
retryDelayMs: 100,
|
|
454
|
+
},
|
|
455
|
+
);
|
|
456
|
+
|
|
457
|
+
mockFetch.mockResolvedValue({
|
|
458
|
+
ok: false,
|
|
459
|
+
status: 429,
|
|
460
|
+
text: async () => "Rate limit exceeded",
|
|
461
|
+
});
|
|
462
|
+
|
|
463
|
+
await expect(rateLimitEmbeddings.embed("test text")).rejects.toThrow(
|
|
464
|
+
"Voyage AI API rate limit exceeded after 2 retry attempts",
|
|
465
|
+
);
|
|
466
|
+
|
|
467
|
+
expect(mockFetch).toHaveBeenCalledTimes(3);
|
|
468
|
+
});
|
|
469
|
+
|
|
470
|
+
it("should handle rate limit errors in batch operations", async () => {
|
|
471
|
+
const mockEmbeddings = [Array(1024).fill(0.1), Array(1024).fill(0.2)];
|
|
472
|
+
|
|
473
|
+
mockFetch
|
|
474
|
+
.mockResolvedValueOnce({
|
|
475
|
+
ok: false,
|
|
476
|
+
status: 429,
|
|
477
|
+
text: async () => "Rate limit",
|
|
478
|
+
})
|
|
479
|
+
.mockResolvedValue({
|
|
480
|
+
ok: true,
|
|
481
|
+
json: async () => ({
|
|
482
|
+
data: mockEmbeddings.map((embedding) => ({ embedding })),
|
|
483
|
+
model: "voyage-2",
|
|
484
|
+
usage: { total_tokens: 20 },
|
|
485
|
+
}),
|
|
486
|
+
});
|
|
487
|
+
|
|
488
|
+
const results = await embeddings.embedBatch(["text1", "text2"]);
|
|
489
|
+
|
|
490
|
+
expect(results).toHaveLength(2);
|
|
491
|
+
expect(mockFetch).toHaveBeenCalledTimes(2);
|
|
492
|
+
});
|
|
493
|
+
|
|
494
|
+
it("should not retry on non-rate-limit errors", async () => {
|
|
495
|
+
mockFetch.mockResolvedValue({
|
|
496
|
+
ok: false,
|
|
497
|
+
status: 401,
|
|
498
|
+
text: async () => "Unauthorized",
|
|
499
|
+
});
|
|
500
|
+
|
|
501
|
+
await expect(embeddings.embed("test text")).rejects.toThrow();
|
|
502
|
+
expect(mockFetch).toHaveBeenCalledTimes(1);
|
|
503
|
+
});
|
|
504
|
+
|
|
505
|
+
it("should accept custom rate limit configuration", () => {
|
|
506
|
+
const customEmbeddings = new VoyageEmbeddings(
|
|
507
|
+
"test-api-key",
|
|
508
|
+
"voyage-2",
|
|
509
|
+
undefined,
|
|
510
|
+
{
|
|
511
|
+
maxRequestsPerMinute: 500,
|
|
512
|
+
retryAttempts: 5,
|
|
513
|
+
retryDelayMs: 2000,
|
|
514
|
+
},
|
|
515
|
+
);
|
|
516
|
+
|
|
517
|
+
expect(customEmbeddings).toBeDefined();
|
|
518
|
+
});
|
|
519
|
+
});
|
|
520
|
+
});
|