@copilotkit/aimock 1.7.0 → 1.9.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude-plugin/marketplace.json +1 -1
- package/.claude-plugin/plugin.json +1 -1
- package/README.md +3 -1
- package/dist/bedrock-converse.cjs +13 -8
- package/dist/bedrock-converse.cjs.map +1 -1
- package/dist/bedrock-converse.d.cts.map +1 -1
- package/dist/bedrock-converse.d.ts.map +1 -1
- package/dist/bedrock-converse.js +14 -9
- package/dist/bedrock-converse.js.map +1 -1
- package/dist/bedrock.cjs +51 -15
- package/dist/bedrock.cjs.map +1 -1
- package/dist/bedrock.d.cts.map +1 -1
- package/dist/bedrock.d.ts.map +1 -1
- package/dist/bedrock.js +52 -16
- package/dist/bedrock.js.map +1 -1
- package/dist/cohere.cjs +3 -2
- package/dist/cohere.cjs.map +1 -1
- package/dist/cohere.d.cts.map +1 -1
- package/dist/cohere.d.ts.map +1 -1
- package/dist/cohere.js +4 -3
- package/dist/cohere.js.map +1 -1
- package/dist/embeddings.cjs +3 -2
- package/dist/embeddings.cjs.map +1 -1
- package/dist/embeddings.d.cts.map +1 -1
- package/dist/embeddings.d.ts.map +1 -1
- package/dist/embeddings.js +4 -3
- package/dist/embeddings.js.map +1 -1
- package/dist/gemini.cjs +129 -35
- package/dist/gemini.cjs.map +1 -1
- package/dist/gemini.d.cts.map +1 -1
- package/dist/gemini.d.ts.map +1 -1
- package/dist/gemini.js +130 -36
- package/dist/gemini.js.map +1 -1
- package/dist/helpers.cjs +154 -3
- package/dist/helpers.cjs.map +1 -1
- package/dist/helpers.d.cts +1 -1
- package/dist/helpers.d.cts.map +1 -1
- package/dist/helpers.d.ts +1 -1
- package/dist/helpers.d.ts.map +1 -1
- package/dist/helpers.js +151 -4
- package/dist/helpers.js.map +1 -1
- package/dist/index.cjs +1 -0
- package/dist/index.d.cts +2 -2
- package/dist/index.d.ts +2 -2
- package/dist/index.js +2 -2
- package/dist/journal.cjs +26 -9
- package/dist/journal.cjs.map +1 -1
- package/dist/journal.d.cts +10 -5
- package/dist/journal.d.cts.map +1 -1
- package/dist/journal.d.ts +10 -5
- package/dist/journal.d.ts.map +1 -1
- package/dist/journal.js +26 -10
- package/dist/journal.js.map +1 -1
- package/dist/llmock.cjs +2 -2
- package/dist/llmock.cjs.map +1 -1
- package/dist/llmock.d.cts +1 -1
- package/dist/llmock.d.ts +1 -1
- package/dist/llmock.js +2 -2
- package/dist/llmock.js.map +1 -1
- package/dist/messages.cjs +192 -2
- package/dist/messages.cjs.map +1 -1
- package/dist/messages.d.cts.map +1 -1
- package/dist/messages.d.ts.map +1 -1
- package/dist/messages.js +193 -3
- package/dist/messages.js.map +1 -1
- package/dist/ollama.cjs +39 -13
- package/dist/ollama.cjs.map +1 -1
- package/dist/ollama.d.cts.map +1 -1
- package/dist/ollama.d.ts.map +1 -1
- package/dist/ollama.js +40 -14
- package/dist/ollama.js.map +1 -1
- package/dist/recorder.cjs +1 -1
- package/dist/recorder.cjs.map +1 -1
- package/dist/recorder.d.cts +1 -0
- package/dist/recorder.d.cts.map +1 -1
- package/dist/recorder.d.ts +1 -0
- package/dist/recorder.d.ts.map +1 -1
- package/dist/recorder.js +1 -1
- package/dist/recorder.js.map +1 -1
- package/dist/responses.cjs +250 -126
- package/dist/responses.cjs.map +1 -1
- package/dist/responses.d.cts.map +1 -1
- package/dist/responses.d.ts.map +1 -1
- package/dist/responses.js +251 -127
- package/dist/responses.js.map +1 -1
- package/dist/router.cjs +16 -10
- package/dist/router.cjs.map +1 -1
- package/dist/router.d.cts +1 -1
- package/dist/router.d.cts.map +1 -1
- package/dist/router.d.ts +1 -1
- package/dist/router.d.ts.map +1 -1
- package/dist/router.js +16 -10
- package/dist/router.js.map +1 -1
- package/dist/server.cjs +47 -7
- package/dist/server.cjs.map +1 -1
- package/dist/server.d.cts.map +1 -1
- package/dist/server.d.ts.map +1 -1
- package/dist/server.js +48 -8
- package/dist/server.js.map +1 -1
- package/dist/stream-collapse.cjs +48 -40
- package/dist/stream-collapse.cjs.map +1 -1
- package/dist/stream-collapse.d.cts.map +1 -1
- package/dist/stream-collapse.d.ts.map +1 -1
- package/dist/stream-collapse.js +48 -40
- package/dist/stream-collapse.js.map +1 -1
- package/dist/types.d.cts +22 -1
- package/dist/types.d.cts.map +1 -1
- package/dist/types.d.ts +22 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/ws-gemini-live.cjs +4 -2
- package/dist/ws-gemini-live.cjs.map +1 -1
- package/dist/ws-gemini-live.d.cts +3 -1
- package/dist/ws-gemini-live.d.cts.map +1 -1
- package/dist/ws-gemini-live.d.ts +3 -1
- package/dist/ws-gemini-live.d.ts.map +1 -1
- package/dist/ws-gemini-live.js +4 -2
- package/dist/ws-gemini-live.js.map +1 -1
- package/dist/ws-realtime.cjs +4 -2
- package/dist/ws-realtime.cjs.map +1 -1
- package/dist/ws-realtime.d.cts +3 -1
- package/dist/ws-realtime.d.cts.map +1 -1
- package/dist/ws-realtime.d.ts +3 -1
- package/dist/ws-realtime.d.ts.map +1 -1
- package/dist/ws-realtime.js +4 -2
- package/dist/ws-realtime.js.map +1 -1
- package/dist/ws-responses.cjs +4 -2
- package/dist/ws-responses.cjs.map +1 -1
- package/dist/ws-responses.d.cts +3 -1
- package/dist/ws-responses.d.cts.map +1 -1
- package/dist/ws-responses.d.ts +3 -1
- package/dist/ws-responses.d.ts.map +1 -1
- package/dist/ws-responses.js +4 -2
- package/dist/ws-responses.js.map +1 -1
- package/package.json +5 -1
package/dist/gemini.cjs
CHANGED
|
@@ -84,8 +84,21 @@ function geminiToCompletionRequest(req, model, stream) {
|
|
|
84
84
|
tools
|
|
85
85
|
};
|
|
86
86
|
}
|
|
87
|
-
function buildGeminiTextStreamChunks(content, chunkSize) {
|
|
87
|
+
function buildGeminiTextStreamChunks(content, chunkSize, reasoning) {
|
|
88
88
|
const chunks = [];
|
|
89
|
+
if (reasoning) for (let i = 0; i < reasoning.length; i += chunkSize) {
|
|
90
|
+
const slice = reasoning.slice(i, i + chunkSize);
|
|
91
|
+
chunks.push({ candidates: [{
|
|
92
|
+
content: {
|
|
93
|
+
role: "model",
|
|
94
|
+
parts: [{
|
|
95
|
+
text: slice,
|
|
96
|
+
thought: true
|
|
97
|
+
}]
|
|
98
|
+
},
|
|
99
|
+
index: 0
|
|
100
|
+
}] });
|
|
101
|
+
}
|
|
89
102
|
for (let i = 0; i < content.length; i += chunkSize) {
|
|
90
103
|
const slice = content.slice(i, i + chunkSize);
|
|
91
104
|
const isLast = i + chunkSize >= content.length;
|
|
@@ -123,25 +136,26 @@ function buildGeminiTextStreamChunks(content, chunkSize) {
|
|
|
123
136
|
});
|
|
124
137
|
return chunks;
|
|
125
138
|
}
|
|
139
|
+
function parseToolCallPart(tc, logger) {
|
|
140
|
+
let argsObj;
|
|
141
|
+
try {
|
|
142
|
+
argsObj = JSON.parse(tc.arguments || "{}");
|
|
143
|
+
} catch {
|
|
144
|
+
logger.warn(`Malformed JSON in fixture tool call arguments for "${tc.name}": ${tc.arguments}`);
|
|
145
|
+
argsObj = {};
|
|
146
|
+
}
|
|
147
|
+
return { functionCall: {
|
|
148
|
+
name: tc.name,
|
|
149
|
+
args: argsObj,
|
|
150
|
+
id: tc.id || require_helpers.generateToolCallId()
|
|
151
|
+
} };
|
|
152
|
+
}
|
|
126
153
|
function buildGeminiToolCallStreamChunks(toolCalls, logger) {
|
|
127
154
|
return [{
|
|
128
155
|
candidates: [{
|
|
129
156
|
content: {
|
|
130
157
|
role: "model",
|
|
131
|
-
parts: toolCalls.map((tc) =>
|
|
132
|
-
let argsObj;
|
|
133
|
-
try {
|
|
134
|
-
argsObj = JSON.parse(tc.arguments || "{}");
|
|
135
|
-
} catch {
|
|
136
|
-
logger.warn(`Malformed JSON in fixture tool call arguments for "${tc.name}": ${tc.arguments}`);
|
|
137
|
-
argsObj = {};
|
|
138
|
-
}
|
|
139
|
-
return { functionCall: {
|
|
140
|
-
name: tc.name,
|
|
141
|
-
args: argsObj,
|
|
142
|
-
id: tc.id || require_helpers.generateToolCallId()
|
|
143
|
-
} };
|
|
144
|
-
})
|
|
158
|
+
parts: toolCalls.map((tc) => parseToolCallPart(tc, logger))
|
|
145
159
|
},
|
|
146
160
|
finishReason: "FUNCTION_CALL",
|
|
147
161
|
index: 0
|
|
@@ -153,12 +167,18 @@ function buildGeminiToolCallStreamChunks(toolCalls, logger) {
|
|
|
153
167
|
}
|
|
154
168
|
}];
|
|
155
169
|
}
|
|
156
|
-
function buildGeminiTextResponse(content) {
|
|
170
|
+
function buildGeminiTextResponse(content, reasoning) {
|
|
171
|
+
const parts = [];
|
|
172
|
+
if (reasoning) parts.push({
|
|
173
|
+
text: reasoning,
|
|
174
|
+
thought: true
|
|
175
|
+
});
|
|
176
|
+
parts.push({ text: content });
|
|
157
177
|
return {
|
|
158
178
|
candidates: [{
|
|
159
179
|
content: {
|
|
160
180
|
role: "model",
|
|
161
|
-
parts
|
|
181
|
+
parts
|
|
162
182
|
},
|
|
163
183
|
finishReason: "STOP",
|
|
164
184
|
index: 0
|
|
@@ -175,20 +195,61 @@ function buildGeminiToolCallResponse(toolCalls, logger) {
|
|
|
175
195
|
candidates: [{
|
|
176
196
|
content: {
|
|
177
197
|
role: "model",
|
|
178
|
-
parts: toolCalls.map((tc) =>
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
|
|
198
|
+
parts: toolCalls.map((tc) => parseToolCallPart(tc, logger))
|
|
199
|
+
},
|
|
200
|
+
finishReason: "FUNCTION_CALL",
|
|
201
|
+
index: 0
|
|
202
|
+
}],
|
|
203
|
+
usageMetadata: {
|
|
204
|
+
promptTokenCount: 0,
|
|
205
|
+
candidatesTokenCount: 0,
|
|
206
|
+
totalTokenCount: 0
|
|
207
|
+
}
|
|
208
|
+
};
|
|
209
|
+
}
|
|
210
|
+
function buildGeminiContentWithToolCallsStreamChunks(content, toolCalls, chunkSize, logger) {
|
|
211
|
+
const chunks = [];
|
|
212
|
+
if (content.length === 0) chunks.push({ candidates: [{
|
|
213
|
+
content: {
|
|
214
|
+
role: "model",
|
|
215
|
+
parts: [{ text: "" }]
|
|
216
|
+
},
|
|
217
|
+
index: 0
|
|
218
|
+
}] });
|
|
219
|
+
else for (let i = 0; i < content.length; i += chunkSize) {
|
|
220
|
+
const slice = content.slice(i, i + chunkSize);
|
|
221
|
+
chunks.push({ candidates: [{
|
|
222
|
+
content: {
|
|
223
|
+
role: "model",
|
|
224
|
+
parts: [{ text: slice }]
|
|
225
|
+
},
|
|
226
|
+
index: 0
|
|
227
|
+
}] });
|
|
228
|
+
}
|
|
229
|
+
const parts = toolCalls.map((tc) => parseToolCallPart(tc, logger));
|
|
230
|
+
chunks.push({
|
|
231
|
+
candidates: [{
|
|
232
|
+
content: {
|
|
233
|
+
role: "model",
|
|
234
|
+
parts
|
|
235
|
+
},
|
|
236
|
+
finishReason: "FUNCTION_CALL",
|
|
237
|
+
index: 0
|
|
238
|
+
}],
|
|
239
|
+
usageMetadata: {
|
|
240
|
+
promptTokenCount: 0,
|
|
241
|
+
candidatesTokenCount: 0,
|
|
242
|
+
totalTokenCount: 0
|
|
243
|
+
}
|
|
244
|
+
});
|
|
245
|
+
return chunks;
|
|
246
|
+
}
|
|
247
|
+
function buildGeminiContentWithToolCallsResponse(content, toolCalls, logger) {
|
|
248
|
+
return {
|
|
249
|
+
candidates: [{
|
|
250
|
+
content: {
|
|
251
|
+
role: "model",
|
|
252
|
+
parts: [{ text: content }, ...toolCalls.map((tc) => parseToolCallPart(tc, logger))]
|
|
192
253
|
},
|
|
193
254
|
finishReason: "FUNCTION_CALL",
|
|
194
255
|
index: 0
|
|
@@ -249,9 +310,10 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
249
310
|
return;
|
|
250
311
|
}
|
|
251
312
|
const completionReq = geminiToCompletionRequest(geminiReq, model, streaming);
|
|
252
|
-
const
|
|
313
|
+
const testId = require_helpers.getTestId(req);
|
|
314
|
+
const fixture = require_router.matchFixture(fixtures, completionReq, journal.getFixtureMatchCountsForTest(testId), defaults.requestTransform);
|
|
253
315
|
const path = req.url ?? `/v1beta/models/${model}:generateContent`;
|
|
254
|
-
if (fixture) journal.incrementFixtureMatchCount(fixture, fixtures);
|
|
316
|
+
if (fixture) journal.incrementFixtureMatchCount(fixture, fixtures, testId);
|
|
255
317
|
if (require_chaos.applyChaos(res, fixture, defaults.chaos, req.headers, journal, {
|
|
256
318
|
method: req.method ?? "POST",
|
|
257
319
|
path,
|
|
@@ -317,6 +379,38 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
317
379
|
require_sse_writer.writeErrorResponse(res, status, JSON.stringify(geminiError));
|
|
318
380
|
return;
|
|
319
381
|
}
|
|
382
|
+
if (require_helpers.isContentWithToolCallsResponse(response)) {
|
|
383
|
+
const journalEntry = journal.add({
|
|
384
|
+
method: req.method ?? "POST",
|
|
385
|
+
path,
|
|
386
|
+
headers: require_helpers.flattenHeaders(req.headers),
|
|
387
|
+
body: completionReq,
|
|
388
|
+
response: {
|
|
389
|
+
status: 200,
|
|
390
|
+
fixture
|
|
391
|
+
}
|
|
392
|
+
});
|
|
393
|
+
if (!streaming) {
|
|
394
|
+
const body = buildGeminiContentWithToolCallsResponse(response.content, response.toolCalls, logger);
|
|
395
|
+
res.writeHead(200, { "Content-Type": "application/json" });
|
|
396
|
+
res.end(JSON.stringify(body));
|
|
397
|
+
} else {
|
|
398
|
+
const chunks = buildGeminiContentWithToolCallsStreamChunks(response.content, response.toolCalls, chunkSize, logger);
|
|
399
|
+
const interruption = require_interruption.createInterruptionSignal(fixture);
|
|
400
|
+
if (!await writeGeminiSSEStream(res, chunks, {
|
|
401
|
+
latency,
|
|
402
|
+
streamingProfile: fixture.streamingProfile,
|
|
403
|
+
signal: interruption?.signal,
|
|
404
|
+
onChunkSent: interruption?.tick
|
|
405
|
+
})) {
|
|
406
|
+
if (!res.writableEnded) res.destroy();
|
|
407
|
+
journalEntry.response.interrupted = true;
|
|
408
|
+
journalEntry.response.interruptReason = interruption?.reason();
|
|
409
|
+
}
|
|
410
|
+
interruption?.cleanup();
|
|
411
|
+
}
|
|
412
|
+
return;
|
|
413
|
+
}
|
|
320
414
|
if (require_helpers.isTextResponse(response)) {
|
|
321
415
|
const journalEntry = journal.add({
|
|
322
416
|
method: req.method ?? "POST",
|
|
@@ -329,11 +423,11 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
329
423
|
}
|
|
330
424
|
});
|
|
331
425
|
if (!streaming) {
|
|
332
|
-
const body = buildGeminiTextResponse(response.content);
|
|
426
|
+
const body = buildGeminiTextResponse(response.content, response.reasoning);
|
|
333
427
|
res.writeHead(200, { "Content-Type": "application/json" });
|
|
334
428
|
res.end(JSON.stringify(body));
|
|
335
429
|
} else {
|
|
336
|
-
const chunks = buildGeminiTextStreamChunks(response.content, chunkSize);
|
|
430
|
+
const chunks = buildGeminiTextStreamChunks(response.content, chunkSize, response.reasoning);
|
|
337
431
|
const interruption = require_interruption.createInterruptionSignal(fixture);
|
|
338
432
|
if (!await writeGeminiSSEStream(res, chunks, {
|
|
339
433
|
latency,
|
package/dist/gemini.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"gemini.cjs","names":["generateToolCallId","calculateDelay","delay","flattenHeaders","matchFixture","applyChaos","proxyAndRecord","isErrorResponse","isTextResponse","createInterruptionSignal","isToolCallResponse"],"sources":["../src/gemini.ts"],"sourcesContent":["/**\n * Google Gemini GenerateContent API support.\n *\n * Translates incoming Gemini requests into the ChatCompletionRequest format\n * used by the fixture router, and converts fixture responses back into the\n * Gemini GenerateContent streaming (or non-streaming) format.\n */\n\nimport type * as http from \"node:http\";\nimport type {\n ChatCompletionRequest,\n ChatMessage,\n Fixture,\n HandlerDefaults,\n RecordProviderKey,\n StreamingProfile,\n ToolCall,\n ToolDefinition,\n} from \"./types.js\";\nimport {\n isTextResponse,\n isToolCallResponse,\n isErrorResponse,\n generateToolCallId,\n flattenHeaders,\n} from \"./helpers.js\";\nimport { matchFixture } from \"./router.js\";\nimport { writeErrorResponse, delay, calculateDelay } from \"./sse-writer.js\";\nimport { createInterruptionSignal } from \"./interruption.js\";\nimport type { Journal } from \"./journal.js\";\nimport type { Logger } from \"./logger.js\";\nimport { applyChaos } from \"./chaos.js\";\nimport { proxyAndRecord } from \"./recorder.js\";\n\n// ─── Gemini request types ───────────────────────────────────────────────────\n\ninterface GeminiPart {\n text?: string;\n functionCall?: { name: string; args: Record<string, unknown>; id?: string };\n functionResponse?: { name: string; response: unknown };\n}\n\ninterface GeminiContent {\n role?: string;\n parts: GeminiPart[];\n}\n\ninterface GeminiFunctionDeclaration {\n name: string;\n description?: string;\n parameters?: object;\n}\n\ninterface GeminiToolDef {\n functionDeclarations?: GeminiFunctionDeclaration[];\n}\n\ninterface GeminiRequest {\n contents?: GeminiContent[];\n systemInstruction?: GeminiContent;\n tools?: GeminiToolDef[];\n generationConfig?: {\n temperature?: number;\n maxOutputTokens?: number;\n [key: string]: unknown;\n };\n [key: string]: unknown;\n}\n\n// ─── Input conversion: Gemini → ChatCompletions messages ────────────────────\n\nexport function geminiToCompletionRequest(\n req: GeminiRequest,\n model: string,\n stream: boolean,\n): ChatCompletionRequest {\n const messages: ChatMessage[] = [];\n\n // systemInstruction → system message\n if (req.systemInstruction) {\n const text = req.systemInstruction.parts\n .filter((p) => p.text !== undefined)\n .map((p) => p.text!)\n .join(\"\");\n if (text) {\n messages.push({ role: \"system\", content: text });\n }\n }\n\n if (req.contents) {\n for (const content of req.contents) {\n const role = content.role ?? \"user\";\n\n if (role === \"user\") {\n // Check for functionResponse parts\n const funcResponses = content.parts.filter((p) => p.functionResponse);\n const textParts = content.parts.filter((p) => p.text !== undefined);\n\n if (funcResponses.length > 0) {\n // functionResponse → tool message\n for (let i = 0; i < funcResponses.length; i++) {\n const part = funcResponses[i];\n messages.push({\n role: \"tool\",\n content:\n typeof part.functionResponse!.response === \"string\"\n ? part.functionResponse!.response\n : JSON.stringify(part.functionResponse!.response),\n tool_call_id: `call_gemini_${part.functionResponse!.name}_${i}`,\n });\n }\n // Any text parts alongside → user message\n if (textParts.length > 0) {\n messages.push({\n role: \"user\",\n content: textParts.map((p) => p.text!).join(\"\"),\n });\n }\n } else {\n // Regular user text\n const text = textParts.map((p) => p.text!).join(\"\");\n messages.push({ role: \"user\", content: text });\n }\n } else if (role === \"model\") {\n // Check for functionCall parts\n const funcCalls = content.parts.filter((p) => p.functionCall);\n const textParts = content.parts.filter((p) => p.text !== undefined);\n\n if (funcCalls.length > 0) {\n messages.push({\n role: \"assistant\",\n content: null,\n tool_calls: funcCalls.map((p, i) => ({\n id: `call_gemini_${p.functionCall!.name}_${i}`,\n type: \"function\" as const,\n function: {\n name: p.functionCall!.name,\n arguments: JSON.stringify(p.functionCall!.args),\n },\n })),\n });\n } else {\n const text = textParts.map((p) => p.text!).join(\"\");\n messages.push({ role: \"assistant\", content: text });\n }\n }\n }\n }\n\n // Convert tools\n let tools: ToolDefinition[] | undefined;\n if (req.tools && req.tools.length > 0) {\n const decls = req.tools.flatMap((t) => t.functionDeclarations ?? []);\n if (decls.length > 0) {\n tools = decls.map((d) => ({\n type: \"function\" as const,\n function: {\n name: d.name,\n description: d.description,\n parameters: d.parameters,\n },\n }));\n }\n }\n\n return {\n model,\n messages,\n stream,\n temperature: req.generationConfig?.temperature,\n tools,\n };\n}\n\n// ─── Response building: fixture → Gemini format ─────────────────────────────\n\ninterface GeminiResponseChunk {\n candidates: {\n content: { role: string; parts: GeminiPart[] };\n finishReason?: string;\n index: number;\n }[];\n usageMetadata?: {\n promptTokenCount: number;\n candidatesTokenCount: number;\n totalTokenCount: number;\n };\n}\n\nfunction buildGeminiTextStreamChunks(content: string, chunkSize: number): GeminiResponseChunk[] {\n const chunks: GeminiResponseChunk[] = [];\n\n // Content chunks\n for (let i = 0; i < content.length; i += chunkSize) {\n const slice = content.slice(i, i + chunkSize);\n const isLast = i + chunkSize >= content.length;\n const chunk: GeminiResponseChunk = {\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: slice }] },\n index: 0,\n ...(isLast ? { finishReason: \"STOP\" } : {}),\n },\n ],\n ...(isLast\n ? {\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n }\n : {}),\n };\n chunks.push(chunk);\n }\n\n // Handle empty content\n if (content.length === 0) {\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: \"\" }] },\n finishReason: \"STOP\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n });\n }\n\n return chunks;\n}\n\nfunction buildGeminiToolCallStreamChunks(\n toolCalls: ToolCall[],\n logger: Logger,\n): GeminiResponseChunk[] {\n const parts: GeminiPart[] = toolCalls.map((tc) => {\n let argsObj: Record<string, unknown>;\n try {\n argsObj = JSON.parse(tc.arguments || \"{}\") as Record<string, unknown>;\n } catch {\n logger.warn(\n `Malformed JSON in fixture tool call arguments for \"${tc.name}\": ${tc.arguments}`,\n );\n argsObj = {};\n }\n return {\n functionCall: { name: tc.name, args: argsObj, id: tc.id || generateToolCallId() },\n };\n });\n\n // Gemini sends all tool calls in a single response chunk\n return [\n {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n },\n ];\n}\n\n// Non-streaming response builders\n\nfunction buildGeminiTextResponse(content: string): GeminiResponseChunk {\n return {\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: content }] },\n finishReason: \"STOP\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n };\n}\n\nfunction buildGeminiToolCallResponse(toolCalls: ToolCall[], logger: Logger): GeminiResponseChunk {\n const parts: GeminiPart[] = toolCalls.map((tc) => {\n let argsObj: Record<string, unknown>;\n try {\n argsObj = JSON.parse(tc.arguments || \"{}\") as Record<string, unknown>;\n } catch {\n logger.warn(\n `Malformed JSON in fixture tool call arguments for \"${tc.name}\": ${tc.arguments}`,\n );\n argsObj = {};\n }\n return {\n functionCall: { name: tc.name, args: argsObj, id: tc.id || generateToolCallId() },\n };\n });\n\n return {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n };\n}\n\n// ─── SSE writer for Gemini streaming ────────────────────────────────────────\n\ninterface GeminiStreamOptions {\n latency?: number;\n streamingProfile?: StreamingProfile;\n signal?: AbortSignal;\n onChunkSent?: () => void;\n}\n\nasync function writeGeminiSSEStream(\n res: http.ServerResponse,\n chunks: GeminiResponseChunk[],\n optionsOrLatency?: number | GeminiStreamOptions,\n): Promise<boolean> {\n const opts: GeminiStreamOptions =\n typeof optionsOrLatency === \"number\" ? { latency: optionsOrLatency } : (optionsOrLatency ?? {});\n const latency = opts.latency ?? 0;\n const profile = opts.streamingProfile;\n const signal = opts.signal;\n const onChunkSent = opts.onChunkSent;\n\n if (res.writableEnded) return true;\n res.setHeader(\"Content-Type\", \"text/event-stream\");\n res.setHeader(\"Cache-Control\", \"no-cache\");\n res.setHeader(\"Connection\", \"keep-alive\");\n\n let chunkIndex = 0;\n for (const chunk of chunks) {\n const chunkDelay = calculateDelay(chunkIndex, profile, latency);\n if (chunkDelay > 0) await delay(chunkDelay, signal);\n if (signal?.aborted) return false;\n if (res.writableEnded) return true;\n // Gemini uses data-only SSE (no event: prefix, no [DONE])\n res.write(`data: ${JSON.stringify(chunk)}\\n\\n`);\n onChunkSent?.();\n if (signal?.aborted) return false;\n chunkIndex++;\n }\n\n if (!res.writableEnded) {\n res.end();\n }\n return true;\n}\n\n// ─── Request handler ────────────────────────────────────────────────────────\n\nexport async function handleGemini(\n req: http.IncomingMessage,\n res: http.ServerResponse,\n raw: string,\n model: string,\n streaming: boolean,\n fixtures: Fixture[],\n journal: Journal,\n defaults: HandlerDefaults,\n setCorsHeaders: (res: http.ServerResponse) => void,\n providerKey: RecordProviderKey = \"gemini\",\n): Promise<void> {\n const { logger } = defaults;\n setCorsHeaders(res);\n\n let geminiReq: GeminiRequest;\n try {\n geminiReq = JSON.parse(raw) as GeminiRequest;\n } catch {\n journal.add({\n method: req.method ?? \"POST\",\n path: req.url ?? `/v1beta/models/${model}:generateContent`,\n headers: flattenHeaders(req.headers),\n body: null,\n response: { status: 400, fixture: null },\n });\n writeErrorResponse(\n res,\n 400,\n JSON.stringify({\n error: {\n message: \"Malformed JSON\",\n code: 400,\n status: \"INVALID_ARGUMENT\",\n },\n }),\n );\n return;\n }\n\n // Convert to ChatCompletionRequest for fixture matching\n const completionReq = geminiToCompletionRequest(geminiReq, model, streaming);\n\n const fixture = matchFixture(fixtures, completionReq, journal.fixtureMatchCounts);\n const path = req.url ?? `/v1beta/models/${model}:generateContent`;\n\n if (fixture) {\n journal.incrementFixtureMatchCount(fixture, fixtures);\n }\n\n if (\n applyChaos(\n res,\n fixture,\n defaults.chaos,\n req.headers,\n journal,\n {\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n },\n defaults.registry,\n defaults.logger,\n )\n )\n return;\n\n if (!fixture) {\n if (defaults.record) {\n const proxied = await proxyAndRecord(\n req,\n res,\n completionReq,\n providerKey,\n path,\n fixtures,\n defaults,\n raw,\n );\n if (proxied) {\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: res.statusCode ?? 200, fixture: null },\n });\n return;\n }\n }\n const strictStatus = defaults.strict ? 503 : 404;\n const strictMessage = defaults.strict\n ? \"Strict mode: no fixture matched\"\n : \"No fixture matched\";\n if (defaults.strict) {\n logger.error(`STRICT: No fixture matched for ${req.method ?? \"POST\"} ${path}`);\n }\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: strictStatus, fixture: null },\n });\n writeErrorResponse(\n res,\n strictStatus,\n JSON.stringify({\n error: {\n message: strictMessage,\n code: strictStatus,\n status: defaults.strict ? \"UNAVAILABLE\" : \"NOT_FOUND\",\n },\n }),\n );\n return;\n }\n\n const response = fixture.response;\n const latency = fixture.latency ?? defaults.latency;\n const chunkSize = Math.max(1, fixture.chunkSize ?? defaults.chunkSize);\n\n // Error response\n if (isErrorResponse(response)) {\n const status = response.status ?? 500;\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status, fixture },\n });\n // Gemini-style error format: { error: { code, message, status } }\n const geminiError = {\n error: {\n code: status,\n message: response.error.message,\n status: response.error.type ?? \"ERROR\",\n },\n };\n writeErrorResponse(res, status, JSON.stringify(geminiError));\n return;\n }\n\n // Text response\n if (isTextResponse(response)) {\n const journalEntry = journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 200, fixture },\n });\n if (!streaming) {\n const body = buildGeminiTextResponse(response.content);\n res.writeHead(200, { \"Content-Type\": \"application/json\" });\n res.end(JSON.stringify(body));\n } else {\n const chunks = buildGeminiTextStreamChunks(response.content, chunkSize);\n const interruption = createInterruptionSignal(fixture);\n const completed = await writeGeminiSSEStream(res, chunks, {\n latency,\n streamingProfile: fixture.streamingProfile,\n signal: interruption?.signal,\n onChunkSent: interruption?.tick,\n });\n if (!completed) {\n if (!res.writableEnded) res.destroy();\n journalEntry.response.interrupted = true;\n journalEntry.response.interruptReason = interruption?.reason();\n }\n interruption?.cleanup();\n }\n return;\n }\n\n // Tool call response\n if (isToolCallResponse(response)) {\n const journalEntry = journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 200, fixture },\n });\n if (!streaming) {\n const body = buildGeminiToolCallResponse(response.toolCalls, logger);\n res.writeHead(200, { \"Content-Type\": \"application/json\" });\n res.end(JSON.stringify(body));\n } else {\n const chunks = buildGeminiToolCallStreamChunks(response.toolCalls, logger);\n const interruption = createInterruptionSignal(fixture);\n const completed = await writeGeminiSSEStream(res, chunks, {\n latency,\n streamingProfile: fixture.streamingProfile,\n signal: interruption?.signal,\n onChunkSent: interruption?.tick,\n });\n if (!completed) {\n if (!res.writableEnded) res.destroy();\n journalEntry.response.interrupted = true;\n journalEntry.response.interruptReason = interruption?.reason();\n }\n interruption?.cleanup();\n }\n return;\n }\n\n // Unknown response type\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 500, fixture },\n });\n writeErrorResponse(\n res,\n 500,\n JSON.stringify({\n error: {\n message: \"Fixture response did not match any known type\",\n code: 500,\n status: \"INTERNAL\",\n },\n }),\n );\n}\n"],"mappings":";;;;;;;;AAuEA,SAAgB,0BACd,KACA,OACA,QACuB;CACvB,MAAM,WAA0B,EAAE;AAGlC,KAAI,IAAI,mBAAmB;EACzB,MAAM,OAAO,IAAI,kBAAkB,MAChC,QAAQ,MAAM,EAAE,SAAS,OAAU,CACnC,KAAK,MAAM,EAAE,KAAM,CACnB,KAAK,GAAG;AACX,MAAI,KACF,UAAS,KAAK;GAAE,MAAM;GAAU,SAAS;GAAM,CAAC;;AAIpD,KAAI,IAAI,SACN,MAAK,MAAM,WAAW,IAAI,UAAU;EAClC,MAAM,OAAO,QAAQ,QAAQ;AAE7B,MAAI,SAAS,QAAQ;GAEnB,MAAM,gBAAgB,QAAQ,MAAM,QAAQ,MAAM,EAAE,iBAAiB;GACrE,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,SAAS,OAAU;AAEnE,OAAI,cAAc,SAAS,GAAG;AAE5B,SAAK,IAAI,IAAI,GAAG,IAAI,cAAc,QAAQ,KAAK;KAC7C,MAAM,OAAO,cAAc;AAC3B,cAAS,KAAK;MACZ,MAAM;MACN,SACE,OAAO,KAAK,iBAAkB,aAAa,WACvC,KAAK,iBAAkB,WACvB,KAAK,UAAU,KAAK,iBAAkB,SAAS;MACrD,cAAc,eAAe,KAAK,iBAAkB,KAAK,GAAG;MAC7D,CAAC;;AAGJ,QAAI,UAAU,SAAS,EACrB,UAAS,KAAK;KACZ,MAAM;KACN,SAAS,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;KAChD,CAAC;UAEC;IAEL,MAAM,OAAO,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;AACnD,aAAS,KAAK;KAAE,MAAM;KAAQ,SAAS;KAAM,CAAC;;aAEvC,SAAS,SAAS;GAE3B,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,aAAa;GAC7D,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,SAAS,OAAU;AAEnE,OAAI,UAAU,SAAS,EACrB,UAAS,KAAK;IACZ,MAAM;IACN,SAAS;IACT,YAAY,UAAU,KAAK,GAAG,OAAO;KACnC,IAAI,eAAe,EAAE,aAAc,KAAK,GAAG;KAC3C,MAAM;KACN,UAAU;MACR,MAAM,EAAE,aAAc;MACtB,WAAW,KAAK,UAAU,EAAE,aAAc,KAAK;MAChD;KACF,EAAE;IACJ,CAAC;QACG;IACL,MAAM,OAAO,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;AACnD,aAAS,KAAK;KAAE,MAAM;KAAa,SAAS;KAAM,CAAC;;;;CAO3D,IAAI;AACJ,KAAI,IAAI,SAAS,IAAI,MAAM,SAAS,GAAG;EACrC,MAAM,QAAQ,IAAI,MAAM,SAAS,MAAM,EAAE,wBAAwB,EAAE,CAAC;AACpE,MAAI,MAAM,SAAS,EACjB,SAAQ,MAAM,KAAK,OAAO;GACxB,MAAM;GACN,UAAU;IACR,MAAM,EAAE;IACR,aAAa,EAAE;IACf,YAAY,EAAE;IACf;GACF,EAAE;;AAIP,QAAO;EACL;EACA;EACA;EACA,aAAa,IAAI,kBAAkB;EACnC;EACD;;AAkBH,SAAS,4BAA4B,SAAiB,WAA0C;CAC9F,MAAM,SAAgC,EAAE;AAGxC,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK,WAAW;EAClD,MAAM,QAAQ,QAAQ,MAAM,GAAG,IAAI,UAAU;EAC7C,MAAM,SAAS,IAAI,aAAa,QAAQ;EACxC,MAAM,QAA6B;GACjC,YAAY,CACV;IACE,SAAS;KAAE,MAAM;KAAS,OAAO,CAAC,EAAE,MAAM,OAAO,CAAC;KAAE;IACpD,OAAO;IACP,GAAI,SAAS,EAAE,cAAc,QAAQ,GAAG,EAAE;IAC3C,CACF;GACD,GAAI,SACA,EACE,eAAe;IACb,kBAAkB;IAClB,sBAAsB;IACtB,iBAAiB;IAClB,EACF,GACD,EAAE;GACP;AACD,SAAO,KAAK,MAAM;;AAIpB,KAAI,QAAQ,WAAW,EACrB,QAAO,KAAK;EACV,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAAO,CAAC,EAAE,MAAM,IAAI,CAAC;IAAE;GACjD,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF,CAAC;AAGJ,QAAO;;AAGT,SAAS,gCACP,WACA,QACuB;AAiBvB,QAAO,CACL;EACE,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OApBN,UAAU,KAAK,OAAO;KAChD,IAAI;AACJ,SAAI;AACF,gBAAU,KAAK,MAAM,GAAG,aAAa,KAAK;aACpC;AACN,aAAO,KACL,sDAAsD,GAAG,KAAK,KAAK,GAAG,YACvE;AACD,gBAAU,EAAE;;AAEd,YAAO,EACL,cAAc;MAAE,MAAM,GAAG;MAAM,MAAM;MAAS,IAAI,GAAG,MAAMA,oCAAoB;MAAE,EAClF;MACD;IAOuC;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF,CACF;;AAKH,SAAS,wBAAwB,SAAsC;AACrE,QAAO;EACL,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAAO,CAAC,EAAE,MAAM,SAAS,CAAC;IAAE;GACtD,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF;;AAGH,SAAS,4BAA4B,WAAuB,QAAqC;AAgB/F,QAAO;EACL,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAlBJ,UAAU,KAAK,OAAO;KAChD,IAAI;AACJ,SAAI;AACF,gBAAU,KAAK,MAAM,GAAG,aAAa,KAAK;aACpC;AACN,aAAO,KACL,sDAAsD,GAAG,KAAK,KAAK,GAAG,YACvE;AACD,gBAAU,EAAE;;AAEd,YAAO,EACL,cAAc;MAAE,MAAM,GAAG;MAAM,MAAM;MAAS,IAAI,GAAG,MAAMA,oCAAoB;MAAE,EAClF;MACD;IAKqC;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF;;AAYH,eAAe,qBACb,KACA,QACA,kBACkB;CAClB,MAAM,OACJ,OAAO,qBAAqB,WAAW,EAAE,SAAS,kBAAkB,GAAI,oBAAoB,EAAE;CAChG,MAAM,UAAU,KAAK,WAAW;CAChC,MAAM,UAAU,KAAK;CACrB,MAAM,SAAS,KAAK;CACpB,MAAM,cAAc,KAAK;AAEzB,KAAI,IAAI,cAAe,QAAO;AAC9B,KAAI,UAAU,gBAAgB,oBAAoB;AAClD,KAAI,UAAU,iBAAiB,WAAW;AAC1C,KAAI,UAAU,cAAc,aAAa;CAEzC,IAAI,aAAa;AACjB,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,aAAaC,kCAAe,YAAY,SAAS,QAAQ;AAC/D,MAAI,aAAa,EAAG,OAAMC,yBAAM,YAAY,OAAO;AACnD,MAAI,QAAQ,QAAS,QAAO;AAC5B,MAAI,IAAI,cAAe,QAAO;AAE9B,MAAI,MAAM,SAAS,KAAK,UAAU,MAAM,CAAC,MAAM;AAC/C,iBAAe;AACf,MAAI,QAAQ,QAAS,QAAO;AAC5B;;AAGF,KAAI,CAAC,IAAI,cACP,KAAI,KAAK;AAEX,QAAO;;AAKT,eAAsB,aACpB,KACA,KACA,KACA,OACA,WACA,UACA,SACA,UACA,gBACA,cAAiC,UAClB;CACf,MAAM,EAAE,WAAW;AACnB,gBAAe,IAAI;CAEnB,IAAI;AACJ,KAAI;AACF,cAAY,KAAK,MAAM,IAAI;SACrB;AACN,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB,MAAM,IAAI,OAAO,kBAAkB,MAAM;GACzC,SAASC,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK,SAAS;IAAM;GACzC,CAAC;AACF,wCACE,KACA,KACA,KAAK,UAAU,EACb,OAAO;GACL,SAAS;GACT,MAAM;GACN,QAAQ;GACT,EACF,CAAC,CACH;AACD;;CAIF,MAAM,gBAAgB,0BAA0B,WAAW,OAAO,UAAU;CAE5E,MAAM,UAAUC,4BAAa,UAAU,eAAe,QAAQ,mBAAmB;CACjF,MAAM,OAAO,IAAI,OAAO,kBAAkB,MAAM;AAEhD,KAAI,QACF,SAAQ,2BAA2B,SAAS,SAAS;AAGvD,KACEC,yBACE,KACA,SACA,SAAS,OACT,IAAI,SACJ,SACA;EACE,QAAQ,IAAI,UAAU;EACtB;EACA,SAASF,+BAAe,IAAI,QAAQ;EACpC,MAAM;EACP,EACD,SAAS,UACT,SAAS,OACV,CAED;AAEF,KAAI,CAAC,SAAS;AACZ,MAAI,SAAS,QAWX;OAVgB,MAAMG,gCACpB,KACA,KACA,eACA,aACA,MACA,UACA,UACA,IACD,EACY;AACX,YAAQ,IAAI;KACV,QAAQ,IAAI,UAAU;KACtB;KACA,SAASH,+BAAe,IAAI,QAAQ;KACpC,MAAM;KACN,UAAU;MAAE,QAAQ,IAAI,cAAc;MAAK,SAAS;MAAM;KAC3D,CAAC;AACF;;;EAGJ,MAAM,eAAe,SAAS,SAAS,MAAM;EAC7C,MAAM,gBAAgB,SAAS,SAC3B,oCACA;AACJ,MAAI,SAAS,OACX,QAAO,MAAM,kCAAkC,IAAI,UAAU,OAAO,GAAG,OAAO;AAEhF,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB;GACA,SAASA,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAc,SAAS;IAAM;GAClD,CAAC;AACF,wCACE,KACA,cACA,KAAK,UAAU,EACb,OAAO;GACL,SAAS;GACT,MAAM;GACN,QAAQ,SAAS,SAAS,gBAAgB;GAC3C,EACF,CAAC,CACH;AACD;;CAGF,MAAM,WAAW,QAAQ;CACzB,MAAM,UAAU,QAAQ,WAAW,SAAS;CAC5C,MAAM,YAAY,KAAK,IAAI,GAAG,QAAQ,aAAa,SAAS,UAAU;AAGtE,KAAII,gCAAgB,SAAS,EAAE;EAC7B,MAAM,SAAS,SAAS,UAAU;AAClC,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB;GACA,SAASJ,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE;IAAQ;IAAS;GAC9B,CAAC;EAEF,MAAM,cAAc,EAClB,OAAO;GACL,MAAM;GACN,SAAS,SAAS,MAAM;GACxB,QAAQ,SAAS,MAAM,QAAQ;GAChC,EACF;AACD,wCAAmB,KAAK,QAAQ,KAAK,UAAU,YAAY,CAAC;AAC5D;;AAIF,KAAIK,+BAAe,SAAS,EAAE;EAC5B,MAAM,eAAe,QAAQ,IAAI;GAC/B,QAAQ,IAAI,UAAU;GACtB;GACA,SAASL,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK;IAAS;GACnC,CAAC;AACF,MAAI,CAAC,WAAW;GACd,MAAM,OAAO,wBAAwB,SAAS,QAAQ;AACtD,OAAI,UAAU,KAAK,EAAE,gBAAgB,oBAAoB,CAAC;AAC1D,OAAI,IAAI,KAAK,UAAU,KAAK,CAAC;SACxB;GACL,MAAM,SAAS,4BAA4B,SAAS,SAAS,UAAU;GACvE,MAAM,eAAeM,8CAAyB,QAAQ;AAOtD,OAAI,CANc,MAAM,qBAAqB,KAAK,QAAQ;IACxD;IACA,kBAAkB,QAAQ;IAC1B,QAAQ,cAAc;IACtB,aAAa,cAAc;IAC5B,CAAC,EACc;AACd,QAAI,CAAC,IAAI,cAAe,KAAI,SAAS;AACrC,iBAAa,SAAS,cAAc;AACpC,iBAAa,SAAS,kBAAkB,cAAc,QAAQ;;AAEhE,iBAAc,SAAS;;AAEzB;;AAIF,KAAIC,mCAAmB,SAAS,EAAE;EAChC,MAAM,eAAe,QAAQ,IAAI;GAC/B,QAAQ,IAAI,UAAU;GACtB;GACA,SAASP,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK;IAAS;GACnC,CAAC;AACF,MAAI,CAAC,WAAW;GACd,MAAM,OAAO,4BAA4B,SAAS,WAAW,OAAO;AACpE,OAAI,UAAU,KAAK,EAAE,gBAAgB,oBAAoB,CAAC;AAC1D,OAAI,IAAI,KAAK,UAAU,KAAK,CAAC;SACxB;GACL,MAAM,SAAS,gCAAgC,SAAS,WAAW,OAAO;GAC1E,MAAM,eAAeM,8CAAyB,QAAQ;AAOtD,OAAI,CANc,MAAM,qBAAqB,KAAK,QAAQ;IACxD;IACA,kBAAkB,QAAQ;IAC1B,QAAQ,cAAc;IACtB,aAAa,cAAc;IAC5B,CAAC,EACc;AACd,QAAI,CAAC,IAAI,cAAe,KAAI,SAAS;AACrC,iBAAa,SAAS,cAAc;AACpC,iBAAa,SAAS,kBAAkB,cAAc,QAAQ;;AAEhE,iBAAc,SAAS;;AAEzB;;AAIF,SAAQ,IAAI;EACV,QAAQ,IAAI,UAAU;EACtB;EACA,SAASN,+BAAe,IAAI,QAAQ;EACpC,MAAM;EACN,UAAU;GAAE,QAAQ;GAAK;GAAS;EACnC,CAAC;AACF,uCACE,KACA,KACA,KAAK,UAAU,EACb,OAAO;EACL,SAAS;EACT,MAAM;EACN,QAAQ;EACT,EACF,CAAC,CACH"}
|
|
1
|
+
{"version":3,"file":"gemini.cjs","names":["generateToolCallId","calculateDelay","delay","flattenHeaders","getTestId","matchFixture","applyChaos","proxyAndRecord","isErrorResponse","isContentWithToolCallsResponse","createInterruptionSignal","isTextResponse","isToolCallResponse"],"sources":["../src/gemini.ts"],"sourcesContent":["/**\n * Google Gemini GenerateContent API support.\n *\n * Translates incoming Gemini requests into the ChatCompletionRequest format\n * used by the fixture router, and converts fixture responses back into the\n * Gemini GenerateContent streaming (or non-streaming) format.\n */\n\nimport type * as http from \"node:http\";\nimport type {\n ChatCompletionRequest,\n ChatMessage,\n Fixture,\n HandlerDefaults,\n RecordProviderKey,\n StreamingProfile,\n ToolCall,\n ToolDefinition,\n} from \"./types.js\";\nimport {\n isTextResponse,\n isToolCallResponse,\n isContentWithToolCallsResponse,\n isErrorResponse,\n generateToolCallId,\n flattenHeaders,\n getTestId,\n} from \"./helpers.js\";\nimport { matchFixture } from \"./router.js\";\nimport { writeErrorResponse, delay, calculateDelay } from \"./sse-writer.js\";\nimport { createInterruptionSignal } from \"./interruption.js\";\nimport type { Journal } from \"./journal.js\";\nimport type { Logger } from \"./logger.js\";\nimport { applyChaos } from \"./chaos.js\";\nimport { proxyAndRecord } from \"./recorder.js\";\n\n// ─── Gemini request types ───────────────────────────────────────────────────\n\ninterface GeminiPart {\n text?: string;\n thought?: boolean;\n functionCall?: { name: string; args: Record<string, unknown>; id?: string };\n functionResponse?: { name: string; response: unknown };\n}\n\ninterface GeminiContent {\n role?: string;\n parts: GeminiPart[];\n}\n\ninterface GeminiFunctionDeclaration {\n name: string;\n description?: string;\n parameters?: object;\n}\n\ninterface GeminiToolDef {\n functionDeclarations?: GeminiFunctionDeclaration[];\n}\n\ninterface GeminiRequest {\n contents?: GeminiContent[];\n systemInstruction?: GeminiContent;\n tools?: GeminiToolDef[];\n generationConfig?: {\n temperature?: number;\n maxOutputTokens?: number;\n [key: string]: unknown;\n };\n [key: string]: unknown;\n}\n\n// ─── Input conversion: Gemini → ChatCompletions messages ────────────────────\n\nexport function geminiToCompletionRequest(\n req: GeminiRequest,\n model: string,\n stream: boolean,\n): ChatCompletionRequest {\n const messages: ChatMessage[] = [];\n\n // systemInstruction → system message\n if (req.systemInstruction) {\n const text = req.systemInstruction.parts\n .filter((p) => p.text !== undefined)\n .map((p) => p.text!)\n .join(\"\");\n if (text) {\n messages.push({ role: \"system\", content: text });\n }\n }\n\n if (req.contents) {\n for (const content of req.contents) {\n const role = content.role ?? \"user\";\n\n if (role === \"user\") {\n // Check for functionResponse parts\n const funcResponses = content.parts.filter((p) => p.functionResponse);\n const textParts = content.parts.filter((p) => p.text !== undefined);\n\n if (funcResponses.length > 0) {\n // functionResponse → tool message\n for (let i = 0; i < funcResponses.length; i++) {\n const part = funcResponses[i];\n messages.push({\n role: \"tool\",\n content:\n typeof part.functionResponse!.response === \"string\"\n ? part.functionResponse!.response\n : JSON.stringify(part.functionResponse!.response),\n tool_call_id: `call_gemini_${part.functionResponse!.name}_${i}`,\n });\n }\n // Any text parts alongside → user message\n if (textParts.length > 0) {\n messages.push({\n role: \"user\",\n content: textParts.map((p) => p.text!).join(\"\"),\n });\n }\n } else {\n // Regular user text\n const text = textParts.map((p) => p.text!).join(\"\");\n messages.push({ role: \"user\", content: text });\n }\n } else if (role === \"model\") {\n // Check for functionCall parts\n const funcCalls = content.parts.filter((p) => p.functionCall);\n const textParts = content.parts.filter((p) => p.text !== undefined);\n\n if (funcCalls.length > 0) {\n messages.push({\n role: \"assistant\",\n content: null,\n tool_calls: funcCalls.map((p, i) => ({\n id: `call_gemini_${p.functionCall!.name}_${i}`,\n type: \"function\" as const,\n function: {\n name: p.functionCall!.name,\n arguments: JSON.stringify(p.functionCall!.args),\n },\n })),\n });\n } else {\n const text = textParts.map((p) => p.text!).join(\"\");\n messages.push({ role: \"assistant\", content: text });\n }\n }\n }\n }\n\n // Convert tools\n let tools: ToolDefinition[] | undefined;\n if (req.tools && req.tools.length > 0) {\n const decls = req.tools.flatMap((t) => t.functionDeclarations ?? []);\n if (decls.length > 0) {\n tools = decls.map((d) => ({\n type: \"function\" as const,\n function: {\n name: d.name,\n description: d.description,\n parameters: d.parameters,\n },\n }));\n }\n }\n\n return {\n model,\n messages,\n stream,\n temperature: req.generationConfig?.temperature,\n tools,\n };\n}\n\n// ─── Response building: fixture → Gemini format ─────────────────────────────\n\ninterface GeminiResponseChunk {\n candidates: {\n content: { role: string; parts: GeminiPart[] };\n finishReason?: string;\n index: number;\n }[];\n usageMetadata?: {\n promptTokenCount: number;\n candidatesTokenCount: number;\n totalTokenCount: number;\n };\n}\n\nfunction buildGeminiTextStreamChunks(\n content: string,\n chunkSize: number,\n reasoning?: string,\n): GeminiResponseChunk[] {\n const chunks: GeminiResponseChunk[] = [];\n\n // Reasoning chunks (thought: true)\n if (reasoning) {\n for (let i = 0; i < reasoning.length; i += chunkSize) {\n const slice = reasoning.slice(i, i + chunkSize);\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: slice, thought: true }] },\n index: 0,\n },\n ],\n });\n }\n }\n\n // Content chunks (original logic unchanged)\n for (let i = 0; i < content.length; i += chunkSize) {\n const slice = content.slice(i, i + chunkSize);\n const isLast = i + chunkSize >= content.length;\n const chunk: GeminiResponseChunk = {\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: slice }] },\n index: 0,\n ...(isLast ? { finishReason: \"STOP\" } : {}),\n },\n ],\n ...(isLast\n ? {\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n }\n : {}),\n };\n chunks.push(chunk);\n }\n\n // Handle empty content (original logic unchanged)\n if (content.length === 0) {\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: \"\" }] },\n finishReason: \"STOP\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n });\n }\n\n return chunks;\n}\n\nfunction parseToolCallPart(tc: ToolCall, logger: Logger): GeminiPart {\n let argsObj: Record<string, unknown>;\n try {\n argsObj = JSON.parse(tc.arguments || \"{}\") as Record<string, unknown>;\n } catch {\n logger.warn(`Malformed JSON in fixture tool call arguments for \"${tc.name}\": ${tc.arguments}`);\n argsObj = {};\n }\n return { functionCall: { name: tc.name, args: argsObj, id: tc.id || generateToolCallId() } };\n}\n\nfunction buildGeminiToolCallStreamChunks(\n toolCalls: ToolCall[],\n logger: Logger,\n): GeminiResponseChunk[] {\n const parts: GeminiPart[] = toolCalls.map((tc) => parseToolCallPart(tc, logger));\n\n // Gemini sends all tool calls in a single response chunk\n return [\n {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n },\n ];\n}\n\n// Non-streaming response builders\n\nfunction buildGeminiTextResponse(content: string, reasoning?: string): GeminiResponseChunk {\n const parts: GeminiPart[] = [];\n if (reasoning) {\n parts.push({ text: reasoning, thought: true });\n }\n parts.push({ text: content });\n\n return {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"STOP\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n };\n}\n\nfunction buildGeminiToolCallResponse(toolCalls: ToolCall[], logger: Logger): GeminiResponseChunk {\n const parts: GeminiPart[] = toolCalls.map((tc) => parseToolCallPart(tc, logger));\n\n return {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n };\n}\n\nfunction buildGeminiContentWithToolCallsStreamChunks(\n content: string,\n toolCalls: ToolCall[],\n chunkSize: number,\n logger: Logger,\n): GeminiResponseChunk[] {\n const chunks: GeminiResponseChunk[] = [];\n\n if (content.length === 0) {\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: \"\" }] },\n index: 0,\n },\n ],\n });\n } else {\n for (let i = 0; i < content.length; i += chunkSize) {\n const slice = content.slice(i, i + chunkSize);\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts: [{ text: slice }] },\n index: 0,\n },\n ],\n });\n }\n }\n\n const parts: GeminiPart[] = toolCalls.map((tc) => parseToolCallPart(tc, logger));\n\n chunks.push({\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n });\n\n return chunks;\n}\n\nfunction buildGeminiContentWithToolCallsResponse(\n content: string,\n toolCalls: ToolCall[],\n logger: Logger,\n): GeminiResponseChunk {\n const parts: GeminiPart[] = [\n { text: content },\n ...toolCalls.map((tc) => parseToolCallPart(tc, logger)),\n ];\n\n return {\n candidates: [\n {\n content: { role: \"model\", parts },\n finishReason: \"FUNCTION_CALL\",\n index: 0,\n },\n ],\n usageMetadata: {\n promptTokenCount: 0,\n candidatesTokenCount: 0,\n totalTokenCount: 0,\n },\n };\n}\n\n// ─── SSE writer for Gemini streaming ────────────────────────────────────────\n\ninterface GeminiStreamOptions {\n latency?: number;\n streamingProfile?: StreamingProfile;\n signal?: AbortSignal;\n onChunkSent?: () => void;\n}\n\nasync function writeGeminiSSEStream(\n res: http.ServerResponse,\n chunks: GeminiResponseChunk[],\n optionsOrLatency?: number | GeminiStreamOptions,\n): Promise<boolean> {\n const opts: GeminiStreamOptions =\n typeof optionsOrLatency === \"number\" ? { latency: optionsOrLatency } : (optionsOrLatency ?? {});\n const latency = opts.latency ?? 0;\n const profile = opts.streamingProfile;\n const signal = opts.signal;\n const onChunkSent = opts.onChunkSent;\n\n if (res.writableEnded) return true;\n res.setHeader(\"Content-Type\", \"text/event-stream\");\n res.setHeader(\"Cache-Control\", \"no-cache\");\n res.setHeader(\"Connection\", \"keep-alive\");\n\n let chunkIndex = 0;\n for (const chunk of chunks) {\n const chunkDelay = calculateDelay(chunkIndex, profile, latency);\n if (chunkDelay > 0) await delay(chunkDelay, signal);\n if (signal?.aborted) return false;\n if (res.writableEnded) return true;\n // Gemini uses data-only SSE (no event: prefix, no [DONE])\n res.write(`data: ${JSON.stringify(chunk)}\\n\\n`);\n onChunkSent?.();\n if (signal?.aborted) return false;\n chunkIndex++;\n }\n\n if (!res.writableEnded) {\n res.end();\n }\n return true;\n}\n\n// ─── Request handler ────────────────────────────────────────────────────────\n\nexport async function handleGemini(\n req: http.IncomingMessage,\n res: http.ServerResponse,\n raw: string,\n model: string,\n streaming: boolean,\n fixtures: Fixture[],\n journal: Journal,\n defaults: HandlerDefaults,\n setCorsHeaders: (res: http.ServerResponse) => void,\n providerKey: RecordProviderKey = \"gemini\",\n): Promise<void> {\n const { logger } = defaults;\n setCorsHeaders(res);\n\n let geminiReq: GeminiRequest;\n try {\n geminiReq = JSON.parse(raw) as GeminiRequest;\n } catch {\n journal.add({\n method: req.method ?? \"POST\",\n path: req.url ?? `/v1beta/models/${model}:generateContent`,\n headers: flattenHeaders(req.headers),\n body: null,\n response: { status: 400, fixture: null },\n });\n writeErrorResponse(\n res,\n 400,\n JSON.stringify({\n error: {\n message: \"Malformed JSON\",\n code: 400,\n status: \"INVALID_ARGUMENT\",\n },\n }),\n );\n return;\n }\n\n // Convert to ChatCompletionRequest for fixture matching\n const completionReq = geminiToCompletionRequest(geminiReq, model, streaming);\n\n const testId = getTestId(req);\n const fixture = matchFixture(\n fixtures,\n completionReq,\n journal.getFixtureMatchCountsForTest(testId),\n defaults.requestTransform,\n );\n const path = req.url ?? `/v1beta/models/${model}:generateContent`;\n\n if (fixture) {\n journal.incrementFixtureMatchCount(fixture, fixtures, testId);\n }\n\n if (\n applyChaos(\n res,\n fixture,\n defaults.chaos,\n req.headers,\n journal,\n {\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n },\n defaults.registry,\n defaults.logger,\n )\n )\n return;\n\n if (!fixture) {\n if (defaults.record) {\n const proxied = await proxyAndRecord(\n req,\n res,\n completionReq,\n providerKey,\n path,\n fixtures,\n defaults,\n raw,\n );\n if (proxied) {\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: res.statusCode ?? 200, fixture: null },\n });\n return;\n }\n }\n const strictStatus = defaults.strict ? 503 : 404;\n const strictMessage = defaults.strict\n ? \"Strict mode: no fixture matched\"\n : \"No fixture matched\";\n if (defaults.strict) {\n logger.error(`STRICT: No fixture matched for ${req.method ?? \"POST\"} ${path}`);\n }\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: strictStatus, fixture: null },\n });\n writeErrorResponse(\n res,\n strictStatus,\n JSON.stringify({\n error: {\n message: strictMessage,\n code: strictStatus,\n status: defaults.strict ? \"UNAVAILABLE\" : \"NOT_FOUND\",\n },\n }),\n );\n return;\n }\n\n const response = fixture.response;\n const latency = fixture.latency ?? defaults.latency;\n const chunkSize = Math.max(1, fixture.chunkSize ?? defaults.chunkSize);\n\n // Error response\n if (isErrorResponse(response)) {\n const status = response.status ?? 500;\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status, fixture },\n });\n // Gemini-style error format: { error: { code, message, status } }\n const geminiError = {\n error: {\n code: status,\n message: response.error.message,\n status: response.error.type ?? \"ERROR\",\n },\n };\n writeErrorResponse(res, status, JSON.stringify(geminiError));\n return;\n }\n\n // Content + tool calls response (must be checked before isTextResponse / isToolCallResponse)\n if (isContentWithToolCallsResponse(response)) {\n const journalEntry = journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 200, fixture },\n });\n if (!streaming) {\n const body = buildGeminiContentWithToolCallsResponse(\n response.content,\n response.toolCalls,\n logger,\n );\n res.writeHead(200, { \"Content-Type\": \"application/json\" });\n res.end(JSON.stringify(body));\n } else {\n const chunks = buildGeminiContentWithToolCallsStreamChunks(\n response.content,\n response.toolCalls,\n chunkSize,\n logger,\n );\n const interruption = createInterruptionSignal(fixture);\n const completed = await writeGeminiSSEStream(res, chunks, {\n latency,\n streamingProfile: fixture.streamingProfile,\n signal: interruption?.signal,\n onChunkSent: interruption?.tick,\n });\n if (!completed) {\n if (!res.writableEnded) res.destroy();\n journalEntry.response.interrupted = true;\n journalEntry.response.interruptReason = interruption?.reason();\n }\n interruption?.cleanup();\n }\n return;\n }\n\n // Text response\n if (isTextResponse(response)) {\n const journalEntry = journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 200, fixture },\n });\n if (!streaming) {\n const body = buildGeminiTextResponse(response.content, response.reasoning);\n res.writeHead(200, { \"Content-Type\": \"application/json\" });\n res.end(JSON.stringify(body));\n } else {\n const chunks = buildGeminiTextStreamChunks(response.content, chunkSize, response.reasoning);\n const interruption = createInterruptionSignal(fixture);\n const completed = await writeGeminiSSEStream(res, chunks, {\n latency,\n streamingProfile: fixture.streamingProfile,\n signal: interruption?.signal,\n onChunkSent: interruption?.tick,\n });\n if (!completed) {\n if (!res.writableEnded) res.destroy();\n journalEntry.response.interrupted = true;\n journalEntry.response.interruptReason = interruption?.reason();\n }\n interruption?.cleanup();\n }\n return;\n }\n\n // Tool call response\n if (isToolCallResponse(response)) {\n const journalEntry = journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 200, fixture },\n });\n if (!streaming) {\n const body = buildGeminiToolCallResponse(response.toolCalls, logger);\n res.writeHead(200, { \"Content-Type\": \"application/json\" });\n res.end(JSON.stringify(body));\n } else {\n const chunks = buildGeminiToolCallStreamChunks(response.toolCalls, logger);\n const interruption = createInterruptionSignal(fixture);\n const completed = await writeGeminiSSEStream(res, chunks, {\n latency,\n streamingProfile: fixture.streamingProfile,\n signal: interruption?.signal,\n onChunkSent: interruption?.tick,\n });\n if (!completed) {\n if (!res.writableEnded) res.destroy();\n journalEntry.response.interrupted = true;\n journalEntry.response.interruptReason = interruption?.reason();\n }\n interruption?.cleanup();\n }\n return;\n }\n\n // Unknown response type\n journal.add({\n method: req.method ?? \"POST\",\n path,\n headers: flattenHeaders(req.headers),\n body: completionReq,\n response: { status: 500, fixture },\n });\n writeErrorResponse(\n res,\n 500,\n JSON.stringify({\n error: {\n message: \"Fixture response did not match any known type\",\n code: 500,\n status: \"INTERNAL\",\n },\n }),\n );\n}\n"],"mappings":";;;;;;;;AA0EA,SAAgB,0BACd,KACA,OACA,QACuB;CACvB,MAAM,WAA0B,EAAE;AAGlC,KAAI,IAAI,mBAAmB;EACzB,MAAM,OAAO,IAAI,kBAAkB,MAChC,QAAQ,MAAM,EAAE,SAAS,OAAU,CACnC,KAAK,MAAM,EAAE,KAAM,CACnB,KAAK,GAAG;AACX,MAAI,KACF,UAAS,KAAK;GAAE,MAAM;GAAU,SAAS;GAAM,CAAC;;AAIpD,KAAI,IAAI,SACN,MAAK,MAAM,WAAW,IAAI,UAAU;EAClC,MAAM,OAAO,QAAQ,QAAQ;AAE7B,MAAI,SAAS,QAAQ;GAEnB,MAAM,gBAAgB,QAAQ,MAAM,QAAQ,MAAM,EAAE,iBAAiB;GACrE,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,SAAS,OAAU;AAEnE,OAAI,cAAc,SAAS,GAAG;AAE5B,SAAK,IAAI,IAAI,GAAG,IAAI,cAAc,QAAQ,KAAK;KAC7C,MAAM,OAAO,cAAc;AAC3B,cAAS,KAAK;MACZ,MAAM;MACN,SACE,OAAO,KAAK,iBAAkB,aAAa,WACvC,KAAK,iBAAkB,WACvB,KAAK,UAAU,KAAK,iBAAkB,SAAS;MACrD,cAAc,eAAe,KAAK,iBAAkB,KAAK,GAAG;MAC7D,CAAC;;AAGJ,QAAI,UAAU,SAAS,EACrB,UAAS,KAAK;KACZ,MAAM;KACN,SAAS,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;KAChD,CAAC;UAEC;IAEL,MAAM,OAAO,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;AACnD,aAAS,KAAK;KAAE,MAAM;KAAQ,SAAS;KAAM,CAAC;;aAEvC,SAAS,SAAS;GAE3B,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,aAAa;GAC7D,MAAM,YAAY,QAAQ,MAAM,QAAQ,MAAM,EAAE,SAAS,OAAU;AAEnE,OAAI,UAAU,SAAS,EACrB,UAAS,KAAK;IACZ,MAAM;IACN,SAAS;IACT,YAAY,UAAU,KAAK,GAAG,OAAO;KACnC,IAAI,eAAe,EAAE,aAAc,KAAK,GAAG;KAC3C,MAAM;KACN,UAAU;MACR,MAAM,EAAE,aAAc;MACtB,WAAW,KAAK,UAAU,EAAE,aAAc,KAAK;MAChD;KACF,EAAE;IACJ,CAAC;QACG;IACL,MAAM,OAAO,UAAU,KAAK,MAAM,EAAE,KAAM,CAAC,KAAK,GAAG;AACnD,aAAS,KAAK;KAAE,MAAM;KAAa,SAAS;KAAM,CAAC;;;;CAO3D,IAAI;AACJ,KAAI,IAAI,SAAS,IAAI,MAAM,SAAS,GAAG;EACrC,MAAM,QAAQ,IAAI,MAAM,SAAS,MAAM,EAAE,wBAAwB,EAAE,CAAC;AACpE,MAAI,MAAM,SAAS,EACjB,SAAQ,MAAM,KAAK,OAAO;GACxB,MAAM;GACN,UAAU;IACR,MAAM,EAAE;IACR,aAAa,EAAE;IACf,YAAY,EAAE;IACf;GACF,EAAE;;AAIP,QAAO;EACL;EACA;EACA;EACA,aAAa,IAAI,kBAAkB;EACnC;EACD;;AAkBH,SAAS,4BACP,SACA,WACA,WACuB;CACvB,MAAM,SAAgC,EAAE;AAGxC,KAAI,UACF,MAAK,IAAI,IAAI,GAAG,IAAI,UAAU,QAAQ,KAAK,WAAW;EACpD,MAAM,QAAQ,UAAU,MAAM,GAAG,IAAI,UAAU;AAC/C,SAAO,KAAK,EACV,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAAO,CAAC;KAAE,MAAM;KAAO,SAAS;KAAM,CAAC;IAAE;GACnE,OAAO;GACR,CACF,EACF,CAAC;;AAKN,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK,WAAW;EAClD,MAAM,QAAQ,QAAQ,MAAM,GAAG,IAAI,UAAU;EAC7C,MAAM,SAAS,IAAI,aAAa,QAAQ;EACxC,MAAM,QAA6B;GACjC,YAAY,CACV;IACE,SAAS;KAAE,MAAM;KAAS,OAAO,CAAC,EAAE,MAAM,OAAO,CAAC;KAAE;IACpD,OAAO;IACP,GAAI,SAAS,EAAE,cAAc,QAAQ,GAAG,EAAE;IAC3C,CACF;GACD,GAAI,SACA,EACE,eAAe;IACb,kBAAkB;IAClB,sBAAsB;IACtB,iBAAiB;IAClB,EACF,GACD,EAAE;GACP;AACD,SAAO,KAAK,MAAM;;AAIpB,KAAI,QAAQ,WAAW,EACrB,QAAO,KAAK;EACV,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAAO,CAAC,EAAE,MAAM,IAAI,CAAC;IAAE;GACjD,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF,CAAC;AAGJ,QAAO;;AAGT,SAAS,kBAAkB,IAAc,QAA4B;CACnE,IAAI;AACJ,KAAI;AACF,YAAU,KAAK,MAAM,GAAG,aAAa,KAAK;SACpC;AACN,SAAO,KAAK,sDAAsD,GAAG,KAAK,KAAK,GAAG,YAAY;AAC9F,YAAU,EAAE;;AAEd,QAAO,EAAE,cAAc;EAAE,MAAM,GAAG;EAAM,MAAM;EAAS,IAAI,GAAG,MAAMA,oCAAoB;EAAE,EAAE;;AAG9F,SAAS,gCACP,WACA,QACuB;AAIvB,QAAO,CACL;EACE,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAPN,UAAU,KAAK,OAAO,kBAAkB,IAAI,OAAO,CAAC;IAOvC;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF,CACF;;AAKH,SAAS,wBAAwB,SAAiB,WAAyC;CACzF,MAAM,QAAsB,EAAE;AAC9B,KAAI,UACF,OAAM,KAAK;EAAE,MAAM;EAAW,SAAS;EAAM,CAAC;AAEhD,OAAM,KAAK,EAAE,MAAM,SAAS,CAAC;AAE7B,QAAO;EACL,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS;IAAO;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF;;AAGH,SAAS,4BAA4B,WAAuB,QAAqC;AAG/F,QAAO;EACL,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OALJ,UAAU,KAAK,OAAO,kBAAkB,IAAI,OAAO,CAAC;IAKzC;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF;;AAGH,SAAS,4CACP,SACA,WACA,WACA,QACuB;CACvB,MAAM,SAAgC,EAAE;AAExC,KAAI,QAAQ,WAAW,EACrB,QAAO,KAAK,EACV,YAAY,CACV;EACE,SAAS;GAAE,MAAM;GAAS,OAAO,CAAC,EAAE,MAAM,IAAI,CAAC;GAAE;EACjD,OAAO;EACR,CACF,EACF,CAAC;KAEF,MAAK,IAAI,IAAI,GAAG,IAAI,QAAQ,QAAQ,KAAK,WAAW;EAClD,MAAM,QAAQ,QAAQ,MAAM,GAAG,IAAI,UAAU;AAC7C,SAAO,KAAK,EACV,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OAAO,CAAC,EAAE,MAAM,OAAO,CAAC;IAAE;GACpD,OAAO;GACR,CACF,EACF,CAAC;;CAIN,MAAM,QAAsB,UAAU,KAAK,OAAO,kBAAkB,IAAI,OAAO,CAAC;AAEhF,QAAO,KAAK;EACV,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS;IAAO;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF,CAAC;AAEF,QAAO;;AAGT,SAAS,wCACP,SACA,WACA,QACqB;AAMrB,QAAO;EACL,YAAY,CACV;GACE,SAAS;IAAE,MAAM;IAAS,OARJ,CAC1B,EAAE,MAAM,SAAS,EACjB,GAAG,UAAU,KAAK,OAAO,kBAAkB,IAAI,OAAO,CAAC,CACxD;IAKsC;GACjC,cAAc;GACd,OAAO;GACR,CACF;EACD,eAAe;GACb,kBAAkB;GAClB,sBAAsB;GACtB,iBAAiB;GAClB;EACF;;AAYH,eAAe,qBACb,KACA,QACA,kBACkB;CAClB,MAAM,OACJ,OAAO,qBAAqB,WAAW,EAAE,SAAS,kBAAkB,GAAI,oBAAoB,EAAE;CAChG,MAAM,UAAU,KAAK,WAAW;CAChC,MAAM,UAAU,KAAK;CACrB,MAAM,SAAS,KAAK;CACpB,MAAM,cAAc,KAAK;AAEzB,KAAI,IAAI,cAAe,QAAO;AAC9B,KAAI,UAAU,gBAAgB,oBAAoB;AAClD,KAAI,UAAU,iBAAiB,WAAW;AAC1C,KAAI,UAAU,cAAc,aAAa;CAEzC,IAAI,aAAa;AACjB,MAAK,MAAM,SAAS,QAAQ;EAC1B,MAAM,aAAaC,kCAAe,YAAY,SAAS,QAAQ;AAC/D,MAAI,aAAa,EAAG,OAAMC,yBAAM,YAAY,OAAO;AACnD,MAAI,QAAQ,QAAS,QAAO;AAC5B,MAAI,IAAI,cAAe,QAAO;AAE9B,MAAI,MAAM,SAAS,KAAK,UAAU,MAAM,CAAC,MAAM;AAC/C,iBAAe;AACf,MAAI,QAAQ,QAAS,QAAO;AAC5B;;AAGF,KAAI,CAAC,IAAI,cACP,KAAI,KAAK;AAEX,QAAO;;AAKT,eAAsB,aACpB,KACA,KACA,KACA,OACA,WACA,UACA,SACA,UACA,gBACA,cAAiC,UAClB;CACf,MAAM,EAAE,WAAW;AACnB,gBAAe,IAAI;CAEnB,IAAI;AACJ,KAAI;AACF,cAAY,KAAK,MAAM,IAAI;SACrB;AACN,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB,MAAM,IAAI,OAAO,kBAAkB,MAAM;GACzC,SAASC,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK,SAAS;IAAM;GACzC,CAAC;AACF,wCACE,KACA,KACA,KAAK,UAAU,EACb,OAAO;GACL,SAAS;GACT,MAAM;GACN,QAAQ;GACT,EACF,CAAC,CACH;AACD;;CAIF,MAAM,gBAAgB,0BAA0B,WAAW,OAAO,UAAU;CAE5E,MAAM,SAASC,0BAAU,IAAI;CAC7B,MAAM,UAAUC,4BACd,UACA,eACA,QAAQ,6BAA6B,OAAO,EAC5C,SAAS,iBACV;CACD,MAAM,OAAO,IAAI,OAAO,kBAAkB,MAAM;AAEhD,KAAI,QACF,SAAQ,2BAA2B,SAAS,UAAU,OAAO;AAG/D,KACEC,yBACE,KACA,SACA,SAAS,OACT,IAAI,SACJ,SACA;EACE,QAAQ,IAAI,UAAU;EACtB;EACA,SAASH,+BAAe,IAAI,QAAQ;EACpC,MAAM;EACP,EACD,SAAS,UACT,SAAS,OACV,CAED;AAEF,KAAI,CAAC,SAAS;AACZ,MAAI,SAAS,QAWX;OAVgB,MAAMI,gCACpB,KACA,KACA,eACA,aACA,MACA,UACA,UACA,IACD,EACY;AACX,YAAQ,IAAI;KACV,QAAQ,IAAI,UAAU;KACtB;KACA,SAASJ,+BAAe,IAAI,QAAQ;KACpC,MAAM;KACN,UAAU;MAAE,QAAQ,IAAI,cAAc;MAAK,SAAS;MAAM;KAC3D,CAAC;AACF;;;EAGJ,MAAM,eAAe,SAAS,SAAS,MAAM;EAC7C,MAAM,gBAAgB,SAAS,SAC3B,oCACA;AACJ,MAAI,SAAS,OACX,QAAO,MAAM,kCAAkC,IAAI,UAAU,OAAO,GAAG,OAAO;AAEhF,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB;GACA,SAASA,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAc,SAAS;IAAM;GAClD,CAAC;AACF,wCACE,KACA,cACA,KAAK,UAAU,EACb,OAAO;GACL,SAAS;GACT,MAAM;GACN,QAAQ,SAAS,SAAS,gBAAgB;GAC3C,EACF,CAAC,CACH;AACD;;CAGF,MAAM,WAAW,QAAQ;CACzB,MAAM,UAAU,QAAQ,WAAW,SAAS;CAC5C,MAAM,YAAY,KAAK,IAAI,GAAG,QAAQ,aAAa,SAAS,UAAU;AAGtE,KAAIK,gCAAgB,SAAS,EAAE;EAC7B,MAAM,SAAS,SAAS,UAAU;AAClC,UAAQ,IAAI;GACV,QAAQ,IAAI,UAAU;GACtB;GACA,SAASL,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE;IAAQ;IAAS;GAC9B,CAAC;EAEF,MAAM,cAAc,EAClB,OAAO;GACL,MAAM;GACN,SAAS,SAAS,MAAM;GACxB,QAAQ,SAAS,MAAM,QAAQ;GAChC,EACF;AACD,wCAAmB,KAAK,QAAQ,KAAK,UAAU,YAAY,CAAC;AAC5D;;AAIF,KAAIM,+CAA+B,SAAS,EAAE;EAC5C,MAAM,eAAe,QAAQ,IAAI;GAC/B,QAAQ,IAAI,UAAU;GACtB;GACA,SAASN,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK;IAAS;GACnC,CAAC;AACF,MAAI,CAAC,WAAW;GACd,MAAM,OAAO,wCACX,SAAS,SACT,SAAS,WACT,OACD;AACD,OAAI,UAAU,KAAK,EAAE,gBAAgB,oBAAoB,CAAC;AAC1D,OAAI,IAAI,KAAK,UAAU,KAAK,CAAC;SACxB;GACL,MAAM,SAAS,4CACb,SAAS,SACT,SAAS,WACT,WACA,OACD;GACD,MAAM,eAAeO,8CAAyB,QAAQ;AAOtD,OAAI,CANc,MAAM,qBAAqB,KAAK,QAAQ;IACxD;IACA,kBAAkB,QAAQ;IAC1B,QAAQ,cAAc;IACtB,aAAa,cAAc;IAC5B,CAAC,EACc;AACd,QAAI,CAAC,IAAI,cAAe,KAAI,SAAS;AACrC,iBAAa,SAAS,cAAc;AACpC,iBAAa,SAAS,kBAAkB,cAAc,QAAQ;;AAEhE,iBAAc,SAAS;;AAEzB;;AAIF,KAAIC,+BAAe,SAAS,EAAE;EAC5B,MAAM,eAAe,QAAQ,IAAI;GAC/B,QAAQ,IAAI,UAAU;GACtB;GACA,SAASR,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK;IAAS;GACnC,CAAC;AACF,MAAI,CAAC,WAAW;GACd,MAAM,OAAO,wBAAwB,SAAS,SAAS,SAAS,UAAU;AAC1E,OAAI,UAAU,KAAK,EAAE,gBAAgB,oBAAoB,CAAC;AAC1D,OAAI,IAAI,KAAK,UAAU,KAAK,CAAC;SACxB;GACL,MAAM,SAAS,4BAA4B,SAAS,SAAS,WAAW,SAAS,UAAU;GAC3F,MAAM,eAAeO,8CAAyB,QAAQ;AAOtD,OAAI,CANc,MAAM,qBAAqB,KAAK,QAAQ;IACxD;IACA,kBAAkB,QAAQ;IAC1B,QAAQ,cAAc;IACtB,aAAa,cAAc;IAC5B,CAAC,EACc;AACd,QAAI,CAAC,IAAI,cAAe,KAAI,SAAS;AACrC,iBAAa,SAAS,cAAc;AACpC,iBAAa,SAAS,kBAAkB,cAAc,QAAQ;;AAEhE,iBAAc,SAAS;;AAEzB;;AAIF,KAAIE,mCAAmB,SAAS,EAAE;EAChC,MAAM,eAAe,QAAQ,IAAI;GAC/B,QAAQ,IAAI,UAAU;GACtB;GACA,SAAST,+BAAe,IAAI,QAAQ;GACpC,MAAM;GACN,UAAU;IAAE,QAAQ;IAAK;IAAS;GACnC,CAAC;AACF,MAAI,CAAC,WAAW;GACd,MAAM,OAAO,4BAA4B,SAAS,WAAW,OAAO;AACpE,OAAI,UAAU,KAAK,EAAE,gBAAgB,oBAAoB,CAAC;AAC1D,OAAI,IAAI,KAAK,UAAU,KAAK,CAAC;SACxB;GACL,MAAM,SAAS,gCAAgC,SAAS,WAAW,OAAO;GAC1E,MAAM,eAAeO,8CAAyB,QAAQ;AAOtD,OAAI,CANc,MAAM,qBAAqB,KAAK,QAAQ;IACxD;IACA,kBAAkB,QAAQ;IAC1B,QAAQ,cAAc;IACtB,aAAa,cAAc;IAC5B,CAAC,EACc;AACd,QAAI,CAAC,IAAI,cAAe,KAAI,SAAS;AACrC,iBAAa,SAAS,cAAc;AACpC,iBAAa,SAAS,kBAAkB,cAAc,QAAQ;;AAEhE,iBAAc,SAAS;;AAEzB;;AAIF,SAAQ,IAAI;EACV,QAAQ,IAAI,UAAU;EACtB;EACA,SAASP,+BAAe,IAAI,QAAQ;EACpC,MAAM;EACN,UAAU;GAAE,QAAQ;GAAK;GAAS;EACnC,CAAC;AACF,uCACE,KACA,KACA,KAAK,UAAU,EACb,OAAO;EACL,SAAS;EACT,MAAM;EACN,QAAQ;EACT,EACF,CAAC,CACH"}
|
package/dist/gemini.d.cts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"gemini.d.cts","names":[],"sources":["../src/gemini.ts"],"sourcesContent":[],"mappings":";;;;;;
|
|
1
|
+
{"version":3,"file":"gemini.d.cts","names":[],"sources":["../src/gemini.ts"],"sourcesContent":[],"mappings":";;;;;;iBAgdsB,YAAA,MACf,IAAA,CAAK,sBACL,IAAA,CAAK,0EAIA,oBACD,mBACC,uCACY,IAAA,CAAK,uCACd,oBACZ"}
|
package/dist/gemini.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"gemini.d.ts","names":[],"sources":["../src/gemini.ts"],"sourcesContent":[],"mappings":";;;;;;
|
|
1
|
+
{"version":3,"file":"gemini.d.ts","names":[],"sources":["../src/gemini.ts"],"sourcesContent":[],"mappings":";;;;;;iBAgdsB,YAAA,MACf,IAAA,CAAK,sBACL,IAAA,CAAK,0EAIA,oBACD,mBACC,uCACY,IAAA,CAAK,uCACd,oBACZ"}
|
package/dist/gemini.js
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { flattenHeaders, generateToolCallId, isErrorResponse, isTextResponse, isToolCallResponse } from "./helpers.js";
|
|
1
|
+
import { flattenHeaders, generateToolCallId, getTestId, isContentWithToolCallsResponse, isErrorResponse, isTextResponse, isToolCallResponse } from "./helpers.js";
|
|
2
2
|
import { matchFixture } from "./router.js";
|
|
3
3
|
import { calculateDelay, delay, writeErrorResponse } from "./sse-writer.js";
|
|
4
4
|
import { createInterruptionSignal } from "./interruption.js";
|
|
@@ -84,8 +84,21 @@ function geminiToCompletionRequest(req, model, stream) {
|
|
|
84
84
|
tools
|
|
85
85
|
};
|
|
86
86
|
}
|
|
87
|
-
function buildGeminiTextStreamChunks(content, chunkSize) {
|
|
87
|
+
function buildGeminiTextStreamChunks(content, chunkSize, reasoning) {
|
|
88
88
|
const chunks = [];
|
|
89
|
+
if (reasoning) for (let i = 0; i < reasoning.length; i += chunkSize) {
|
|
90
|
+
const slice = reasoning.slice(i, i + chunkSize);
|
|
91
|
+
chunks.push({ candidates: [{
|
|
92
|
+
content: {
|
|
93
|
+
role: "model",
|
|
94
|
+
parts: [{
|
|
95
|
+
text: slice,
|
|
96
|
+
thought: true
|
|
97
|
+
}]
|
|
98
|
+
},
|
|
99
|
+
index: 0
|
|
100
|
+
}] });
|
|
101
|
+
}
|
|
89
102
|
for (let i = 0; i < content.length; i += chunkSize) {
|
|
90
103
|
const slice = content.slice(i, i + chunkSize);
|
|
91
104
|
const isLast = i + chunkSize >= content.length;
|
|
@@ -123,25 +136,26 @@ function buildGeminiTextStreamChunks(content, chunkSize) {
|
|
|
123
136
|
});
|
|
124
137
|
return chunks;
|
|
125
138
|
}
|
|
139
|
+
function parseToolCallPart(tc, logger) {
|
|
140
|
+
let argsObj;
|
|
141
|
+
try {
|
|
142
|
+
argsObj = JSON.parse(tc.arguments || "{}");
|
|
143
|
+
} catch {
|
|
144
|
+
logger.warn(`Malformed JSON in fixture tool call arguments for "${tc.name}": ${tc.arguments}`);
|
|
145
|
+
argsObj = {};
|
|
146
|
+
}
|
|
147
|
+
return { functionCall: {
|
|
148
|
+
name: tc.name,
|
|
149
|
+
args: argsObj,
|
|
150
|
+
id: tc.id || generateToolCallId()
|
|
151
|
+
} };
|
|
152
|
+
}
|
|
126
153
|
function buildGeminiToolCallStreamChunks(toolCalls, logger) {
|
|
127
154
|
return [{
|
|
128
155
|
candidates: [{
|
|
129
156
|
content: {
|
|
130
157
|
role: "model",
|
|
131
|
-
parts: toolCalls.map((tc) =>
|
|
132
|
-
let argsObj;
|
|
133
|
-
try {
|
|
134
|
-
argsObj = JSON.parse(tc.arguments || "{}");
|
|
135
|
-
} catch {
|
|
136
|
-
logger.warn(`Malformed JSON in fixture tool call arguments for "${tc.name}": ${tc.arguments}`);
|
|
137
|
-
argsObj = {};
|
|
138
|
-
}
|
|
139
|
-
return { functionCall: {
|
|
140
|
-
name: tc.name,
|
|
141
|
-
args: argsObj,
|
|
142
|
-
id: tc.id || generateToolCallId()
|
|
143
|
-
} };
|
|
144
|
-
})
|
|
158
|
+
parts: toolCalls.map((tc) => parseToolCallPart(tc, logger))
|
|
145
159
|
},
|
|
146
160
|
finishReason: "FUNCTION_CALL",
|
|
147
161
|
index: 0
|
|
@@ -153,12 +167,18 @@ function buildGeminiToolCallStreamChunks(toolCalls, logger) {
|
|
|
153
167
|
}
|
|
154
168
|
}];
|
|
155
169
|
}
|
|
156
|
-
function buildGeminiTextResponse(content) {
|
|
170
|
+
function buildGeminiTextResponse(content, reasoning) {
|
|
171
|
+
const parts = [];
|
|
172
|
+
if (reasoning) parts.push({
|
|
173
|
+
text: reasoning,
|
|
174
|
+
thought: true
|
|
175
|
+
});
|
|
176
|
+
parts.push({ text: content });
|
|
157
177
|
return {
|
|
158
178
|
candidates: [{
|
|
159
179
|
content: {
|
|
160
180
|
role: "model",
|
|
161
|
-
parts
|
|
181
|
+
parts
|
|
162
182
|
},
|
|
163
183
|
finishReason: "STOP",
|
|
164
184
|
index: 0
|
|
@@ -175,20 +195,61 @@ function buildGeminiToolCallResponse(toolCalls, logger) {
|
|
|
175
195
|
candidates: [{
|
|
176
196
|
content: {
|
|
177
197
|
role: "model",
|
|
178
|
-
parts: toolCalls.map((tc) =>
|
|
179
|
-
|
|
180
|
-
|
|
181
|
-
|
|
182
|
-
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
|
|
198
|
+
parts: toolCalls.map((tc) => parseToolCallPart(tc, logger))
|
|
199
|
+
},
|
|
200
|
+
finishReason: "FUNCTION_CALL",
|
|
201
|
+
index: 0
|
|
202
|
+
}],
|
|
203
|
+
usageMetadata: {
|
|
204
|
+
promptTokenCount: 0,
|
|
205
|
+
candidatesTokenCount: 0,
|
|
206
|
+
totalTokenCount: 0
|
|
207
|
+
}
|
|
208
|
+
};
|
|
209
|
+
}
|
|
210
|
+
function buildGeminiContentWithToolCallsStreamChunks(content, toolCalls, chunkSize, logger) {
|
|
211
|
+
const chunks = [];
|
|
212
|
+
if (content.length === 0) chunks.push({ candidates: [{
|
|
213
|
+
content: {
|
|
214
|
+
role: "model",
|
|
215
|
+
parts: [{ text: "" }]
|
|
216
|
+
},
|
|
217
|
+
index: 0
|
|
218
|
+
}] });
|
|
219
|
+
else for (let i = 0; i < content.length; i += chunkSize) {
|
|
220
|
+
const slice = content.slice(i, i + chunkSize);
|
|
221
|
+
chunks.push({ candidates: [{
|
|
222
|
+
content: {
|
|
223
|
+
role: "model",
|
|
224
|
+
parts: [{ text: slice }]
|
|
225
|
+
},
|
|
226
|
+
index: 0
|
|
227
|
+
}] });
|
|
228
|
+
}
|
|
229
|
+
const parts = toolCalls.map((tc) => parseToolCallPart(tc, logger));
|
|
230
|
+
chunks.push({
|
|
231
|
+
candidates: [{
|
|
232
|
+
content: {
|
|
233
|
+
role: "model",
|
|
234
|
+
parts
|
|
235
|
+
},
|
|
236
|
+
finishReason: "FUNCTION_CALL",
|
|
237
|
+
index: 0
|
|
238
|
+
}],
|
|
239
|
+
usageMetadata: {
|
|
240
|
+
promptTokenCount: 0,
|
|
241
|
+
candidatesTokenCount: 0,
|
|
242
|
+
totalTokenCount: 0
|
|
243
|
+
}
|
|
244
|
+
});
|
|
245
|
+
return chunks;
|
|
246
|
+
}
|
|
247
|
+
function buildGeminiContentWithToolCallsResponse(content, toolCalls, logger) {
|
|
248
|
+
return {
|
|
249
|
+
candidates: [{
|
|
250
|
+
content: {
|
|
251
|
+
role: "model",
|
|
252
|
+
parts: [{ text: content }, ...toolCalls.map((tc) => parseToolCallPart(tc, logger))]
|
|
192
253
|
},
|
|
193
254
|
finishReason: "FUNCTION_CALL",
|
|
194
255
|
index: 0
|
|
@@ -249,9 +310,10 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
249
310
|
return;
|
|
250
311
|
}
|
|
251
312
|
const completionReq = geminiToCompletionRequest(geminiReq, model, streaming);
|
|
252
|
-
const
|
|
313
|
+
const testId = getTestId(req);
|
|
314
|
+
const fixture = matchFixture(fixtures, completionReq, journal.getFixtureMatchCountsForTest(testId), defaults.requestTransform);
|
|
253
315
|
const path = req.url ?? `/v1beta/models/${model}:generateContent`;
|
|
254
|
-
if (fixture) journal.incrementFixtureMatchCount(fixture, fixtures);
|
|
316
|
+
if (fixture) journal.incrementFixtureMatchCount(fixture, fixtures, testId);
|
|
255
317
|
if (applyChaos(res, fixture, defaults.chaos, req.headers, journal, {
|
|
256
318
|
method: req.method ?? "POST",
|
|
257
319
|
path,
|
|
@@ -317,6 +379,38 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
317
379
|
writeErrorResponse(res, status, JSON.stringify(geminiError));
|
|
318
380
|
return;
|
|
319
381
|
}
|
|
382
|
+
if (isContentWithToolCallsResponse(response)) {
|
|
383
|
+
const journalEntry = journal.add({
|
|
384
|
+
method: req.method ?? "POST",
|
|
385
|
+
path,
|
|
386
|
+
headers: flattenHeaders(req.headers),
|
|
387
|
+
body: completionReq,
|
|
388
|
+
response: {
|
|
389
|
+
status: 200,
|
|
390
|
+
fixture
|
|
391
|
+
}
|
|
392
|
+
});
|
|
393
|
+
if (!streaming) {
|
|
394
|
+
const body = buildGeminiContentWithToolCallsResponse(response.content, response.toolCalls, logger);
|
|
395
|
+
res.writeHead(200, { "Content-Type": "application/json" });
|
|
396
|
+
res.end(JSON.stringify(body));
|
|
397
|
+
} else {
|
|
398
|
+
const chunks = buildGeminiContentWithToolCallsStreamChunks(response.content, response.toolCalls, chunkSize, logger);
|
|
399
|
+
const interruption = createInterruptionSignal(fixture);
|
|
400
|
+
if (!await writeGeminiSSEStream(res, chunks, {
|
|
401
|
+
latency,
|
|
402
|
+
streamingProfile: fixture.streamingProfile,
|
|
403
|
+
signal: interruption?.signal,
|
|
404
|
+
onChunkSent: interruption?.tick
|
|
405
|
+
})) {
|
|
406
|
+
if (!res.writableEnded) res.destroy();
|
|
407
|
+
journalEntry.response.interrupted = true;
|
|
408
|
+
journalEntry.response.interruptReason = interruption?.reason();
|
|
409
|
+
}
|
|
410
|
+
interruption?.cleanup();
|
|
411
|
+
}
|
|
412
|
+
return;
|
|
413
|
+
}
|
|
320
414
|
if (isTextResponse(response)) {
|
|
321
415
|
const journalEntry = journal.add({
|
|
322
416
|
method: req.method ?? "POST",
|
|
@@ -329,11 +423,11 @@ async function handleGemini(req, res, raw, model, streaming, fixtures, journal,
|
|
|
329
423
|
}
|
|
330
424
|
});
|
|
331
425
|
if (!streaming) {
|
|
332
|
-
const body = buildGeminiTextResponse(response.content);
|
|
426
|
+
const body = buildGeminiTextResponse(response.content, response.reasoning);
|
|
333
427
|
res.writeHead(200, { "Content-Type": "application/json" });
|
|
334
428
|
res.end(JSON.stringify(body));
|
|
335
429
|
} else {
|
|
336
|
-
const chunks = buildGeminiTextStreamChunks(response.content, chunkSize);
|
|
430
|
+
const chunks = buildGeminiTextStreamChunks(response.content, chunkSize, response.reasoning);
|
|
337
431
|
const interruption = createInterruptionSignal(fixture);
|
|
338
432
|
if (!await writeGeminiSSEStream(res, chunks, {
|
|
339
433
|
latency,
|