@genui-a3/providers 0.0.4 → 0.0.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -0
- package/dist/anthropic/index.cjs +115 -12
- package/dist/anthropic/index.cjs.map +1 -1
- package/dist/anthropic/index.d.cts +3 -1
- package/dist/anthropic/index.d.ts +3 -1
- package/dist/anthropic/index.js +115 -12
- package/dist/anthropic/index.js.map +1 -1
- package/dist/bedrock/index.cjs +115 -12
- package/dist/bedrock/index.cjs.map +1 -1
- package/dist/bedrock/index.d.cts +3 -1
- package/dist/bedrock/index.d.ts +3 -1
- package/dist/bedrock/index.js +115 -12
- package/dist/bedrock/index.js.map +1 -1
- package/dist/openai/index.cjs +115 -12
- package/dist/openai/index.cjs.map +1 -1
- package/dist/openai/index.d.cts +3 -1
- package/dist/openai/index.d.ts +3 -1
- package/dist/openai/index.js +115 -12
- package/dist/openai/index.js.map +1 -1
- package/package.json +1 -1
package/dist/openai/index.cjs
CHANGED
|
@@ -2,6 +2,7 @@
|
|
|
2
2
|
|
|
3
3
|
var openai = require('@ai-sdk/openai');
|
|
4
4
|
var ai = require('ai');
|
|
5
|
+
var core = require('@genui-a3/core');
|
|
5
6
|
var client = require('@ag-ui/client');
|
|
6
7
|
|
|
7
8
|
// openai/index.ts
|
|
@@ -69,20 +70,119 @@ function extractDelta(partial, prevLength) {
|
|
|
69
70
|
return chatbotMessage.slice(prevLength);
|
|
70
71
|
}
|
|
71
72
|
|
|
73
|
+
// utils/backoff.ts
|
|
74
|
+
function calculateBackoff(attempt, config) {
|
|
75
|
+
let delay;
|
|
76
|
+
switch (config.strategy) {
|
|
77
|
+
case "exponential":
|
|
78
|
+
delay = config.baseDelayMs * Math.pow(2, attempt);
|
|
79
|
+
break;
|
|
80
|
+
case "linear":
|
|
81
|
+
delay = config.baseDelayMs * (attempt + 1);
|
|
82
|
+
break;
|
|
83
|
+
case "fixed":
|
|
84
|
+
delay = config.baseDelayMs;
|
|
85
|
+
break;
|
|
86
|
+
}
|
|
87
|
+
delay = Math.min(delay, config.maxDelayMs);
|
|
88
|
+
if (config.jitter) {
|
|
89
|
+
delay = Math.random() * delay;
|
|
90
|
+
}
|
|
91
|
+
return delay;
|
|
92
|
+
}
|
|
93
|
+
function sleep(ms, signal) {
|
|
94
|
+
return new Promise((resolve, reject) => {
|
|
95
|
+
if (signal?.aborted) {
|
|
96
|
+
reject(signal.reason);
|
|
97
|
+
return;
|
|
98
|
+
}
|
|
99
|
+
const timer = setTimeout(resolve, ms);
|
|
100
|
+
signal?.addEventListener(
|
|
101
|
+
"abort",
|
|
102
|
+
() => {
|
|
103
|
+
clearTimeout(timer);
|
|
104
|
+
reject(signal.reason);
|
|
105
|
+
},
|
|
106
|
+
{ once: true }
|
|
107
|
+
);
|
|
108
|
+
});
|
|
109
|
+
}
|
|
110
|
+
|
|
72
111
|
// utils/executeWithFallback.ts
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
112
|
+
function buildSignal(requestTimeoutMs, totalAbort) {
|
|
113
|
+
const signals = [];
|
|
114
|
+
if (requestTimeoutMs !== void 0) {
|
|
115
|
+
signals.push(AbortSignal.timeout(requestTimeoutMs));
|
|
116
|
+
}
|
|
117
|
+
if (totalAbort) {
|
|
118
|
+
signals.push(totalAbort);
|
|
119
|
+
}
|
|
120
|
+
if (signals.length === 0) return void 0;
|
|
121
|
+
if (signals.length === 1) return signals[0];
|
|
122
|
+
return AbortSignal.any(signals);
|
|
123
|
+
}
|
|
124
|
+
async function attemptAction(action, model, signal, attempt, errors) {
|
|
125
|
+
try {
|
|
126
|
+
const value = await action(model, signal);
|
|
127
|
+
return { ok: true, value };
|
|
128
|
+
} catch (error) {
|
|
129
|
+
const errorObj = error;
|
|
130
|
+
errors.push({ model, attempt, error: errorObj });
|
|
131
|
+
return { ok: false, error: errorObj };
|
|
132
|
+
}
|
|
133
|
+
}
|
|
134
|
+
function checkTotalTimeout(totalAbort, timeoutMs, errors) {
|
|
135
|
+
if (totalAbort?.signal.aborted) {
|
|
136
|
+
throw new core.A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors);
|
|
137
|
+
}
|
|
138
|
+
}
|
|
139
|
+
async function handleAttemptError(errorObj, attempt, maxRetries, retryAll, resolved, totalAbort) {
|
|
140
|
+
const isLastAttempt = attempt === 1 + maxRetries;
|
|
141
|
+
const isRetryable = retryAll || resolved.isRetryableError(errorObj);
|
|
142
|
+
if (isRetryable && !isLastAttempt) {
|
|
143
|
+
const delay = calculateBackoff(attempt - 1, resolved.backoff);
|
|
144
|
+
await sleep(delay, totalAbort?.signal).catch(() => {
|
|
145
|
+
});
|
|
146
|
+
return "retry";
|
|
147
|
+
}
|
|
148
|
+
return "next-model";
|
|
149
|
+
}
|
|
150
|
+
async function executeWithFallback(models, action, config) {
|
|
151
|
+
const resolved = config ?? core.DEFAULT_RESILIENCE_CONFIG;
|
|
152
|
+
const errors = [];
|
|
153
|
+
const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts;
|
|
154
|
+
const retryAll = resolved.retry !== false && resolved.retry.retryOn === "all";
|
|
155
|
+
let totalAbort;
|
|
156
|
+
let totalTimer;
|
|
157
|
+
if (resolved.timeout.totalTimeoutMs !== void 0) {
|
|
158
|
+
totalAbort = new AbortController();
|
|
159
|
+
totalTimer = setTimeout(
|
|
160
|
+
() => totalAbort.abort(new Error("Total timeout exceeded")),
|
|
161
|
+
resolved.timeout.totalTimeoutMs
|
|
162
|
+
);
|
|
163
|
+
}
|
|
164
|
+
try {
|
|
165
|
+
for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {
|
|
166
|
+
const model = models[modelIndex];
|
|
167
|
+
for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {
|
|
168
|
+
checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
|
|
169
|
+
const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal);
|
|
170
|
+
const result = await attemptAction(action, model, signal, attempt, errors);
|
|
171
|
+
if (result.ok) return result.value;
|
|
172
|
+
checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
|
|
173
|
+
const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort);
|
|
174
|
+
if (decision === "next-model") break;
|
|
82
175
|
}
|
|
83
176
|
}
|
|
177
|
+
throw new core.A3ResilienceError(
|
|
178
|
+
`All models failed after ${errors.length} total attempt(s): ${models.join(", ")}`,
|
|
179
|
+
errors
|
|
180
|
+
);
|
|
181
|
+
} finally {
|
|
182
|
+
if (totalTimer !== void 0) {
|
|
183
|
+
clearTimeout(totalTimer);
|
|
184
|
+
}
|
|
84
185
|
}
|
|
85
|
-
throw new Error("All models failed");
|
|
86
186
|
}
|
|
87
187
|
|
|
88
188
|
// openai/index.ts
|
|
@@ -158,20 +258,23 @@ function createOpenAIProvider(config) {
|
|
|
158
258
|
organization: config.organization
|
|
159
259
|
});
|
|
160
260
|
const models = config.models;
|
|
261
|
+
const resilience = core.resolveResilienceConfig(config.resilience);
|
|
161
262
|
return {
|
|
162
263
|
name: "openai",
|
|
163
264
|
async sendRequest(request) {
|
|
164
265
|
const messages = toAIMessages(request.messages);
|
|
165
266
|
return executeWithFallback(
|
|
166
267
|
models,
|
|
167
|
-
(model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
|
|
268
|
+
(model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
|
|
269
|
+
resilience
|
|
168
270
|
);
|
|
169
271
|
},
|
|
170
272
|
async *sendRequestStream(request) {
|
|
171
273
|
const messages = toAIMessages(request.messages);
|
|
172
274
|
const { result, reader, first } = await executeWithFallback(
|
|
173
275
|
models,
|
|
174
|
-
(model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
|
|
276
|
+
(model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
|
|
277
|
+
resilience
|
|
175
278
|
);
|
|
176
279
|
yield* processOpenAIStream(result, reader, first, "openai", request.responseSchema);
|
|
177
280
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":["EventType","jsonSchema","generateText","Output","streamText","createOpenAI"],"mappings":";;;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;ACnFA,eAAsB,mBAAA,CAAuB,QAAkB,MAAA,EAAmD;AAGhH,EAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,MAAA,CAAO,QAAQ,CAAA,EAAA,EAAK;AACtC,IAAA,MAAM,KAAA,GAAQ,OAAO,CAAC,CAAA;AAEtB,IAAA,IAAI;AAEF,MAAA,OAAO,MAAM,OAAO,KAAK,CAAA;AAAA,IAC3B,SAAS,KAAA,EAAO;AACd,MAAA,MAAM,QAAA,GAAW,KAAA;AAGjB,MAAA,IAAI,CAAA,KAAM,MAAA,CAAO,MAAA,GAAS,CAAA,EAAG;AAC3B,QAAA,MAAM,QAAA;AAAA,MACR;AAAA,IACF;AAAA,EACF;AAEA,EAAA,MAAM,IAAI,MAAM,mBAAmB,CAAA;AACrC;;;ACEA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAOC,cAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAMC,eAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQC,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAASC,aAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQD,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiBE,mBAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AAEtB,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UAClC,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OAC7F;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UACnE,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OACnG;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.cjs","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","/**\n * Executes an action with model fallback support.\n * Tries each model in order; if one fails, falls back to the next.\n * Throws the last error if all models fail.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model\n * @returns The result from the first successful model\n * @throws The error from the last model if all fail\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model) => provider.call(model, params),\n * )\n * ```\n */\nexport async function executeWithFallback<T>(models: string[], action: (model: string) => Promise<T>): Promise<T> {\n const errors: Array<{ model: string; error: Error }> = []\n\n for (let i = 0; i < models.length; i++) {\n const model = models[i]\n\n try {\n // eslint-disable-next-line no-await-in-loop\n return await action(model)\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, error: errorObj })\n\n if (i === models.length - 1) {\n throw errorObj\n }\n }\n }\n\n throw new Error('All models failed')\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport type {\n Provider,\n ProviderRequest,\n ProviderResponse,\n ProviderMessage,\n BaseState,\n StreamEvent,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '../utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(models, (model) =>\n sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(models, (model) =>\n sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/backoff.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":["EventType","A3TimeoutError","DEFAULT_RESILIENCE_CONFIG","A3ResilienceError","jsonSchema","generateText","Output","streamText","createOpenAI","resolveResilienceConfig"],"mappings":";;;;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;AC5FO,SAAS,gBAAA,CAAiB,SAAiB,MAAA,EAAyC;AACzF,EAAA,IAAI,KAAA;AAEJ,EAAA,QAAQ,OAAO,QAAA;AAAU,IACvB,KAAK,aAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA,GAAc,IAAA,CAAK,GAAA,CAAI,GAAG,OAAO,CAAA;AAChD,MAAA;AAAA,IACF,KAAK,QAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,eAAe,OAAA,GAAU,CAAA,CAAA;AACxC,MAAA;AAAA,IACF,KAAK,OAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA;AACf,MAAA;AAAA;AAGJ,EAAA,KAAA,GAAQ,IAAA,CAAK,GAAA,CAAI,KAAA,EAAO,MAAA,CAAO,UAAU,CAAA;AAEzC,EAAA,IAAI,OAAO,MAAA,EAAQ;AACjB,IAAA,KAAA,GAAQ,IAAA,CAAK,QAAO,GAAI,KAAA;AAAA,EAC1B;AAEA,EAAA,OAAO,KAAA;AACT;AAQO,SAAS,KAAA,CAAM,IAAY,MAAA,EAAqC;AACrE,EAAA,OAAO,IAAI,OAAA,CAAc,CAAC,OAAA,EAAS,MAAA,KAAW;AAC5C,IAAA,IAAI,QAAQ,OAAA,EAAS;AACnB,MAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAC7B,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,KAAA,GAAQ,UAAA,CAAW,OAAA,EAAS,EAAE,CAAA;AAEpC,IAAA,MAAA,EAAQ,gBAAA;AAAA,MACN,OAAA;AAAA,MACA,MAAM;AACJ,QAAA,YAAA,CAAa,KAAK,CAAA;AAClB,QAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAAA,MAC/B,CAAA;AAAA,MACA,EAAE,MAAM,IAAA;AAAK,KACf;AAAA,EACF,CAAC,CAAA;AACH;;;AC7CA,SAAS,WAAA,CAAY,kBAA2B,UAAA,EAAmD;AACjG,EAAA,MAAM,UAAyB,EAAC;AAEhC,EAAA,IAAI,qBAAqB,MAAA,EAAW;AAClC,IAAA,OAAA,CAAQ,IAAA,CAAK,WAAA,CAAY,OAAA,CAAQ,gBAAgB,CAAC,CAAA;AAAA,EACpD;AAEA,EAAA,IAAI,UAAA,EAAY;AACd,IAAA,OAAA,CAAQ,KAAK,UAAU,CAAA;AAAA,EACzB;AAEA,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,MAAA;AACjC,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,QAAQ,CAAC,CAAA;AAC1C,EAAA,OAAO,WAAA,CAAY,IAAI,OAAO,CAAA;AAChC;AAIA,eAAe,aAAA,CACb,MAAA,EACA,KAAA,EACA,MAAA,EACA,SACA,MAAA,EAC2B;AAC3B,EAAA,IAAI;AACF,IAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,KAAA,EAAO,MAAM,CAAA;AACxC,IAAA,OAAO,EAAE,EAAA,EAAI,IAAA,EAAM,KAAA,EAAM;AAAA,EAC3B,SAAS,KAAA,EAAO;AACd,IAAA,MAAM,QAAA,GAAW,KAAA;AACjB,IAAA,MAAA,CAAO,KAAK,EAAE,KAAA,EAAO,OAAA,EAAS,KAAA,EAAO,UAAU,CAAA;AAC/C,IAAA,OAAO,EAAE,EAAA,EAAI,KAAA,EAAO,KAAA,EAAO,QAAA,EAAS;AAAA,EACtC;AACF;AAEA,SAAS,iBAAA,CAAkB,UAAA,EAAyC,SAAA,EAA+B,MAAA,EAAsC;AACvI,EAAA,IAAI,UAAA,EAAY,OAAO,OAAA,EAAS;AAC9B,IAAA,MAAM,IAAIC,mBAAA,CAAe,CAAA,iBAAA,EAAoB,SAAS,eAAe,MAAM,CAAA;AAAA,EAC7E;AACF;AAEA,eAAe,mBACb,QAAA,EACA,OAAA,EACA,UAAA,EACA,QAAA,EACA,UACA,UAAA,EACiC;AACjC,EAAA,MAAM,aAAA,GAAgB,YAAY,CAAA,GAAI,UAAA;AACtC,EAAA,MAAM,WAAA,GAAc,QAAA,IAAY,QAAA,CAAS,gBAAA,CAAiB,QAAQ,CAAA;AAElE,EAAA,IAAI,WAAA,IAAe,CAAC,aAAA,EAAe;AACjC,IAAA,MAAM,KAAA,GAAQ,gBAAA,CAAiB,OAAA,GAAU,CAAA,EAAG,SAAS,OAAO,CAAA;AAC5D,IAAA,MAAM,MAAM,KAAA,EAAO,UAAA,EAAY,MAAM,CAAA,CAAE,MAAM,MAAM;AAAA,IAEnD,CAAC,CAAA;AACD,IAAA,OAAO,OAAA;AAAA,EACT;AAEA,EAAA,OAAO,YAAA;AACT;AA6BA,eAAsB,mBAAA,CACpB,MAAA,EACA,MAAA,EACA,MAAA,EACY;AACZ,EAAA,MAAM,WAAW,MAAA,IAAUC,8BAAA;AAC3B,EAAA,MAAM,SAAiC,EAAC;AACxC,EAAA,MAAM,aAAa,QAAA,CAAS,KAAA,KAAU,KAAA,GAAQ,CAAA,GAAI,SAAS,KAAA,CAAM,WAAA;AACjE,EAAA,MAAM,WAAW,QAAA,CAAS,KAAA,KAAU,KAAA,IAAS,QAAA,CAAS,MAAM,OAAA,KAAY,KAAA;AAGxE,EAAA,IAAI,UAAA;AACJ,EAAA,IAAI,UAAA;AAEJ,EAAA,IAAI,QAAA,CAAS,OAAA,CAAQ,cAAA,KAAmB,MAAA,EAAW;AACjD,IAAA,UAAA,GAAa,IAAI,eAAA,EAAgB;AACjC,IAAA,UAAA,GAAa,UAAA;AAAA,MACX,MAAM,UAAA,CAAY,KAAA,CAAM,IAAI,KAAA,CAAM,wBAAwB,CAAC,CAAA;AAAA,MAC3D,SAAS,OAAA,CAAQ;AAAA,KACnB;AAAA,EACF;AAEA,EAAA,IAAI;AACF,IAAA,KAAA,IAAS,UAAA,GAAa,CAAA,EAAG,UAAA,GAAa,MAAA,CAAO,QAAQ,UAAA,EAAA,EAAc;AACjE,MAAA,MAAM,KAAA,GAAQ,OAAO,UAAU,CAAA;AAE/B,MAAA,KAAA,IAAS,OAAA,GAAU,CAAA,EAAG,OAAA,IAAW,CAAA,GAAI,YAAY,OAAA,EAAA,EAAW;AAC1D,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAErE,QAAA,MAAM,SAAS,WAAA,CAAY,QAAA,CAAS,OAAA,CAAQ,gBAAA,EAAkB,YAAY,MAAM,CAAA;AAEhF,QAAA,MAAM,SAAS,MAAM,aAAA,CAAc,QAAQ,KAAA,EAAO,MAAA,EAAQ,SAAS,MAAM,CAAA;AACzE,QAAA,IAAI,MAAA,CAAO,EAAA,EAAI,OAAO,MAAA,CAAO,KAAA;AAE7B,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAGrE,QAAA,MAAM,QAAA,GAAW,MAAM,kBAAA,CAAmB,MAAA,CAAO,OAAO,OAAA,EAAS,UAAA,EAAY,QAAA,EAAU,QAAA,EAAU,UAAU,CAAA;AAC3G,QAAA,IAAI,aAAa,YAAA,EAAc;AAAA,MACjC;AAAA,IACF;AAGA,IAAA,MAAM,IAAIC,sBAAA;AAAA,MACR,2BAA2B,MAAA,CAAO,MAAM,sBAAsB,MAAA,CAAO,IAAA,CAAK,IAAI,CAAC,CAAA,CAAA;AAAA,MAC/E;AAAA,KACF;AAAA,EACF,CAAA,SAAE;AACA,IAAA,IAAI,eAAe,MAAA,EAAW;AAC5B,MAAA,YAAA,CAAa,UAAU,CAAA;AAAA,IACzB;AAAA,EACF;AACF;;;AC7GA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAOC,cAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAMC,eAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQC,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAASC,aAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQD,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiBE,mBAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AACtB,EAAA,MAAM,UAAA,GAAuCC,4BAAA,CAAwB,MAAA,CAAO,UAAU,CAAA;AAEtF,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QACL,MAAA;AAAA,QACA,CAAC,UAAU,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QACtG;AAAA,OACF;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QACtC,MAAA;AAAA,QACA,CAAC,UAAU,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QAC5G;AAAA,OACF;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.cjs","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","import type { BackoffConfig } from '@genui-a3/core'\n\n/**\n * Calculates the backoff delay for a given retry attempt.\n *\n * @param attempt - Zero-based attempt index (0 = first retry)\n * @param config - Backoff configuration with all fields required\n * @returns Delay in milliseconds\n */\nexport function calculateBackoff(attempt: number, config: Required<BackoffConfig>): number {\n let delay: number\n\n switch (config.strategy) {\n case 'exponential':\n delay = config.baseDelayMs * Math.pow(2, attempt)\n break\n case 'linear':\n delay = config.baseDelayMs * (attempt + 1)\n break\n case 'fixed':\n delay = config.baseDelayMs\n break\n }\n\n delay = Math.min(delay, config.maxDelayMs)\n\n if (config.jitter) {\n delay = Math.random() * delay\n }\n\n return delay\n}\n\n/**\n * Sleeps for the specified duration. Can be aborted via an AbortSignal.\n *\n * @param ms - Duration in milliseconds\n * @param signal - Optional AbortSignal to cancel the sleep early\n */\nexport function sleep(ms: number, signal?: AbortSignal): Promise<void> {\n return new Promise<void>((resolve, reject) => {\n if (signal?.aborted) {\n reject(signal.reason as Error)\n return\n }\n\n const timer = setTimeout(resolve, ms)\n\n signal?.addEventListener(\n 'abort',\n () => {\n clearTimeout(timer)\n reject(signal.reason as Error)\n },\n { once: true },\n )\n })\n}\n","import {\n A3ResilienceError,\n A3TimeoutError,\n DEFAULT_RESILIENCE_CONFIG,\n type ResilienceErrorEntry,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { calculateBackoff, sleep } from './backoff'\n\n/**\n * Builds an AbortSignal that fires when either the per-request timeout or total timeout expires.\n */\nfunction buildSignal(requestTimeoutMs?: number, totalAbort?: AbortSignal): AbortSignal | undefined {\n const signals: AbortSignal[] = []\n\n if (requestTimeoutMs !== undefined) {\n signals.push(AbortSignal.timeout(requestTimeoutMs))\n }\n\n if (totalAbort) {\n signals.push(totalAbort)\n }\n\n if (signals.length === 0) return undefined\n if (signals.length === 1) return signals[0]\n return AbortSignal.any(signals)\n}\n\ntype AttemptResult<T> = { ok: true; value: T } | { ok: false; error: Error }\n\nasync function attemptAction<T>(\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n model: string,\n signal: AbortSignal | undefined,\n attempt: number,\n errors: ResilienceErrorEntry[],\n): Promise<AttemptResult<T>> {\n try {\n const value = await action(model, signal)\n return { ok: true, value }\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, attempt, error: errorObj })\n return { ok: false, error: errorObj }\n }\n}\n\nfunction checkTotalTimeout(totalAbort: AbortController | undefined, timeoutMs: number | undefined, errors: ResilienceErrorEntry[]): void {\n if (totalAbort?.signal.aborted) {\n throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors)\n }\n}\n\nasync function handleAttemptError(\n errorObj: Error,\n attempt: number,\n maxRetries: number,\n retryAll: boolean,\n resolved: ResolvedResilienceConfig,\n totalAbort: AbortController | undefined,\n): Promise<'retry' | 'next-model'> {\n const isLastAttempt = attempt === 1 + maxRetries\n const isRetryable = retryAll || resolved.isRetryableError(errorObj)\n\n if (isRetryable && !isLastAttempt) {\n const delay = calculateBackoff(attempt - 1, resolved.backoff)\n await sleep(delay, totalAbort?.signal).catch(() => {\n // Sleep was aborted by total timeout — will be caught at top of loop\n })\n return 'retry'\n }\n\n return 'next-model'\n}\n\n/**\n * Executes an action with model fallback, retry, backoff, and timeout support.\n *\n * For each model (in priority order):\n * 1. Attempts the action up to `1 + maxAttempts` times\n * 2. On transient errors, waits with backoff before retrying\n * 3. On non-retryable errors (or after exhausting retries), falls back to the next model\n *\n * Throws `A3ResilienceError` with full error history when all models are exhausted.\n * Throws `A3TimeoutError` when the total timeout is exceeded.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model. Receives an optional AbortSignal.\n * @param config - Resolved resilience configuration (defaults applied if omitted)\n * @returns The result from the first successful attempt\n * @throws {A3ResilienceError} When all models and retries are exhausted\n * @throws {A3TimeoutError} When the total timeout is exceeded\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model, signal) => provider.call(model, params, { abortSignal: signal }),\n * resolvedConfig,\n * )\n * ```\n */\nexport async function executeWithFallback<T>(\n models: string[],\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n config?: ResolvedResilienceConfig,\n): Promise<T> {\n const resolved = config ?? DEFAULT_RESILIENCE_CONFIG\n const errors: ResilienceErrorEntry[] = []\n const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts\n const retryAll = resolved.retry !== false && resolved.retry.retryOn === 'all'\n\n // Total timeout controller\n let totalAbort: AbortController | undefined\n let totalTimer: ReturnType<typeof setTimeout> | undefined\n\n if (resolved.timeout.totalTimeoutMs !== undefined) {\n totalAbort = new AbortController()\n totalTimer = setTimeout(\n () => totalAbort!.abort(new Error('Total timeout exceeded')),\n resolved.timeout.totalTimeoutMs,\n )\n }\n\n try {\n for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {\n const model = models[modelIndex]\n\n for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal)\n // eslint-disable-next-line no-await-in-loop\n const result = await attemptAction(action, model, signal, attempt, errors)\n if (result.ok) return result.value\n\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n // eslint-disable-next-line no-await-in-loop\n const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort)\n if (decision === 'next-model') break\n }\n }\n\n // All models exhausted\n throw new A3ResilienceError(\n `All models failed after ${errors.length} total attempt(s): ${models.join(', ')}`,\n errors,\n )\n } finally {\n if (totalTimer !== undefined) {\n clearTimeout(totalTimer)\n }\n }\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport {\n resolveResilienceConfig,\n type Provider,\n type ProviderRequest,\n type ProviderResponse,\n type ProviderMessage,\n type BaseState,\n type StreamEvent,\n type ResilienceConfig,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '@providers/utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */\n resilience?: ResilienceConfig\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n const resilience: ResolvedResilienceConfig = resolveResilienceConfig(config.resilience)\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(\n models,\n (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(\n models,\n (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
|
package/dist/openai/index.d.cts
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { Provider } from '@genui-a3/core';
|
|
1
|
+
import { ResilienceConfig, Provider } from '@genui-a3/core';
|
|
2
2
|
|
|
3
3
|
/**
|
|
4
4
|
* Configuration for creating an OpenAI provider.
|
|
@@ -15,6 +15,8 @@ interface OpenAIProviderConfig {
|
|
|
15
15
|
baseURL?: string;
|
|
16
16
|
/** Optional OpenAI organization ID */
|
|
17
17
|
organization?: string;
|
|
18
|
+
/** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */
|
|
19
|
+
resilience?: ResilienceConfig;
|
|
18
20
|
}
|
|
19
21
|
/**
|
|
20
22
|
* Creates an OpenAI provider instance.
|
package/dist/openai/index.d.ts
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { Provider } from '@genui-a3/core';
|
|
1
|
+
import { ResilienceConfig, Provider } from '@genui-a3/core';
|
|
2
2
|
|
|
3
3
|
/**
|
|
4
4
|
* Configuration for creating an OpenAI provider.
|
|
@@ -15,6 +15,8 @@ interface OpenAIProviderConfig {
|
|
|
15
15
|
baseURL?: string;
|
|
16
16
|
/** Optional OpenAI organization ID */
|
|
17
17
|
organization?: string;
|
|
18
|
+
/** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */
|
|
19
|
+
resilience?: ResilienceConfig;
|
|
18
20
|
}
|
|
19
21
|
/**
|
|
20
22
|
* Creates an OpenAI provider instance.
|
package/dist/openai/index.js
CHANGED
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import { createOpenAI } from '@ai-sdk/openai';
|
|
2
2
|
import { streamText, Output, jsonSchema, generateText } from 'ai';
|
|
3
|
+
import { resolveResilienceConfig, DEFAULT_RESILIENCE_CONFIG, A3ResilienceError, A3TimeoutError } from '@genui-a3/core';
|
|
3
4
|
import { EventType } from '@ag-ui/client';
|
|
4
5
|
|
|
5
6
|
// openai/index.ts
|
|
@@ -67,20 +68,119 @@ function extractDelta(partial, prevLength) {
|
|
|
67
68
|
return chatbotMessage.slice(prevLength);
|
|
68
69
|
}
|
|
69
70
|
|
|
71
|
+
// utils/backoff.ts
|
|
72
|
+
function calculateBackoff(attempt, config) {
|
|
73
|
+
let delay;
|
|
74
|
+
switch (config.strategy) {
|
|
75
|
+
case "exponential":
|
|
76
|
+
delay = config.baseDelayMs * Math.pow(2, attempt);
|
|
77
|
+
break;
|
|
78
|
+
case "linear":
|
|
79
|
+
delay = config.baseDelayMs * (attempt + 1);
|
|
80
|
+
break;
|
|
81
|
+
case "fixed":
|
|
82
|
+
delay = config.baseDelayMs;
|
|
83
|
+
break;
|
|
84
|
+
}
|
|
85
|
+
delay = Math.min(delay, config.maxDelayMs);
|
|
86
|
+
if (config.jitter) {
|
|
87
|
+
delay = Math.random() * delay;
|
|
88
|
+
}
|
|
89
|
+
return delay;
|
|
90
|
+
}
|
|
91
|
+
function sleep(ms, signal) {
|
|
92
|
+
return new Promise((resolve, reject) => {
|
|
93
|
+
if (signal?.aborted) {
|
|
94
|
+
reject(signal.reason);
|
|
95
|
+
return;
|
|
96
|
+
}
|
|
97
|
+
const timer = setTimeout(resolve, ms);
|
|
98
|
+
signal?.addEventListener(
|
|
99
|
+
"abort",
|
|
100
|
+
() => {
|
|
101
|
+
clearTimeout(timer);
|
|
102
|
+
reject(signal.reason);
|
|
103
|
+
},
|
|
104
|
+
{ once: true }
|
|
105
|
+
);
|
|
106
|
+
});
|
|
107
|
+
}
|
|
108
|
+
|
|
70
109
|
// utils/executeWithFallback.ts
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
110
|
+
function buildSignal(requestTimeoutMs, totalAbort) {
|
|
111
|
+
const signals = [];
|
|
112
|
+
if (requestTimeoutMs !== void 0) {
|
|
113
|
+
signals.push(AbortSignal.timeout(requestTimeoutMs));
|
|
114
|
+
}
|
|
115
|
+
if (totalAbort) {
|
|
116
|
+
signals.push(totalAbort);
|
|
117
|
+
}
|
|
118
|
+
if (signals.length === 0) return void 0;
|
|
119
|
+
if (signals.length === 1) return signals[0];
|
|
120
|
+
return AbortSignal.any(signals);
|
|
121
|
+
}
|
|
122
|
+
async function attemptAction(action, model, signal, attempt, errors) {
|
|
123
|
+
try {
|
|
124
|
+
const value = await action(model, signal);
|
|
125
|
+
return { ok: true, value };
|
|
126
|
+
} catch (error) {
|
|
127
|
+
const errorObj = error;
|
|
128
|
+
errors.push({ model, attempt, error: errorObj });
|
|
129
|
+
return { ok: false, error: errorObj };
|
|
130
|
+
}
|
|
131
|
+
}
|
|
132
|
+
function checkTotalTimeout(totalAbort, timeoutMs, errors) {
|
|
133
|
+
if (totalAbort?.signal.aborted) {
|
|
134
|
+
throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors);
|
|
135
|
+
}
|
|
136
|
+
}
|
|
137
|
+
async function handleAttemptError(errorObj, attempt, maxRetries, retryAll, resolved, totalAbort) {
|
|
138
|
+
const isLastAttempt = attempt === 1 + maxRetries;
|
|
139
|
+
const isRetryable = retryAll || resolved.isRetryableError(errorObj);
|
|
140
|
+
if (isRetryable && !isLastAttempt) {
|
|
141
|
+
const delay = calculateBackoff(attempt - 1, resolved.backoff);
|
|
142
|
+
await sleep(delay, totalAbort?.signal).catch(() => {
|
|
143
|
+
});
|
|
144
|
+
return "retry";
|
|
145
|
+
}
|
|
146
|
+
return "next-model";
|
|
147
|
+
}
|
|
148
|
+
async function executeWithFallback(models, action, config) {
|
|
149
|
+
const resolved = config ?? DEFAULT_RESILIENCE_CONFIG;
|
|
150
|
+
const errors = [];
|
|
151
|
+
const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts;
|
|
152
|
+
const retryAll = resolved.retry !== false && resolved.retry.retryOn === "all";
|
|
153
|
+
let totalAbort;
|
|
154
|
+
let totalTimer;
|
|
155
|
+
if (resolved.timeout.totalTimeoutMs !== void 0) {
|
|
156
|
+
totalAbort = new AbortController();
|
|
157
|
+
totalTimer = setTimeout(
|
|
158
|
+
() => totalAbort.abort(new Error("Total timeout exceeded")),
|
|
159
|
+
resolved.timeout.totalTimeoutMs
|
|
160
|
+
);
|
|
161
|
+
}
|
|
162
|
+
try {
|
|
163
|
+
for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {
|
|
164
|
+
const model = models[modelIndex];
|
|
165
|
+
for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {
|
|
166
|
+
checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
|
|
167
|
+
const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal);
|
|
168
|
+
const result = await attemptAction(action, model, signal, attempt, errors);
|
|
169
|
+
if (result.ok) return result.value;
|
|
170
|
+
checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
|
|
171
|
+
const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort);
|
|
172
|
+
if (decision === "next-model") break;
|
|
80
173
|
}
|
|
81
174
|
}
|
|
175
|
+
throw new A3ResilienceError(
|
|
176
|
+
`All models failed after ${errors.length} total attempt(s): ${models.join(", ")}`,
|
|
177
|
+
errors
|
|
178
|
+
);
|
|
179
|
+
} finally {
|
|
180
|
+
if (totalTimer !== void 0) {
|
|
181
|
+
clearTimeout(totalTimer);
|
|
182
|
+
}
|
|
82
183
|
}
|
|
83
|
-
throw new Error("All models failed");
|
|
84
184
|
}
|
|
85
185
|
|
|
86
186
|
// openai/index.ts
|
|
@@ -156,20 +256,23 @@ function createOpenAIProvider(config) {
|
|
|
156
256
|
organization: config.organization
|
|
157
257
|
});
|
|
158
258
|
const models = config.models;
|
|
259
|
+
const resilience = resolveResilienceConfig(config.resilience);
|
|
159
260
|
return {
|
|
160
261
|
name: "openai",
|
|
161
262
|
async sendRequest(request) {
|
|
162
263
|
const messages = toAIMessages(request.messages);
|
|
163
264
|
return executeWithFallback(
|
|
164
265
|
models,
|
|
165
|
-
(model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
|
|
266
|
+
(model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
|
|
267
|
+
resilience
|
|
166
268
|
);
|
|
167
269
|
},
|
|
168
270
|
async *sendRequestStream(request) {
|
|
169
271
|
const messages = toAIMessages(request.messages);
|
|
170
272
|
const { result, reader, first } = await executeWithFallback(
|
|
171
273
|
models,
|
|
172
|
-
(model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
|
|
274
|
+
(model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
|
|
275
|
+
resilience
|
|
173
276
|
);
|
|
174
277
|
yield* processOpenAIStream(result, reader, first, "openai", request.responseSchema);
|
|
175
278
|
}
|
package/dist/openai/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":[],"mappings":";;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAM,SAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;ACnFA,eAAsB,mBAAA,CAAuB,QAAkB,MAAA,EAAmD;AAGhH,EAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,MAAA,CAAO,QAAQ,CAAA,EAAA,EAAK;AACtC,IAAA,MAAM,KAAA,GAAQ,OAAO,CAAC,CAAA;AAEtB,IAAA,IAAI;AAEF,MAAA,OAAO,MAAM,OAAO,KAAK,CAAA;AAAA,IAC3B,SAAS,KAAA,EAAO;AACd,MAAA,MAAM,QAAA,GAAW,KAAA;AAGjB,MAAA,IAAI,CAAA,KAAM,MAAA,CAAO,MAAA,GAAS,CAAA,EAAG;AAC3B,QAAA,MAAM,QAAA;AAAA,MACR;AAAA,IACF;AAAA,EACF;AAEA,EAAA,MAAM,IAAI,MAAM,mBAAmB,CAAA;AACrC;;;ACEA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAO,WAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAM,YAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAAS,UAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiB,YAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AAEtB,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UAClC,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OAC7F;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UACnE,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OACnG;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.js","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","/**\n * Executes an action with model fallback support.\n * Tries each model in order; if one fails, falls back to the next.\n * Throws the last error if all models fail.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model\n * @returns The result from the first successful model\n * @throws The error from the last model if all fail\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model) => provider.call(model, params),\n * )\n * ```\n */\nexport async function executeWithFallback<T>(models: string[], action: (model: string) => Promise<T>): Promise<T> {\n const errors: Array<{ model: string; error: Error }> = []\n\n for (let i = 0; i < models.length; i++) {\n const model = models[i]\n\n try {\n // eslint-disable-next-line no-await-in-loop\n return await action(model)\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, error: errorObj })\n\n if (i === models.length - 1) {\n throw errorObj\n }\n }\n }\n\n throw new Error('All models failed')\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport type {\n Provider,\n ProviderRequest,\n ProviderResponse,\n ProviderMessage,\n BaseState,\n StreamEvent,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '../utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(models, (model) =>\n sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(models, (model) =>\n sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/backoff.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":[],"mappings":";;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAM,SAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;AC5FO,SAAS,gBAAA,CAAiB,SAAiB,MAAA,EAAyC;AACzF,EAAA,IAAI,KAAA;AAEJ,EAAA,QAAQ,OAAO,QAAA;AAAU,IACvB,KAAK,aAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA,GAAc,IAAA,CAAK,GAAA,CAAI,GAAG,OAAO,CAAA;AAChD,MAAA;AAAA,IACF,KAAK,QAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,eAAe,OAAA,GAAU,CAAA,CAAA;AACxC,MAAA;AAAA,IACF,KAAK,OAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA;AACf,MAAA;AAAA;AAGJ,EAAA,KAAA,GAAQ,IAAA,CAAK,GAAA,CAAI,KAAA,EAAO,MAAA,CAAO,UAAU,CAAA;AAEzC,EAAA,IAAI,OAAO,MAAA,EAAQ;AACjB,IAAA,KAAA,GAAQ,IAAA,CAAK,QAAO,GAAI,KAAA;AAAA,EAC1B;AAEA,EAAA,OAAO,KAAA;AACT;AAQO,SAAS,KAAA,CAAM,IAAY,MAAA,EAAqC;AACrE,EAAA,OAAO,IAAI,OAAA,CAAc,CAAC,OAAA,EAAS,MAAA,KAAW;AAC5C,IAAA,IAAI,QAAQ,OAAA,EAAS;AACnB,MAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAC7B,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,KAAA,GAAQ,UAAA,CAAW,OAAA,EAAS,EAAE,CAAA;AAEpC,IAAA,MAAA,EAAQ,gBAAA;AAAA,MACN,OAAA;AAAA,MACA,MAAM;AACJ,QAAA,YAAA,CAAa,KAAK,CAAA;AAClB,QAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAAA,MAC/B,CAAA;AAAA,MACA,EAAE,MAAM,IAAA;AAAK,KACf;AAAA,EACF,CAAC,CAAA;AACH;;;AC7CA,SAAS,WAAA,CAAY,kBAA2B,UAAA,EAAmD;AACjG,EAAA,MAAM,UAAyB,EAAC;AAEhC,EAAA,IAAI,qBAAqB,MAAA,EAAW;AAClC,IAAA,OAAA,CAAQ,IAAA,CAAK,WAAA,CAAY,OAAA,CAAQ,gBAAgB,CAAC,CAAA;AAAA,EACpD;AAEA,EAAA,IAAI,UAAA,EAAY;AACd,IAAA,OAAA,CAAQ,KAAK,UAAU,CAAA;AAAA,EACzB;AAEA,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,MAAA;AACjC,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,QAAQ,CAAC,CAAA;AAC1C,EAAA,OAAO,WAAA,CAAY,IAAI,OAAO,CAAA;AAChC;AAIA,eAAe,aAAA,CACb,MAAA,EACA,KAAA,EACA,MAAA,EACA,SACA,MAAA,EAC2B;AAC3B,EAAA,IAAI;AACF,IAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,KAAA,EAAO,MAAM,CAAA;AACxC,IAAA,OAAO,EAAE,EAAA,EAAI,IAAA,EAAM,KAAA,EAAM;AAAA,EAC3B,SAAS,KAAA,EAAO;AACd,IAAA,MAAM,QAAA,GAAW,KAAA;AACjB,IAAA,MAAA,CAAO,KAAK,EAAE,KAAA,EAAO,OAAA,EAAS,KAAA,EAAO,UAAU,CAAA;AAC/C,IAAA,OAAO,EAAE,EAAA,EAAI,KAAA,EAAO,KAAA,EAAO,QAAA,EAAS;AAAA,EACtC;AACF;AAEA,SAAS,iBAAA,CAAkB,UAAA,EAAyC,SAAA,EAA+B,MAAA,EAAsC;AACvI,EAAA,IAAI,UAAA,EAAY,OAAO,OAAA,EAAS;AAC9B,IAAA,MAAM,IAAI,cAAA,CAAe,CAAA,iBAAA,EAAoB,SAAS,eAAe,MAAM,CAAA;AAAA,EAC7E;AACF;AAEA,eAAe,mBACb,QAAA,EACA,OAAA,EACA,UAAA,EACA,QAAA,EACA,UACA,UAAA,EACiC;AACjC,EAAA,MAAM,aAAA,GAAgB,YAAY,CAAA,GAAI,UAAA;AACtC,EAAA,MAAM,WAAA,GAAc,QAAA,IAAY,QAAA,CAAS,gBAAA,CAAiB,QAAQ,CAAA;AAElE,EAAA,IAAI,WAAA,IAAe,CAAC,aAAA,EAAe;AACjC,IAAA,MAAM,KAAA,GAAQ,gBAAA,CAAiB,OAAA,GAAU,CAAA,EAAG,SAAS,OAAO,CAAA;AAC5D,IAAA,MAAM,MAAM,KAAA,EAAO,UAAA,EAAY,MAAM,CAAA,CAAE,MAAM,MAAM;AAAA,IAEnD,CAAC,CAAA;AACD,IAAA,OAAO,OAAA;AAAA,EACT;AAEA,EAAA,OAAO,YAAA;AACT;AA6BA,eAAsB,mBAAA,CACpB,MAAA,EACA,MAAA,EACA,MAAA,EACY;AACZ,EAAA,MAAM,WAAW,MAAA,IAAU,yBAAA;AAC3B,EAAA,MAAM,SAAiC,EAAC;AACxC,EAAA,MAAM,aAAa,QAAA,CAAS,KAAA,KAAU,KAAA,GAAQ,CAAA,GAAI,SAAS,KAAA,CAAM,WAAA;AACjE,EAAA,MAAM,WAAW,QAAA,CAAS,KAAA,KAAU,KAAA,IAAS,QAAA,CAAS,MAAM,OAAA,KAAY,KAAA;AAGxE,EAAA,IAAI,UAAA;AACJ,EAAA,IAAI,UAAA;AAEJ,EAAA,IAAI,QAAA,CAAS,OAAA,CAAQ,cAAA,KAAmB,MAAA,EAAW;AACjD,IAAA,UAAA,GAAa,IAAI,eAAA,EAAgB;AACjC,IAAA,UAAA,GAAa,UAAA;AAAA,MACX,MAAM,UAAA,CAAY,KAAA,CAAM,IAAI,KAAA,CAAM,wBAAwB,CAAC,CAAA;AAAA,MAC3D,SAAS,OAAA,CAAQ;AAAA,KACnB;AAAA,EACF;AAEA,EAAA,IAAI;AACF,IAAA,KAAA,IAAS,UAAA,GAAa,CAAA,EAAG,UAAA,GAAa,MAAA,CAAO,QAAQ,UAAA,EAAA,EAAc;AACjE,MAAA,MAAM,KAAA,GAAQ,OAAO,UAAU,CAAA;AAE/B,MAAA,KAAA,IAAS,OAAA,GAAU,CAAA,EAAG,OAAA,IAAW,CAAA,GAAI,YAAY,OAAA,EAAA,EAAW;AAC1D,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAErE,QAAA,MAAM,SAAS,WAAA,CAAY,QAAA,CAAS,OAAA,CAAQ,gBAAA,EAAkB,YAAY,MAAM,CAAA;AAEhF,QAAA,MAAM,SAAS,MAAM,aAAA,CAAc,QAAQ,KAAA,EAAO,MAAA,EAAQ,SAAS,MAAM,CAAA;AACzE,QAAA,IAAI,MAAA,CAAO,EAAA,EAAI,OAAO,MAAA,CAAO,KAAA;AAE7B,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAGrE,QAAA,MAAM,QAAA,GAAW,MAAM,kBAAA,CAAmB,MAAA,CAAO,OAAO,OAAA,EAAS,UAAA,EAAY,QAAA,EAAU,QAAA,EAAU,UAAU,CAAA;AAC3G,QAAA,IAAI,aAAa,YAAA,EAAc;AAAA,MACjC;AAAA,IACF;AAGA,IAAA,MAAM,IAAI,iBAAA;AAAA,MACR,2BAA2B,MAAA,CAAO,MAAM,sBAAsB,MAAA,CAAO,IAAA,CAAK,IAAI,CAAC,CAAA,CAAA;AAAA,MAC/E;AAAA,KACF;AAAA,EACF,CAAA,SAAE;AACA,IAAA,IAAI,eAAe,MAAA,EAAW;AAC5B,MAAA,YAAA,CAAa,UAAU,CAAA;AAAA,IACzB;AAAA,EACF;AACF;;;AC7GA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAO,WAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAM,YAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAAS,UAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiB,YAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AACtB,EAAA,MAAM,UAAA,GAAuC,uBAAA,CAAwB,MAAA,CAAO,UAAU,CAAA;AAEtF,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QACL,MAAA;AAAA,QACA,CAAC,UAAU,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QACtG;AAAA,OACF;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QACtC,MAAA;AAAA,QACA,CAAC,UAAU,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QAC5G;AAAA,OACF;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.js","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","import type { BackoffConfig } from '@genui-a3/core'\n\n/**\n * Calculates the backoff delay for a given retry attempt.\n *\n * @param attempt - Zero-based attempt index (0 = first retry)\n * @param config - Backoff configuration with all fields required\n * @returns Delay in milliseconds\n */\nexport function calculateBackoff(attempt: number, config: Required<BackoffConfig>): number {\n let delay: number\n\n switch (config.strategy) {\n case 'exponential':\n delay = config.baseDelayMs * Math.pow(2, attempt)\n break\n case 'linear':\n delay = config.baseDelayMs * (attempt + 1)\n break\n case 'fixed':\n delay = config.baseDelayMs\n break\n }\n\n delay = Math.min(delay, config.maxDelayMs)\n\n if (config.jitter) {\n delay = Math.random() * delay\n }\n\n return delay\n}\n\n/**\n * Sleeps for the specified duration. Can be aborted via an AbortSignal.\n *\n * @param ms - Duration in milliseconds\n * @param signal - Optional AbortSignal to cancel the sleep early\n */\nexport function sleep(ms: number, signal?: AbortSignal): Promise<void> {\n return new Promise<void>((resolve, reject) => {\n if (signal?.aborted) {\n reject(signal.reason as Error)\n return\n }\n\n const timer = setTimeout(resolve, ms)\n\n signal?.addEventListener(\n 'abort',\n () => {\n clearTimeout(timer)\n reject(signal.reason as Error)\n },\n { once: true },\n )\n })\n}\n","import {\n A3ResilienceError,\n A3TimeoutError,\n DEFAULT_RESILIENCE_CONFIG,\n type ResilienceErrorEntry,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { calculateBackoff, sleep } from './backoff'\n\n/**\n * Builds an AbortSignal that fires when either the per-request timeout or total timeout expires.\n */\nfunction buildSignal(requestTimeoutMs?: number, totalAbort?: AbortSignal): AbortSignal | undefined {\n const signals: AbortSignal[] = []\n\n if (requestTimeoutMs !== undefined) {\n signals.push(AbortSignal.timeout(requestTimeoutMs))\n }\n\n if (totalAbort) {\n signals.push(totalAbort)\n }\n\n if (signals.length === 0) return undefined\n if (signals.length === 1) return signals[0]\n return AbortSignal.any(signals)\n}\n\ntype AttemptResult<T> = { ok: true; value: T } | { ok: false; error: Error }\n\nasync function attemptAction<T>(\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n model: string,\n signal: AbortSignal | undefined,\n attempt: number,\n errors: ResilienceErrorEntry[],\n): Promise<AttemptResult<T>> {\n try {\n const value = await action(model, signal)\n return { ok: true, value }\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, attempt, error: errorObj })\n return { ok: false, error: errorObj }\n }\n}\n\nfunction checkTotalTimeout(totalAbort: AbortController | undefined, timeoutMs: number | undefined, errors: ResilienceErrorEntry[]): void {\n if (totalAbort?.signal.aborted) {\n throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors)\n }\n}\n\nasync function handleAttemptError(\n errorObj: Error,\n attempt: number,\n maxRetries: number,\n retryAll: boolean,\n resolved: ResolvedResilienceConfig,\n totalAbort: AbortController | undefined,\n): Promise<'retry' | 'next-model'> {\n const isLastAttempt = attempt === 1 + maxRetries\n const isRetryable = retryAll || resolved.isRetryableError(errorObj)\n\n if (isRetryable && !isLastAttempt) {\n const delay = calculateBackoff(attempt - 1, resolved.backoff)\n await sleep(delay, totalAbort?.signal).catch(() => {\n // Sleep was aborted by total timeout — will be caught at top of loop\n })\n return 'retry'\n }\n\n return 'next-model'\n}\n\n/**\n * Executes an action with model fallback, retry, backoff, and timeout support.\n *\n * For each model (in priority order):\n * 1. Attempts the action up to `1 + maxAttempts` times\n * 2. On transient errors, waits with backoff before retrying\n * 3. On non-retryable errors (or after exhausting retries), falls back to the next model\n *\n * Throws `A3ResilienceError` with full error history when all models are exhausted.\n * Throws `A3TimeoutError` when the total timeout is exceeded.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model. Receives an optional AbortSignal.\n * @param config - Resolved resilience configuration (defaults applied if omitted)\n * @returns The result from the first successful attempt\n * @throws {A3ResilienceError} When all models and retries are exhausted\n * @throws {A3TimeoutError} When the total timeout is exceeded\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model, signal) => provider.call(model, params, { abortSignal: signal }),\n * resolvedConfig,\n * )\n * ```\n */\nexport async function executeWithFallback<T>(\n models: string[],\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n config?: ResolvedResilienceConfig,\n): Promise<T> {\n const resolved = config ?? DEFAULT_RESILIENCE_CONFIG\n const errors: ResilienceErrorEntry[] = []\n const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts\n const retryAll = resolved.retry !== false && resolved.retry.retryOn === 'all'\n\n // Total timeout controller\n let totalAbort: AbortController | undefined\n let totalTimer: ReturnType<typeof setTimeout> | undefined\n\n if (resolved.timeout.totalTimeoutMs !== undefined) {\n totalAbort = new AbortController()\n totalTimer = setTimeout(\n () => totalAbort!.abort(new Error('Total timeout exceeded')),\n resolved.timeout.totalTimeoutMs,\n )\n }\n\n try {\n for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {\n const model = models[modelIndex]\n\n for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal)\n // eslint-disable-next-line no-await-in-loop\n const result = await attemptAction(action, model, signal, attempt, errors)\n if (result.ok) return result.value\n\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n // eslint-disable-next-line no-await-in-loop\n const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort)\n if (decision === 'next-model') break\n }\n }\n\n // All models exhausted\n throw new A3ResilienceError(\n `All models failed after ${errors.length} total attempt(s): ${models.join(', ')}`,\n errors,\n )\n } finally {\n if (totalTimer !== undefined) {\n clearTimeout(totalTimer)\n }\n }\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport {\n resolveResilienceConfig,\n type Provider,\n type ProviderRequest,\n type ProviderResponse,\n type ProviderMessage,\n type BaseState,\n type StreamEvent,\n type ResilienceConfig,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '@providers/utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */\n resilience?: ResilienceConfig\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n const resilience: ResolvedResilienceConfig = resolveResilienceConfig(config.resilience)\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(\n models,\n (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(\n models,\n (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
|