@genui-a3/providers 0.0.4 → 0.0.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2,6 +2,7 @@
2
2
 
3
3
  var openai = require('@ai-sdk/openai');
4
4
  var ai = require('ai');
5
+ var core = require('@genui-a3/core');
5
6
  var client = require('@ag-ui/client');
6
7
 
7
8
  // openai/index.ts
@@ -69,20 +70,119 @@ function extractDelta(partial, prevLength) {
69
70
  return chatbotMessage.slice(prevLength);
70
71
  }
71
72
 
73
+ // utils/backoff.ts
74
+ function calculateBackoff(attempt, config) {
75
+ let delay;
76
+ switch (config.strategy) {
77
+ case "exponential":
78
+ delay = config.baseDelayMs * Math.pow(2, attempt);
79
+ break;
80
+ case "linear":
81
+ delay = config.baseDelayMs * (attempt + 1);
82
+ break;
83
+ case "fixed":
84
+ delay = config.baseDelayMs;
85
+ break;
86
+ }
87
+ delay = Math.min(delay, config.maxDelayMs);
88
+ if (config.jitter) {
89
+ delay = Math.random() * delay;
90
+ }
91
+ return delay;
92
+ }
93
+ function sleep(ms, signal) {
94
+ return new Promise((resolve, reject) => {
95
+ if (signal?.aborted) {
96
+ reject(signal.reason);
97
+ return;
98
+ }
99
+ const timer = setTimeout(resolve, ms);
100
+ signal?.addEventListener(
101
+ "abort",
102
+ () => {
103
+ clearTimeout(timer);
104
+ reject(signal.reason);
105
+ },
106
+ { once: true }
107
+ );
108
+ });
109
+ }
110
+
72
111
  // utils/executeWithFallback.ts
73
- async function executeWithFallback(models, action) {
74
- for (let i = 0; i < models.length; i++) {
75
- const model = models[i];
76
- try {
77
- return await action(model);
78
- } catch (error) {
79
- const errorObj = error;
80
- if (i === models.length - 1) {
81
- throw errorObj;
112
+ function buildSignal(requestTimeoutMs, totalAbort) {
113
+ const signals = [];
114
+ if (requestTimeoutMs !== void 0) {
115
+ signals.push(AbortSignal.timeout(requestTimeoutMs));
116
+ }
117
+ if (totalAbort) {
118
+ signals.push(totalAbort);
119
+ }
120
+ if (signals.length === 0) return void 0;
121
+ if (signals.length === 1) return signals[0];
122
+ return AbortSignal.any(signals);
123
+ }
124
+ async function attemptAction(action, model, signal, attempt, errors) {
125
+ try {
126
+ const value = await action(model, signal);
127
+ return { ok: true, value };
128
+ } catch (error) {
129
+ const errorObj = error;
130
+ errors.push({ model, attempt, error: errorObj });
131
+ return { ok: false, error: errorObj };
132
+ }
133
+ }
134
+ function checkTotalTimeout(totalAbort, timeoutMs, errors) {
135
+ if (totalAbort?.signal.aborted) {
136
+ throw new core.A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors);
137
+ }
138
+ }
139
+ async function handleAttemptError(errorObj, attempt, maxRetries, retryAll, resolved, totalAbort) {
140
+ const isLastAttempt = attempt === 1 + maxRetries;
141
+ const isRetryable = retryAll || resolved.isRetryableError(errorObj);
142
+ if (isRetryable && !isLastAttempt) {
143
+ const delay = calculateBackoff(attempt - 1, resolved.backoff);
144
+ await sleep(delay, totalAbort?.signal).catch(() => {
145
+ });
146
+ return "retry";
147
+ }
148
+ return "next-model";
149
+ }
150
+ async function executeWithFallback(models, action, config) {
151
+ const resolved = config ?? core.DEFAULT_RESILIENCE_CONFIG;
152
+ const errors = [];
153
+ const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts;
154
+ const retryAll = resolved.retry !== false && resolved.retry.retryOn === "all";
155
+ let totalAbort;
156
+ let totalTimer;
157
+ if (resolved.timeout.totalTimeoutMs !== void 0) {
158
+ totalAbort = new AbortController();
159
+ totalTimer = setTimeout(
160
+ () => totalAbort.abort(new Error("Total timeout exceeded")),
161
+ resolved.timeout.totalTimeoutMs
162
+ );
163
+ }
164
+ try {
165
+ for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {
166
+ const model = models[modelIndex];
167
+ for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {
168
+ checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
169
+ const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal);
170
+ const result = await attemptAction(action, model, signal, attempt, errors);
171
+ if (result.ok) return result.value;
172
+ checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
173
+ const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort);
174
+ if (decision === "next-model") break;
82
175
  }
83
176
  }
177
+ throw new core.A3ResilienceError(
178
+ `All models failed after ${errors.length} total attempt(s): ${models.join(", ")}`,
179
+ errors
180
+ );
181
+ } finally {
182
+ if (totalTimer !== void 0) {
183
+ clearTimeout(totalTimer);
184
+ }
84
185
  }
85
- throw new Error("All models failed");
86
186
  }
87
187
 
88
188
  // openai/index.ts
@@ -158,20 +258,23 @@ function createOpenAIProvider(config) {
158
258
  organization: config.organization
159
259
  });
160
260
  const models = config.models;
261
+ const resilience = core.resolveResilienceConfig(config.resilience);
161
262
  return {
162
263
  name: "openai",
163
264
  async sendRequest(request) {
164
265
  const messages = toAIMessages(request.messages);
165
266
  return executeWithFallback(
166
267
  models,
167
- (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
268
+ (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
269
+ resilience
168
270
  );
169
271
  },
170
272
  async *sendRequestStream(request) {
171
273
  const messages = toAIMessages(request.messages);
172
274
  const { result, reader, first } = await executeWithFallback(
173
275
  models,
174
- (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
276
+ (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
277
+ resilience
175
278
  );
176
279
  yield* processOpenAIStream(result, reader, first, "openai", request.responseSchema);
177
280
  }
@@ -1 +1 @@
1
- {"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":["EventType","jsonSchema","generateText","Output","streamText","createOpenAI"],"mappings":";;;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;ACnFA,eAAsB,mBAAA,CAAuB,QAAkB,MAAA,EAAmD;AAGhH,EAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,MAAA,CAAO,QAAQ,CAAA,EAAA,EAAK;AACtC,IAAA,MAAM,KAAA,GAAQ,OAAO,CAAC,CAAA;AAEtB,IAAA,IAAI;AAEF,MAAA,OAAO,MAAM,OAAO,KAAK,CAAA;AAAA,IAC3B,SAAS,KAAA,EAAO;AACd,MAAA,MAAM,QAAA,GAAW,KAAA;AAGjB,MAAA,IAAI,CAAA,KAAM,MAAA,CAAO,MAAA,GAAS,CAAA,EAAG;AAC3B,QAAA,MAAM,QAAA;AAAA,MACR;AAAA,IACF;AAAA,EACF;AAEA,EAAA,MAAM,IAAI,MAAM,mBAAmB,CAAA;AACrC;;;ACEA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAOC,cAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAMC,eAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQC,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAASC,aAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQD,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiBE,mBAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AAEtB,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UAClC,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OAC7F;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UACnE,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OACnG;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.cjs","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","/**\n * Executes an action with model fallback support.\n * Tries each model in order; if one fails, falls back to the next.\n * Throws the last error if all models fail.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model\n * @returns The result from the first successful model\n * @throws The error from the last model if all fail\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model) => provider.call(model, params),\n * )\n * ```\n */\nexport async function executeWithFallback<T>(models: string[], action: (model: string) => Promise<T>): Promise<T> {\n const errors: Array<{ model: string; error: Error }> = []\n\n for (let i = 0; i < models.length; i++) {\n const model = models[i]\n\n try {\n // eslint-disable-next-line no-await-in-loop\n return await action(model)\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, error: errorObj })\n\n if (i === models.length - 1) {\n throw errorObj\n }\n }\n }\n\n throw new Error('All models failed')\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport type {\n Provider,\n ProviderRequest,\n ProviderResponse,\n ProviderMessage,\n BaseState,\n StreamEvent,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '../utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(models, (model) =>\n sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(models, (model) =>\n sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
1
+ {"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/backoff.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":["EventType","A3TimeoutError","DEFAULT_RESILIENCE_CONFIG","A3ResilienceError","jsonSchema","generateText","Output","streamText","createOpenAI","resolveResilienceConfig"],"mappings":";;;;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAMA,gBAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAMA,gBAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;AC5FO,SAAS,gBAAA,CAAiB,SAAiB,MAAA,EAAyC;AACzF,EAAA,IAAI,KAAA;AAEJ,EAAA,QAAQ,OAAO,QAAA;AAAU,IACvB,KAAK,aAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA,GAAc,IAAA,CAAK,GAAA,CAAI,GAAG,OAAO,CAAA;AAChD,MAAA;AAAA,IACF,KAAK,QAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,eAAe,OAAA,GAAU,CAAA,CAAA;AACxC,MAAA;AAAA,IACF,KAAK,OAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA;AACf,MAAA;AAAA;AAGJ,EAAA,KAAA,GAAQ,IAAA,CAAK,GAAA,CAAI,KAAA,EAAO,MAAA,CAAO,UAAU,CAAA;AAEzC,EAAA,IAAI,OAAO,MAAA,EAAQ;AACjB,IAAA,KAAA,GAAQ,IAAA,CAAK,QAAO,GAAI,KAAA;AAAA,EAC1B;AAEA,EAAA,OAAO,KAAA;AACT;AAQO,SAAS,KAAA,CAAM,IAAY,MAAA,EAAqC;AACrE,EAAA,OAAO,IAAI,OAAA,CAAc,CAAC,OAAA,EAAS,MAAA,KAAW;AAC5C,IAAA,IAAI,QAAQ,OAAA,EAAS;AACnB,MAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAC7B,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,KAAA,GAAQ,UAAA,CAAW,OAAA,EAAS,EAAE,CAAA;AAEpC,IAAA,MAAA,EAAQ,gBAAA;AAAA,MACN,OAAA;AAAA,MACA,MAAM;AACJ,QAAA,YAAA,CAAa,KAAK,CAAA;AAClB,QAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAAA,MAC/B,CAAA;AAAA,MACA,EAAE,MAAM,IAAA;AAAK,KACf;AAAA,EACF,CAAC,CAAA;AACH;;;AC7CA,SAAS,WAAA,CAAY,kBAA2B,UAAA,EAAmD;AACjG,EAAA,MAAM,UAAyB,EAAC;AAEhC,EAAA,IAAI,qBAAqB,MAAA,EAAW;AAClC,IAAA,OAAA,CAAQ,IAAA,CAAK,WAAA,CAAY,OAAA,CAAQ,gBAAgB,CAAC,CAAA;AAAA,EACpD;AAEA,EAAA,IAAI,UAAA,EAAY;AACd,IAAA,OAAA,CAAQ,KAAK,UAAU,CAAA;AAAA,EACzB;AAEA,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,MAAA;AACjC,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,QAAQ,CAAC,CAAA;AAC1C,EAAA,OAAO,WAAA,CAAY,IAAI,OAAO,CAAA;AAChC;AAIA,eAAe,aAAA,CACb,MAAA,EACA,KAAA,EACA,MAAA,EACA,SACA,MAAA,EAC2B;AAC3B,EAAA,IAAI;AACF,IAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,KAAA,EAAO,MAAM,CAAA;AACxC,IAAA,OAAO,EAAE,EAAA,EAAI,IAAA,EAAM,KAAA,EAAM;AAAA,EAC3B,SAAS,KAAA,EAAO;AACd,IAAA,MAAM,QAAA,GAAW,KAAA;AACjB,IAAA,MAAA,CAAO,KAAK,EAAE,KAAA,EAAO,OAAA,EAAS,KAAA,EAAO,UAAU,CAAA;AAC/C,IAAA,OAAO,EAAE,EAAA,EAAI,KAAA,EAAO,KAAA,EAAO,QAAA,EAAS;AAAA,EACtC;AACF;AAEA,SAAS,iBAAA,CAAkB,UAAA,EAAyC,SAAA,EAA+B,MAAA,EAAsC;AACvI,EAAA,IAAI,UAAA,EAAY,OAAO,OAAA,EAAS;AAC9B,IAAA,MAAM,IAAIC,mBAAA,CAAe,CAAA,iBAAA,EAAoB,SAAS,eAAe,MAAM,CAAA;AAAA,EAC7E;AACF;AAEA,eAAe,mBACb,QAAA,EACA,OAAA,EACA,UAAA,EACA,QAAA,EACA,UACA,UAAA,EACiC;AACjC,EAAA,MAAM,aAAA,GAAgB,YAAY,CAAA,GAAI,UAAA;AACtC,EAAA,MAAM,WAAA,GAAc,QAAA,IAAY,QAAA,CAAS,gBAAA,CAAiB,QAAQ,CAAA;AAElE,EAAA,IAAI,WAAA,IAAe,CAAC,aAAA,EAAe;AACjC,IAAA,MAAM,KAAA,GAAQ,gBAAA,CAAiB,OAAA,GAAU,CAAA,EAAG,SAAS,OAAO,CAAA;AAC5D,IAAA,MAAM,MAAM,KAAA,EAAO,UAAA,EAAY,MAAM,CAAA,CAAE,MAAM,MAAM;AAAA,IAEnD,CAAC,CAAA;AACD,IAAA,OAAO,OAAA;AAAA,EACT;AAEA,EAAA,OAAO,YAAA;AACT;AA6BA,eAAsB,mBAAA,CACpB,MAAA,EACA,MAAA,EACA,MAAA,EACY;AACZ,EAAA,MAAM,WAAW,MAAA,IAAUC,8BAAA;AAC3B,EAAA,MAAM,SAAiC,EAAC;AACxC,EAAA,MAAM,aAAa,QAAA,CAAS,KAAA,KAAU,KAAA,GAAQ,CAAA,GAAI,SAAS,KAAA,CAAM,WAAA;AACjE,EAAA,MAAM,WAAW,QAAA,CAAS,KAAA,KAAU,KAAA,IAAS,QAAA,CAAS,MAAM,OAAA,KAAY,KAAA;AAGxE,EAAA,IAAI,UAAA;AACJ,EAAA,IAAI,UAAA;AAEJ,EAAA,IAAI,QAAA,CAAS,OAAA,CAAQ,cAAA,KAAmB,MAAA,EAAW;AACjD,IAAA,UAAA,GAAa,IAAI,eAAA,EAAgB;AACjC,IAAA,UAAA,GAAa,UAAA;AAAA,MACX,MAAM,UAAA,CAAY,KAAA,CAAM,IAAI,KAAA,CAAM,wBAAwB,CAAC,CAAA;AAAA,MAC3D,SAAS,OAAA,CAAQ;AAAA,KACnB;AAAA,EACF;AAEA,EAAA,IAAI;AACF,IAAA,KAAA,IAAS,UAAA,GAAa,CAAA,EAAG,UAAA,GAAa,MAAA,CAAO,QAAQ,UAAA,EAAA,EAAc;AACjE,MAAA,MAAM,KAAA,GAAQ,OAAO,UAAU,CAAA;AAE/B,MAAA,KAAA,IAAS,OAAA,GAAU,CAAA,EAAG,OAAA,IAAW,CAAA,GAAI,YAAY,OAAA,EAAA,EAAW;AAC1D,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAErE,QAAA,MAAM,SAAS,WAAA,CAAY,QAAA,CAAS,OAAA,CAAQ,gBAAA,EAAkB,YAAY,MAAM,CAAA;AAEhF,QAAA,MAAM,SAAS,MAAM,aAAA,CAAc,QAAQ,KAAA,EAAO,MAAA,EAAQ,SAAS,MAAM,CAAA;AACzE,QAAA,IAAI,MAAA,CAAO,EAAA,EAAI,OAAO,MAAA,CAAO,KAAA;AAE7B,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAGrE,QAAA,MAAM,QAAA,GAAW,MAAM,kBAAA,CAAmB,MAAA,CAAO,OAAO,OAAA,EAAS,UAAA,EAAY,QAAA,EAAU,QAAA,EAAU,UAAU,CAAA;AAC3G,QAAA,IAAI,aAAa,YAAA,EAAc;AAAA,MACjC;AAAA,IACF;AAGA,IAAA,MAAM,IAAIC,sBAAA;AAAA,MACR,2BAA2B,MAAA,CAAO,MAAM,sBAAsB,MAAA,CAAO,IAAA,CAAK,IAAI,CAAC,CAAA,CAAA;AAAA,MAC/E;AAAA,KACF;AAAA,EACF,CAAA,SAAE;AACA,IAAA,IAAI,eAAe,MAAA,EAAW;AAC5B,MAAA,YAAA,CAAa,UAAU,CAAA;AAAA,IACzB;AAAA,EACF;AACF;;;AC7GA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAOC,cAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAMC,eAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQC,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAASC,aAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQD,UAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiBE,mBAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AACtB,EAAA,MAAM,UAAA,GAAuCC,4BAAA,CAAwB,MAAA,CAAO,UAAU,CAAA;AAEtF,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QACL,MAAA;AAAA,QACA,CAAC,UAAU,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QACtG;AAAA,OACF;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QACtC,MAAA;AAAA,QACA,CAAC,UAAU,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QAC5G;AAAA,OACF;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.cjs","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","import type { BackoffConfig } from '@genui-a3/core'\n\n/**\n * Calculates the backoff delay for a given retry attempt.\n *\n * @param attempt - Zero-based attempt index (0 = first retry)\n * @param config - Backoff configuration with all fields required\n * @returns Delay in milliseconds\n */\nexport function calculateBackoff(attempt: number, config: Required<BackoffConfig>): number {\n let delay: number\n\n switch (config.strategy) {\n case 'exponential':\n delay = config.baseDelayMs * Math.pow(2, attempt)\n break\n case 'linear':\n delay = config.baseDelayMs * (attempt + 1)\n break\n case 'fixed':\n delay = config.baseDelayMs\n break\n }\n\n delay = Math.min(delay, config.maxDelayMs)\n\n if (config.jitter) {\n delay = Math.random() * delay\n }\n\n return delay\n}\n\n/**\n * Sleeps for the specified duration. Can be aborted via an AbortSignal.\n *\n * @param ms - Duration in milliseconds\n * @param signal - Optional AbortSignal to cancel the sleep early\n */\nexport function sleep(ms: number, signal?: AbortSignal): Promise<void> {\n return new Promise<void>((resolve, reject) => {\n if (signal?.aborted) {\n reject(signal.reason as Error)\n return\n }\n\n const timer = setTimeout(resolve, ms)\n\n signal?.addEventListener(\n 'abort',\n () => {\n clearTimeout(timer)\n reject(signal.reason as Error)\n },\n { once: true },\n )\n })\n}\n","import {\n A3ResilienceError,\n A3TimeoutError,\n DEFAULT_RESILIENCE_CONFIG,\n type ResilienceErrorEntry,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { calculateBackoff, sleep } from './backoff'\n\n/**\n * Builds an AbortSignal that fires when either the per-request timeout or total timeout expires.\n */\nfunction buildSignal(requestTimeoutMs?: number, totalAbort?: AbortSignal): AbortSignal | undefined {\n const signals: AbortSignal[] = []\n\n if (requestTimeoutMs !== undefined) {\n signals.push(AbortSignal.timeout(requestTimeoutMs))\n }\n\n if (totalAbort) {\n signals.push(totalAbort)\n }\n\n if (signals.length === 0) return undefined\n if (signals.length === 1) return signals[0]\n return AbortSignal.any(signals)\n}\n\ntype AttemptResult<T> = { ok: true; value: T } | { ok: false; error: Error }\n\nasync function attemptAction<T>(\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n model: string,\n signal: AbortSignal | undefined,\n attempt: number,\n errors: ResilienceErrorEntry[],\n): Promise<AttemptResult<T>> {\n try {\n const value = await action(model, signal)\n return { ok: true, value }\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, attempt, error: errorObj })\n return { ok: false, error: errorObj }\n }\n}\n\nfunction checkTotalTimeout(totalAbort: AbortController | undefined, timeoutMs: number | undefined, errors: ResilienceErrorEntry[]): void {\n if (totalAbort?.signal.aborted) {\n throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors)\n }\n}\n\nasync function handleAttemptError(\n errorObj: Error,\n attempt: number,\n maxRetries: number,\n retryAll: boolean,\n resolved: ResolvedResilienceConfig,\n totalAbort: AbortController | undefined,\n): Promise<'retry' | 'next-model'> {\n const isLastAttempt = attempt === 1 + maxRetries\n const isRetryable = retryAll || resolved.isRetryableError(errorObj)\n\n if (isRetryable && !isLastAttempt) {\n const delay = calculateBackoff(attempt - 1, resolved.backoff)\n await sleep(delay, totalAbort?.signal).catch(() => {\n // Sleep was aborted by total timeout — will be caught at top of loop\n })\n return 'retry'\n }\n\n return 'next-model'\n}\n\n/**\n * Executes an action with model fallback, retry, backoff, and timeout support.\n *\n * For each model (in priority order):\n * 1. Attempts the action up to `1 + maxAttempts` times\n * 2. On transient errors, waits with backoff before retrying\n * 3. On non-retryable errors (or after exhausting retries), falls back to the next model\n *\n * Throws `A3ResilienceError` with full error history when all models are exhausted.\n * Throws `A3TimeoutError` when the total timeout is exceeded.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model. Receives an optional AbortSignal.\n * @param config - Resolved resilience configuration (defaults applied if omitted)\n * @returns The result from the first successful attempt\n * @throws {A3ResilienceError} When all models and retries are exhausted\n * @throws {A3TimeoutError} When the total timeout is exceeded\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model, signal) => provider.call(model, params, { abortSignal: signal }),\n * resolvedConfig,\n * )\n * ```\n */\nexport async function executeWithFallback<T>(\n models: string[],\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n config?: ResolvedResilienceConfig,\n): Promise<T> {\n const resolved = config ?? DEFAULT_RESILIENCE_CONFIG\n const errors: ResilienceErrorEntry[] = []\n const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts\n const retryAll = resolved.retry !== false && resolved.retry.retryOn === 'all'\n\n // Total timeout controller\n let totalAbort: AbortController | undefined\n let totalTimer: ReturnType<typeof setTimeout> | undefined\n\n if (resolved.timeout.totalTimeoutMs !== undefined) {\n totalAbort = new AbortController()\n totalTimer = setTimeout(\n () => totalAbort!.abort(new Error('Total timeout exceeded')),\n resolved.timeout.totalTimeoutMs,\n )\n }\n\n try {\n for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {\n const model = models[modelIndex]\n\n for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal)\n // eslint-disable-next-line no-await-in-loop\n const result = await attemptAction(action, model, signal, attempt, errors)\n if (result.ok) return result.value\n\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n // eslint-disable-next-line no-await-in-loop\n const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort)\n if (decision === 'next-model') break\n }\n }\n\n // All models exhausted\n throw new A3ResilienceError(\n `All models failed after ${errors.length} total attempt(s): ${models.join(', ')}`,\n errors,\n )\n } finally {\n if (totalTimer !== undefined) {\n clearTimeout(totalTimer)\n }\n }\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport {\n resolveResilienceConfig,\n type Provider,\n type ProviderRequest,\n type ProviderResponse,\n type ProviderMessage,\n type BaseState,\n type StreamEvent,\n type ResilienceConfig,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '@providers/utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */\n resilience?: ResilienceConfig\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n const resilience: ResolvedResilienceConfig = resolveResilienceConfig(config.resilience)\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(\n models,\n (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(\n models,\n (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
@@ -1,4 +1,4 @@
1
- import { Provider } from '@genui-a3/core';
1
+ import { ResilienceConfig, Provider } from '@genui-a3/core';
2
2
 
3
3
  /**
4
4
  * Configuration for creating an OpenAI provider.
@@ -15,6 +15,8 @@ interface OpenAIProviderConfig {
15
15
  baseURL?: string;
16
16
  /** Optional OpenAI organization ID */
17
17
  organization?: string;
18
+ /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */
19
+ resilience?: ResilienceConfig;
18
20
  }
19
21
  /**
20
22
  * Creates an OpenAI provider instance.
@@ -1,4 +1,4 @@
1
- import { Provider } from '@genui-a3/core';
1
+ import { ResilienceConfig, Provider } from '@genui-a3/core';
2
2
 
3
3
  /**
4
4
  * Configuration for creating an OpenAI provider.
@@ -15,6 +15,8 @@ interface OpenAIProviderConfig {
15
15
  baseURL?: string;
16
16
  /** Optional OpenAI organization ID */
17
17
  organization?: string;
18
+ /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */
19
+ resilience?: ResilienceConfig;
18
20
  }
19
21
  /**
20
22
  * Creates an OpenAI provider instance.
@@ -1,5 +1,6 @@
1
1
  import { createOpenAI } from '@ai-sdk/openai';
2
2
  import { streamText, Output, jsonSchema, generateText } from 'ai';
3
+ import { resolveResilienceConfig, DEFAULT_RESILIENCE_CONFIG, A3ResilienceError, A3TimeoutError } from '@genui-a3/core';
3
4
  import { EventType } from '@ag-ui/client';
4
5
 
5
6
  // openai/index.ts
@@ -67,20 +68,119 @@ function extractDelta(partial, prevLength) {
67
68
  return chatbotMessage.slice(prevLength);
68
69
  }
69
70
 
71
+ // utils/backoff.ts
72
+ function calculateBackoff(attempt, config) {
73
+ let delay;
74
+ switch (config.strategy) {
75
+ case "exponential":
76
+ delay = config.baseDelayMs * Math.pow(2, attempt);
77
+ break;
78
+ case "linear":
79
+ delay = config.baseDelayMs * (attempt + 1);
80
+ break;
81
+ case "fixed":
82
+ delay = config.baseDelayMs;
83
+ break;
84
+ }
85
+ delay = Math.min(delay, config.maxDelayMs);
86
+ if (config.jitter) {
87
+ delay = Math.random() * delay;
88
+ }
89
+ return delay;
90
+ }
91
+ function sleep(ms, signal) {
92
+ return new Promise((resolve, reject) => {
93
+ if (signal?.aborted) {
94
+ reject(signal.reason);
95
+ return;
96
+ }
97
+ const timer = setTimeout(resolve, ms);
98
+ signal?.addEventListener(
99
+ "abort",
100
+ () => {
101
+ clearTimeout(timer);
102
+ reject(signal.reason);
103
+ },
104
+ { once: true }
105
+ );
106
+ });
107
+ }
108
+
70
109
  // utils/executeWithFallback.ts
71
- async function executeWithFallback(models, action) {
72
- for (let i = 0; i < models.length; i++) {
73
- const model = models[i];
74
- try {
75
- return await action(model);
76
- } catch (error) {
77
- const errorObj = error;
78
- if (i === models.length - 1) {
79
- throw errorObj;
110
+ function buildSignal(requestTimeoutMs, totalAbort) {
111
+ const signals = [];
112
+ if (requestTimeoutMs !== void 0) {
113
+ signals.push(AbortSignal.timeout(requestTimeoutMs));
114
+ }
115
+ if (totalAbort) {
116
+ signals.push(totalAbort);
117
+ }
118
+ if (signals.length === 0) return void 0;
119
+ if (signals.length === 1) return signals[0];
120
+ return AbortSignal.any(signals);
121
+ }
122
+ async function attemptAction(action, model, signal, attempt, errors) {
123
+ try {
124
+ const value = await action(model, signal);
125
+ return { ok: true, value };
126
+ } catch (error) {
127
+ const errorObj = error;
128
+ errors.push({ model, attempt, error: errorObj });
129
+ return { ok: false, error: errorObj };
130
+ }
131
+ }
132
+ function checkTotalTimeout(totalAbort, timeoutMs, errors) {
133
+ if (totalAbort?.signal.aborted) {
134
+ throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors);
135
+ }
136
+ }
137
+ async function handleAttemptError(errorObj, attempt, maxRetries, retryAll, resolved, totalAbort) {
138
+ const isLastAttempt = attempt === 1 + maxRetries;
139
+ const isRetryable = retryAll || resolved.isRetryableError(errorObj);
140
+ if (isRetryable && !isLastAttempt) {
141
+ const delay = calculateBackoff(attempt - 1, resolved.backoff);
142
+ await sleep(delay, totalAbort?.signal).catch(() => {
143
+ });
144
+ return "retry";
145
+ }
146
+ return "next-model";
147
+ }
148
+ async function executeWithFallback(models, action, config) {
149
+ const resolved = config ?? DEFAULT_RESILIENCE_CONFIG;
150
+ const errors = [];
151
+ const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts;
152
+ const retryAll = resolved.retry !== false && resolved.retry.retryOn === "all";
153
+ let totalAbort;
154
+ let totalTimer;
155
+ if (resolved.timeout.totalTimeoutMs !== void 0) {
156
+ totalAbort = new AbortController();
157
+ totalTimer = setTimeout(
158
+ () => totalAbort.abort(new Error("Total timeout exceeded")),
159
+ resolved.timeout.totalTimeoutMs
160
+ );
161
+ }
162
+ try {
163
+ for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {
164
+ const model = models[modelIndex];
165
+ for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {
166
+ checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
167
+ const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal);
168
+ const result = await attemptAction(action, model, signal, attempt, errors);
169
+ if (result.ok) return result.value;
170
+ checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors);
171
+ const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort);
172
+ if (decision === "next-model") break;
80
173
  }
81
174
  }
175
+ throw new A3ResilienceError(
176
+ `All models failed after ${errors.length} total attempt(s): ${models.join(", ")}`,
177
+ errors
178
+ );
179
+ } finally {
180
+ if (totalTimer !== void 0) {
181
+ clearTimeout(totalTimer);
182
+ }
82
183
  }
83
- throw new Error("All models failed");
84
184
  }
85
185
 
86
186
  // openai/index.ts
@@ -156,20 +256,23 @@ function createOpenAIProvider(config) {
156
256
  organization: config.organization
157
257
  });
158
258
  const models = config.models;
259
+ const resilience = resolveResilienceConfig(config.resilience);
159
260
  return {
160
261
  name: "openai",
161
262
  async sendRequest(request) {
162
263
  const messages = toAIMessages(request.messages);
163
264
  return executeWithFallback(
164
265
  models,
165
- (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
266
+ (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
267
+ resilience
166
268
  );
167
269
  },
168
270
  async *sendRequestStream(request) {
169
271
  const messages = toAIMessages(request.messages);
170
272
  const { result, reader, first } = await executeWithFallback(
171
273
  models,
172
- (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema)
274
+ (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),
275
+ resilience
173
276
  );
174
277
  yield* processOpenAIStream(result, reader, first, "openai", request.responseSchema);
175
278
  }
@@ -1 +1 @@
1
- {"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":[],"mappings":";;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAM,SAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;ACnFA,eAAsB,mBAAA,CAAuB,QAAkB,MAAA,EAAmD;AAGhH,EAAA,KAAA,IAAS,CAAA,GAAI,CAAA,EAAG,CAAA,GAAI,MAAA,CAAO,QAAQ,CAAA,EAAA,EAAK;AACtC,IAAA,MAAM,KAAA,GAAQ,OAAO,CAAC,CAAA;AAEtB,IAAA,IAAI;AAEF,MAAA,OAAO,MAAM,OAAO,KAAK,CAAA;AAAA,IAC3B,SAAS,KAAA,EAAO;AACd,MAAA,MAAM,QAAA,GAAW,KAAA;AAGjB,MAAA,IAAI,CAAA,KAAM,MAAA,CAAO,MAAA,GAAS,CAAA,EAAG;AAC3B,QAAA,MAAM,QAAA;AAAA,MACR;AAAA,IACF;AAAA,EACF;AAEA,EAAA,MAAM,IAAI,MAAM,mBAAmB,CAAA;AACrC;;;ACEA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAO,WAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAM,YAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAAS,UAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiB,YAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AAEtB,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UAClC,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OAC7F;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QAAoB,MAAA;AAAA,QAAQ,CAAC,UACnE,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc;AAAA,OACnG;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.js","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","/**\n * Executes an action with model fallback support.\n * Tries each model in order; if one fails, falls back to the next.\n * Throws the last error if all models fail.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model\n * @returns The result from the first successful model\n * @throws The error from the last model if all fail\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model) => provider.call(model, params),\n * )\n * ```\n */\nexport async function executeWithFallback<T>(models: string[], action: (model: string) => Promise<T>): Promise<T> {\n const errors: Array<{ model: string; error: Error }> = []\n\n for (let i = 0; i < models.length; i++) {\n const model = models[i]\n\n try {\n // eslint-disable-next-line no-await-in-loop\n return await action(model)\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, error: errorObj })\n\n if (i === models.length - 1) {\n throw errorObj\n }\n }\n }\n\n throw new Error('All models failed')\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport type {\n Provider,\n ProviderRequest,\n ProviderResponse,\n ProviderMessage,\n BaseState,\n StreamEvent,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '../utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(models, (model) =>\n sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(models, (model) =>\n sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
1
+ {"version":3,"sources":["../../openai/streamProcessor.ts","../../utils/backoff.ts","../../utils/executeWithFallback.ts","../../openai/index.ts"],"names":[],"mappings":";;;;;;AAoBA,gBAAuB,mBAAA,CACrB,YAAA,EACA,MAAA,EACA,KAAA,EACA,SACA,MAAA,EACqC;AACrC,EAAA,IAAI,iBAAA,GAAoB,CAAA;AAExB,EAAA,IAAI;AAEF,IAAA,IAAI,CAAC,MAAM,IAAA,EAAM;AACf,MAAA,MAAM,UAAU,KAAA,CAAM,KAAA;AACtB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAAA,IACF;AAGA,IAAA,IAAI,IAAA,GAAO,MAAM,MAAA,CAAO,IAAA,EAAK;AAC7B,IAAA,OAAO,CAAC,KAAK,IAAA,EAAM;AACjB,MAAA,MAAM,UAAU,IAAA,CAAK,KAAA;AACrB,MAAA,MAAM,KAAA,GAAQ,YAAA,CAAa,OAAA,EAAS,iBAAiB,CAAA;AACrD,MAAA,IAAI,KAAA,EAAO;AACT,QAAA,iBAAA,IAAqB,KAAA,CAAM,MAAA;AAC3B,QAAA,MAAM;AAAA,UACJ,MAAM,SAAA,CAAU,oBAAA;AAAA,UAChB,SAAA,EAAW,EAAA;AAAA,UACX,KAAA;AAAA,UACA;AAAA,SACF;AAAA,MACF;AAEA,MAAA,IAAA,GAAO,MAAM,OAAO,IAAA,EAAK;AAAA,IAC3B;AAGA,IAAA,MAAM,WAAA,GAAc,MAAM,YAAA,CAAa,MAAA;AAEvC,IAAA,IAAI,gBAAgB,IAAA,EAAM;AACxB,MAAA,MAAM;AAAA,QACJ,MAAM,SAAA,CAAU,SAAA;AAAA,QAChB,OAAA,EAAS,0CAAA;AAAA,QACT;AAAA,OACF;AACA,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,SAAA,GAAY,MAAA,CAAO,KAAA,CAAM,WAAW,CAAA;AAC1C,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,gBAAA;AAAA,MAChB,UAAA,EAAY,EAAA;AAAA,MACZ,SAAA,EAAW,EAAA;AAAA,MACX,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,SAAS,CAAA;AAAA,MACjC;AAAA,KACF;AAAA,EACF,SAAS,GAAA,EAAK;AACZ,IAAA,MAAM;AAAA,MACJ,MAAM,SAAA,CAAU,SAAA;AAAA,MAChB,OAAA,EAAS,CAAA,qBAAA,EAAyB,GAAA,CAAc,OAAO,CAAA,CAAA;AAAA,MACvD;AAAA,KACF;AAAA,EACF;AACF;AAKA,SAAS,YAAA,CAAa,SAAkC,UAAA,EAAmC;AACzF,EAAA,MAAM,iBAAiB,OAAA,CAAQ,cAAA;AAC/B,EAAA,IAAI,OAAO,cAAA,KAAmB,QAAA,IAAY,cAAA,CAAe,UAAU,UAAA,EAAY;AAC7E,IAAA,OAAO,IAAA;AAAA,EACT;AACA,EAAA,OAAO,cAAA,CAAe,MAAM,UAAU,CAAA;AACxC;;;AC5FO,SAAS,gBAAA,CAAiB,SAAiB,MAAA,EAAyC;AACzF,EAAA,IAAI,KAAA;AAEJ,EAAA,QAAQ,OAAO,QAAA;AAAU,IACvB,KAAK,aAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA,GAAc,IAAA,CAAK,GAAA,CAAI,GAAG,OAAO,CAAA;AAChD,MAAA;AAAA,IACF,KAAK,QAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,eAAe,OAAA,GAAU,CAAA,CAAA;AACxC,MAAA;AAAA,IACF,KAAK,OAAA;AACH,MAAA,KAAA,GAAQ,MAAA,CAAO,WAAA;AACf,MAAA;AAAA;AAGJ,EAAA,KAAA,GAAQ,IAAA,CAAK,GAAA,CAAI,KAAA,EAAO,MAAA,CAAO,UAAU,CAAA;AAEzC,EAAA,IAAI,OAAO,MAAA,EAAQ;AACjB,IAAA,KAAA,GAAQ,IAAA,CAAK,QAAO,GAAI,KAAA;AAAA,EAC1B;AAEA,EAAA,OAAO,KAAA;AACT;AAQO,SAAS,KAAA,CAAM,IAAY,MAAA,EAAqC;AACrE,EAAA,OAAO,IAAI,OAAA,CAAc,CAAC,OAAA,EAAS,MAAA,KAAW;AAC5C,IAAA,IAAI,QAAQ,OAAA,EAAS;AACnB,MAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAC7B,MAAA;AAAA,IACF;AAEA,IAAA,MAAM,KAAA,GAAQ,UAAA,CAAW,OAAA,EAAS,EAAE,CAAA;AAEpC,IAAA,MAAA,EAAQ,gBAAA;AAAA,MACN,OAAA;AAAA,MACA,MAAM;AACJ,QAAA,YAAA,CAAa,KAAK,CAAA;AAClB,QAAA,MAAA,CAAO,OAAO,MAAe,CAAA;AAAA,MAC/B,CAAA;AAAA,MACA,EAAE,MAAM,IAAA;AAAK,KACf;AAAA,EACF,CAAC,CAAA;AACH;;;AC7CA,SAAS,WAAA,CAAY,kBAA2B,UAAA,EAAmD;AACjG,EAAA,MAAM,UAAyB,EAAC;AAEhC,EAAA,IAAI,qBAAqB,MAAA,EAAW;AAClC,IAAA,OAAA,CAAQ,IAAA,CAAK,WAAA,CAAY,OAAA,CAAQ,gBAAgB,CAAC,CAAA;AAAA,EACpD;AAEA,EAAA,IAAI,UAAA,EAAY;AACd,IAAA,OAAA,CAAQ,KAAK,UAAU,CAAA;AAAA,EACzB;AAEA,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,MAAA;AACjC,EAAA,IAAI,OAAA,CAAQ,MAAA,KAAW,CAAA,EAAG,OAAO,QAAQ,CAAC,CAAA;AAC1C,EAAA,OAAO,WAAA,CAAY,IAAI,OAAO,CAAA;AAChC;AAIA,eAAe,aAAA,CACb,MAAA,EACA,KAAA,EACA,MAAA,EACA,SACA,MAAA,EAC2B;AAC3B,EAAA,IAAI;AACF,IAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,KAAA,EAAO,MAAM,CAAA;AACxC,IAAA,OAAO,EAAE,EAAA,EAAI,IAAA,EAAM,KAAA,EAAM;AAAA,EAC3B,SAAS,KAAA,EAAO;AACd,IAAA,MAAM,QAAA,GAAW,KAAA;AACjB,IAAA,MAAA,CAAO,KAAK,EAAE,KAAA,EAAO,OAAA,EAAS,KAAA,EAAO,UAAU,CAAA;AAC/C,IAAA,OAAO,EAAE,EAAA,EAAI,KAAA,EAAO,KAAA,EAAO,QAAA,EAAS;AAAA,EACtC;AACF;AAEA,SAAS,iBAAA,CAAkB,UAAA,EAAyC,SAAA,EAA+B,MAAA,EAAsC;AACvI,EAAA,IAAI,UAAA,EAAY,OAAO,OAAA,EAAS;AAC9B,IAAA,MAAM,IAAI,cAAA,CAAe,CAAA,iBAAA,EAAoB,SAAS,eAAe,MAAM,CAAA;AAAA,EAC7E;AACF;AAEA,eAAe,mBACb,QAAA,EACA,OAAA,EACA,UAAA,EACA,QAAA,EACA,UACA,UAAA,EACiC;AACjC,EAAA,MAAM,aAAA,GAAgB,YAAY,CAAA,GAAI,UAAA;AACtC,EAAA,MAAM,WAAA,GAAc,QAAA,IAAY,QAAA,CAAS,gBAAA,CAAiB,QAAQ,CAAA;AAElE,EAAA,IAAI,WAAA,IAAe,CAAC,aAAA,EAAe;AACjC,IAAA,MAAM,KAAA,GAAQ,gBAAA,CAAiB,OAAA,GAAU,CAAA,EAAG,SAAS,OAAO,CAAA;AAC5D,IAAA,MAAM,MAAM,KAAA,EAAO,UAAA,EAAY,MAAM,CAAA,CAAE,MAAM,MAAM;AAAA,IAEnD,CAAC,CAAA;AACD,IAAA,OAAO,OAAA;AAAA,EACT;AAEA,EAAA,OAAO,YAAA;AACT;AA6BA,eAAsB,mBAAA,CACpB,MAAA,EACA,MAAA,EACA,MAAA,EACY;AACZ,EAAA,MAAM,WAAW,MAAA,IAAU,yBAAA;AAC3B,EAAA,MAAM,SAAiC,EAAC;AACxC,EAAA,MAAM,aAAa,QAAA,CAAS,KAAA,KAAU,KAAA,GAAQ,CAAA,GAAI,SAAS,KAAA,CAAM,WAAA;AACjE,EAAA,MAAM,WAAW,QAAA,CAAS,KAAA,KAAU,KAAA,IAAS,QAAA,CAAS,MAAM,OAAA,KAAY,KAAA;AAGxE,EAAA,IAAI,UAAA;AACJ,EAAA,IAAI,UAAA;AAEJ,EAAA,IAAI,QAAA,CAAS,OAAA,CAAQ,cAAA,KAAmB,MAAA,EAAW;AACjD,IAAA,UAAA,GAAa,IAAI,eAAA,EAAgB;AACjC,IAAA,UAAA,GAAa,UAAA;AAAA,MACX,MAAM,UAAA,CAAY,KAAA,CAAM,IAAI,KAAA,CAAM,wBAAwB,CAAC,CAAA;AAAA,MAC3D,SAAS,OAAA,CAAQ;AAAA,KACnB;AAAA,EACF;AAEA,EAAA,IAAI;AACF,IAAA,KAAA,IAAS,UAAA,GAAa,CAAA,EAAG,UAAA,GAAa,MAAA,CAAO,QAAQ,UAAA,EAAA,EAAc;AACjE,MAAA,MAAM,KAAA,GAAQ,OAAO,UAAU,CAAA;AAE/B,MAAA,KAAA,IAAS,OAAA,GAAU,CAAA,EAAG,OAAA,IAAW,CAAA,GAAI,YAAY,OAAA,EAAA,EAAW;AAC1D,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAErE,QAAA,MAAM,SAAS,WAAA,CAAY,QAAA,CAAS,OAAA,CAAQ,gBAAA,EAAkB,YAAY,MAAM,CAAA;AAEhF,QAAA,MAAM,SAAS,MAAM,aAAA,CAAc,QAAQ,KAAA,EAAO,MAAA,EAAQ,SAAS,MAAM,CAAA;AACzE,QAAA,IAAI,MAAA,CAAO,EAAA,EAAI,OAAO,MAAA,CAAO,KAAA;AAE7B,QAAA,iBAAA,CAAkB,UAAA,EAAY,QAAA,CAAS,OAAA,CAAQ,cAAA,EAAgB,MAAM,CAAA;AAGrE,QAAA,MAAM,QAAA,GAAW,MAAM,kBAAA,CAAmB,MAAA,CAAO,OAAO,OAAA,EAAS,UAAA,EAAY,QAAA,EAAU,QAAA,EAAU,UAAU,CAAA;AAC3G,QAAA,IAAI,aAAa,YAAA,EAAc;AAAA,MACjC;AAAA,IACF;AAGA,IAAA,MAAM,IAAI,iBAAA;AAAA,MACR,2BAA2B,MAAA,CAAO,MAAM,sBAAsB,MAAA,CAAO,IAAA,CAAK,IAAI,CAAC,CAAA,CAAA;AAAA,MAC/E;AAAA,KACF;AAAA,EACF,CAAA,SAAE;AACA,IAAA,IAAI,eAAe,MAAA,EAAW;AAC5B,MAAA,YAAA,CAAa,UAAU,CAAA;AAAA,IACzB;AAAA,EACF;AACF;;;AC7GA,SAAS,oBAAoB,MAAA,EAAgC;AAC3D,EAAA,MAAM,MAAA,GAAS,EAAE,GAAG,MAAA,EAAO;AAC3B,EAAA,IAAI,MAAA,CAAO,IAAA,KAAS,QAAA,IAAY,MAAA,CAAO,UAAA,EAAY;AACjD,IAAA,MAAA,CAAO,oBAAA,GAAuB,KAAA;AAC9B,IAAA,MAAA,CAAO,QAAA,GAAW,MAAA,CAAO,IAAA,CAAK,MAAA,CAAO,UAAqC,CAAA;AAC1E,IAAA,MAAM,QAAQ,MAAA,CAAO,UAAA;AACrB,IAAA,MAAM,cAA0C,EAAC;AACjD,IAAA,KAAA,MAAW,CAAC,GAAA,EAAK,KAAK,KAAK,MAAA,CAAO,OAAA,CAAQ,KAAK,CAAA,EAAG;AAChD,MAAA,WAAA,CAAY,GAAG,CAAA,GAAI,mBAAA,CAAoB,KAAK,CAAA;AAAA,IAC9C;AACA,IAAA,MAAA,CAAO,UAAA,GAAa,WAAA;AAAA,EACtB;AACA,EAAA,IAAI,MAAA,CAAO,KAAA,IAAS,OAAO,MAAA,CAAO,UAAU,QAAA,EAAU;AACpD,IAAA,MAAA,CAAO,KAAA,GAAQ,mBAAA,CAAoB,MAAA,CAAO,KAAmB,CAAA;AAAA,EAC/D;AACA,EAAA,KAAA,MAAW,OAAA,IAAW,CAAC,OAAA,EAAS,OAAA,EAAS,OAAO,CAAA,EAAY;AAC1D,IAAA,IAAI,KAAA,CAAM,OAAA,CAAQ,MAAA,CAAO,OAAO,CAAC,CAAA,EAAG;AAClC,MAAA,MAAA,CAAO,OAAO,CAAA,GAAK,MAAA,CAAO,OAAO,CAAA,CAAmB,IAAI,CAAC,CAAA,KAAM,mBAAA,CAAoB,CAAC,CAAC,CAAA;AAAA,IACvF;AAAA,EACF;AACA,EAAA,OAAO,MAAA;AACT;AAQA,SAAS,eAAe,SAAA,EAA8C;AACpE,EAAA,MAAM,MAAA,GAAS,mBAAA,CAAoB,SAAA,CAAU,YAAA,EAA4B,CAAA;AACzE,EAAA,OAAO,WAAW,MAAA,EAAiB;AAAA,IACjC,QAAA,EAAU,CAAC,KAAA,KAAmB;AAC5B,MAAA,MAAM,MAAA,GAAS,SAAA,CAAU,SAAA,CAAU,KAAK,CAAA;AACxC,MAAA,OAAO,MAAA,CAAO,OAAA,GACV,EAAE,OAAA,EAAS,MAAe,KAAA,EAAO,MAAA,CAAO,IAAA,EAAK,GAC7C,EAAE,OAAA,EAAS,KAAA,EAAgB,KAAA,EAAO,OAAO,KAAA,EAAM;AAAA,IACrD;AAAA,GACD,CAAA;AACH;AAEA,SAAS,aAAa,QAAA,EAA6C;AACjE,EAAA,OAAO,QAAA,CAAS,GAAA,CAAI,CAAC,GAAA,MAAS;AAAA,IAC5B,MAAM,GAAA,CAAI,IAAA;AAAA,IACV,SAAS,GAAA,CAAI;AAAA,GACf,CAAE,CAAA;AACJ;AAEA,eAAe,aAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EAC2B;AAC3B,EAAA,MAAM,MAAA,GAAS,MAAM,YAAA,CAAa;AAAA,IAChC,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAED,EAAA,OAAO;AAAA,IACL,OAAA,EAAS,IAAA,CAAK,SAAA,CAAU,MAAA,CAAO,MAAM,CAAA;AAAA,IACrC,KAAA,EAAO;AAAA,MACL,WAAA,EAAa,MAAA,CAAO,KAAA,CAAM,WAAA,IAAe,CAAA;AAAA,MACzC,YAAA,EAAc,MAAA,CAAO,KAAA,CAAM,YAAA,IAAgB,CAAA;AAAA,MAC3C,cAAc,MAAA,CAAO,KAAA,CAAM,eAAe,CAAA,KAAM,MAAA,CAAO,MAAM,YAAA,IAAgB,CAAA;AAAA;AAC/E,GACF;AACF;AAEA,eAAe,mBAAA,CACb,cAAA,EACA,KAAA,EACA,MAAA,EACA,UACA,MAAA,EACA;AACA,EAAA,MAAM,SAAS,UAAA,CAAW;AAAA,IACxB,KAAA,EAAO,eAAe,KAAK,CAAA;AAAA,IAC3B,MAAA;AAAA,IACA,QAAA;AAAA,IACA,MAAA,EAAQ,OAAO,MAAA,CAAO,EAAE,QAAQ,cAAA,CAAe,MAAM,GAAG;AAAA,GACzD,CAAA;AAGD,EAAA,MAAM,gBAAgB,MAAA,CAAO,mBAAA;AAC7B,EAAA,MAAM,MAAA,GAAS,aAAA,CAAc,MAAA,CAAO,aAAa,CAAA,EAAE;AACnD,EAAA,MAAM,KAAA,GAAQ,MAAM,MAAA,CAAO,IAAA,EAAK;AAEhC,EAAA,OAAO,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAM;AACjC;AAoBO,SAAS,qBAAqB,MAAA,EAAwC;AAC3E,EAAA,MAAM,iBAAiB,YAAA,CAAa;AAAA,IAClC,QAAQ,MAAA,CAAO,MAAA;AAAA,IACf,SAAS,MAAA,CAAO,OAAA;AAAA,IAChB,cAAc,MAAA,CAAO;AAAA,GACtB,CAAA;AACD,EAAA,MAAM,SAAS,MAAA,CAAO,MAAA;AACtB,EAAA,MAAM,UAAA,GAAuC,uBAAA,CAAwB,MAAA,CAAO,UAAU,CAAA;AAEtF,EAAA,OAAO;AAAA,IACL,IAAA,EAAM,QAAA;AAAA,IAEN,MAAM,YAAY,OAAA,EAAqD;AACrE,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,OAAO,mBAAA;AAAA,QACL,MAAA;AAAA,QACA,CAAC,UAAU,aAAA,CAAc,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QACtG;AAAA,OACF;AAAA,IACF,CAAA;AAAA,IAEA,OAAO,kBACL,OAAA,EACqC;AACrC,MAAA,MAAM,QAAA,GAAW,YAAA,CAAa,OAAA,CAAQ,QAAQ,CAAA;AAE9C,MAAA,MAAM,EAAE,MAAA,EAAQ,MAAA,EAAQ,KAAA,KAAU,MAAM,mBAAA;AAAA,QACtC,MAAA;AAAA,QACA,CAAC,UAAU,mBAAA,CAAoB,cAAA,EAAgB,OAAO,OAAA,CAAQ,YAAA,EAAc,QAAA,EAAU,OAAA,CAAQ,cAAc,CAAA;AAAA,QAC5G;AAAA,OACF;AAEA,MAAA,OAAO,oBAA4B,MAAA,EAAQ,MAAA,EAAQ,KAAA,EAAO,QAAA,EAAU,QAAQ,cAAc,CAAA;AAAA,IAC5F;AAAA,GACF;AACF","file":"index.js","sourcesContent":["import { EventType } from '@ag-ui/client'\nimport { ZodType } from 'zod'\nimport type { AgentId, StreamEvent, BaseState } from '@genui-a3/core'\nimport type { StreamTextResult, ToolSet } from 'ai'\n\n/**\n * Processes an OpenAI streaming response (via Vercel AI SDK) into AG-UI events.\n *\n * Uses `partialOutputStream` from `streamText` + `Output.object()` to receive\n * progressively-built partial objects. Tracks `chatbotMessage` growth to yield\n * TEXT_MESSAGE_CONTENT deltas. After the stream completes, validates the final\n * object and yields TOOL_CALL_RESULT.\n *\n * @param streamResult - The streamText result containing partialOutputStream and output promise\n * @param reader - Pre-started async iterator for the partial object stream\n * @param first - The first iteration result (already consumed to trigger the API call)\n * @param agentId - Agent identifier for event tagging\n * @param schema - Zod schema for final response validation\n * @returns Async generator of AG-UI stream events\n */\nexport async function* processOpenAIStream<TState extends BaseState = BaseState>(\n streamResult: StreamTextResult<ToolSet, never>,\n reader: AsyncIterator<unknown>,\n first: IteratorResult<unknown>,\n agentId: AgentId,\n schema: ZodType,\n): AsyncGenerator<StreamEvent<TState>> {\n let prevMessageLength = 0\n\n try {\n // Process the first partial (already consumed to trigger the API call)\n if (!first.done) {\n const partial = first.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n }\n\n // Process remaining partials\n let next = await reader.next()\n while (!next.done) {\n const partial = next.value as Record<string, unknown>\n const delta = extractDelta(partial, prevMessageLength)\n if (delta) {\n prevMessageLength += delta.length\n yield {\n type: EventType.TEXT_MESSAGE_CONTENT,\n messageId: '',\n delta,\n agentId,\n } as StreamEvent<TState>\n }\n // eslint-disable-next-line no-await-in-loop\n next = await reader.next()\n }\n\n // Stream complete — await and validate the final object\n const finalObject = await streamResult.output\n\n if (finalObject === null) {\n yield {\n type: EventType.RUN_ERROR,\n message: 'OpenAI stream completed with null output',\n agentId,\n } as StreamEvent<TState>\n return\n }\n\n const validated = schema.parse(finalObject)\n yield {\n type: EventType.TOOL_CALL_RESULT,\n toolCallId: '',\n messageId: '',\n content: JSON.stringify(validated),\n agentId,\n } as StreamEvent<TState>\n } catch (err) {\n yield {\n type: EventType.RUN_ERROR,\n message: `OpenAI stream error: ${(err as Error).message}`,\n agentId,\n } as StreamEvent<TState>\n }\n}\n\n/**\n * Extracts the new portion of chatbotMessage from a partial object.\n */\nfunction extractDelta(partial: Record<string, unknown>, prevLength: number): string | null {\n const chatbotMessage = partial.chatbotMessage\n if (typeof chatbotMessage !== 'string' || chatbotMessage.length <= prevLength) {\n return null\n }\n return chatbotMessage.slice(prevLength)\n}\n","import type { BackoffConfig } from '@genui-a3/core'\n\n/**\n * Calculates the backoff delay for a given retry attempt.\n *\n * @param attempt - Zero-based attempt index (0 = first retry)\n * @param config - Backoff configuration with all fields required\n * @returns Delay in milliseconds\n */\nexport function calculateBackoff(attempt: number, config: Required<BackoffConfig>): number {\n let delay: number\n\n switch (config.strategy) {\n case 'exponential':\n delay = config.baseDelayMs * Math.pow(2, attempt)\n break\n case 'linear':\n delay = config.baseDelayMs * (attempt + 1)\n break\n case 'fixed':\n delay = config.baseDelayMs\n break\n }\n\n delay = Math.min(delay, config.maxDelayMs)\n\n if (config.jitter) {\n delay = Math.random() * delay\n }\n\n return delay\n}\n\n/**\n * Sleeps for the specified duration. Can be aborted via an AbortSignal.\n *\n * @param ms - Duration in milliseconds\n * @param signal - Optional AbortSignal to cancel the sleep early\n */\nexport function sleep(ms: number, signal?: AbortSignal): Promise<void> {\n return new Promise<void>((resolve, reject) => {\n if (signal?.aborted) {\n reject(signal.reason as Error)\n return\n }\n\n const timer = setTimeout(resolve, ms)\n\n signal?.addEventListener(\n 'abort',\n () => {\n clearTimeout(timer)\n reject(signal.reason as Error)\n },\n { once: true },\n )\n })\n}\n","import {\n A3ResilienceError,\n A3TimeoutError,\n DEFAULT_RESILIENCE_CONFIG,\n type ResilienceErrorEntry,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { calculateBackoff, sleep } from './backoff'\n\n/**\n * Builds an AbortSignal that fires when either the per-request timeout or total timeout expires.\n */\nfunction buildSignal(requestTimeoutMs?: number, totalAbort?: AbortSignal): AbortSignal | undefined {\n const signals: AbortSignal[] = []\n\n if (requestTimeoutMs !== undefined) {\n signals.push(AbortSignal.timeout(requestTimeoutMs))\n }\n\n if (totalAbort) {\n signals.push(totalAbort)\n }\n\n if (signals.length === 0) return undefined\n if (signals.length === 1) return signals[0]\n return AbortSignal.any(signals)\n}\n\ntype AttemptResult<T> = { ok: true; value: T } | { ok: false; error: Error }\n\nasync function attemptAction<T>(\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n model: string,\n signal: AbortSignal | undefined,\n attempt: number,\n errors: ResilienceErrorEntry[],\n): Promise<AttemptResult<T>> {\n try {\n const value = await action(model, signal)\n return { ok: true, value }\n } catch (error) {\n const errorObj = error as Error\n errors.push({ model, attempt, error: errorObj })\n return { ok: false, error: errorObj }\n }\n}\n\nfunction checkTotalTimeout(totalAbort: AbortController | undefined, timeoutMs: number | undefined, errors: ResilienceErrorEntry[]): void {\n if (totalAbort?.signal.aborted) {\n throw new A3TimeoutError(`Total timeout of ${timeoutMs}ms exceeded`, errors)\n }\n}\n\nasync function handleAttemptError(\n errorObj: Error,\n attempt: number,\n maxRetries: number,\n retryAll: boolean,\n resolved: ResolvedResilienceConfig,\n totalAbort: AbortController | undefined,\n): Promise<'retry' | 'next-model'> {\n const isLastAttempt = attempt === 1 + maxRetries\n const isRetryable = retryAll || resolved.isRetryableError(errorObj)\n\n if (isRetryable && !isLastAttempt) {\n const delay = calculateBackoff(attempt - 1, resolved.backoff)\n await sleep(delay, totalAbort?.signal).catch(() => {\n // Sleep was aborted by total timeout — will be caught at top of loop\n })\n return 'retry'\n }\n\n return 'next-model'\n}\n\n/**\n * Executes an action with model fallback, retry, backoff, and timeout support.\n *\n * For each model (in priority order):\n * 1. Attempts the action up to `1 + maxAttempts` times\n * 2. On transient errors, waits with backoff before retrying\n * 3. On non-retryable errors (or after exhausting retries), falls back to the next model\n *\n * Throws `A3ResilienceError` with full error history when all models are exhausted.\n * Throws `A3TimeoutError` when the total timeout is exceeded.\n *\n * @param models - Model identifiers in priority order\n * @param action - Async action to attempt with each model. Receives an optional AbortSignal.\n * @param config - Resolved resilience configuration (defaults applied if omitted)\n * @returns The result from the first successful attempt\n * @throws {A3ResilienceError} When all models and retries are exhausted\n * @throws {A3TimeoutError} When the total timeout is exceeded\n *\n * @example\n * ```typescript\n * const result = await executeWithFallback(\n * ['model-primary', 'model-fallback'],\n * (model, signal) => provider.call(model, params, { abortSignal: signal }),\n * resolvedConfig,\n * )\n * ```\n */\nexport async function executeWithFallback<T>(\n models: string[],\n action: (model: string, signal?: AbortSignal) => Promise<T>,\n config?: ResolvedResilienceConfig,\n): Promise<T> {\n const resolved = config ?? DEFAULT_RESILIENCE_CONFIG\n const errors: ResilienceErrorEntry[] = []\n const maxRetries = resolved.retry === false ? 0 : resolved.retry.maxAttempts\n const retryAll = resolved.retry !== false && resolved.retry.retryOn === 'all'\n\n // Total timeout controller\n let totalAbort: AbortController | undefined\n let totalTimer: ReturnType<typeof setTimeout> | undefined\n\n if (resolved.timeout.totalTimeoutMs !== undefined) {\n totalAbort = new AbortController()\n totalTimer = setTimeout(\n () => totalAbort!.abort(new Error('Total timeout exceeded')),\n resolved.timeout.totalTimeoutMs,\n )\n }\n\n try {\n for (let modelIndex = 0; modelIndex < models.length; modelIndex++) {\n const model = models[modelIndex]\n\n for (let attempt = 1; attempt <= 1 + maxRetries; attempt++) {\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n const signal = buildSignal(resolved.timeout.requestTimeoutMs, totalAbort?.signal)\n // eslint-disable-next-line no-await-in-loop\n const result = await attemptAction(action, model, signal, attempt, errors)\n if (result.ok) return result.value\n\n checkTotalTimeout(totalAbort, resolved.timeout.totalTimeoutMs, errors)\n\n // eslint-disable-next-line no-await-in-loop\n const decision = await handleAttemptError(result.error, attempt, maxRetries, retryAll, resolved, totalAbort)\n if (decision === 'next-model') break\n }\n }\n\n // All models exhausted\n throw new A3ResilienceError(\n `All models failed after ${errors.length} total attempt(s): ${models.join(', ')}`,\n errors,\n )\n } finally {\n if (totalTimer !== undefined) {\n clearTimeout(totalTimer)\n }\n }\n}\n","import { createOpenAI } from '@ai-sdk/openai'\nimport { generateText, streamText, Output, ModelMessage, jsonSchema } from 'ai'\nimport {\n resolveResilienceConfig,\n type Provider,\n type ProviderRequest,\n type ProviderResponse,\n type ProviderMessage,\n type BaseState,\n type StreamEvent,\n type ResilienceConfig,\n type ResolvedResilienceConfig,\n} from '@genui-a3/core'\nimport { processOpenAIStream } from './streamProcessor'\nimport { executeWithFallback } from '@providers/utils/executeWithFallback'\n\n/**\n * Configuration for creating an OpenAI provider.\n */\nexport interface OpenAIProviderConfig {\n /** OpenAI API key. Defaults to OPENAI_API_KEY env var (OpenAI SDK default). */\n apiKey?: string\n /**\n * Model identifiers in order of preference (first = primary, rest = fallbacks).\n * e.g. ['gpt-4o', 'gpt-4o-mini']\n */\n models: string[]\n /** Optional base URL for Azure OpenAI or compatible endpoints */\n baseURL?: string\n /** Optional OpenAI organization ID */\n organization?: string\n /** Resilience configuration (retry, backoff, timeout). Uses industry-standard defaults if omitted. */\n resilience?: ResilienceConfig\n}\n\ntype JsonSchema = Record<string, unknown>\n\n/**\n * Enforces strict JSON schema constraints required by OpenAI's structured output API.\n * OpenAI requires all object properties — including optional ones — to be in the `required` array.\n * This function recursively adds `required` and `additionalProperties: false` to all objects.\n *\n * @param schema - JSON schema to enforce\n * @returns Enforced JSON schema\n */\nfunction enforceStrictSchema(schema: JsonSchema): JsonSchema {\n const result = { ...schema }\n if (result.type === 'object' && result.properties) {\n result.additionalProperties = false\n result.required = Object.keys(result.properties as Record<string, unknown>)\n const props = result.properties as Record<string, JsonSchema>\n const strictProps: Record<string, JsonSchema> = {}\n for (const [key, value] of Object.entries(props)) {\n strictProps[key] = enforceStrictSchema(value)\n }\n result.properties = strictProps\n }\n if (result.items && typeof result.items === 'object') {\n result.items = enforceStrictSchema(result.items as JsonSchema)\n }\n for (const keyword of ['anyOf', 'oneOf', 'allOf'] as const) {\n if (Array.isArray(result[keyword])) {\n result[keyword] = (result[keyword] as JsonSchema[]).map((s) => enforceStrictSchema(s))\n }\n }\n return result\n}\n\n/**\n * Converts a Zod schema to an OpenAI-compatible strict JSON schema wrapped for the Vercel AI SDK.\n *\n * @param zodSchema - Zod schema to convert\n * @returns JSON schema wrapped for Vercel AI SDK\n */\nfunction toOpenAISchema(zodSchema: ProviderRequest['responseSchema']) {\n const strict = enforceStrictSchema(zodSchema.toJSONSchema() as JsonSchema)\n return jsonSchema(strict as never, {\n validate: (value: unknown) => {\n const result = zodSchema.safeParse(value)\n return result.success\n ? { success: true as const, value: result.data }\n : { success: false as const, error: result.error }\n },\n })\n}\n\nfunction toAIMessages(messages: ProviderMessage[]): ModelMessage[] {\n return messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n }))\n}\n\nasync function sendWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n): Promise<ProviderResponse> {\n const result = await generateText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n return {\n content: JSON.stringify(result.output),\n usage: {\n inputTokens: result.usage.inputTokens ?? 0,\n outputTokens: result.usage.outputTokens ?? 0,\n totalTokens: (result.usage.inputTokens ?? 0) + (result.usage.outputTokens ?? 0),\n },\n }\n}\n\nasync function sendStreamWithModel(\n openaiProvider: ReturnType<typeof createOpenAI>,\n model: string,\n system: string,\n messages: ModelMessage[],\n schema: ProviderRequest['responseSchema'],\n) {\n const result = streamText({\n model: openaiProvider(model),\n system,\n messages,\n output: Output.object({ schema: toOpenAISchema(schema) }),\n })\n\n // Force the API call to start so executeWithFallback can catch connection errors\n const partialStream = result.partialOutputStream\n const reader = partialStream[Symbol.asyncIterator]()\n const first = await reader.next()\n\n return { result, reader, first }\n}\n\n/**\n * Creates an OpenAI provider instance.\n *\n * Uses the Vercel AI SDK (`ai` + `@ai-sdk/openai`) for structured output via\n * `generateText` + `Output.object()` (blocking) and `streamText` + `Output.object()`\n * (streaming). The AI SDK handles Zod-to-JSON-schema conversion, partial JSON\n * parsing, and validation internally.\n *\n * @param config - OpenAI provider configuration\n * @returns A Provider implementation using OpenAI\n *\n * @example\n * ```typescript\n * const provider = createOpenAIProvider({\n * models: ['gpt-4o', 'gpt-4o-mini'],\n * })\n * ```\n */\nexport function createOpenAIProvider(config: OpenAIProviderConfig): Provider {\n const openaiProvider = createOpenAI({\n apiKey: config.apiKey,\n baseURL: config.baseURL,\n organization: config.organization,\n })\n const models = config.models\n const resilience: ResolvedResilienceConfig = resolveResilienceConfig(config.resilience)\n\n return {\n name: 'openai',\n\n async sendRequest(request: ProviderRequest): Promise<ProviderResponse> {\n const messages = toAIMessages(request.messages)\n\n return executeWithFallback(\n models,\n (model) => sendWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n },\n\n async *sendRequestStream<TState extends BaseState = BaseState>(\n request: ProviderRequest,\n ): AsyncGenerator<StreamEvent<TState>> {\n const messages = toAIMessages(request.messages)\n\n const { result, reader, first } = await executeWithFallback(\n models,\n (model) => sendStreamWithModel(openaiProvider, model, request.systemPrompt, messages, request.responseSchema),\n resilience,\n )\n\n yield* processOpenAIStream<TState>(result, reader, first, 'openai', request.responseSchema)\n },\n }\n}\n"]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@genui-a3/providers",
3
- "version": "0.0.4",
3
+ "version": "0.0.6",
4
4
  "description": "Provider implementations for the A3 agentic framework",
5
5
  "type": "module",
6
6
  "exports": {