@ai-sdk/openai 3.0.33 → 3.0.34
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.d.mts +3 -0
- package/dist/index.d.ts +3 -0
- package/dist/index.js +45 -29
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +45 -29
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +3 -0
- package/dist/internal/index.d.ts +3 -0
- package/dist/internal/index.js +44 -28
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +44 -28
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +3 -3
- package/src/image/openai-image-options.ts +1 -0
- package/src/responses/convert-to-openai-responses-input.ts +7 -2
- package/src/responses/openai-responses-api.ts +4 -0
- package/src/responses/openai-responses-language-model.ts +11 -0
- package/src/responses/openai-responses-provider-metadata.ts +1 -0
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,11 @@
|
|
|
1
1
|
# @ai-sdk/openai
|
|
2
2
|
|
|
3
|
+
## 3.0.34
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- 66a374c: Support `phase` parameter on Responses API message items. The `phase` field (`'commentary'` or `'final_answer'`) is returned by models like `gpt-5.3-codex` on assistant message output items and must be preserved when sending follow-up requests. The phase value is available in `providerMetadata.openai.phase` on text parts and is automatically included on assistant messages sent back to the API.
|
|
8
|
+
|
|
3
9
|
## 3.0.33
|
|
4
10
|
|
|
5
11
|
### Patch Changes
|
package/dist/index.d.mts
CHANGED
|
@@ -220,6 +220,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
220
220
|
item: {
|
|
221
221
|
type: "message";
|
|
222
222
|
id: string;
|
|
223
|
+
phase?: "commentary" | "final_answer" | null | undefined;
|
|
223
224
|
} | {
|
|
224
225
|
type: "reasoning";
|
|
225
226
|
id: string;
|
|
@@ -315,6 +316,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
315
316
|
item: {
|
|
316
317
|
type: "message";
|
|
317
318
|
id: string;
|
|
319
|
+
phase?: "commentary" | "final_answer" | null | undefined;
|
|
318
320
|
} | {
|
|
319
321
|
type: "reasoning";
|
|
320
322
|
id: string;
|
|
@@ -1027,6 +1029,7 @@ type OpenaiResponsesProviderMetadata = {
|
|
|
1027
1029
|
};
|
|
1028
1030
|
type ResponsesTextProviderMetadata = {
|
|
1029
1031
|
itemId: string;
|
|
1032
|
+
phase?: 'commentary' | 'final_answer' | null;
|
|
1030
1033
|
annotations?: Array<ResponsesOutputTextAnnotationProviderMetadata>;
|
|
1031
1034
|
};
|
|
1032
1035
|
type OpenaiResponsesTextProviderMetadata = {
|
package/dist/index.d.ts
CHANGED
|
@@ -220,6 +220,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
220
220
|
item: {
|
|
221
221
|
type: "message";
|
|
222
222
|
id: string;
|
|
223
|
+
phase?: "commentary" | "final_answer" | null | undefined;
|
|
223
224
|
} | {
|
|
224
225
|
type: "reasoning";
|
|
225
226
|
id: string;
|
|
@@ -315,6 +316,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
315
316
|
item: {
|
|
316
317
|
type: "message";
|
|
317
318
|
id: string;
|
|
319
|
+
phase?: "commentary" | "final_answer" | null | undefined;
|
|
318
320
|
} | {
|
|
319
321
|
type: "reasoning";
|
|
320
322
|
id: string;
|
|
@@ -1027,6 +1029,7 @@ type OpenaiResponsesProviderMetadata = {
|
|
|
1027
1029
|
};
|
|
1028
1030
|
type ResponsesTextProviderMetadata = {
|
|
1029
1031
|
itemId: string;
|
|
1032
|
+
phase?: 'commentary' | 'final_answer' | null;
|
|
1030
1033
|
annotations?: Array<ResponsesOutputTextAnnotationProviderMetadata>;
|
|
1031
1034
|
};
|
|
1032
1035
|
type OpenaiResponsesTextProviderMetadata = {
|
package/dist/index.js
CHANGED
|
@@ -1753,6 +1753,7 @@ var modelMaxImagesPerCall = {
|
|
|
1753
1753
|
"chatgpt-image-latest": 10
|
|
1754
1754
|
};
|
|
1755
1755
|
var defaultResponseFormatPrefixes = [
|
|
1756
|
+
"chatgpt-image-",
|
|
1756
1757
|
"gpt-image-1-mini",
|
|
1757
1758
|
"gpt-image-1.5",
|
|
1758
1759
|
"gpt-image-1"
|
|
@@ -2574,7 +2575,7 @@ async function convertToOpenAIResponsesInput({
|
|
|
2574
2575
|
hasShellTool = false,
|
|
2575
2576
|
hasApplyPatchTool = false
|
|
2576
2577
|
}) {
|
|
2577
|
-
var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l
|
|
2578
|
+
var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l;
|
|
2578
2579
|
const input = [];
|
|
2579
2580
|
const warnings = [];
|
|
2580
2581
|
const processedApprovalIds = /* @__PURE__ */ new Set();
|
|
@@ -2655,7 +2656,9 @@ async function convertToOpenAIResponsesInput({
|
|
|
2655
2656
|
for (const part of content) {
|
|
2656
2657
|
switch (part.type) {
|
|
2657
2658
|
case "text": {
|
|
2658
|
-
const
|
|
2659
|
+
const providerOpts = (_a = part.providerOptions) == null ? void 0 : _a[providerOptionsName];
|
|
2660
|
+
const id = providerOpts == null ? void 0 : providerOpts.itemId;
|
|
2661
|
+
const phase = providerOpts == null ? void 0 : providerOpts.phase;
|
|
2659
2662
|
if (hasConversation && id != null) {
|
|
2660
2663
|
break;
|
|
2661
2664
|
}
|
|
@@ -2666,12 +2669,13 @@ async function convertToOpenAIResponsesInput({
|
|
|
2666
2669
|
input.push({
|
|
2667
2670
|
role: "assistant",
|
|
2668
2671
|
content: [{ type: "output_text", text: part.text }],
|
|
2669
|
-
id
|
|
2672
|
+
id,
|
|
2673
|
+
...phase != null && { phase }
|
|
2670
2674
|
});
|
|
2671
2675
|
break;
|
|
2672
2676
|
}
|
|
2673
2677
|
case "tool-call": {
|
|
2674
|
-
const id = (
|
|
2678
|
+
const id = (_f = (_c = (_b = part.providerOptions) == null ? void 0 : _b[providerOptionsName]) == null ? void 0 : _c.itemId) != null ? _f : (_e = (_d = part.providerMetadata) == null ? void 0 : _d[providerOptionsName]) == null ? void 0 : _e.itemId;
|
|
2675
2679
|
if (hasConversation && id != null) {
|
|
2676
2680
|
break;
|
|
2677
2681
|
}
|
|
@@ -2782,7 +2786,7 @@ async function convertToOpenAIResponsesInput({
|
|
|
2782
2786
|
break;
|
|
2783
2787
|
}
|
|
2784
2788
|
if (store) {
|
|
2785
|
-
const itemId = (
|
|
2789
|
+
const itemId = (_i = (_h = (_g = part.providerOptions) == null ? void 0 : _g[providerOptionsName]) == null ? void 0 : _h.itemId) != null ? _i : part.toolCallId;
|
|
2786
2790
|
input.push({ type: "item_reference", id: itemId });
|
|
2787
2791
|
} else {
|
|
2788
2792
|
warnings.push({
|
|
@@ -2876,7 +2880,7 @@ async function convertToOpenAIResponsesInput({
|
|
|
2876
2880
|
}
|
|
2877
2881
|
const output = part.output;
|
|
2878
2882
|
if (output.type === "execution-denied") {
|
|
2879
|
-
const approvalId = (
|
|
2883
|
+
const approvalId = (_k = (_j = output.providerOptions) == null ? void 0 : _j.openai) == null ? void 0 : _k.approvalId;
|
|
2880
2884
|
if (approvalId) {
|
|
2881
2885
|
continue;
|
|
2882
2886
|
}
|
|
@@ -2935,7 +2939,7 @@ async function convertToOpenAIResponsesInput({
|
|
|
2935
2939
|
contentValue = output.value;
|
|
2936
2940
|
break;
|
|
2937
2941
|
case "execution-denied":
|
|
2938
|
-
contentValue = (
|
|
2942
|
+
contentValue = (_l = output.reason) != null ? _l : "Tool execution denied.";
|
|
2939
2943
|
break;
|
|
2940
2944
|
case "json":
|
|
2941
2945
|
case "error-json":
|
|
@@ -3068,7 +3072,8 @@ var openaiResponsesChunkSchema = (0, import_provider_utils24.lazySchema)(
|
|
|
3068
3072
|
item: import_v419.z.discriminatedUnion("type", [
|
|
3069
3073
|
import_v419.z.object({
|
|
3070
3074
|
type: import_v419.z.literal("message"),
|
|
3071
|
-
id: import_v419.z.string()
|
|
3075
|
+
id: import_v419.z.string(),
|
|
3076
|
+
phase: import_v419.z.enum(["commentary", "final_answer"]).nullish()
|
|
3072
3077
|
}),
|
|
3073
3078
|
import_v419.z.object({
|
|
3074
3079
|
type: import_v419.z.literal("reasoning"),
|
|
@@ -3185,7 +3190,8 @@ var openaiResponsesChunkSchema = (0, import_provider_utils24.lazySchema)(
|
|
|
3185
3190
|
item: import_v419.z.discriminatedUnion("type", [
|
|
3186
3191
|
import_v419.z.object({
|
|
3187
3192
|
type: import_v419.z.literal("message"),
|
|
3188
|
-
id: import_v419.z.string()
|
|
3193
|
+
id: import_v419.z.string(),
|
|
3194
|
+
phase: import_v419.z.enum(["commentary", "final_answer"]).nullish()
|
|
3189
3195
|
}),
|
|
3190
3196
|
import_v419.z.object({
|
|
3191
3197
|
type: import_v419.z.literal("reasoning"),
|
|
@@ -3497,6 +3503,7 @@ var openaiResponsesResponseSchema = (0, import_provider_utils24.lazySchema)(
|
|
|
3497
3503
|
type: import_v419.z.literal("message"),
|
|
3498
3504
|
role: import_v419.z.literal("assistant"),
|
|
3499
3505
|
id: import_v419.z.string(),
|
|
3506
|
+
phase: import_v419.z.enum(["commentary", "final_answer"]).nullish(),
|
|
3500
3507
|
content: import_v419.z.array(
|
|
3501
3508
|
import_v419.z.object({
|
|
3502
3509
|
type: import_v419.z.literal("output_text"),
|
|
@@ -4629,6 +4636,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4629
4636
|
}
|
|
4630
4637
|
const providerMetadata2 = {
|
|
4631
4638
|
itemId: part.id,
|
|
4639
|
+
...part.phase != null && { phase: part.phase },
|
|
4632
4640
|
...contentPart.annotations.length > 0 && {
|
|
4633
4641
|
annotations: contentPart.annotations
|
|
4634
4642
|
}
|
|
@@ -4943,6 +4951,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4943
4951
|
let responseId = null;
|
|
4944
4952
|
const ongoingToolCalls = {};
|
|
4945
4953
|
const ongoingAnnotations = [];
|
|
4954
|
+
let activeMessagePhase;
|
|
4946
4955
|
let hasFunctionCall = false;
|
|
4947
4956
|
const activeReasoning = {};
|
|
4948
4957
|
let serviceTier;
|
|
@@ -4953,7 +4962,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4953
4962
|
controller.enqueue({ type: "stream-start", warnings });
|
|
4954
4963
|
},
|
|
4955
4964
|
transform(chunk, controller) {
|
|
4956
|
-
var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m, _n, _o, _p, _q, _r, _s, _t, _u, _v, _w, _x, _y, _z, _A, _B, _C, _D;
|
|
4965
|
+
var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m, _n, _o, _p, _q, _r, _s, _t, _u, _v, _w, _x, _y, _z, _A, _B, _C, _D, _E, _F;
|
|
4957
4966
|
if (options.includeRawChunks) {
|
|
4958
4967
|
controller.enqueue({ type: "raw", rawValue: chunk.rawValue });
|
|
4959
4968
|
}
|
|
@@ -5094,12 +5103,16 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5094
5103
|
} else if (value.item.type === "shell_call_output") {
|
|
5095
5104
|
} else if (value.item.type === "message") {
|
|
5096
5105
|
ongoingAnnotations.splice(0, ongoingAnnotations.length);
|
|
5106
|
+
activeMessagePhase = (_a = value.item.phase) != null ? _a : void 0;
|
|
5097
5107
|
controller.enqueue({
|
|
5098
5108
|
type: "text-start",
|
|
5099
5109
|
id: value.item.id,
|
|
5100
5110
|
providerMetadata: {
|
|
5101
5111
|
[providerOptionsName]: {
|
|
5102
|
-
itemId: value.item.id
|
|
5112
|
+
itemId: value.item.id,
|
|
5113
|
+
...value.item.phase != null && {
|
|
5114
|
+
phase: value.item.phase
|
|
5115
|
+
}
|
|
5103
5116
|
}
|
|
5104
5117
|
}
|
|
5105
5118
|
});
|
|
@@ -5114,19 +5127,22 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5114
5127
|
providerMetadata: {
|
|
5115
5128
|
[providerOptionsName]: {
|
|
5116
5129
|
itemId: value.item.id,
|
|
5117
|
-
reasoningEncryptedContent: (
|
|
5130
|
+
reasoningEncryptedContent: (_b = value.item.encrypted_content) != null ? _b : null
|
|
5118
5131
|
}
|
|
5119
5132
|
}
|
|
5120
5133
|
});
|
|
5121
5134
|
}
|
|
5122
5135
|
} else if (isResponseOutputItemDoneChunk(value)) {
|
|
5123
5136
|
if (value.item.type === "message") {
|
|
5137
|
+
const phase = (_c = value.item.phase) != null ? _c : activeMessagePhase;
|
|
5138
|
+
activeMessagePhase = void 0;
|
|
5124
5139
|
controller.enqueue({
|
|
5125
5140
|
type: "text-end",
|
|
5126
5141
|
id: value.item.id,
|
|
5127
5142
|
providerMetadata: {
|
|
5128
5143
|
[providerOptionsName]: {
|
|
5129
5144
|
itemId: value.item.id,
|
|
5145
|
+
...phase != null && { phase },
|
|
5130
5146
|
...ongoingAnnotations.length > 0 && {
|
|
5131
5147
|
annotations: ongoingAnnotations
|
|
5132
5148
|
}
|
|
@@ -5191,13 +5207,13 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5191
5207
|
toolName: toolNameMapping.toCustomToolName("file_search"),
|
|
5192
5208
|
result: {
|
|
5193
5209
|
queries: value.item.queries,
|
|
5194
|
-
results: (
|
|
5210
|
+
results: (_e = (_d = value.item.results) == null ? void 0 : _d.map((result) => ({
|
|
5195
5211
|
attributes: result.attributes,
|
|
5196
5212
|
fileId: result.file_id,
|
|
5197
5213
|
filename: result.filename,
|
|
5198
5214
|
score: result.score,
|
|
5199
5215
|
text: result.text
|
|
5200
|
-
}))) != null ?
|
|
5216
|
+
}))) != null ? _e : null
|
|
5201
5217
|
}
|
|
5202
5218
|
});
|
|
5203
5219
|
} else if (value.item.type === "code_interpreter_call") {
|
|
@@ -5221,10 +5237,10 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5221
5237
|
});
|
|
5222
5238
|
} else if (value.item.type === "mcp_call") {
|
|
5223
5239
|
ongoingToolCalls[value.output_index] = void 0;
|
|
5224
|
-
const approvalRequestId = (
|
|
5225
|
-
const aliasedToolCallId = approvalRequestId != null ? (
|
|
5240
|
+
const approvalRequestId = (_f = value.item.approval_request_id) != null ? _f : void 0;
|
|
5241
|
+
const aliasedToolCallId = approvalRequestId != null ? (_h = (_g = approvalRequestIdToDummyToolCallIdFromStream.get(
|
|
5226
5242
|
approvalRequestId
|
|
5227
|
-
)) != null ?
|
|
5243
|
+
)) != null ? _g : approvalRequestIdToDummyToolCallIdFromPrompt[approvalRequestId]) != null ? _h : value.item.id : value.item.id;
|
|
5228
5244
|
const toolName = `mcp.${value.item.name}`;
|
|
5229
5245
|
controller.enqueue({
|
|
5230
5246
|
type: "tool-call",
|
|
@@ -5294,8 +5310,8 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5294
5310
|
ongoingToolCalls[value.output_index] = void 0;
|
|
5295
5311
|
} else if (value.item.type === "mcp_approval_request") {
|
|
5296
5312
|
ongoingToolCalls[value.output_index] = void 0;
|
|
5297
|
-
const dummyToolCallId = (
|
|
5298
|
-
const approvalRequestId = (
|
|
5313
|
+
const dummyToolCallId = (_k = (_j = (_i = self.config).generateId) == null ? void 0 : _j.call(_i)) != null ? _k : (0, import_provider_utils27.generateId)();
|
|
5314
|
+
const approvalRequestId = (_l = value.item.approval_request_id) != null ? _l : value.item.id;
|
|
5299
5315
|
approvalRequestIdToDummyToolCallIdFromStream.set(
|
|
5300
5316
|
approvalRequestId,
|
|
5301
5317
|
dummyToolCallId
|
|
@@ -5384,7 +5400,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5384
5400
|
providerMetadata: {
|
|
5385
5401
|
[providerOptionsName]: {
|
|
5386
5402
|
itemId: value.item.id,
|
|
5387
|
-
reasoningEncryptedContent: (
|
|
5403
|
+
reasoningEncryptedContent: (_m = value.item.encrypted_content) != null ? _m : null
|
|
5388
5404
|
}
|
|
5389
5405
|
}
|
|
5390
5406
|
});
|
|
@@ -5488,7 +5504,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5488
5504
|
id: value.item_id,
|
|
5489
5505
|
delta: value.delta
|
|
5490
5506
|
});
|
|
5491
|
-
if (((
|
|
5507
|
+
if (((_o = (_n = options.providerOptions) == null ? void 0 : _n[providerOptionsName]) == null ? void 0 : _o.logprobs) && value.logprobs) {
|
|
5492
5508
|
logprobs.push(value.logprobs);
|
|
5493
5509
|
}
|
|
5494
5510
|
} else if (value.type === "response.reasoning_summary_part.added") {
|
|
@@ -5517,7 +5533,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5517
5533
|
providerMetadata: {
|
|
5518
5534
|
[providerOptionsName]: {
|
|
5519
5535
|
itemId: value.item_id,
|
|
5520
|
-
reasoningEncryptedContent: (
|
|
5536
|
+
reasoningEncryptedContent: (_q = (_p = activeReasoning[value.item_id]) == null ? void 0 : _p.encryptedContent) != null ? _q : null
|
|
5521
5537
|
}
|
|
5522
5538
|
}
|
|
5523
5539
|
});
|
|
@@ -5551,10 +5567,10 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5551
5567
|
} else if (isResponseFinishedChunk(value)) {
|
|
5552
5568
|
finishReason = {
|
|
5553
5569
|
unified: mapOpenAIResponseFinishReason({
|
|
5554
|
-
finishReason: (
|
|
5570
|
+
finishReason: (_r = value.response.incomplete_details) == null ? void 0 : _r.reason,
|
|
5555
5571
|
hasFunctionCall
|
|
5556
5572
|
}),
|
|
5557
|
-
raw: (
|
|
5573
|
+
raw: (_t = (_s = value.response.incomplete_details) == null ? void 0 : _s.reason) != null ? _t : void 0
|
|
5558
5574
|
};
|
|
5559
5575
|
usage = value.response.usage;
|
|
5560
5576
|
if (typeof value.response.service_tier === "string") {
|
|
@@ -5566,7 +5582,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5566
5582
|
controller.enqueue({
|
|
5567
5583
|
type: "source",
|
|
5568
5584
|
sourceType: "url",
|
|
5569
|
-
id: (
|
|
5585
|
+
id: (_w = (_v = (_u = self.config).generateId) == null ? void 0 : _v.call(_u)) != null ? _w : (0, import_provider_utils27.generateId)(),
|
|
5570
5586
|
url: value.annotation.url,
|
|
5571
5587
|
title: value.annotation.title
|
|
5572
5588
|
});
|
|
@@ -5574,7 +5590,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5574
5590
|
controller.enqueue({
|
|
5575
5591
|
type: "source",
|
|
5576
5592
|
sourceType: "document",
|
|
5577
|
-
id: (
|
|
5593
|
+
id: (_z = (_y = (_x = self.config).generateId) == null ? void 0 : _y.call(_x)) != null ? _z : (0, import_provider_utils27.generateId)(),
|
|
5578
5594
|
mediaType: "text/plain",
|
|
5579
5595
|
title: value.annotation.filename,
|
|
5580
5596
|
filename: value.annotation.filename,
|
|
@@ -5590,7 +5606,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5590
5606
|
controller.enqueue({
|
|
5591
5607
|
type: "source",
|
|
5592
5608
|
sourceType: "document",
|
|
5593
|
-
id: (
|
|
5609
|
+
id: (_C = (_B = (_A = self.config).generateId) == null ? void 0 : _B.call(_A)) != null ? _C : (0, import_provider_utils27.generateId)(),
|
|
5594
5610
|
mediaType: "text/plain",
|
|
5595
5611
|
title: value.annotation.filename,
|
|
5596
5612
|
filename: value.annotation.filename,
|
|
@@ -5606,7 +5622,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5606
5622
|
controller.enqueue({
|
|
5607
5623
|
type: "source",
|
|
5608
5624
|
sourceType: "document",
|
|
5609
|
-
id: (
|
|
5625
|
+
id: (_F = (_E = (_D = self.config).generateId) == null ? void 0 : _E.call(_D)) != null ? _F : (0, import_provider_utils27.generateId)(),
|
|
5610
5626
|
mediaType: "application/octet-stream",
|
|
5611
5627
|
title: value.annotation.file_id,
|
|
5612
5628
|
filename: value.annotation.file_id,
|
|
@@ -6065,7 +6081,7 @@ var OpenAITranscriptionModel = class {
|
|
|
6065
6081
|
};
|
|
6066
6082
|
|
|
6067
6083
|
// src/version.ts
|
|
6068
|
-
var VERSION = true ? "3.0.
|
|
6084
|
+
var VERSION = true ? "3.0.34" : "0.0.0-test";
|
|
6069
6085
|
|
|
6070
6086
|
// src/openai-provider.ts
|
|
6071
6087
|
function createOpenAI(options = {}) {
|