@ai-sdk/openai 3.0.0-beta.52 → 3.0.0-beta.54
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/dist/index.js +18 -11
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +18 -11
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +17 -10
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +17 -10
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +3 -3
package/dist/internal/index.js
CHANGED
|
@@ -271,7 +271,7 @@ function getResponseMetadata({
|
|
|
271
271
|
return {
|
|
272
272
|
id: id != null ? id : void 0,
|
|
273
273
|
modelId: model != null ? model : void 0,
|
|
274
|
-
timestamp: created
|
|
274
|
+
timestamp: created ? new Date(created * 1e3) : void 0
|
|
275
275
|
};
|
|
276
276
|
}
|
|
277
277
|
|
|
@@ -896,7 +896,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
896
896
|
outputTokens: void 0,
|
|
897
897
|
totalTokens: void 0
|
|
898
898
|
};
|
|
899
|
-
let
|
|
899
|
+
let metadataExtracted = false;
|
|
900
900
|
let isActiveText = false;
|
|
901
901
|
const providerMetadata = { openai: {} };
|
|
902
902
|
return {
|
|
@@ -921,12 +921,15 @@ var OpenAIChatLanguageModel = class {
|
|
|
921
921
|
controller.enqueue({ type: "error", error: value.error });
|
|
922
922
|
return;
|
|
923
923
|
}
|
|
924
|
-
if (
|
|
925
|
-
|
|
926
|
-
|
|
927
|
-
|
|
928
|
-
|
|
929
|
-
|
|
924
|
+
if (!metadataExtracted) {
|
|
925
|
+
const metadata = getResponseMetadata(value);
|
|
926
|
+
if (Object.values(metadata).some(Boolean)) {
|
|
927
|
+
metadataExtracted = true;
|
|
928
|
+
controller.enqueue({
|
|
929
|
+
type: "response-metadata",
|
|
930
|
+
...getResponseMetadata(value)
|
|
931
|
+
});
|
|
932
|
+
}
|
|
930
933
|
}
|
|
931
934
|
if (value.usage != null) {
|
|
932
935
|
usage.inputTokens = (_a = value.usage.prompt_tokens) != null ? _a : void 0;
|
|
@@ -2918,12 +2921,16 @@ var TOP_LOGPROBS_MAX = 20;
|
|
|
2918
2921
|
var openaiResponsesReasoningModelIds = [
|
|
2919
2922
|
"o1",
|
|
2920
2923
|
"o1-2024-12-17",
|
|
2921
|
-
"o3-mini",
|
|
2922
|
-
"o3-mini-2025-01-31",
|
|
2923
2924
|
"o3",
|
|
2924
2925
|
"o3-2025-04-16",
|
|
2926
|
+
"o3-deep-research",
|
|
2927
|
+
"o3-deep-research-2025-06-26",
|
|
2928
|
+
"o3-mini",
|
|
2929
|
+
"o3-mini-2025-01-31",
|
|
2925
2930
|
"o4-mini",
|
|
2926
2931
|
"o4-mini-2025-04-16",
|
|
2932
|
+
"o4-mini-deep-research",
|
|
2933
|
+
"o4-mini-deep-research-2025-06-26",
|
|
2927
2934
|
"codex-mini-latest",
|
|
2928
2935
|
"computer-use-preview",
|
|
2929
2936
|
"gpt-5",
|