@ai-sdk/openai 2.0.62 → 2.0.64
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/dist/index.js +18 -11
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +18 -11
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +17 -10
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +17 -10
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/internal/index.js
CHANGED
|
@@ -258,7 +258,7 @@ function getResponseMetadata({
|
|
|
258
258
|
return {
|
|
259
259
|
id: id != null ? id : void 0,
|
|
260
260
|
modelId: model != null ? model : void 0,
|
|
261
|
-
timestamp: created
|
|
261
|
+
timestamp: created ? new Date(created * 1e3) : void 0
|
|
262
262
|
};
|
|
263
263
|
}
|
|
264
264
|
|
|
@@ -883,7 +883,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
883
883
|
outputTokens: void 0,
|
|
884
884
|
totalTokens: void 0
|
|
885
885
|
};
|
|
886
|
-
let
|
|
886
|
+
let metadataExtracted = false;
|
|
887
887
|
let isActiveText = false;
|
|
888
888
|
const providerMetadata = { openai: {} };
|
|
889
889
|
return {
|
|
@@ -908,12 +908,15 @@ var OpenAIChatLanguageModel = class {
|
|
|
908
908
|
controller.enqueue({ type: "error", error: value.error });
|
|
909
909
|
return;
|
|
910
910
|
}
|
|
911
|
-
if (
|
|
912
|
-
|
|
913
|
-
|
|
914
|
-
|
|
915
|
-
|
|
916
|
-
|
|
911
|
+
if (!metadataExtracted) {
|
|
912
|
+
const metadata = getResponseMetadata(value);
|
|
913
|
+
if (Object.values(metadata).some(Boolean)) {
|
|
914
|
+
metadataExtracted = true;
|
|
915
|
+
controller.enqueue({
|
|
916
|
+
type: "response-metadata",
|
|
917
|
+
...getResponseMetadata(value)
|
|
918
|
+
});
|
|
919
|
+
}
|
|
917
920
|
}
|
|
918
921
|
if (value.usage != null) {
|
|
919
922
|
usage.inputTokens = (_a = value.usage.prompt_tokens) != null ? _a : void 0;
|
|
@@ -2887,12 +2890,16 @@ var TOP_LOGPROBS_MAX = 20;
|
|
|
2887
2890
|
var openaiResponsesReasoningModelIds = [
|
|
2888
2891
|
"o1",
|
|
2889
2892
|
"o1-2024-12-17",
|
|
2890
|
-
"o3-mini",
|
|
2891
|
-
"o3-mini-2025-01-31",
|
|
2892
2893
|
"o3",
|
|
2893
2894
|
"o3-2025-04-16",
|
|
2895
|
+
"o3-deep-research",
|
|
2896
|
+
"o3-deep-research-2025-06-26",
|
|
2897
|
+
"o3-mini",
|
|
2898
|
+
"o3-mini-2025-01-31",
|
|
2894
2899
|
"o4-mini",
|
|
2895
2900
|
"o4-mini-2025-04-16",
|
|
2901
|
+
"o4-mini-deep-research",
|
|
2902
|
+
"o4-mini-deep-research-2025-06-26",
|
|
2896
2903
|
"codex-mini-latest",
|
|
2897
2904
|
"computer-use-preview",
|
|
2898
2905
|
"gpt-5",
|