@ai-sdk/openai 2.0.62 → 2.0.64
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/dist/index.js +18 -11
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +18 -11
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +17 -10
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +17 -10
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/internal/index.mjs
CHANGED
|
@@ -220,7 +220,7 @@ function getResponseMetadata({
|
|
|
220
220
|
return {
|
|
221
221
|
id: id != null ? id : void 0,
|
|
222
222
|
modelId: model != null ? model : void 0,
|
|
223
|
-
timestamp: created
|
|
223
|
+
timestamp: created ? new Date(created * 1e3) : void 0
|
|
224
224
|
};
|
|
225
225
|
}
|
|
226
226
|
|
|
@@ -853,7 +853,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
853
853
|
outputTokens: void 0,
|
|
854
854
|
totalTokens: void 0
|
|
855
855
|
};
|
|
856
|
-
let
|
|
856
|
+
let metadataExtracted = false;
|
|
857
857
|
let isActiveText = false;
|
|
858
858
|
const providerMetadata = { openai: {} };
|
|
859
859
|
return {
|
|
@@ -878,12 +878,15 @@ var OpenAIChatLanguageModel = class {
|
|
|
878
878
|
controller.enqueue({ type: "error", error: value.error });
|
|
879
879
|
return;
|
|
880
880
|
}
|
|
881
|
-
if (
|
|
882
|
-
|
|
883
|
-
|
|
884
|
-
|
|
885
|
-
|
|
886
|
-
|
|
881
|
+
if (!metadataExtracted) {
|
|
882
|
+
const metadata = getResponseMetadata(value);
|
|
883
|
+
if (Object.values(metadata).some(Boolean)) {
|
|
884
|
+
metadataExtracted = true;
|
|
885
|
+
controller.enqueue({
|
|
886
|
+
type: "response-metadata",
|
|
887
|
+
...getResponseMetadata(value)
|
|
888
|
+
});
|
|
889
|
+
}
|
|
887
890
|
}
|
|
888
891
|
if (value.usage != null) {
|
|
889
892
|
usage.inputTokens = (_a = value.usage.prompt_tokens) != null ? _a : void 0;
|
|
@@ -2929,12 +2932,16 @@ var TOP_LOGPROBS_MAX = 20;
|
|
|
2929
2932
|
var openaiResponsesReasoningModelIds = [
|
|
2930
2933
|
"o1",
|
|
2931
2934
|
"o1-2024-12-17",
|
|
2932
|
-
"o3-mini",
|
|
2933
|
-
"o3-mini-2025-01-31",
|
|
2934
2935
|
"o3",
|
|
2935
2936
|
"o3-2025-04-16",
|
|
2937
|
+
"o3-deep-research",
|
|
2938
|
+
"o3-deep-research-2025-06-26",
|
|
2939
|
+
"o3-mini",
|
|
2940
|
+
"o3-mini-2025-01-31",
|
|
2936
2941
|
"o4-mini",
|
|
2937
2942
|
"o4-mini-2025-04-16",
|
|
2943
|
+
"o4-mini-deep-research",
|
|
2944
|
+
"o4-mini-deep-research-2025-06-26",
|
|
2938
2945
|
"codex-mini-latest",
|
|
2939
2946
|
"computer-use-preview",
|
|
2940
2947
|
"gpt-5",
|