@ai-sdk/openai 3.0.18 → 3.0.20
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +19 -0
- package/dist/index.d.mts +20 -1
- package/dist/index.d.ts +20 -1
- package/dist/index.js +15 -17
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +15 -17
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +20 -1
- package/dist/internal/index.d.ts +20 -1
- package/dist/internal/index.js +14 -16
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +14 -16
- package/dist/internal/index.mjs.map +1 -1
- package/docs/03-openai.mdx +67 -11
- package/package.json +2 -2
- package/src/index.ts +2 -0
- package/src/responses/openai-responses-language-model.ts +24 -26
- package/src/responses/openai-responses-provider-metadata.ts +23 -1
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,24 @@
|
|
|
1
1
|
# @ai-sdk/openai
|
|
2
2
|
|
|
3
|
+
## 3.0.20
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- Updated dependencies [462ad00]
|
|
8
|
+
- @ai-sdk/provider-utils@4.0.10
|
|
9
|
+
|
|
10
|
+
## 3.0.19
|
|
11
|
+
|
|
12
|
+
### Patch Changes
|
|
13
|
+
|
|
14
|
+
- 04c89b1: Provide Responses API providerMetadata types at the message / reasoning level.
|
|
15
|
+
|
|
16
|
+
- Export the following types for use in client code:
|
|
17
|
+
- `OpenaiResponsesProviderMetadata`
|
|
18
|
+
- `OpenaiResponsesReasoningProviderMetadata`
|
|
19
|
+
- `AzureResponsesProviderMetadata`
|
|
20
|
+
- `AzureResponsesReasoningProviderMetadata`
|
|
21
|
+
|
|
3
22
|
## 3.0.18
|
|
4
23
|
|
|
5
24
|
### Patch Changes
|
package/dist/index.d.mts
CHANGED
|
@@ -506,6 +506,10 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
506
506
|
param?: string | null | undefined;
|
|
507
507
|
};
|
|
508
508
|
}>;
|
|
509
|
+
type OpenAIResponsesChunk = InferSchema<typeof openaiResponsesChunkSchema>;
|
|
510
|
+
type OpenAIResponsesLogprobs = NonNullable<(OpenAIResponsesChunk & {
|
|
511
|
+
type: 'response.output_text.delta';
|
|
512
|
+
})['logprobs']> | null;
|
|
509
513
|
|
|
510
514
|
/**
|
|
511
515
|
* Type definitions for the apply_patch operations.
|
|
@@ -909,6 +913,21 @@ type OpenaiResponsesChunk = InferSchema<typeof openaiResponsesChunkSchema>;
|
|
|
909
913
|
type ResponsesOutputTextAnnotationProviderMetadata = Extract<OpenaiResponsesChunk, {
|
|
910
914
|
type: 'response.output_text.annotation.added';
|
|
911
915
|
}>['annotation'];
|
|
916
|
+
type ResponsesProviderMetadata = {
|
|
917
|
+
responseId: string | null | undefined;
|
|
918
|
+
logprobs?: Array<OpenAIResponsesLogprobs>;
|
|
919
|
+
serviceTier?: string;
|
|
920
|
+
};
|
|
921
|
+
type ResponsesReasoningProviderMetadata = {
|
|
922
|
+
itemId: string;
|
|
923
|
+
reasoningEncryptedContent?: string | null;
|
|
924
|
+
};
|
|
925
|
+
type OpenaiResponsesReasoningProviderMetadata = {
|
|
926
|
+
openai: ResponsesReasoningProviderMetadata;
|
|
927
|
+
};
|
|
928
|
+
type OpenaiResponsesProviderMetadata = {
|
|
929
|
+
openai: ResponsesProviderMetadata;
|
|
930
|
+
};
|
|
912
931
|
type ResponsesTextProviderMetadata = {
|
|
913
932
|
itemId: string;
|
|
914
933
|
annotations?: Array<ResponsesOutputTextAnnotationProviderMetadata>;
|
|
@@ -935,4 +954,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
|
|
|
935
954
|
|
|
936
955
|
declare const VERSION: string;
|
|
937
956
|
|
|
938
|
-
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
|
957
|
+
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
package/dist/index.d.ts
CHANGED
|
@@ -506,6 +506,10 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
506
506
|
param?: string | null | undefined;
|
|
507
507
|
};
|
|
508
508
|
}>;
|
|
509
|
+
type OpenAIResponsesChunk = InferSchema<typeof openaiResponsesChunkSchema>;
|
|
510
|
+
type OpenAIResponsesLogprobs = NonNullable<(OpenAIResponsesChunk & {
|
|
511
|
+
type: 'response.output_text.delta';
|
|
512
|
+
})['logprobs']> | null;
|
|
509
513
|
|
|
510
514
|
/**
|
|
511
515
|
* Type definitions for the apply_patch operations.
|
|
@@ -909,6 +913,21 @@ type OpenaiResponsesChunk = InferSchema<typeof openaiResponsesChunkSchema>;
|
|
|
909
913
|
type ResponsesOutputTextAnnotationProviderMetadata = Extract<OpenaiResponsesChunk, {
|
|
910
914
|
type: 'response.output_text.annotation.added';
|
|
911
915
|
}>['annotation'];
|
|
916
|
+
type ResponsesProviderMetadata = {
|
|
917
|
+
responseId: string | null | undefined;
|
|
918
|
+
logprobs?: Array<OpenAIResponsesLogprobs>;
|
|
919
|
+
serviceTier?: string;
|
|
920
|
+
};
|
|
921
|
+
type ResponsesReasoningProviderMetadata = {
|
|
922
|
+
itemId: string;
|
|
923
|
+
reasoningEncryptedContent?: string | null;
|
|
924
|
+
};
|
|
925
|
+
type OpenaiResponsesReasoningProviderMetadata = {
|
|
926
|
+
openai: ResponsesReasoningProviderMetadata;
|
|
927
|
+
};
|
|
928
|
+
type OpenaiResponsesProviderMetadata = {
|
|
929
|
+
openai: ResponsesProviderMetadata;
|
|
930
|
+
};
|
|
912
931
|
type ResponsesTextProviderMetadata = {
|
|
913
932
|
itemId: string;
|
|
914
933
|
annotations?: Array<ResponsesOutputTextAnnotationProviderMetadata>;
|
|
@@ -935,4 +954,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
|
|
|
935
954
|
|
|
936
955
|
declare const VERSION: string;
|
|
937
956
|
|
|
938
|
-
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
|
957
|
+
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
package/dist/index.js
CHANGED
|
@@ -4622,14 +4622,12 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4622
4622
|
}
|
|
4623
4623
|
}
|
|
4624
4624
|
const providerMetadata = {
|
|
4625
|
-
[providerOptionsName]: {
|
|
4625
|
+
[providerOptionsName]: {
|
|
4626
|
+
responseId: response.id,
|
|
4627
|
+
...logprobs.length > 0 ? { logprobs } : {},
|
|
4628
|
+
...typeof response.service_tier === "string" ? { serviceTier: response.service_tier } : {}
|
|
4629
|
+
}
|
|
4626
4630
|
};
|
|
4627
|
-
if (logprobs.length > 0) {
|
|
4628
|
-
providerMetadata[providerOptionsName].logprobs = logprobs;
|
|
4629
|
-
}
|
|
4630
|
-
if (typeof response.service_tier === "string") {
|
|
4631
|
-
providerMetadata[providerOptionsName].serviceTier = response.service_tier;
|
|
4632
|
-
}
|
|
4633
4631
|
const usage = response.usage;
|
|
4634
4632
|
return {
|
|
4635
4633
|
content,
|
|
@@ -5229,7 +5227,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5229
5227
|
type: "reasoning-end",
|
|
5230
5228
|
id: `${value.item_id}:${summaryIndex}`,
|
|
5231
5229
|
providerMetadata: {
|
|
5232
|
-
[providerOptionsName]: {
|
|
5230
|
+
[providerOptionsName]: {
|
|
5231
|
+
itemId: value.item_id
|
|
5232
|
+
}
|
|
5233
5233
|
}
|
|
5234
5234
|
});
|
|
5235
5235
|
activeReasoningPart.summaryParts[summaryIndex] = "concluded";
|
|
@@ -5263,7 +5263,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5263
5263
|
type: "reasoning-end",
|
|
5264
5264
|
id: `${value.item_id}:${value.summary_index}`,
|
|
5265
5265
|
providerMetadata: {
|
|
5266
|
-
[providerOptionsName]: {
|
|
5266
|
+
[providerOptionsName]: {
|
|
5267
|
+
itemId: value.item_id
|
|
5268
|
+
}
|
|
5267
5269
|
}
|
|
5268
5270
|
});
|
|
5269
5271
|
activeReasoning[value.item_id].summaryParts[value.summary_index] = "concluded";
|
|
@@ -5348,15 +5350,11 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
5348
5350
|
flush(controller) {
|
|
5349
5351
|
const providerMetadata = {
|
|
5350
5352
|
[providerOptionsName]: {
|
|
5351
|
-
responseId
|
|
5353
|
+
responseId,
|
|
5354
|
+
...logprobs.length > 0 ? { logprobs } : {},
|
|
5355
|
+
...serviceTier !== void 0 ? { serviceTier } : {}
|
|
5352
5356
|
}
|
|
5353
5357
|
};
|
|
5354
|
-
if (logprobs.length > 0) {
|
|
5355
|
-
providerMetadata[providerOptionsName].logprobs = logprobs;
|
|
5356
|
-
}
|
|
5357
|
-
if (serviceTier !== void 0) {
|
|
5358
|
-
providerMetadata[providerOptionsName].serviceTier = serviceTier;
|
|
5359
|
-
}
|
|
5360
5358
|
controller.enqueue({
|
|
5361
5359
|
type: "finish",
|
|
5362
5360
|
finishReason,
|
|
@@ -5788,7 +5786,7 @@ var OpenAITranscriptionModel = class {
|
|
|
5788
5786
|
};
|
|
5789
5787
|
|
|
5790
5788
|
// src/version.ts
|
|
5791
|
-
var VERSION = true ? "3.0.
|
|
5789
|
+
var VERSION = true ? "3.0.20" : "0.0.0-test";
|
|
5792
5790
|
|
|
5793
5791
|
// src/openai-provider.ts
|
|
5794
5792
|
function createOpenAI(options = {}) {
|