@ai-sdk/openai 3.0.55 → 3.0.58

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,24 @@
1
1
  # @ai-sdk/openai
2
2
 
3
+ ## 3.0.58
4
+
5
+ ### Patch Changes
6
+
7
+ - 2370948: feat(openai): preserve `namespace` on function_call output items
8
+
9
+ ## 3.0.57
10
+
11
+ ### Patch Changes
12
+
13
+ - d33e7cc: chore(provider/openai): add type for image model options for type-safe processing
14
+
15
+ ## 3.0.56
16
+
17
+ ### Patch Changes
18
+
19
+ - Updated dependencies [7beadf0]
20
+ - @ai-sdk/provider-utils@4.0.26
21
+
3
22
  ## 3.0.55
4
23
 
5
24
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -43,6 +43,33 @@ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
43
43
  type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
44
44
 
45
45
  type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | 'gpt-image-2' | 'chatgpt-image-latest' | (string & {});
46
+ declare const openaiImageModelOptions: _ai_sdk_provider_utils.LazySchema<{
47
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
48
+ background?: "auto" | "transparent" | "opaque" | undefined;
49
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
50
+ outputCompression?: number | undefined;
51
+ user?: string | undefined;
52
+ }>;
53
+ type OpenAIImageModelOptions = InferSchema<typeof openaiImageModelOptions>;
54
+ declare const openaiImageModelGenerationOptions: _ai_sdk_provider_utils.LazySchema<{
55
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
56
+ background?: "auto" | "transparent" | "opaque" | undefined;
57
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
58
+ outputCompression?: number | undefined;
59
+ user?: string | undefined;
60
+ style?: "vivid" | "natural" | undefined;
61
+ moderation?: "auto" | "low" | undefined;
62
+ }>;
63
+ type OpenAIImageModelGenerationOptions = InferSchema<typeof openaiImageModelGenerationOptions>;
64
+ declare const openaiImageModelEditOptions: _ai_sdk_provider_utils.LazySchema<{
65
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
66
+ background?: "auto" | "transparent" | "opaque" | undefined;
67
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
68
+ outputCompression?: number | undefined;
69
+ user?: string | undefined;
70
+ inputFidelity?: "low" | "high" | undefined;
71
+ }>;
72
+ type OpenAIImageModelEditOptions = InferSchema<typeof openaiImageModelEditOptions>;
46
73
 
47
74
  declare const webSearchToolFactory: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSchema<{}, {
48
75
  /**
@@ -253,6 +280,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
253
280
  call_id: string;
254
281
  name: string;
255
282
  arguments: string;
283
+ namespace?: string | null | undefined;
256
284
  } | {
257
285
  type: "web_search_call";
258
286
  id: string;
@@ -370,6 +398,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
370
398
  name: string;
371
399
  arguments: string;
372
400
  status: "completed";
401
+ namespace?: string | null | undefined;
373
402
  } | {
374
403
  type: "custom_tool_call";
375
404
  id: string;
@@ -1190,4 +1219,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
1190
1219
 
1191
1220
  declare const VERSION: string;
1192
1221
 
1193
- export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
1222
+ export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAIImageModelEditOptions, type OpenAIImageModelGenerationOptions, type OpenAIImageModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
package/dist/index.d.ts CHANGED
@@ -43,6 +43,33 @@ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
43
43
  type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
44
44
 
45
45
  type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | 'gpt-image-2' | 'chatgpt-image-latest' | (string & {});
46
+ declare const openaiImageModelOptions: _ai_sdk_provider_utils.LazySchema<{
47
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
48
+ background?: "auto" | "transparent" | "opaque" | undefined;
49
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
50
+ outputCompression?: number | undefined;
51
+ user?: string | undefined;
52
+ }>;
53
+ type OpenAIImageModelOptions = InferSchema<typeof openaiImageModelOptions>;
54
+ declare const openaiImageModelGenerationOptions: _ai_sdk_provider_utils.LazySchema<{
55
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
56
+ background?: "auto" | "transparent" | "opaque" | undefined;
57
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
58
+ outputCompression?: number | undefined;
59
+ user?: string | undefined;
60
+ style?: "vivid" | "natural" | undefined;
61
+ moderation?: "auto" | "low" | undefined;
62
+ }>;
63
+ type OpenAIImageModelGenerationOptions = InferSchema<typeof openaiImageModelGenerationOptions>;
64
+ declare const openaiImageModelEditOptions: _ai_sdk_provider_utils.LazySchema<{
65
+ quality?: "auto" | "low" | "medium" | "high" | "standard" | "hd" | undefined;
66
+ background?: "auto" | "transparent" | "opaque" | undefined;
67
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
68
+ outputCompression?: number | undefined;
69
+ user?: string | undefined;
70
+ inputFidelity?: "low" | "high" | undefined;
71
+ }>;
72
+ type OpenAIImageModelEditOptions = InferSchema<typeof openaiImageModelEditOptions>;
46
73
 
47
74
  declare const webSearchToolFactory: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSchema<{}, {
48
75
  /**
@@ -253,6 +280,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
253
280
  call_id: string;
254
281
  name: string;
255
282
  arguments: string;
283
+ namespace?: string | null | undefined;
256
284
  } | {
257
285
  type: "web_search_call";
258
286
  id: string;
@@ -370,6 +398,7 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
370
398
  name: string;
371
399
  arguments: string;
372
400
  status: "completed";
401
+ namespace?: string | null | undefined;
373
402
  } | {
374
403
  type: "custom_tool_call";
375
404
  id: string;
@@ -1190,4 +1219,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
1190
1219
 
1191
1220
  declare const VERSION: string;
1192
1221
 
1193
- export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
1222
+ export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAIImageModelEditOptions, type OpenAIImageModelGenerationOptions, type OpenAIImageModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };