@ai-sdk/openai 3.0.0-beta.99 → 3.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,4 +1,4 @@
1
- import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
1
+ import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
2
2
  import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
3
3
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
4
4
 
@@ -19,6 +19,8 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
19
19
  promptCacheKey?: string | undefined;
20
20
  promptCacheRetention?: "in_memory" | "24h" | undefined;
21
21
  safetyIdentifier?: string | undefined;
22
+ systemMessageMode?: "remove" | "system" | "developer" | undefined;
23
+ forceReasoning?: boolean | undefined;
22
24
  }>;
23
25
  type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
24
26
 
@@ -41,8 +43,8 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
41
43
  constructor(modelId: OpenAIChatModelId, config: OpenAIChatConfig);
42
44
  get provider(): string;
43
45
  private getArgs;
44
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
45
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
46
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
47
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
46
48
  }
47
49
 
48
50
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
@@ -73,8 +75,8 @@ declare class OpenAICompletionLanguageModel implements LanguageModelV3 {
73
75
  get provider(): string;
74
76
  readonly supportedUrls: Record<string, RegExp[]>;
75
77
  private getArgs;
76
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
77
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
78
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
79
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
78
80
  }
79
81
 
80
82
  type OpenAIConfig = {
@@ -115,7 +117,7 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
115
117
  doEmbed({ values, headers, abortSignal, providerOptions, }: Parameters<EmbeddingModelV3['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV3['doEmbed']>>>;
116
118
  }
117
119
 
118
- type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | (string & {});
120
+ type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
119
121
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
120
122
  declare const hasDefaultResponseFormat: Set<string>;
121
123
 
@@ -131,7 +133,7 @@ declare class OpenAIImageModel implements ImageModelV3 {
131
133
  get maxImagesPerCall(): number;
132
134
  get provider(): string;
133
135
  constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
134
- doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
136
+ doGenerate({ prompt, files, mask, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
135
137
  }
136
138
 
137
139
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
@@ -196,8 +198,8 @@ declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
196
198
  readonly supportedUrls: Record<string, RegExp[]>;
197
199
  get provider(): string;
198
200
  private getArgs;
199
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
200
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
201
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
202
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
201
203
  }
202
204
 
203
205
  /**
@@ -525,7 +527,7 @@ declare const fileSearch: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSc
525
527
  }>;
526
528
 
527
529
  declare const imageGenerationArgsSchema: _ai_sdk_provider_utils.LazySchema<{
528
- background?: "auto" | "opaque" | "transparent" | undefined;
530
+ background?: "auto" | "transparent" | "opaque" | undefined;
529
531
  inputFidelity?: "low" | "high" | undefined;
530
532
  inputImageMask?: {
531
533
  fileId?: string | undefined;
@@ -1,4 +1,4 @@
1
- import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
1
+ import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
2
2
  import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
3
3
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
4
4
 
@@ -19,6 +19,8 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
19
19
  promptCacheKey?: string | undefined;
20
20
  promptCacheRetention?: "in_memory" | "24h" | undefined;
21
21
  safetyIdentifier?: string | undefined;
22
+ systemMessageMode?: "remove" | "system" | "developer" | undefined;
23
+ forceReasoning?: boolean | undefined;
22
24
  }>;
23
25
  type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
24
26
 
@@ -41,8 +43,8 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
41
43
  constructor(modelId: OpenAIChatModelId, config: OpenAIChatConfig);
42
44
  get provider(): string;
43
45
  private getArgs;
44
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
45
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
46
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
47
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
46
48
  }
47
49
 
48
50
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
@@ -73,8 +75,8 @@ declare class OpenAICompletionLanguageModel implements LanguageModelV3 {
73
75
  get provider(): string;
74
76
  readonly supportedUrls: Record<string, RegExp[]>;
75
77
  private getArgs;
76
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
77
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
78
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
79
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
78
80
  }
79
81
 
80
82
  type OpenAIConfig = {
@@ -115,7 +117,7 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
115
117
  doEmbed({ values, headers, abortSignal, providerOptions, }: Parameters<EmbeddingModelV3['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV3['doEmbed']>>>;
116
118
  }
117
119
 
118
- type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | (string & {});
120
+ type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
119
121
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
120
122
  declare const hasDefaultResponseFormat: Set<string>;
121
123
 
@@ -131,7 +133,7 @@ declare class OpenAIImageModel implements ImageModelV3 {
131
133
  get maxImagesPerCall(): number;
132
134
  get provider(): string;
133
135
  constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
134
- doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
136
+ doGenerate({ prompt, files, mask, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
135
137
  }
136
138
 
137
139
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
@@ -196,8 +198,8 @@ declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
196
198
  readonly supportedUrls: Record<string, RegExp[]>;
197
199
  get provider(): string;
198
200
  private getArgs;
199
- doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
200
- doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
201
+ doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
202
+ doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
201
203
  }
202
204
 
203
205
  /**
@@ -525,7 +527,7 @@ declare const fileSearch: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSc
525
527
  }>;
526
528
 
527
529
  declare const imageGenerationArgsSchema: _ai_sdk_provider_utils.LazySchema<{
528
- background?: "auto" | "opaque" | "transparent" | undefined;
530
+ background?: "auto" | "transparent" | "opaque" | undefined;
529
531
  inputFidelity?: "low" | "high" | undefined;
530
532
  inputImageMask?: {
531
533
  fileId?: string | undefined;