@ai-sdk/openai 3.0.0-beta.98 → 3.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +299 -0
- package/dist/index.d.mts +18 -17
- package/dist/index.d.ts +18 -17
- package/dist/index.js +342 -200
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +350 -204
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +12 -10
- package/dist/internal/index.d.ts +12 -10
- package/dist/internal/index.js +339 -199
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +347 -203
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +4 -4
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
|
|
1
|
+
import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
|
|
2
2
|
import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
3
3
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
4
4
|
|
|
@@ -19,6 +19,8 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
|
|
|
19
19
|
promptCacheKey?: string | undefined;
|
|
20
20
|
promptCacheRetention?: "in_memory" | "24h" | undefined;
|
|
21
21
|
safetyIdentifier?: string | undefined;
|
|
22
|
+
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
23
|
+
forceReasoning?: boolean | undefined;
|
|
22
24
|
}>;
|
|
23
25
|
type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
|
|
24
26
|
|
|
@@ -41,8 +43,8 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
|
|
|
41
43
|
constructor(modelId: OpenAIChatModelId, config: OpenAIChatConfig);
|
|
42
44
|
get provider(): string;
|
|
43
45
|
private getArgs;
|
|
44
|
-
doGenerate(options:
|
|
45
|
-
doStream(options:
|
|
46
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
47
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
46
48
|
}
|
|
47
49
|
|
|
48
50
|
type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
|
|
@@ -73,8 +75,8 @@ declare class OpenAICompletionLanguageModel implements LanguageModelV3 {
|
|
|
73
75
|
get provider(): string;
|
|
74
76
|
readonly supportedUrls: Record<string, RegExp[]>;
|
|
75
77
|
private getArgs;
|
|
76
|
-
doGenerate(options:
|
|
77
|
-
doStream(options:
|
|
78
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
79
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
78
80
|
}
|
|
79
81
|
|
|
80
82
|
type OpenAIConfig = {
|
|
@@ -115,7 +117,7 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
|
|
|
115
117
|
doEmbed({ values, headers, abortSignal, providerOptions, }: Parameters<EmbeddingModelV3['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV3['doEmbed']>>>;
|
|
116
118
|
}
|
|
117
119
|
|
|
118
|
-
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | (string & {});
|
|
120
|
+
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
|
|
119
121
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
120
122
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
121
123
|
|
|
@@ -131,7 +133,7 @@ declare class OpenAIImageModel implements ImageModelV3 {
|
|
|
131
133
|
get maxImagesPerCall(): number;
|
|
132
134
|
get provider(): string;
|
|
133
135
|
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
134
|
-
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
136
|
+
doGenerate({ prompt, files, mask, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
135
137
|
}
|
|
136
138
|
|
|
137
139
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
@@ -196,8 +198,8 @@ declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
|
|
|
196
198
|
readonly supportedUrls: Record<string, RegExp[]>;
|
|
197
199
|
get provider(): string;
|
|
198
200
|
private getArgs;
|
|
199
|
-
doGenerate(options:
|
|
200
|
-
doStream(options:
|
|
201
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
202
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
201
203
|
}
|
|
202
204
|
|
|
203
205
|
/**
|
|
@@ -525,7 +527,7 @@ declare const fileSearch: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSc
|
|
|
525
527
|
}>;
|
|
526
528
|
|
|
527
529
|
declare const imageGenerationArgsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
528
|
-
background?: "auto" | "
|
|
530
|
+
background?: "auto" | "transparent" | "opaque" | undefined;
|
|
529
531
|
inputFidelity?: "low" | "high" | undefined;
|
|
530
532
|
inputImageMask?: {
|
|
531
533
|
fileId?: string | undefined;
|
package/dist/internal/index.d.ts
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
|
|
1
|
+
import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
|
|
2
2
|
import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
3
3
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
4
4
|
|
|
@@ -19,6 +19,8 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
|
|
|
19
19
|
promptCacheKey?: string | undefined;
|
|
20
20
|
promptCacheRetention?: "in_memory" | "24h" | undefined;
|
|
21
21
|
safetyIdentifier?: string | undefined;
|
|
22
|
+
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
23
|
+
forceReasoning?: boolean | undefined;
|
|
22
24
|
}>;
|
|
23
25
|
type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
|
|
24
26
|
|
|
@@ -41,8 +43,8 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
|
|
|
41
43
|
constructor(modelId: OpenAIChatModelId, config: OpenAIChatConfig);
|
|
42
44
|
get provider(): string;
|
|
43
45
|
private getArgs;
|
|
44
|
-
doGenerate(options:
|
|
45
|
-
doStream(options:
|
|
46
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
47
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
46
48
|
}
|
|
47
49
|
|
|
48
50
|
type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
|
|
@@ -73,8 +75,8 @@ declare class OpenAICompletionLanguageModel implements LanguageModelV3 {
|
|
|
73
75
|
get provider(): string;
|
|
74
76
|
readonly supportedUrls: Record<string, RegExp[]>;
|
|
75
77
|
private getArgs;
|
|
76
|
-
doGenerate(options:
|
|
77
|
-
doStream(options:
|
|
78
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
79
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
78
80
|
}
|
|
79
81
|
|
|
80
82
|
type OpenAIConfig = {
|
|
@@ -115,7 +117,7 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
|
|
|
115
117
|
doEmbed({ values, headers, abortSignal, providerOptions, }: Parameters<EmbeddingModelV3['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV3['doEmbed']>>>;
|
|
116
118
|
}
|
|
117
119
|
|
|
118
|
-
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | (string & {});
|
|
120
|
+
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
|
|
119
121
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
120
122
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
121
123
|
|
|
@@ -131,7 +133,7 @@ declare class OpenAIImageModel implements ImageModelV3 {
|
|
|
131
133
|
get maxImagesPerCall(): number;
|
|
132
134
|
get provider(): string;
|
|
133
135
|
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
134
|
-
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
136
|
+
doGenerate({ prompt, files, mask, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
135
137
|
}
|
|
136
138
|
|
|
137
139
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
@@ -196,8 +198,8 @@ declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
|
|
|
196
198
|
readonly supportedUrls: Record<string, RegExp[]>;
|
|
197
199
|
get provider(): string;
|
|
198
200
|
private getArgs;
|
|
199
|
-
doGenerate(options:
|
|
200
|
-
doStream(options:
|
|
201
|
+
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
|
|
202
|
+
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
|
|
201
203
|
}
|
|
202
204
|
|
|
203
205
|
/**
|
|
@@ -525,7 +527,7 @@ declare const fileSearch: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSc
|
|
|
525
527
|
}>;
|
|
526
528
|
|
|
527
529
|
declare const imageGenerationArgsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
528
|
-
background?: "auto" | "
|
|
530
|
+
background?: "auto" | "transparent" | "opaque" | undefined;
|
|
529
531
|
inputFidelity?: "low" | "high" | undefined;
|
|
530
532
|
inputImageMask?: {
|
|
531
533
|
fileId?: string | undefined;
|