@ai-sdk/openai 3.0.36 → 3.0.38
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/README.md +1 -1
- package/dist/index.d.mts +53 -3
- package/dist/index.d.ts +53 -3
- package/dist/index.js +1168 -966
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +1118 -912
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +20 -2
- package/dist/internal/index.d.ts +20 -2
- package/dist/internal/index.js +349 -158
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +327 -132
- package/dist/internal/index.mjs.map +1 -1
- package/docs/03-openai.mdx +79 -1
- package/package.json +2 -2
- package/src/chat/openai-chat-options.ts +0 -5
- package/src/openai-language-model-capabilities.ts +0 -2
- package/src/openai-tools.ts +13 -1
- package/src/responses/convert-to-openai-responses-input.ts +74 -0
- package/src/responses/openai-responses-api.ts +58 -0
- package/src/responses/openai-responses-language-model.ts +94 -9
- package/src/responses/openai-responses-options.ts +0 -22
- package/src/responses/openai-responses-prepare-tools.ts +41 -11
- package/src/tool/custom.ts +64 -0
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,19 @@
|
|
|
1
1
|
# @ai-sdk/openai
|
|
2
2
|
|
|
3
|
+
## 3.0.38
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- 64a8fae: chore: remove obsolete model IDs for Anthropic, Google, OpenAI, xAI
|
|
8
|
+
|
|
9
|
+
## 3.0.37
|
|
10
|
+
|
|
11
|
+
### Patch Changes
|
|
12
|
+
|
|
13
|
+
- 58bc42d: feat(provider/openai): support custom tools with alias mapping
|
|
14
|
+
- Updated dependencies [58bc42d]
|
|
15
|
+
- @ai-sdk/provider-utils@4.0.16
|
|
16
|
+
|
|
3
17
|
## 3.0.36
|
|
4
18
|
|
|
5
19
|
### Patch Changes
|
package/README.md
CHANGED
|
@@ -34,7 +34,7 @@ import { openai } from '@ai-sdk/openai';
|
|
|
34
34
|
import { generateText } from 'ai';
|
|
35
35
|
|
|
36
36
|
const { text } = await generateText({
|
|
37
|
-
model: openai('gpt-
|
|
37
|
+
model: openai('gpt-5-mini'),
|
|
38
38
|
prompt: 'Write a vegetarian lasagna recipe for 4 people.',
|
|
39
39
|
});
|
|
40
40
|
```
|
package/dist/index.d.mts
CHANGED
|
@@ -3,7 +3,7 @@ import { ProviderV3, LanguageModelV3, EmbeddingModelV3, ImageModelV3, Transcript
|
|
|
3
3
|
import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
4
4
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
5
5
|
|
|
6
|
-
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-audio-preview' | 'gpt-4o-audio-preview-2024-12-17' | 'gpt-4o-audio-preview-2025-06-03' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini-audio-preview' | 'gpt-4o-mini-audio-preview-2024-12-17' | 'gpt-4o-search-preview' | 'gpt-4o-search-preview-2025-03-11' | 'gpt-4o-mini-search-preview' | 'gpt-4o-mini-search-preview-2025-03-11' | 'gpt-
|
|
6
|
+
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-audio-preview' | 'gpt-4o-audio-preview-2024-12-17' | 'gpt-4o-audio-preview-2025-06-03' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini-audio-preview' | 'gpt-4o-mini-audio-preview-2024-12-17' | 'gpt-4o-search-preview' | 'gpt-4o-search-preview-2025-03-11' | 'gpt-4o-mini-search-preview' | 'gpt-4o-mini-search-preview-2025-03-11' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-16k' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-2025-11-13' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-2025-12-11' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5.2-pro-2025-12-11' | (string & {});
|
|
7
7
|
declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
8
8
|
logitBias?: Record<number, number> | undefined;
|
|
9
9
|
logprobs?: number | boolean | undefined;
|
|
@@ -286,6 +286,12 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
286
286
|
path: string;
|
|
287
287
|
diff: string;
|
|
288
288
|
};
|
|
289
|
+
} | {
|
|
290
|
+
type: "custom_tool_call";
|
|
291
|
+
id: string;
|
|
292
|
+
call_id: string;
|
|
293
|
+
name: string;
|
|
294
|
+
input: string;
|
|
289
295
|
} | {
|
|
290
296
|
type: "shell_call";
|
|
291
297
|
id: string;
|
|
@@ -328,6 +334,13 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
328
334
|
name: string;
|
|
329
335
|
arguments: string;
|
|
330
336
|
status: "completed";
|
|
337
|
+
} | {
|
|
338
|
+
type: "custom_tool_call";
|
|
339
|
+
id: string;
|
|
340
|
+
call_id: string;
|
|
341
|
+
name: string;
|
|
342
|
+
input: string;
|
|
343
|
+
status: "completed";
|
|
331
344
|
} | {
|
|
332
345
|
type: "code_interpreter_call";
|
|
333
346
|
id: string;
|
|
@@ -477,6 +490,11 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
477
490
|
item_id: string;
|
|
478
491
|
output_index: number;
|
|
479
492
|
delta: string;
|
|
493
|
+
} | {
|
|
494
|
+
type: "response.custom_tool_call_input.delta";
|
|
495
|
+
item_id: string;
|
|
496
|
+
output_index: number;
|
|
497
|
+
delta: string;
|
|
480
498
|
} | {
|
|
481
499
|
type: "response.image_generation_call.partial_image";
|
|
482
500
|
item_id: string;
|
|
@@ -556,6 +574,28 @@ type OpenAIResponsesLogprobs = NonNullable<(OpenAIResponsesChunk & {
|
|
|
556
574
|
type: 'response.output_text.delta';
|
|
557
575
|
})['logprobs']> | null;
|
|
558
576
|
|
|
577
|
+
declare const customToolFactory: _ai_sdk_provider_utils.ProviderToolFactory<string, {
|
|
578
|
+
/**
|
|
579
|
+
* The name of the custom tool, used to identify it in the API.
|
|
580
|
+
*/
|
|
581
|
+
name: string;
|
|
582
|
+
/**
|
|
583
|
+
* An optional description of what the tool does.
|
|
584
|
+
*/
|
|
585
|
+
description?: string;
|
|
586
|
+
/**
|
|
587
|
+
* The output format specification for the tool.
|
|
588
|
+
* Omit for unconstrained text output.
|
|
589
|
+
*/
|
|
590
|
+
format?: {
|
|
591
|
+
type: "grammar";
|
|
592
|
+
syntax: "regex" | "lark";
|
|
593
|
+
definition: string;
|
|
594
|
+
} | {
|
|
595
|
+
type: "text";
|
|
596
|
+
};
|
|
597
|
+
}>;
|
|
598
|
+
|
|
559
599
|
/**
|
|
560
600
|
* Type definitions for the apply_patch operations.
|
|
561
601
|
*/
|
|
@@ -602,6 +642,16 @@ declare const openaiTools: {
|
|
|
602
642
|
status: "completed" | "failed";
|
|
603
643
|
output?: string;
|
|
604
644
|
}, {}>;
|
|
645
|
+
/**
|
|
646
|
+
* Custom tools let callers constrain model output to a grammar (regex or
|
|
647
|
+
* Lark syntax). The model returns a `custom_tool_call` output item whose
|
|
648
|
+
* `input` field is a string matching the specified grammar.
|
|
649
|
+
*
|
|
650
|
+
* @param name - The name of the custom tool.
|
|
651
|
+
* @param description - An optional description of the tool.
|
|
652
|
+
* @param format - The output format constraint (grammar type, syntax, and definition).
|
|
653
|
+
*/
|
|
654
|
+
customTool: (args: Parameters<typeof customToolFactory>[0]) => _ai_sdk_provider_utils.Tool<string, unknown>;
|
|
605
655
|
/**
|
|
606
656
|
* The Code Interpreter tool allows models to write and run Python code in a
|
|
607
657
|
* sandboxed environment to solve complex problems in domains like data analysis,
|
|
@@ -690,7 +740,7 @@ declare const openaiTools: {
|
|
|
690
740
|
* Local shell is a tool that allows agents to run shell commands locally
|
|
691
741
|
* on a machine you or the user provides.
|
|
692
742
|
*
|
|
693
|
-
* Supported models: `gpt-5-codex`
|
|
743
|
+
* Supported models: `gpt-5-codex`
|
|
694
744
|
*/
|
|
695
745
|
localShell: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSchema<{
|
|
696
746
|
action: {
|
|
@@ -870,7 +920,7 @@ declare const openaiTools: {
|
|
|
870
920
|
}>;
|
|
871
921
|
};
|
|
872
922
|
|
|
873
|
-
type OpenAIResponsesModelId = '
|
|
923
|
+
type OpenAIResponsesModelId = 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-2025-11-13' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-2025-12-11' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5.2-pro-2025-12-11' | 'gpt-5.2-codex' | 'gpt-5.3-codex' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | 'o4-mini' | 'o4-mini-2025-04-16' | (string & {});
|
|
874
924
|
declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
875
925
|
conversation?: string | null | undefined;
|
|
876
926
|
include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
|
package/dist/index.d.ts
CHANGED
|
@@ -3,7 +3,7 @@ import { ProviderV3, LanguageModelV3, EmbeddingModelV3, ImageModelV3, Transcript
|
|
|
3
3
|
import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
4
4
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
5
5
|
|
|
6
|
-
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-audio-preview' | 'gpt-4o-audio-preview-2024-12-17' | 'gpt-4o-audio-preview-2025-06-03' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini-audio-preview' | 'gpt-4o-mini-audio-preview-2024-12-17' | 'gpt-4o-search-preview' | 'gpt-4o-search-preview-2025-03-11' | 'gpt-4o-mini-search-preview' | 'gpt-4o-mini-search-preview-2025-03-11' | 'gpt-
|
|
6
|
+
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-audio-preview' | 'gpt-4o-audio-preview-2024-12-17' | 'gpt-4o-audio-preview-2025-06-03' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini-audio-preview' | 'gpt-4o-mini-audio-preview-2024-12-17' | 'gpt-4o-search-preview' | 'gpt-4o-search-preview-2025-03-11' | 'gpt-4o-mini-search-preview' | 'gpt-4o-mini-search-preview-2025-03-11' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo-16k' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-2025-11-13' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-2025-12-11' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5.2-pro-2025-12-11' | (string & {});
|
|
7
7
|
declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
8
8
|
logitBias?: Record<number, number> | undefined;
|
|
9
9
|
logprobs?: number | boolean | undefined;
|
|
@@ -286,6 +286,12 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
286
286
|
path: string;
|
|
287
287
|
diff: string;
|
|
288
288
|
};
|
|
289
|
+
} | {
|
|
290
|
+
type: "custom_tool_call";
|
|
291
|
+
id: string;
|
|
292
|
+
call_id: string;
|
|
293
|
+
name: string;
|
|
294
|
+
input: string;
|
|
289
295
|
} | {
|
|
290
296
|
type: "shell_call";
|
|
291
297
|
id: string;
|
|
@@ -328,6 +334,13 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
328
334
|
name: string;
|
|
329
335
|
arguments: string;
|
|
330
336
|
status: "completed";
|
|
337
|
+
} | {
|
|
338
|
+
type: "custom_tool_call";
|
|
339
|
+
id: string;
|
|
340
|
+
call_id: string;
|
|
341
|
+
name: string;
|
|
342
|
+
input: string;
|
|
343
|
+
status: "completed";
|
|
331
344
|
} | {
|
|
332
345
|
type: "code_interpreter_call";
|
|
333
346
|
id: string;
|
|
@@ -477,6 +490,11 @@ declare const openaiResponsesChunkSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
|
477
490
|
item_id: string;
|
|
478
491
|
output_index: number;
|
|
479
492
|
delta: string;
|
|
493
|
+
} | {
|
|
494
|
+
type: "response.custom_tool_call_input.delta";
|
|
495
|
+
item_id: string;
|
|
496
|
+
output_index: number;
|
|
497
|
+
delta: string;
|
|
480
498
|
} | {
|
|
481
499
|
type: "response.image_generation_call.partial_image";
|
|
482
500
|
item_id: string;
|
|
@@ -556,6 +574,28 @@ type OpenAIResponsesLogprobs = NonNullable<(OpenAIResponsesChunk & {
|
|
|
556
574
|
type: 'response.output_text.delta';
|
|
557
575
|
})['logprobs']> | null;
|
|
558
576
|
|
|
577
|
+
declare const customToolFactory: _ai_sdk_provider_utils.ProviderToolFactory<string, {
|
|
578
|
+
/**
|
|
579
|
+
* The name of the custom tool, used to identify it in the API.
|
|
580
|
+
*/
|
|
581
|
+
name: string;
|
|
582
|
+
/**
|
|
583
|
+
* An optional description of what the tool does.
|
|
584
|
+
*/
|
|
585
|
+
description?: string;
|
|
586
|
+
/**
|
|
587
|
+
* The output format specification for the tool.
|
|
588
|
+
* Omit for unconstrained text output.
|
|
589
|
+
*/
|
|
590
|
+
format?: {
|
|
591
|
+
type: "grammar";
|
|
592
|
+
syntax: "regex" | "lark";
|
|
593
|
+
definition: string;
|
|
594
|
+
} | {
|
|
595
|
+
type: "text";
|
|
596
|
+
};
|
|
597
|
+
}>;
|
|
598
|
+
|
|
559
599
|
/**
|
|
560
600
|
* Type definitions for the apply_patch operations.
|
|
561
601
|
*/
|
|
@@ -602,6 +642,16 @@ declare const openaiTools: {
|
|
|
602
642
|
status: "completed" | "failed";
|
|
603
643
|
output?: string;
|
|
604
644
|
}, {}>;
|
|
645
|
+
/**
|
|
646
|
+
* Custom tools let callers constrain model output to a grammar (regex or
|
|
647
|
+
* Lark syntax). The model returns a `custom_tool_call` output item whose
|
|
648
|
+
* `input` field is a string matching the specified grammar.
|
|
649
|
+
*
|
|
650
|
+
* @param name - The name of the custom tool.
|
|
651
|
+
* @param description - An optional description of the tool.
|
|
652
|
+
* @param format - The output format constraint (grammar type, syntax, and definition).
|
|
653
|
+
*/
|
|
654
|
+
customTool: (args: Parameters<typeof customToolFactory>[0]) => _ai_sdk_provider_utils.Tool<string, unknown>;
|
|
605
655
|
/**
|
|
606
656
|
* The Code Interpreter tool allows models to write and run Python code in a
|
|
607
657
|
* sandboxed environment to solve complex problems in domains like data analysis,
|
|
@@ -690,7 +740,7 @@ declare const openaiTools: {
|
|
|
690
740
|
* Local shell is a tool that allows agents to run shell commands locally
|
|
691
741
|
* on a machine you or the user provides.
|
|
692
742
|
*
|
|
693
|
-
* Supported models: `gpt-5-codex`
|
|
743
|
+
* Supported models: `gpt-5-codex`
|
|
694
744
|
*/
|
|
695
745
|
localShell: _ai_sdk_provider_utils.ProviderToolFactoryWithOutputSchema<{
|
|
696
746
|
action: {
|
|
@@ -870,7 +920,7 @@ declare const openaiTools: {
|
|
|
870
920
|
}>;
|
|
871
921
|
};
|
|
872
922
|
|
|
873
|
-
type OpenAIResponsesModelId = '
|
|
923
|
+
type OpenAIResponsesModelId = 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-2025-11-13' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-2025-12-11' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5.2-pro-2025-12-11' | 'gpt-5.2-codex' | 'gpt-5.3-codex' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | 'o4-mini' | 'o4-mini-2025-04-16' | (string & {});
|
|
874
924
|
declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
875
925
|
conversation?: string | null | undefined;
|
|
876
926
|
include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
|