openai 0.14.0 → 0.16.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +41 -0
- data/README.md +3 -3
- data/lib/openai/helpers/structured_output/json_schema_converter.rb +20 -21
- data/lib/openai/helpers/structured_output/union_of.rb +11 -1
- data/lib/openai/models/audio/speech_create_params.rb +0 -9
- data/lib/openai/models/chat/chat_completion.rb +2 -2
- data/lib/openai/models/chat/chat_completion_audio_param.rb +0 -9
- data/lib/openai/models/chat/chat_completion_chunk.rb +2 -2
- data/lib/openai/models/chat/chat_completion_store_message.rb +32 -1
- data/lib/openai/models/chat/completion_create_params.rb +33 -7
- data/lib/openai/models/function_definition.rb +1 -1
- data/lib/openai/models/image_edit_params.rb +4 -1
- data/lib/openai/models/image_generate_params.rb +4 -1
- data/lib/openai/models/images_response.rb +2 -5
- data/lib/openai/models/responses/response.rb +52 -6
- data/lib/openai/models/responses/response_code_interpreter_tool_call.rb +5 -3
- data/lib/openai/models/responses/response_create_params.rb +33 -7
- data/lib/openai/models/responses/response_mcp_call_arguments_delta_event.rb +9 -4
- data/lib/openai/models/responses/response_mcp_call_arguments_done_event.rb +7 -4
- data/lib/openai/models/responses/response_mcp_call_completed_event.rb +17 -1
- data/lib/openai/models/responses/response_mcp_call_failed_event.rb +17 -1
- data/lib/openai/models/responses/response_mcp_list_tools_completed_event.rb +17 -1
- data/lib/openai/models/responses/response_mcp_list_tools_failed_event.rb +17 -1
- data/lib/openai/models/responses/response_mcp_list_tools_in_progress_event.rb +17 -1
- data/lib/openai/models/responses/response_stream_event.rb +1 -7
- data/lib/openai/models/responses/response_text_delta_event.rb +66 -1
- data/lib/openai/models/responses/response_text_done_event.rb +66 -1
- data/lib/openai/resources/chat/completions.rb +12 -4
- data/lib/openai/resources/images.rb +6 -6
- data/lib/openai/resources/responses.rb +42 -17
- data/lib/openai/version.rb +1 -1
- data/lib/openai.rb +0 -2
- data/rbi/openai/models/audio/speech_create_params.rbi +0 -9
- data/rbi/openai/models/chat/chat_completion.rbi +3 -3
- data/rbi/openai/models/chat/chat_completion_audio_param.rbi +0 -15
- data/rbi/openai/models/chat/chat_completion_chunk.rbi +3 -3
- data/rbi/openai/models/chat/chat_completion_store_message.rbi +68 -3
- data/rbi/openai/models/chat/completion_create_params.rbi +47 -9
- data/rbi/openai/models/function_definition.rbi +2 -2
- data/rbi/openai/models/image_edit_params.rbi +6 -0
- data/rbi/openai/models/image_generate_params.rbi +6 -0
- data/rbi/openai/models/images_response.rbi +2 -2
- data/rbi/openai/models/responses/response.rbi +47 -9
- data/rbi/openai/models/responses/response_code_interpreter_tool_call.rbi +6 -3
- data/rbi/openai/models/responses/response_create_params.rbi +47 -9
- data/rbi/openai/models/responses/response_mcp_call_arguments_delta_event.rbi +7 -5
- data/rbi/openai/models/responses/response_mcp_call_arguments_done_event.rbi +5 -5
- data/rbi/openai/models/responses/response_mcp_call_completed_event.rbi +28 -4
- data/rbi/openai/models/responses/response_mcp_call_failed_event.rbi +28 -4
- data/rbi/openai/models/responses/response_mcp_list_tools_completed_event.rbi +28 -4
- data/rbi/openai/models/responses/response_mcp_list_tools_failed_event.rbi +28 -4
- data/rbi/openai/models/responses/response_mcp_list_tools_in_progress_event.rbi +28 -4
- data/rbi/openai/models/responses/response_stream_event.rbi +0 -2
- data/rbi/openai/models/responses/response_text_delta_event.rbi +131 -0
- data/rbi/openai/models/responses/response_text_done_event.rbi +131 -0
- data/rbi/openai/resources/chat/completions.rbi +36 -8
- data/rbi/openai/resources/images.rbi +22 -10
- data/rbi/openai/resources/responses.rbi +36 -8
- data/sig/openai/models/audio/speech_create_params.rbs +0 -6
- data/sig/openai/models/chat/chat_completion_audio_param.rbs +0 -6
- data/sig/openai/models/chat/chat_completion_store_message.rbs +29 -3
- data/sig/openai/models/chat/completion_create_params.rbs +14 -0
- data/sig/openai/models/responses/response.rbs +14 -0
- data/sig/openai/models/responses/response_create_params.rbs +14 -0
- data/sig/openai/models/responses/response_mcp_call_arguments_delta_event.rbs +4 -4
- data/sig/openai/models/responses/response_mcp_call_arguments_done_event.rbs +4 -4
- data/sig/openai/models/responses/response_mcp_call_completed_event.rbs +14 -1
- data/sig/openai/models/responses/response_mcp_call_failed_event.rbs +14 -1
- data/sig/openai/models/responses/response_mcp_list_tools_completed_event.rbs +14 -1
- data/sig/openai/models/responses/response_mcp_list_tools_failed_event.rbs +14 -1
- data/sig/openai/models/responses/response_mcp_list_tools_in_progress_event.rbs +10 -0
- data/sig/openai/models/responses/response_stream_event.rbs +0 -2
- data/sig/openai/models/responses/response_text_delta_event.rbs +52 -0
- data/sig/openai/models/responses/response_text_done_event.rbs +52 -0
- data/sig/openai/resources/chat/completions.rbs +4 -0
- data/sig/openai/resources/responses.rbs +4 -0
- metadata +2 -8
- data/lib/openai/models/responses/response_reasoning_delta_event.rb +0 -60
- data/lib/openai/models/responses/response_reasoning_done_event.rb +0 -60
- data/rbi/openai/models/responses/response_reasoning_delta_event.rbi +0 -83
- data/rbi/openai/models/responses/response_reasoning_done_event.rbi +0 -83
- data/sig/openai/models/responses/response_reasoning_delta_event.rbs +0 -47
- data/sig/openai/models/responses/response_reasoning_done_event.rbs +0 -47
@@ -46,9 +46,6 @@ module OpenAI
|
|
46
46
|
| :ballad
|
47
47
|
| :coral
|
48
48
|
| :echo
|
49
|
-
| :fable
|
50
|
-
| :onyx
|
51
|
-
| :nova
|
52
49
|
| :sage
|
53
50
|
| :shimmer
|
54
51
|
| :verse
|
@@ -63,9 +60,6 @@ module OpenAI
|
|
63
60
|
BALLAD: :ballad
|
64
61
|
CORAL: :coral
|
65
62
|
ECHO: :echo
|
66
|
-
FABLE: :fable
|
67
|
-
ONYX: :onyx
|
68
|
-
NOVA: :nova
|
69
63
|
SAGE: :sage
|
70
64
|
SHIMMER: :shimmer
|
71
65
|
VERSE: :verse
|
@@ -3,16 +3,42 @@ module OpenAI
|
|
3
3
|
class ChatCompletionStoreMessage = Chat::ChatCompletionStoreMessage
|
4
4
|
|
5
5
|
module Chat
|
6
|
-
type chat_completion_store_message =
|
6
|
+
type chat_completion_store_message =
|
7
|
+
{
|
8
|
+
id: String,
|
9
|
+
content_parts: ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]?
|
10
|
+
}
|
7
11
|
|
8
12
|
class ChatCompletionStoreMessage < OpenAI::Models::Chat::ChatCompletionMessage
|
9
13
|
def id: -> String
|
10
14
|
|
11
15
|
def id=: (String _) -> String
|
12
16
|
|
13
|
-
def
|
17
|
+
def content_parts: -> ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]?
|
14
18
|
|
15
|
-
def
|
19
|
+
def content_parts=: (
|
20
|
+
::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]? _
|
21
|
+
) -> ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]?
|
22
|
+
|
23
|
+
def initialize: (
|
24
|
+
id: String,
|
25
|
+
?content_parts: ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]?
|
26
|
+
) -> void
|
27
|
+
|
28
|
+
def to_hash: -> {
|
29
|
+
id: String,
|
30
|
+
content_parts: ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]?
|
31
|
+
}
|
32
|
+
|
33
|
+
type content_part =
|
34
|
+
OpenAI::Chat::ChatCompletionContentPartText
|
35
|
+
| OpenAI::Chat::ChatCompletionContentPartImage
|
36
|
+
|
37
|
+
module ContentPart
|
38
|
+
extend OpenAI::Internal::Type::Union
|
39
|
+
|
40
|
+
def self?.variants: -> ::Array[OpenAI::Models::Chat::ChatCompletionStoreMessage::content_part]
|
41
|
+
end
|
16
42
|
end
|
17
43
|
end
|
18
44
|
end
|
@@ -19,8 +19,10 @@ module OpenAI
|
|
19
19
|
parallel_tool_calls: bool,
|
20
20
|
prediction: OpenAI::Chat::ChatCompletionPredictionContent?,
|
21
21
|
presence_penalty: Float?,
|
22
|
+
prompt_cache_key: String,
|
22
23
|
reasoning_effort: OpenAI::Models::reasoning_effort?,
|
23
24
|
response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format,
|
25
|
+
safety_identifier: String,
|
24
26
|
seed: Integer?,
|
25
27
|
service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?,
|
26
28
|
stop: OpenAI::Models::Chat::CompletionCreateParams::stop?,
|
@@ -82,6 +84,10 @@ module OpenAI
|
|
82
84
|
|
83
85
|
attr_accessor presence_penalty: Float?
|
84
86
|
|
87
|
+
attr_reader prompt_cache_key: String?
|
88
|
+
|
89
|
+
def prompt_cache_key=: (String) -> String
|
90
|
+
|
85
91
|
attr_accessor reasoning_effort: OpenAI::Models::reasoning_effort?
|
86
92
|
|
87
93
|
attr_reader response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format?
|
@@ -90,6 +96,10 @@ module OpenAI
|
|
90
96
|
OpenAI::Models::Chat::CompletionCreateParams::response_format
|
91
97
|
) -> OpenAI::Models::Chat::CompletionCreateParams::response_format
|
92
98
|
|
99
|
+
attr_reader safety_identifier: String?
|
100
|
+
|
101
|
+
def safety_identifier=: (String) -> String
|
102
|
+
|
93
103
|
attr_accessor seed: Integer?
|
94
104
|
|
95
105
|
attr_accessor service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?
|
@@ -145,8 +155,10 @@ module OpenAI
|
|
145
155
|
?parallel_tool_calls: bool,
|
146
156
|
?prediction: OpenAI::Chat::ChatCompletionPredictionContent?,
|
147
157
|
?presence_penalty: Float?,
|
158
|
+
?prompt_cache_key: String,
|
148
159
|
?reasoning_effort: OpenAI::Models::reasoning_effort?,
|
149
160
|
?response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format,
|
161
|
+
?safety_identifier: String,
|
150
162
|
?seed: Integer?,
|
151
163
|
?service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?,
|
152
164
|
?stop: OpenAI::Models::Chat::CompletionCreateParams::stop?,
|
@@ -179,8 +191,10 @@ module OpenAI
|
|
179
191
|
parallel_tool_calls: bool,
|
180
192
|
prediction: OpenAI::Chat::ChatCompletionPredictionContent?,
|
181
193
|
presence_penalty: Float?,
|
194
|
+
prompt_cache_key: String,
|
182
195
|
reasoning_effort: OpenAI::Models::reasoning_effort?,
|
183
196
|
response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format,
|
197
|
+
safety_identifier: String,
|
184
198
|
seed: Integer?,
|
185
199
|
service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?,
|
186
200
|
stop: OpenAI::Models::Chat::CompletionCreateParams::stop?,
|
@@ -22,7 +22,9 @@ module OpenAI
|
|
22
22
|
max_tool_calls: Integer?,
|
23
23
|
previous_response_id: String?,
|
24
24
|
prompt: OpenAI::Responses::ResponsePrompt?,
|
25
|
+
prompt_cache_key: String,
|
25
26
|
reasoning: OpenAI::Reasoning?,
|
27
|
+
safety_identifier: String,
|
26
28
|
service_tier: OpenAI::Models::Responses::Response::service_tier?,
|
27
29
|
status: OpenAI::Models::Responses::response_status,
|
28
30
|
text: OpenAI::Responses::ResponseTextConfig,
|
@@ -71,8 +73,16 @@ module OpenAI
|
|
71
73
|
|
72
74
|
attr_accessor prompt: OpenAI::Responses::ResponsePrompt?
|
73
75
|
|
76
|
+
attr_reader prompt_cache_key: String?
|
77
|
+
|
78
|
+
def prompt_cache_key=: (String) -> String
|
79
|
+
|
74
80
|
attr_accessor reasoning: OpenAI::Reasoning?
|
75
81
|
|
82
|
+
attr_reader safety_identifier: String?
|
83
|
+
|
84
|
+
def safety_identifier=: (String) -> String
|
85
|
+
|
76
86
|
attr_accessor service_tier: OpenAI::Models::Responses::Response::service_tier?
|
77
87
|
|
78
88
|
attr_reader status: OpenAI::Models::Responses::response_status?
|
@@ -120,7 +130,9 @@ module OpenAI
|
|
120
130
|
?max_tool_calls: Integer?,
|
121
131
|
?previous_response_id: String?,
|
122
132
|
?prompt: OpenAI::Responses::ResponsePrompt?,
|
133
|
+
?prompt_cache_key: String,
|
123
134
|
?reasoning: OpenAI::Reasoning?,
|
135
|
+
?safety_identifier: String,
|
124
136
|
?service_tier: OpenAI::Models::Responses::Response::service_tier?,
|
125
137
|
?status: OpenAI::Models::Responses::response_status,
|
126
138
|
?text: OpenAI::Responses::ResponseTextConfig,
|
@@ -151,7 +163,9 @@ module OpenAI
|
|
151
163
|
max_tool_calls: Integer?,
|
152
164
|
previous_response_id: String?,
|
153
165
|
prompt: OpenAI::Responses::ResponsePrompt?,
|
166
|
+
prompt_cache_key: String,
|
154
167
|
reasoning: OpenAI::Reasoning?,
|
168
|
+
safety_identifier: String,
|
155
169
|
service_tier: OpenAI::Models::Responses::Response::service_tier?,
|
156
170
|
status: OpenAI::Models::Responses::response_status,
|
157
171
|
text: OpenAI::Responses::ResponseTextConfig,
|
@@ -14,7 +14,9 @@ module OpenAI
|
|
14
14
|
parallel_tool_calls: bool?,
|
15
15
|
previous_response_id: String?,
|
16
16
|
prompt: OpenAI::Responses::ResponsePrompt?,
|
17
|
+
prompt_cache_key: String,
|
17
18
|
reasoning: OpenAI::Reasoning?,
|
19
|
+
safety_identifier: String,
|
18
20
|
service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?,
|
19
21
|
store: bool?,
|
20
22
|
temperature: Float?,
|
@@ -62,8 +64,16 @@ module OpenAI
|
|
62
64
|
|
63
65
|
attr_accessor prompt: OpenAI::Responses::ResponsePrompt?
|
64
66
|
|
67
|
+
attr_reader prompt_cache_key: String?
|
68
|
+
|
69
|
+
def prompt_cache_key=: (String) -> String
|
70
|
+
|
65
71
|
attr_accessor reasoning: OpenAI::Reasoning?
|
66
72
|
|
73
|
+
attr_reader safety_identifier: String?
|
74
|
+
|
75
|
+
def safety_identifier=: (String) -> String
|
76
|
+
|
67
77
|
attr_accessor service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?
|
68
78
|
|
69
79
|
attr_accessor store: bool?
|
@@ -110,7 +120,9 @@ module OpenAI
|
|
110
120
|
?parallel_tool_calls: bool?,
|
111
121
|
?previous_response_id: String?,
|
112
122
|
?prompt: OpenAI::Responses::ResponsePrompt?,
|
123
|
+
?prompt_cache_key: String,
|
113
124
|
?reasoning: OpenAI::Reasoning?,
|
125
|
+
?safety_identifier: String,
|
114
126
|
?service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?,
|
115
127
|
?store: bool?,
|
116
128
|
?temperature: Float?,
|
@@ -136,7 +148,9 @@ module OpenAI
|
|
136
148
|
parallel_tool_calls: bool?,
|
137
149
|
previous_response_id: String?,
|
138
150
|
prompt: OpenAI::Responses::ResponsePrompt?,
|
151
|
+
prompt_cache_key: String,
|
139
152
|
reasoning: OpenAI::Reasoning?,
|
153
|
+
safety_identifier: String,
|
140
154
|
service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?,
|
141
155
|
store: bool?,
|
142
156
|
temperature: Float?,
|
@@ -3,7 +3,7 @@ module OpenAI
|
|
3
3
|
module Responses
|
4
4
|
type response_mcp_call_arguments_delta_event =
|
5
5
|
{
|
6
|
-
delta:
|
6
|
+
delta: String,
|
7
7
|
item_id: String,
|
8
8
|
output_index: Integer,
|
9
9
|
sequence_number: Integer,
|
@@ -11,7 +11,7 @@ module OpenAI
|
|
11
11
|
}
|
12
12
|
|
13
13
|
class ResponseMcpCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel
|
14
|
-
attr_accessor delta:
|
14
|
+
attr_accessor delta: String
|
15
15
|
|
16
16
|
attr_accessor item_id: String
|
17
17
|
|
@@ -22,7 +22,7 @@ module OpenAI
|
|
22
22
|
attr_accessor type: :"response.mcp_call_arguments.delta"
|
23
23
|
|
24
24
|
def initialize: (
|
25
|
-
delta:
|
25
|
+
delta: String,
|
26
26
|
item_id: String,
|
27
27
|
output_index: Integer,
|
28
28
|
sequence_number: Integer,
|
@@ -30,7 +30,7 @@ module OpenAI
|
|
30
30
|
) -> void
|
31
31
|
|
32
32
|
def to_hash: -> {
|
33
|
-
delta:
|
33
|
+
delta: String,
|
34
34
|
item_id: String,
|
35
35
|
output_index: Integer,
|
36
36
|
sequence_number: Integer,
|
@@ -3,7 +3,7 @@ module OpenAI
|
|
3
3
|
module Responses
|
4
4
|
type response_mcp_call_arguments_done_event =
|
5
5
|
{
|
6
|
-
arguments:
|
6
|
+
arguments: String,
|
7
7
|
item_id: String,
|
8
8
|
output_index: Integer,
|
9
9
|
sequence_number: Integer,
|
@@ -11,7 +11,7 @@ module OpenAI
|
|
11
11
|
}
|
12
12
|
|
13
13
|
class ResponseMcpCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel
|
14
|
-
attr_accessor arguments:
|
14
|
+
attr_accessor arguments: String
|
15
15
|
|
16
16
|
attr_accessor item_id: String
|
17
17
|
|
@@ -22,7 +22,7 @@ module OpenAI
|
|
22
22
|
attr_accessor type: :"response.mcp_call_arguments.done"
|
23
23
|
|
24
24
|
def initialize: (
|
25
|
-
arguments:
|
25
|
+
arguments: String,
|
26
26
|
item_id: String,
|
27
27
|
output_index: Integer,
|
28
28
|
sequence_number: Integer,
|
@@ -30,7 +30,7 @@ module OpenAI
|
|
30
30
|
) -> void
|
31
31
|
|
32
32
|
def to_hash: -> {
|
33
|
-
arguments:
|
33
|
+
arguments: String,
|
34
34
|
item_id: String,
|
35
35
|
output_index: Integer,
|
36
36
|
sequence_number: Integer,
|
@@ -2,19 +2,32 @@ module OpenAI
|
|
2
2
|
module Models
|
3
3
|
module Responses
|
4
4
|
type response_mcp_call_completed_event =
|
5
|
-
{
|
5
|
+
{
|
6
|
+
item_id: String,
|
7
|
+
output_index: Integer,
|
8
|
+
sequence_number: Integer,
|
9
|
+
type: :"response.mcp_call.completed"
|
10
|
+
}
|
6
11
|
|
7
12
|
class ResponseMcpCallCompletedEvent < OpenAI::Internal::Type::BaseModel
|
13
|
+
attr_accessor item_id: String
|
14
|
+
|
15
|
+
attr_accessor output_index: Integer
|
16
|
+
|
8
17
|
attr_accessor sequence_number: Integer
|
9
18
|
|
10
19
|
attr_accessor type: :"response.mcp_call.completed"
|
11
20
|
|
12
21
|
def initialize: (
|
22
|
+
item_id: String,
|
23
|
+
output_index: Integer,
|
13
24
|
sequence_number: Integer,
|
14
25
|
?type: :"response.mcp_call.completed"
|
15
26
|
) -> void
|
16
27
|
|
17
28
|
def to_hash: -> {
|
29
|
+
item_id: String,
|
30
|
+
output_index: Integer,
|
18
31
|
sequence_number: Integer,
|
19
32
|
type: :"response.mcp_call.completed"
|
20
33
|
}
|
@@ -2,19 +2,32 @@ module OpenAI
|
|
2
2
|
module Models
|
3
3
|
module Responses
|
4
4
|
type response_mcp_call_failed_event =
|
5
|
-
{
|
5
|
+
{
|
6
|
+
item_id: String,
|
7
|
+
output_index: Integer,
|
8
|
+
sequence_number: Integer,
|
9
|
+
type: :"response.mcp_call.failed"
|
10
|
+
}
|
6
11
|
|
7
12
|
class ResponseMcpCallFailedEvent < OpenAI::Internal::Type::BaseModel
|
13
|
+
attr_accessor item_id: String
|
14
|
+
|
15
|
+
attr_accessor output_index: Integer
|
16
|
+
|
8
17
|
attr_accessor sequence_number: Integer
|
9
18
|
|
10
19
|
attr_accessor type: :"response.mcp_call.failed"
|
11
20
|
|
12
21
|
def initialize: (
|
22
|
+
item_id: String,
|
23
|
+
output_index: Integer,
|
13
24
|
sequence_number: Integer,
|
14
25
|
?type: :"response.mcp_call.failed"
|
15
26
|
) -> void
|
16
27
|
|
17
28
|
def to_hash: -> {
|
29
|
+
item_id: String,
|
30
|
+
output_index: Integer,
|
18
31
|
sequence_number: Integer,
|
19
32
|
type: :"response.mcp_call.failed"
|
20
33
|
}
|
@@ -2,19 +2,32 @@ module OpenAI
|
|
2
2
|
module Models
|
3
3
|
module Responses
|
4
4
|
type response_mcp_list_tools_completed_event =
|
5
|
-
{
|
5
|
+
{
|
6
|
+
item_id: String,
|
7
|
+
output_index: Integer,
|
8
|
+
sequence_number: Integer,
|
9
|
+
type: :"response.mcp_list_tools.completed"
|
10
|
+
}
|
6
11
|
|
7
12
|
class ResponseMcpListToolsCompletedEvent < OpenAI::Internal::Type::BaseModel
|
13
|
+
attr_accessor item_id: String
|
14
|
+
|
15
|
+
attr_accessor output_index: Integer
|
16
|
+
|
8
17
|
attr_accessor sequence_number: Integer
|
9
18
|
|
10
19
|
attr_accessor type: :"response.mcp_list_tools.completed"
|
11
20
|
|
12
21
|
def initialize: (
|
22
|
+
item_id: String,
|
23
|
+
output_index: Integer,
|
13
24
|
sequence_number: Integer,
|
14
25
|
?type: :"response.mcp_list_tools.completed"
|
15
26
|
) -> void
|
16
27
|
|
17
28
|
def to_hash: -> {
|
29
|
+
item_id: String,
|
30
|
+
output_index: Integer,
|
18
31
|
sequence_number: Integer,
|
19
32
|
type: :"response.mcp_list_tools.completed"
|
20
33
|
}
|
@@ -2,19 +2,32 @@ module OpenAI
|
|
2
2
|
module Models
|
3
3
|
module Responses
|
4
4
|
type response_mcp_list_tools_failed_event =
|
5
|
-
{
|
5
|
+
{
|
6
|
+
item_id: String,
|
7
|
+
output_index: Integer,
|
8
|
+
sequence_number: Integer,
|
9
|
+
type: :"response.mcp_list_tools.failed"
|
10
|
+
}
|
6
11
|
|
7
12
|
class ResponseMcpListToolsFailedEvent < OpenAI::Internal::Type::BaseModel
|
13
|
+
attr_accessor item_id: String
|
14
|
+
|
15
|
+
attr_accessor output_index: Integer
|
16
|
+
|
8
17
|
attr_accessor sequence_number: Integer
|
9
18
|
|
10
19
|
attr_accessor type: :"response.mcp_list_tools.failed"
|
11
20
|
|
12
21
|
def initialize: (
|
22
|
+
item_id: String,
|
23
|
+
output_index: Integer,
|
13
24
|
sequence_number: Integer,
|
14
25
|
?type: :"response.mcp_list_tools.failed"
|
15
26
|
) -> void
|
16
27
|
|
17
28
|
def to_hash: -> {
|
29
|
+
item_id: String,
|
30
|
+
output_index: Integer,
|
18
31
|
sequence_number: Integer,
|
19
32
|
type: :"response.mcp_list_tools.failed"
|
20
33
|
}
|
@@ -3,21 +3,31 @@ module OpenAI
|
|
3
3
|
module Responses
|
4
4
|
type response_mcp_list_tools_in_progress_event =
|
5
5
|
{
|
6
|
+
item_id: String,
|
7
|
+
output_index: Integer,
|
6
8
|
sequence_number: Integer,
|
7
9
|
type: :"response.mcp_list_tools.in_progress"
|
8
10
|
}
|
9
11
|
|
10
12
|
class ResponseMcpListToolsInProgressEvent < OpenAI::Internal::Type::BaseModel
|
13
|
+
attr_accessor item_id: String
|
14
|
+
|
15
|
+
attr_accessor output_index: Integer
|
16
|
+
|
11
17
|
attr_accessor sequence_number: Integer
|
12
18
|
|
13
19
|
attr_accessor type: :"response.mcp_list_tools.in_progress"
|
14
20
|
|
15
21
|
def initialize: (
|
22
|
+
item_id: String,
|
23
|
+
output_index: Integer,
|
16
24
|
sequence_number: Integer,
|
17
25
|
?type: :"response.mcp_list_tools.in_progress"
|
18
26
|
) -> void
|
19
27
|
|
20
28
|
def to_hash: -> {
|
29
|
+
item_id: String,
|
30
|
+
output_index: Integer,
|
21
31
|
sequence_number: Integer,
|
22
32
|
type: :"response.mcp_list_tools.in_progress"
|
23
33
|
}
|
@@ -51,8 +51,6 @@ module OpenAI
|
|
51
51
|
| OpenAI::Responses::ResponseMcpListToolsInProgressEvent
|
52
52
|
| OpenAI::Responses::ResponseOutputTextAnnotationAddedEvent
|
53
53
|
| OpenAI::Responses::ResponseQueuedEvent
|
54
|
-
| OpenAI::Responses::ResponseReasoningDeltaEvent
|
55
|
-
| OpenAI::Responses::ResponseReasoningDoneEvent
|
56
54
|
| OpenAI::Responses::ResponseReasoningSummaryDeltaEvent
|
57
55
|
| OpenAI::Responses::ResponseReasoningSummaryDoneEvent
|
58
56
|
|
@@ -6,6 +6,7 @@ module OpenAI
|
|
6
6
|
content_index: Integer,
|
7
7
|
delta: String,
|
8
8
|
item_id: String,
|
9
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob],
|
9
10
|
output_index: Integer,
|
10
11
|
sequence_number: Integer,
|
11
12
|
type: :"response.output_text.delta"
|
@@ -18,6 +19,8 @@ module OpenAI
|
|
18
19
|
|
19
20
|
attr_accessor item_id: String
|
20
21
|
|
22
|
+
attr_accessor logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob]
|
23
|
+
|
21
24
|
attr_accessor output_index: Integer
|
22
25
|
|
23
26
|
attr_accessor sequence_number: Integer
|
@@ -28,6 +31,7 @@ module OpenAI
|
|
28
31
|
content_index: Integer,
|
29
32
|
delta: String,
|
30
33
|
item_id: String,
|
34
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob],
|
31
35
|
output_index: Integer,
|
32
36
|
sequence_number: Integer,
|
33
37
|
?type: :"response.output_text.delta"
|
@@ -37,10 +41,58 @@ module OpenAI
|
|
37
41
|
content_index: Integer,
|
38
42
|
delta: String,
|
39
43
|
item_id: String,
|
44
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob],
|
40
45
|
output_index: Integer,
|
41
46
|
sequence_number: Integer,
|
42
47
|
type: :"response.output_text.delta"
|
43
48
|
}
|
49
|
+
|
50
|
+
type logprob =
|
51
|
+
{
|
52
|
+
token: String,
|
53
|
+
logprob: Float,
|
54
|
+
top_logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]
|
55
|
+
}
|
56
|
+
|
57
|
+
class Logprob < OpenAI::Internal::Type::BaseModel
|
58
|
+
attr_accessor token: String
|
59
|
+
|
60
|
+
attr_accessor logprob: Float
|
61
|
+
|
62
|
+
attr_reader top_logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]?
|
63
|
+
|
64
|
+
def top_logprobs=: (
|
65
|
+
::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]
|
66
|
+
) -> ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]
|
67
|
+
|
68
|
+
def initialize: (
|
69
|
+
token: String,
|
70
|
+
logprob: Float,
|
71
|
+
?top_logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]
|
72
|
+
) -> void
|
73
|
+
|
74
|
+
def to_hash: -> {
|
75
|
+
token: String,
|
76
|
+
logprob: Float,
|
77
|
+
top_logprobs: ::Array[OpenAI::Responses::ResponseTextDeltaEvent::Logprob::TopLogprob]
|
78
|
+
}
|
79
|
+
|
80
|
+
type top_logprob = { token: String, logprob: Float }
|
81
|
+
|
82
|
+
class TopLogprob < OpenAI::Internal::Type::BaseModel
|
83
|
+
attr_reader token: String?
|
84
|
+
|
85
|
+
def token=: (String) -> String
|
86
|
+
|
87
|
+
attr_reader logprob: Float?
|
88
|
+
|
89
|
+
def logprob=: (Float) -> Float
|
90
|
+
|
91
|
+
def initialize: (?token: String, ?logprob: Float) -> void
|
92
|
+
|
93
|
+
def to_hash: -> { token: String, logprob: Float }
|
94
|
+
end
|
95
|
+
end
|
44
96
|
end
|
45
97
|
end
|
46
98
|
end
|
@@ -5,6 +5,7 @@ module OpenAI
|
|
5
5
|
{
|
6
6
|
content_index: Integer,
|
7
7
|
item_id: String,
|
8
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob],
|
8
9
|
output_index: Integer,
|
9
10
|
sequence_number: Integer,
|
10
11
|
text: String,
|
@@ -16,6 +17,8 @@ module OpenAI
|
|
16
17
|
|
17
18
|
attr_accessor item_id: String
|
18
19
|
|
20
|
+
attr_accessor logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob]
|
21
|
+
|
19
22
|
attr_accessor output_index: Integer
|
20
23
|
|
21
24
|
attr_accessor sequence_number: Integer
|
@@ -27,6 +30,7 @@ module OpenAI
|
|
27
30
|
def initialize: (
|
28
31
|
content_index: Integer,
|
29
32
|
item_id: String,
|
33
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob],
|
30
34
|
output_index: Integer,
|
31
35
|
sequence_number: Integer,
|
32
36
|
text: String,
|
@@ -36,11 +40,59 @@ module OpenAI
|
|
36
40
|
def to_hash: -> {
|
37
41
|
content_index: Integer,
|
38
42
|
item_id: String,
|
43
|
+
logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob],
|
39
44
|
output_index: Integer,
|
40
45
|
sequence_number: Integer,
|
41
46
|
text: String,
|
42
47
|
type: :"response.output_text.done"
|
43
48
|
}
|
49
|
+
|
50
|
+
type logprob =
|
51
|
+
{
|
52
|
+
token: String,
|
53
|
+
logprob: Float,
|
54
|
+
top_logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]
|
55
|
+
}
|
56
|
+
|
57
|
+
class Logprob < OpenAI::Internal::Type::BaseModel
|
58
|
+
attr_accessor token: String
|
59
|
+
|
60
|
+
attr_accessor logprob: Float
|
61
|
+
|
62
|
+
attr_reader top_logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]?
|
63
|
+
|
64
|
+
def top_logprobs=: (
|
65
|
+
::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]
|
66
|
+
) -> ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]
|
67
|
+
|
68
|
+
def initialize: (
|
69
|
+
token: String,
|
70
|
+
logprob: Float,
|
71
|
+
?top_logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]
|
72
|
+
) -> void
|
73
|
+
|
74
|
+
def to_hash: -> {
|
75
|
+
token: String,
|
76
|
+
logprob: Float,
|
77
|
+
top_logprobs: ::Array[OpenAI::Responses::ResponseTextDoneEvent::Logprob::TopLogprob]
|
78
|
+
}
|
79
|
+
|
80
|
+
type top_logprob = { token: String, logprob: Float }
|
81
|
+
|
82
|
+
class TopLogprob < OpenAI::Internal::Type::BaseModel
|
83
|
+
attr_reader token: String?
|
84
|
+
|
85
|
+
def token=: (String) -> String
|
86
|
+
|
87
|
+
attr_reader logprob: Float?
|
88
|
+
|
89
|
+
def logprob=: (Float) -> Float
|
90
|
+
|
91
|
+
def initialize: (?token: String, ?logprob: Float) -> void
|
92
|
+
|
93
|
+
def to_hash: -> { token: String, logprob: Float }
|
94
|
+
end
|
95
|
+
end
|
44
96
|
end
|
45
97
|
end
|
46
98
|
end
|
@@ -21,8 +21,10 @@ module OpenAI
|
|
21
21
|
?parallel_tool_calls: bool,
|
22
22
|
?prediction: OpenAI::Chat::ChatCompletionPredictionContent?,
|
23
23
|
?presence_penalty: Float?,
|
24
|
+
?prompt_cache_key: String,
|
24
25
|
?reasoning_effort: OpenAI::Models::reasoning_effort?,
|
25
26
|
?response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format,
|
27
|
+
?safety_identifier: String,
|
26
28
|
?seed: Integer?,
|
27
29
|
?service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?,
|
28
30
|
?stop: OpenAI::Models::Chat::CompletionCreateParams::stop?,
|
@@ -55,8 +57,10 @@ module OpenAI
|
|
55
57
|
?parallel_tool_calls: bool,
|
56
58
|
?prediction: OpenAI::Chat::ChatCompletionPredictionContent?,
|
57
59
|
?presence_penalty: Float?,
|
60
|
+
?prompt_cache_key: String,
|
58
61
|
?reasoning_effort: OpenAI::Models::reasoning_effort?,
|
59
62
|
?response_format: OpenAI::Models::Chat::CompletionCreateParams::response_format,
|
63
|
+
?safety_identifier: String,
|
60
64
|
?seed: Integer?,
|
61
65
|
?service_tier: OpenAI::Models::Chat::CompletionCreateParams::service_tier?,
|
62
66
|
?stop: OpenAI::Models::Chat::CompletionCreateParams::stop?,
|
@@ -15,7 +15,9 @@ module OpenAI
|
|
15
15
|
?parallel_tool_calls: bool?,
|
16
16
|
?previous_response_id: String?,
|
17
17
|
?prompt: OpenAI::Responses::ResponsePrompt?,
|
18
|
+
?prompt_cache_key: String,
|
18
19
|
?reasoning: OpenAI::Reasoning?,
|
20
|
+
?safety_identifier: String,
|
19
21
|
?service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?,
|
20
22
|
?store: bool?,
|
21
23
|
?temperature: Float?,
|
@@ -41,7 +43,9 @@ module OpenAI
|
|
41
43
|
?parallel_tool_calls: bool?,
|
42
44
|
?previous_response_id: String?,
|
43
45
|
?prompt: OpenAI::Responses::ResponsePrompt?,
|
46
|
+
?prompt_cache_key: String,
|
44
47
|
?reasoning: OpenAI::Reasoning?,
|
48
|
+
?safety_identifier: String,
|
45
49
|
?service_tier: OpenAI::Models::Responses::ResponseCreateParams::service_tier?,
|
46
50
|
?store: bool?,
|
47
51
|
?temperature: Float?,
|