openai 0.55.0 → 0.56.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +16 -0
- data/README.md +1 -1
- data/lib/openai/internal/type/enum.rb +4 -4
- data/lib/openai/models/batch_create_params.rb +9 -6
- data/lib/openai/models/chat_model.rb +4 -0
- data/lib/openai/models/comparison_filter.rb +2 -0
- data/lib/openai/models/responses/namespace_tool.rb +13 -1
- data/lib/openai/models/responses/response_compact_params.rb +12 -0
- data/lib/openai/version.rb +1 -1
- data/rbi/openai/models/batch_create_params.rbi +17 -9
- data/rbi/openai/models/chat_model.rbi +6 -0
- data/rbi/openai/models/comparison_filter.rbi +2 -0
- data/rbi/openai/models/responses/namespace_tool.rbi +11 -0
- data/rbi/openai/models/responses/response_compact_params.rbi +20 -0
- data/rbi/openai/resources/batches.rbi +4 -3
- data/sig/openai/models/batch_create_params.rbs +2 -0
- data/sig/openai/models/chat_model.rbs +8 -0
- data/sig/openai/models/comparison_filter.rbs +3 -1
- data/sig/openai/models/responses/namespace_tool.rbs +7 -0
- data/sig/openai/models/responses/response_compact_params.rbs +8 -0
- metadata +2 -2
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: 8a0399e48fa95f65032c615357845d2a29d1c0ce34677cf08fc446657061fcc7
|
|
4
|
+
data.tar.gz: 52a29c39bcea24e9d9cac99bf840a7f5f4de0ca2ff20b897607dc86e5396eae4
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 4efb1d88dae705b22655d2ba203fa0d05a6106f142683dcebf02de9002dece4dedf55422bb84cc3be58f73e2c6fe4a38dc5641abae5bc5a2ce9361e23737fec3
|
|
7
|
+
data.tar.gz: '091021a04c89b45d3ffc6cd8ee096adb98d7ef0bb886c6db3c48a66178d833044539664f4fdd9848326550b6c07f5123f5daeca737972523356b50368d558535'
|
data/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,21 @@
|
|
|
1
1
|
# Changelog
|
|
2
2
|
|
|
3
|
+
## 0.56.0 (2026-03-17)
|
|
4
|
+
|
|
5
|
+
Full Changelog: [v0.55.0...v0.56.0](https://github.com/openai/openai-ruby/compare/v0.55.0...v0.56.0)
|
|
6
|
+
|
|
7
|
+
### Features
|
|
8
|
+
|
|
9
|
+
* **api:** 5.4 nano and mini model slugs ([ce7f09d](https://github.com/openai/openai-ruby/commit/ce7f09dcafdb1a8259ce14d0e7c012b872469a32))
|
|
10
|
+
* **api:** add /v1/videos endpoint support to batch create params ([7a661eb](https://github.com/openai/openai-ruby/commit/7a661eb405c72bb2696e0acd327ecd56d7ed6c54))
|
|
11
|
+
* **api:** add defer_loading field to namespace_tool function ([2f3b9b7](https://github.com/openai/openai-ruby/commit/2f3b9b7a2fa7b63ed20d362a993d98d8cec72fc1))
|
|
12
|
+
* **api:** add in/nin operators to ComparisonFilter ([70c8bf5](https://github.com/openai/openai-ruby/commit/70c8bf55c4d0e2054e23972479edb09c0d8f3d8e))
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
### Chores
|
|
16
|
+
|
|
17
|
+
* **internal:** tweak CI branches ([1a5054f](https://github.com/openai/openai-ruby/commit/1a5054f561a63a4b19c0d2a035e1c45dcfbc21fc))
|
|
18
|
+
|
|
3
19
|
## 0.55.0 (2026-03-13)
|
|
4
20
|
|
|
5
21
|
Full Changelog: [v0.54.0...v0.55.0](https://github.com/openai/openai-ruby/compare/v0.54.0...v0.55.0)
|
data/README.md
CHANGED
|
@@ -21,9 +21,9 @@ module OpenAI
|
|
|
21
21
|
# case chat_model
|
|
22
22
|
# when OpenAI::ChatModel::GPT_5_4
|
|
23
23
|
# # ...
|
|
24
|
-
# when OpenAI::ChatModel::
|
|
24
|
+
# when OpenAI::ChatModel::GPT_5_4_MINI
|
|
25
25
|
# # ...
|
|
26
|
-
# when OpenAI::ChatModel::
|
|
26
|
+
# when OpenAI::ChatModel::GPT_5_4_NANO
|
|
27
27
|
# # ...
|
|
28
28
|
# else
|
|
29
29
|
# puts(chat_model)
|
|
@@ -33,9 +33,9 @@ module OpenAI
|
|
|
33
33
|
# case chat_model
|
|
34
34
|
# in :"gpt-5.4"
|
|
35
35
|
# # ...
|
|
36
|
-
# in :"gpt-5.
|
|
36
|
+
# in :"gpt-5.4-mini"
|
|
37
37
|
# # ...
|
|
38
|
-
# in :"gpt-5.
|
|
38
|
+
# in :"gpt-5.4-nano"
|
|
39
39
|
# # ...
|
|
40
40
|
# else
|
|
41
41
|
# puts(chat_model)
|
|
@@ -17,9 +17,10 @@ module OpenAI
|
|
|
17
17
|
# @!attribute endpoint
|
|
18
18
|
# The endpoint to be used for all requests in the batch. Currently
|
|
19
19
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
20
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
21
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
22
|
-
# of 50,000 embedding inputs across all requests in the
|
|
20
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
21
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
22
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
23
|
+
# batch.
|
|
23
24
|
#
|
|
24
25
|
# @return [Symbol, OpenAI::Models::BatchCreateParams::Endpoint]
|
|
25
26
|
required :endpoint, enum: -> { OpenAI::BatchCreateParams::Endpoint }
|
|
@@ -85,9 +86,10 @@ module OpenAI
|
|
|
85
86
|
|
|
86
87
|
# The endpoint to be used for all requests in the batch. Currently
|
|
87
88
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
88
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
89
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
90
|
-
# of 50,000 embedding inputs across all requests in the
|
|
89
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
90
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
91
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
92
|
+
# batch.
|
|
91
93
|
module Endpoint
|
|
92
94
|
extend OpenAI::Internal::Type::Enum
|
|
93
95
|
|
|
@@ -98,6 +100,7 @@ module OpenAI
|
|
|
98
100
|
V1_MODERATIONS = :"/v1/moderations"
|
|
99
101
|
V1_IMAGES_GENERATIONS = :"/v1/images/generations"
|
|
100
102
|
V1_IMAGES_EDITS = :"/v1/images/edits"
|
|
103
|
+
V1_VIDEOS = :"/v1/videos"
|
|
101
104
|
|
|
102
105
|
# @!method self.values
|
|
103
106
|
# @return [Array<Symbol>]
|
|
@@ -6,6 +6,10 @@ module OpenAI
|
|
|
6
6
|
extend OpenAI::Internal::Type::Enum
|
|
7
7
|
|
|
8
8
|
GPT_5_4 = :"gpt-5.4"
|
|
9
|
+
GPT_5_4_MINI = :"gpt-5.4-mini"
|
|
10
|
+
GPT_5_4_NANO = :"gpt-5.4-nano"
|
|
11
|
+
GPT_5_4_MINI_2026_03_17 = :"gpt-5.4-mini-2026-03-17"
|
|
12
|
+
GPT_5_4_NANO_2026_03_17 = :"gpt-5.4-nano-2026-03-17"
|
|
9
13
|
GPT_5_3_CHAT_LATEST = :"gpt-5.3-chat-latest"
|
|
10
14
|
GPT_5_2 = :"gpt-5.2"
|
|
11
15
|
GPT_5_2_2025_12_11 = :"gpt-5.2-2025-12-11"
|
|
@@ -61,6 +61,12 @@ module OpenAI
|
|
|
61
61
|
# @return [Symbol, :function]
|
|
62
62
|
required :type, const: :function
|
|
63
63
|
|
|
64
|
+
# @!attribute defer_loading
|
|
65
|
+
# Whether this function should be deferred and discovered via tool search.
|
|
66
|
+
#
|
|
67
|
+
# @return [Boolean, nil]
|
|
68
|
+
optional :defer_loading, OpenAI::Internal::Type::Boolean
|
|
69
|
+
|
|
64
70
|
# @!attribute description
|
|
65
71
|
#
|
|
66
72
|
# @return [String, nil]
|
|
@@ -76,11 +82,17 @@ module OpenAI
|
|
|
76
82
|
# @return [Boolean, nil]
|
|
77
83
|
optional :strict, OpenAI::Internal::Type::Boolean, nil?: true
|
|
78
84
|
|
|
79
|
-
# @!method initialize(name:, description: nil, parameters: nil, strict: nil, type: :function)
|
|
85
|
+
# @!method initialize(name:, defer_loading: nil, description: nil, parameters: nil, strict: nil, type: :function)
|
|
80
86
|
# @param name [String]
|
|
87
|
+
#
|
|
88
|
+
# @param defer_loading [Boolean] Whether this function should be deferred and discovered via tool search.
|
|
89
|
+
#
|
|
81
90
|
# @param description [String, nil]
|
|
91
|
+
#
|
|
82
92
|
# @param parameters [Object, nil]
|
|
93
|
+
#
|
|
83
94
|
# @param strict [Boolean, nil]
|
|
95
|
+
#
|
|
84
96
|
# @param type [Symbol, :function]
|
|
85
97
|
end
|
|
86
98
|
|
|
@@ -74,6 +74,14 @@ module OpenAI
|
|
|
74
74
|
|
|
75
75
|
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_4 }
|
|
76
76
|
|
|
77
|
+
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_4_MINI }
|
|
78
|
+
|
|
79
|
+
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_4_NANO }
|
|
80
|
+
|
|
81
|
+
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_4_MINI_2026_03_17 }
|
|
82
|
+
|
|
83
|
+
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_4_NANO_2026_03_17 }
|
|
84
|
+
|
|
77
85
|
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_3_CHAT_LATEST }
|
|
78
86
|
|
|
79
87
|
variant const: -> { OpenAI::Models::Responses::ResponseCompactParams::Model::GPT_5_2 }
|
|
@@ -260,6 +268,10 @@ module OpenAI
|
|
|
260
268
|
# @!group
|
|
261
269
|
|
|
262
270
|
GPT_5_4 = :"gpt-5.4"
|
|
271
|
+
GPT_5_4_MINI = :"gpt-5.4-mini"
|
|
272
|
+
GPT_5_4_NANO = :"gpt-5.4-nano"
|
|
273
|
+
GPT_5_4_MINI_2026_03_17 = :"gpt-5.4-mini-2026-03-17"
|
|
274
|
+
GPT_5_4_NANO_2026_03_17 = :"gpt-5.4-nano-2026-03-17"
|
|
263
275
|
GPT_5_3_CHAT_LATEST = :"gpt-5.3-chat-latest"
|
|
264
276
|
GPT_5_2 = :"gpt-5.2"
|
|
265
277
|
GPT_5_2_2025_12_11 = :"gpt-5.2-2025-12-11"
|
data/lib/openai/version.rb
CHANGED
|
@@ -18,9 +18,10 @@ module OpenAI
|
|
|
18
18
|
|
|
19
19
|
# The endpoint to be used for all requests in the batch. Currently
|
|
20
20
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
21
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
22
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
23
|
-
# of 50,000 embedding inputs across all requests in the
|
|
21
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
22
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
23
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
24
|
+
# batch.
|
|
24
25
|
sig { returns(OpenAI::BatchCreateParams::Endpoint::OrSymbol) }
|
|
25
26
|
attr_accessor :endpoint
|
|
26
27
|
|
|
@@ -76,9 +77,10 @@ module OpenAI
|
|
|
76
77
|
completion_window:,
|
|
77
78
|
# The endpoint to be used for all requests in the batch. Currently
|
|
78
79
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
79
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
80
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
81
|
-
# of 50,000 embedding inputs across all requests in the
|
|
80
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
81
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
82
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
83
|
+
# batch.
|
|
82
84
|
endpoint:,
|
|
83
85
|
# The ID of an uploaded file that contains requests for the new batch.
|
|
84
86
|
#
|
|
@@ -148,9 +150,10 @@ module OpenAI
|
|
|
148
150
|
|
|
149
151
|
# The endpoint to be used for all requests in the batch. Currently
|
|
150
152
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
151
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
152
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
153
|
-
# of 50,000 embedding inputs across all requests in the
|
|
153
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
154
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
155
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
156
|
+
# batch.
|
|
154
157
|
module Endpoint
|
|
155
158
|
extend OpenAI::Internal::Type::Enum
|
|
156
159
|
|
|
@@ -193,6 +196,11 @@ module OpenAI
|
|
|
193
196
|
:"/v1/images/edits",
|
|
194
197
|
OpenAI::BatchCreateParams::Endpoint::TaggedSymbol
|
|
195
198
|
)
|
|
199
|
+
V1_VIDEOS =
|
|
200
|
+
T.let(
|
|
201
|
+
:"/v1/videos",
|
|
202
|
+
OpenAI::BatchCreateParams::Endpoint::TaggedSymbol
|
|
203
|
+
)
|
|
196
204
|
|
|
197
205
|
sig do
|
|
198
206
|
override.returns(
|
|
@@ -9,6 +9,12 @@ module OpenAI
|
|
|
9
9
|
OrSymbol = T.type_alias { T.any(Symbol, String) }
|
|
10
10
|
|
|
11
11
|
GPT_5_4 = T.let(:"gpt-5.4", OpenAI::ChatModel::TaggedSymbol)
|
|
12
|
+
GPT_5_4_MINI = T.let(:"gpt-5.4-mini", OpenAI::ChatModel::TaggedSymbol)
|
|
13
|
+
GPT_5_4_NANO = T.let(:"gpt-5.4-nano", OpenAI::ChatModel::TaggedSymbol)
|
|
14
|
+
GPT_5_4_MINI_2026_03_17 =
|
|
15
|
+
T.let(:"gpt-5.4-mini-2026-03-17", OpenAI::ChatModel::TaggedSymbol)
|
|
16
|
+
GPT_5_4_NANO_2026_03_17 =
|
|
17
|
+
T.let(:"gpt-5.4-nano-2026-03-17", OpenAI::ChatModel::TaggedSymbol)
|
|
12
18
|
GPT_5_3_CHAT_LATEST =
|
|
13
19
|
T.let(:"gpt-5.3-chat-latest", OpenAI::ChatModel::TaggedSymbol)
|
|
14
20
|
GPT_5_2 = T.let(:"gpt-5.2", OpenAI::ChatModel::TaggedSymbol)
|
|
@@ -97,6 +97,8 @@ module OpenAI
|
|
|
97
97
|
GTE = T.let(:gte, OpenAI::ComparisonFilter::Type::TaggedSymbol)
|
|
98
98
|
LT = T.let(:lt, OpenAI::ComparisonFilter::Type::TaggedSymbol)
|
|
99
99
|
LTE = T.let(:lte, OpenAI::ComparisonFilter::Type::TaggedSymbol)
|
|
100
|
+
IN = T.let(:in, OpenAI::ComparisonFilter::Type::TaggedSymbol)
|
|
101
|
+
NIN = T.let(:nin, OpenAI::ComparisonFilter::Type::TaggedSymbol)
|
|
100
102
|
|
|
101
103
|
sig do
|
|
102
104
|
override.returns(
|
|
@@ -107,6 +107,13 @@ module OpenAI
|
|
|
107
107
|
sig { returns(Symbol) }
|
|
108
108
|
attr_accessor :type
|
|
109
109
|
|
|
110
|
+
# Whether this function should be deferred and discovered via tool search.
|
|
111
|
+
sig { returns(T.nilable(T::Boolean)) }
|
|
112
|
+
attr_reader :defer_loading
|
|
113
|
+
|
|
114
|
+
sig { params(defer_loading: T::Boolean).void }
|
|
115
|
+
attr_writer :defer_loading
|
|
116
|
+
|
|
110
117
|
sig { returns(T.nilable(String)) }
|
|
111
118
|
attr_accessor :description
|
|
112
119
|
|
|
@@ -119,6 +126,7 @@ module OpenAI
|
|
|
119
126
|
sig do
|
|
120
127
|
params(
|
|
121
128
|
name: String,
|
|
129
|
+
defer_loading: T::Boolean,
|
|
122
130
|
description: T.nilable(String),
|
|
123
131
|
parameters: T.nilable(T.anything),
|
|
124
132
|
strict: T.nilable(T::Boolean),
|
|
@@ -127,6 +135,8 @@ module OpenAI
|
|
|
127
135
|
end
|
|
128
136
|
def self.new(
|
|
129
137
|
name:,
|
|
138
|
+
# Whether this function should be deferred and discovered via tool search.
|
|
139
|
+
defer_loading: nil,
|
|
130
140
|
description: nil,
|
|
131
141
|
parameters: nil,
|
|
132
142
|
strict: nil,
|
|
@@ -139,6 +149,7 @@ module OpenAI
|
|
|
139
149
|
{
|
|
140
150
|
name: String,
|
|
141
151
|
type: Symbol,
|
|
152
|
+
defer_loading: T::Boolean,
|
|
142
153
|
description: T.nilable(String),
|
|
143
154
|
parameters: T.nilable(T.anything),
|
|
144
155
|
strict: T.nilable(T::Boolean)
|
|
@@ -163,6 +163,26 @@ module OpenAI
|
|
|
163
163
|
:"gpt-5.4",
|
|
164
164
|
OpenAI::Responses::ResponseCompactParams::Model::TaggedSymbol
|
|
165
165
|
)
|
|
166
|
+
GPT_5_4_MINI =
|
|
167
|
+
T.let(
|
|
168
|
+
:"gpt-5.4-mini",
|
|
169
|
+
OpenAI::Responses::ResponseCompactParams::Model::TaggedSymbol
|
|
170
|
+
)
|
|
171
|
+
GPT_5_4_NANO =
|
|
172
|
+
T.let(
|
|
173
|
+
:"gpt-5.4-nano",
|
|
174
|
+
OpenAI::Responses::ResponseCompactParams::Model::TaggedSymbol
|
|
175
|
+
)
|
|
176
|
+
GPT_5_4_MINI_2026_03_17 =
|
|
177
|
+
T.let(
|
|
178
|
+
:"gpt-5.4-mini-2026-03-17",
|
|
179
|
+
OpenAI::Responses::ResponseCompactParams::Model::TaggedSymbol
|
|
180
|
+
)
|
|
181
|
+
GPT_5_4_NANO_2026_03_17 =
|
|
182
|
+
T.let(
|
|
183
|
+
:"gpt-5.4-nano-2026-03-17",
|
|
184
|
+
OpenAI::Responses::ResponseCompactParams::Model::TaggedSymbol
|
|
185
|
+
)
|
|
166
186
|
GPT_5_3_CHAT_LATEST =
|
|
167
187
|
T.let(
|
|
168
188
|
:"gpt-5.3-chat-latest",
|
|
@@ -23,9 +23,10 @@ module OpenAI
|
|
|
23
23
|
completion_window:,
|
|
24
24
|
# The endpoint to be used for all requests in the batch. Currently
|
|
25
25
|
# `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
|
|
26
|
-
# `/v1/moderations`, `/v1/images/generations`,
|
|
27
|
-
# supported. Note that `/v1/embeddings` batches are also
|
|
28
|
-
# of 50,000 embedding inputs across all requests in the
|
|
26
|
+
# `/v1/moderations`, `/v1/images/generations`, `/v1/images/edits`, and
|
|
27
|
+
# `/v1/videos` are supported. Note that `/v1/embeddings` batches are also
|
|
28
|
+
# restricted to a maximum of 50,000 embedding inputs across all requests in the
|
|
29
|
+
# batch.
|
|
29
30
|
endpoint:,
|
|
30
31
|
# The ID of an uploaded file that contains requests for the new batch.
|
|
31
32
|
#
|
|
@@ -64,6 +64,7 @@ module OpenAI
|
|
|
64
64
|
| :"/v1/moderations"
|
|
65
65
|
| :"/v1/images/generations"
|
|
66
66
|
| :"/v1/images/edits"
|
|
67
|
+
| :"/v1/videos"
|
|
67
68
|
|
|
68
69
|
module Endpoint
|
|
69
70
|
extend OpenAI::Internal::Type::Enum
|
|
@@ -75,6 +76,7 @@ module OpenAI
|
|
|
75
76
|
V1_MODERATIONS: :"/v1/moderations"
|
|
76
77
|
V1_IMAGES_GENERATIONS: :"/v1/images/generations"
|
|
77
78
|
V1_IMAGES_EDITS: :"/v1/images/edits"
|
|
79
|
+
V1_VIDEOS: :"/v1/videos"
|
|
78
80
|
|
|
79
81
|
def self?.values: -> ::Array[OpenAI::Models::BatchCreateParams::endpoint]
|
|
80
82
|
end
|
|
@@ -2,6 +2,10 @@ module OpenAI
|
|
|
2
2
|
module Models
|
|
3
3
|
type chat_model =
|
|
4
4
|
:"gpt-5.4"
|
|
5
|
+
| :"gpt-5.4-mini"
|
|
6
|
+
| :"gpt-5.4-nano"
|
|
7
|
+
| :"gpt-5.4-mini-2026-03-17"
|
|
8
|
+
| :"gpt-5.4-nano-2026-03-17"
|
|
5
9
|
| :"gpt-5.3-chat-latest"
|
|
6
10
|
| :"gpt-5.2"
|
|
7
11
|
| :"gpt-5.2-2025-12-11"
|
|
@@ -80,6 +84,10 @@ module OpenAI
|
|
|
80
84
|
extend OpenAI::Internal::Type::Enum
|
|
81
85
|
|
|
82
86
|
GPT_5_4: :"gpt-5.4"
|
|
87
|
+
GPT_5_4_MINI: :"gpt-5.4-mini"
|
|
88
|
+
GPT_5_4_NANO: :"gpt-5.4-nano"
|
|
89
|
+
GPT_5_4_MINI_2026_03_17: :"gpt-5.4-mini-2026-03-17"
|
|
90
|
+
GPT_5_4_NANO_2026_03_17: :"gpt-5.4-nano-2026-03-17"
|
|
83
91
|
GPT_5_3_CHAT_LATEST: :"gpt-5.3-chat-latest"
|
|
84
92
|
GPT_5_2: :"gpt-5.2"
|
|
85
93
|
GPT_5_2_2025_12_11: :"gpt-5.2-2025-12-11"
|
|
@@ -26,7 +26,7 @@ module OpenAI
|
|
|
26
26
|
value: OpenAI::Models::ComparisonFilter::value
|
|
27
27
|
}
|
|
28
28
|
|
|
29
|
-
type type_ = :eq | :ne | :gt | :gte | :lt | :lte
|
|
29
|
+
type type_ = :eq | :ne | :gt | :gte | :lt | :lte | :in | :nin
|
|
30
30
|
|
|
31
31
|
module Type
|
|
32
32
|
extend OpenAI::Internal::Type::Enum
|
|
@@ -37,6 +37,8 @@ module OpenAI
|
|
|
37
37
|
GTE: :gte
|
|
38
38
|
LT: :lt
|
|
39
39
|
LTE: :lte
|
|
40
|
+
IN: :in
|
|
41
|
+
NIN: :nin
|
|
40
42
|
|
|
41
43
|
def self?.values: -> ::Array[OpenAI::Models::ComparisonFilter::type_]
|
|
42
44
|
end
|
|
@@ -43,6 +43,7 @@ module OpenAI
|
|
|
43
43
|
{
|
|
44
44
|
name: String,
|
|
45
45
|
type: :function,
|
|
46
|
+
defer_loading: bool,
|
|
46
47
|
description: String?,
|
|
47
48
|
parameters: top?,
|
|
48
49
|
strict: bool?
|
|
@@ -53,6 +54,10 @@ module OpenAI
|
|
|
53
54
|
|
|
54
55
|
attr_accessor type: :function
|
|
55
56
|
|
|
57
|
+
attr_reader defer_loading: bool?
|
|
58
|
+
|
|
59
|
+
def defer_loading=: (bool) -> bool
|
|
60
|
+
|
|
56
61
|
attr_accessor description: String?
|
|
57
62
|
|
|
58
63
|
attr_accessor parameters: top?
|
|
@@ -61,6 +66,7 @@ module OpenAI
|
|
|
61
66
|
|
|
62
67
|
def initialize: (
|
|
63
68
|
name: String,
|
|
69
|
+
?defer_loading: bool,
|
|
64
70
|
?description: String?,
|
|
65
71
|
?parameters: top?,
|
|
66
72
|
?strict: bool?,
|
|
@@ -70,6 +76,7 @@ module OpenAI
|
|
|
70
76
|
def to_hash: -> {
|
|
71
77
|
name: String,
|
|
72
78
|
type: :function,
|
|
79
|
+
defer_loading: bool,
|
|
73
80
|
description: String?,
|
|
74
81
|
parameters: top?,
|
|
75
82
|
strict: bool?
|
|
@@ -45,6 +45,10 @@ module OpenAI
|
|
|
45
45
|
|
|
46
46
|
type model =
|
|
47
47
|
:"gpt-5.4"
|
|
48
|
+
| :"gpt-5.4-mini"
|
|
49
|
+
| :"gpt-5.4-nano"
|
|
50
|
+
| :"gpt-5.4-mini-2026-03-17"
|
|
51
|
+
| :"gpt-5.4-nano-2026-03-17"
|
|
48
52
|
| :"gpt-5.3-chat-latest"
|
|
49
53
|
| :"gpt-5.2"
|
|
50
54
|
| :"gpt-5.2-2025-12-11"
|
|
@@ -140,6 +144,10 @@ module OpenAI
|
|
|
140
144
|
def self?.variants: -> ::Array[OpenAI::Models::Responses::ResponseCompactParams::model]
|
|
141
145
|
|
|
142
146
|
GPT_5_4: :"gpt-5.4"
|
|
147
|
+
GPT_5_4_MINI: :"gpt-5.4-mini"
|
|
148
|
+
GPT_5_4_NANO: :"gpt-5.4-nano"
|
|
149
|
+
GPT_5_4_MINI_2026_03_17: :"gpt-5.4-mini-2026-03-17"
|
|
150
|
+
GPT_5_4_NANO_2026_03_17: :"gpt-5.4-nano-2026-03-17"
|
|
143
151
|
GPT_5_3_CHAT_LATEST: :"gpt-5.3-chat-latest"
|
|
144
152
|
GPT_5_2: :"gpt-5.2"
|
|
145
153
|
GPT_5_2_2025_12_11: :"gpt-5.2-2025-12-11"
|
metadata
CHANGED
|
@@ -1,14 +1,14 @@
|
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
|
2
2
|
name: openai
|
|
3
3
|
version: !ruby/object:Gem::Version
|
|
4
|
-
version: 0.
|
|
4
|
+
version: 0.56.0
|
|
5
5
|
platform: ruby
|
|
6
6
|
authors:
|
|
7
7
|
- OpenAI
|
|
8
8
|
autorequire:
|
|
9
9
|
bindir: bin
|
|
10
10
|
cert_chain: []
|
|
11
|
-
date: 2026-03-
|
|
11
|
+
date: 2026-03-18 00:00:00.000000000 Z
|
|
12
12
|
dependencies:
|
|
13
13
|
- !ruby/object:Gem::Dependency
|
|
14
14
|
name: base64
|