openai 0.29.0 → 0.31.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +22 -0
- data/README.md +1 -1
- data/lib/openai/client.rb +4 -0
- data/lib/openai/internal/conversation_cursor_page.rb +2 -2
- data/lib/openai/models/all_models.rb +2 -0
- data/lib/openai/models/beta/assistant_create_params.rb +3 -0
- data/lib/openai/models/beta/assistant_update_params.rb +3 -0
- data/lib/openai/models/beta/chatkit/chat_session.rb +94 -0
- data/lib/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rb +22 -0
- data/lib/openai/models/beta/chatkit/chat_session_chatkit_configuration.rb +38 -0
- data/lib/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rb +125 -0
- data/lib/openai/models/beta/chatkit/chat_session_expires_after_param.rb +30 -0
- data/lib/openai/models/beta/chatkit/chat_session_file_upload.rb +38 -0
- data/lib/openai/models/beta/chatkit/chat_session_history.rb +34 -0
- data/lib/openai/models/beta/chatkit/chat_session_rate_limits.rb +22 -0
- data/lib/openai/models/beta/chatkit/chat_session_rate_limits_param.rb +22 -0
- data/lib/openai/models/beta/chatkit/chat_session_status.rb +20 -0
- data/lib/openai/models/beta/chatkit/chat_session_workflow_param.rb +80 -0
- data/lib/openai/models/beta/chatkit/chatkit_attachment.rb +69 -0
- data/lib/openai/models/beta/chatkit/chatkit_response_output_text.rb +143 -0
- data/lib/openai/models/beta/chatkit/chatkit_thread.rb +145 -0
- data/lib/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rb +65 -0
- data/lib/openai/models/beta/chatkit/chatkit_thread_item_list.rb +374 -0
- data/lib/openai/models/beta/chatkit/chatkit_thread_user_message_item.rb +183 -0
- data/lib/openai/models/beta/chatkit/chatkit_widget_item.rb +64 -0
- data/lib/openai/models/beta/chatkit/session_cancel_params.rb +18 -0
- data/lib/openai/models/beta/chatkit/session_create_params.rb +63 -0
- data/lib/openai/models/beta/chatkit/thread_delete_params.rb +18 -0
- data/lib/openai/models/beta/chatkit/thread_delete_response.rb +39 -0
- data/lib/openai/models/beta/chatkit/thread_list_items_params.rb +66 -0
- data/lib/openai/models/beta/chatkit/thread_list_params.rb +75 -0
- data/lib/openai/models/beta/chatkit/thread_retrieve_params.rb +18 -0
- data/lib/openai/models/beta/chatkit_upload_file_params.rb +28 -0
- data/lib/openai/models/beta/chatkit_upload_file_response.rb +25 -0
- data/lib/openai/models/beta/chatkit_workflow.rb +78 -0
- data/lib/openai/models/beta/file_part.rb +56 -0
- data/lib/openai/models/beta/image_part.rb +64 -0
- data/lib/openai/models/beta/threads/run_create_params.rb +3 -0
- data/lib/openai/models/chat/completion_create_params.rb +3 -0
- data/lib/openai/models/comparison_filter.rb +29 -6
- data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +3 -0
- data/lib/openai/models/evals/run_cancel_response.rb +6 -0
- data/lib/openai/models/evals/run_create_params.rb +6 -0
- data/lib/openai/models/evals/run_create_response.rb +6 -0
- data/lib/openai/models/evals/run_list_response.rb +6 -0
- data/lib/openai/models/evals/run_retrieve_response.rb +6 -0
- data/lib/openai/models/graders/score_model_grader.rb +3 -0
- data/lib/openai/models/image_edit_params.rb +4 -2
- data/lib/openai/models/image_model.rb +1 -0
- data/lib/openai/models/realtime/realtime_session.rb +4 -0
- data/lib/openai/models/realtime/realtime_session_create_request.rb +12 -0
- data/lib/openai/models/realtime/realtime_session_create_response.rb +12 -0
- data/lib/openai/models/reasoning.rb +3 -0
- data/lib/openai/models/reasoning_effort.rb +3 -0
- data/lib/openai/models/responses/tool.rb +5 -2
- data/lib/openai/models/responses_model.rb +2 -0
- data/lib/openai/models/vector_stores/vector_store_file.rb +3 -3
- data/lib/openai/models/video.rb +122 -0
- data/lib/openai/models/video_create_error.rb +21 -0
- data/lib/openai/models/video_create_params.rb +54 -0
- data/lib/openai/models/video_delete_params.rb +14 -0
- data/lib/openai/models/video_delete_response.rb +35 -0
- data/lib/openai/models/video_download_content_params.rb +34 -0
- data/lib/openai/models/video_list_params.rb +54 -0
- data/lib/openai/models/video_model.rb +15 -0
- data/lib/openai/models/video_remix_params.rb +22 -0
- data/lib/openai/models/video_retrieve_params.rb +14 -0
- data/lib/openai/models/video_seconds.rb +16 -0
- data/lib/openai/models/video_size.rb +17 -0
- data/lib/openai/models.rb +22 -0
- data/lib/openai/resources/beta/chatkit/sessions.rb +71 -0
- data/lib/openai/resources/beta/chatkit/threads.rb +126 -0
- data/lib/openai/resources/beta/chatkit.rb +50 -0
- data/lib/openai/resources/beta.rb +4 -0
- data/lib/openai/resources/files.rb +1 -1
- data/lib/openai/resources/videos.rb +165 -0
- data/lib/openai/version.rb +1 -1
- data/lib/openai.rb +46 -0
- data/rbi/openai/client.rbi +3 -0
- data/rbi/openai/models/all_models.rbi +10 -0
- data/rbi/openai/models/beta/assistant_create_params.rbi +6 -0
- data/rbi/openai/models/beta/assistant_update_params.rbi +6 -0
- data/rbi/openai/models/beta/chatkit/chat_session.rbi +141 -0
- data/rbi/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rbi +35 -0
- data/rbi/openai/models/beta/chatkit/chat_session_chatkit_configuration.rbi +87 -0
- data/rbi/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rbi +256 -0
- data/rbi/openai/models/beta/chatkit/chat_session_expires_after_param.rbi +43 -0
- data/rbi/openai/models/beta/chatkit/chat_session_file_upload.rbi +61 -0
- data/rbi/openai/models/beta/chatkit/chat_session_history.rbi +52 -0
- data/rbi/openai/models/beta/chatkit/chat_session_rate_limits.rbi +37 -0
- data/rbi/openai/models/beta/chatkit/chat_session_rate_limits_param.rbi +40 -0
- data/rbi/openai/models/beta/chatkit/chat_session_status.rbi +43 -0
- data/rbi/openai/models/beta/chatkit/chat_session_workflow_param.rbi +166 -0
- data/rbi/openai/models/beta/chatkit/chatkit_attachment.rbi +116 -0
- data/rbi/openai/models/beta/chatkit/chatkit_response_output_text.rbi +287 -0
- data/rbi/openai/models/beta/chatkit/chatkit_thread.rbi +220 -0
- data/rbi/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rbi +94 -0
- data/rbi/openai/models/beta/chatkit/chatkit_thread_item_list.rbi +590 -0
- data/rbi/openai/models/beta/chatkit/chatkit_thread_user_message_item.rbi +324 -0
- data/rbi/openai/models/beta/chatkit/chatkit_widget_item.rbi +87 -0
- data/rbi/openai/models/beta/chatkit/session_cancel_params.rbi +34 -0
- data/rbi/openai/models/beta/chatkit/session_create_params.rbi +136 -0
- data/rbi/openai/models/beta/chatkit/thread_delete_params.rbi +34 -0
- data/rbi/openai/models/beta/chatkit/thread_delete_response.rbi +55 -0
- data/rbi/openai/models/beta/chatkit/thread_list_items_params.rbi +138 -0
- data/rbi/openai/models/beta/chatkit/thread_list_params.rbi +145 -0
- data/rbi/openai/models/beta/chatkit/thread_retrieve_params.rbi +34 -0
- data/rbi/openai/models/beta/chatkit_upload_file_params.rbi +50 -0
- data/rbi/openai/models/beta/chatkit_upload_file_response.rbi +25 -0
- data/rbi/openai/models/beta/chatkit_workflow.rbi +134 -0
- data/rbi/openai/models/beta/file_part.rbi +74 -0
- data/rbi/openai/models/beta/image_part.rbi +82 -0
- data/rbi/openai/models/beta/threads/run_create_params.rbi +6 -0
- data/rbi/openai/models/chat/completion_create_params.rbi +6 -0
- data/rbi/openai/models/comparison_filter.rbi +43 -4
- data/rbi/openai/models/eval_create_response.rbi +4 -4
- data/rbi/openai/models/eval_list_response.rbi +4 -4
- data/rbi/openai/models/eval_retrieve_response.rbi +4 -4
- data/rbi/openai/models/eval_update_response.rbi +4 -4
- data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +6 -0
- data/rbi/openai/models/evals/run_cancel_response.rbi +12 -0
- data/rbi/openai/models/evals/run_create_params.rbi +12 -0
- data/rbi/openai/models/evals/run_create_response.rbi +12 -0
- data/rbi/openai/models/evals/run_list_response.rbi +12 -0
- data/rbi/openai/models/evals/run_retrieve_response.rbi +12 -0
- data/rbi/openai/models/graders/score_model_grader.rbi +6 -0
- data/rbi/openai/models/image_edit_params.rbi +6 -3
- data/rbi/openai/models/image_model.rbi +2 -0
- data/rbi/openai/models/realtime/realtime_session.rbi +20 -0
- data/rbi/openai/models/realtime/realtime_session_create_request.rbi +20 -0
- data/rbi/openai/models/realtime/realtime_session_create_response.rbi +20 -0
- data/rbi/openai/models/reasoning.rbi +6 -0
- data/rbi/openai/models/reasoning_effort.rbi +3 -0
- data/rbi/openai/models/responses/tool.rbi +11 -3
- data/rbi/openai/models/responses_model.rbi +10 -0
- data/rbi/openai/models/vector_stores/vector_store_file.rbi +3 -3
- data/rbi/openai/models/video.rbi +143 -0
- data/rbi/openai/models/video_create_error.rbi +26 -0
- data/rbi/openai/models/video_create_params.rbi +87 -0
- data/rbi/openai/models/video_delete_params.rbi +27 -0
- data/rbi/openai/models/video_delete_response.rbi +46 -0
- data/rbi/openai/models/video_download_content_params.rbi +89 -0
- data/rbi/openai/models/video_list_params.rbi +91 -0
- data/rbi/openai/models/video_model.rbi +19 -0
- data/rbi/openai/models/video_remix_params.rbi +40 -0
- data/rbi/openai/models/video_retrieve_params.rbi +27 -0
- data/rbi/openai/models/video_seconds.rbi +20 -0
- data/rbi/openai/models/video_size.rbi +23 -0
- data/rbi/openai/models.rbi +22 -0
- data/rbi/openai/resources/beta/assistants.rbi +6 -0
- data/rbi/openai/resources/beta/chatkit/sessions.rbi +61 -0
- data/rbi/openai/resources/beta/chatkit/threads.rbi +110 -0
- data/rbi/openai/resources/beta/chatkit.rbi +35 -0
- data/rbi/openai/resources/beta/threads/runs.rbi +6 -0
- data/rbi/openai/resources/beta.rbi +3 -0
- data/rbi/openai/resources/chat/completions.rbi +6 -0
- data/rbi/openai/resources/files.rbi +1 -1
- data/rbi/openai/resources/images.rbi +4 -2
- data/rbi/openai/resources/videos.rbi +121 -0
- data/sig/openai/client.rbs +2 -0
- data/sig/openai/models/all_models.rbs +4 -0
- data/sig/openai/models/beta/chatkit/chat_session.rbs +69 -0
- data/sig/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rbs +17 -0
- data/sig/openai/models/beta/chatkit/chat_session_chatkit_configuration.rbs +34 -0
- data/sig/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rbs +103 -0
- data/sig/openai/models/beta/chatkit/chat_session_expires_after_param.rbs +20 -0
- data/sig/openai/models/beta/chatkit/chat_session_file_upload.rbs +30 -0
- data/sig/openai/models/beta/chatkit/chat_session_history.rbs +19 -0
- data/sig/openai/models/beta/chatkit/chat_session_rate_limits.rbs +18 -0
- data/sig/openai/models/beta/chatkit/chat_session_rate_limits_param.rbs +20 -0
- data/sig/openai/models/beta/chatkit/chat_session_status.rbs +19 -0
- data/sig/openai/models/beta/chatkit/chat_session_workflow_param.rbs +69 -0
- data/sig/openai/models/beta/chatkit/chatkit_attachment.rbs +57 -0
- data/sig/openai/models/beta/chatkit/chatkit_response_output_text.rbs +114 -0
- data/sig/openai/models/beta/chatkit/chatkit_thread.rbs +96 -0
- data/sig/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rbs +51 -0
- data/sig/openai/models/beta/chatkit/chatkit_thread_item_list.rbs +276 -0
- data/sig/openai/models/beta/chatkit/chatkit_thread_user_message_item.rbs +127 -0
- data/sig/openai/models/beta/chatkit/chatkit_widget_item.rbs +51 -0
- data/sig/openai/models/beta/chatkit/session_cancel_params.rbs +19 -0
- data/sig/openai/models/beta/chatkit/session_create_params.rbs +62 -0
- data/sig/openai/models/beta/chatkit/thread_delete_params.rbs +19 -0
- data/sig/openai/models/beta/chatkit/thread_delete_response.rbs +30 -0
- data/sig/openai/models/beta/chatkit/thread_list_items_params.rbs +66 -0
- data/sig/openai/models/beta/chatkit/thread_list_params.rbs +73 -0
- data/sig/openai/models/beta/chatkit/thread_retrieve_params.rbs +19 -0
- data/sig/openai/models/beta/chatkit_upload_file_params.rbs +26 -0
- data/sig/openai/models/beta/chatkit_upload_file_response.rbs +14 -0
- data/sig/openai/models/beta/chatkit_workflow.rbs +55 -0
- data/sig/openai/models/beta/file_part.rbs +42 -0
- data/sig/openai/models/beta/image_part.rbs +47 -0
- data/sig/openai/models/comparison_filter.rbs +15 -1
- data/sig/openai/models/eval_create_response.rbs +2 -2
- data/sig/openai/models/eval_list_response.rbs +2 -2
- data/sig/openai/models/eval_retrieve_response.rbs +2 -2
- data/sig/openai/models/eval_update_response.rbs +2 -2
- data/sig/openai/models/image_model.rbs +3 -1
- data/sig/openai/models/realtime/realtime_session.rbs +8 -0
- data/sig/openai/models/realtime/realtime_session_create_request.rbs +8 -0
- data/sig/openai/models/realtime/realtime_session_create_response.rbs +8 -0
- data/sig/openai/models/responses/tool.rbs +2 -1
- data/sig/openai/models/responses_model.rbs +4 -0
- data/sig/openai/models/video.rbs +88 -0
- data/sig/openai/models/video_create_error.rbs +15 -0
- data/sig/openai/models/video_create_params.rbs +58 -0
- data/sig/openai/models/video_delete_params.rbs +14 -0
- data/sig/openai/models/video_delete_response.rbs +22 -0
- data/sig/openai/models/video_download_content_params.rbs +40 -0
- data/sig/openai/models/video_list_params.rbs +55 -0
- data/sig/openai/models/video_model.rbs +14 -0
- data/sig/openai/models/video_remix_params.rbs +23 -0
- data/sig/openai/models/video_retrieve_params.rbs +15 -0
- data/sig/openai/models/video_seconds.rbs +15 -0
- data/sig/openai/models/video_size.rbs +16 -0
- data/sig/openai/models.rbs +22 -0
- data/sig/openai/resources/beta/chatkit/sessions.rbs +25 -0
- data/sig/openai/resources/beta/chatkit/threads.rbs +39 -0
- data/sig/openai/resources/beta/chatkit.rbs +18 -0
- data/sig/openai/resources/beta.rbs +2 -0
- data/sig/openai/resources/videos.rbs +45 -0
- metadata +140 -2
@@ -230,6 +230,9 @@ module OpenAI
|
|
230
230
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
231
231
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
232
232
|
# response.
|
233
|
+
#
|
234
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
235
|
+
# effort.
|
233
236
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
234
237
|
attr_accessor :reasoning_effort
|
235
238
|
|
@@ -667,6 +670,9 @@ module OpenAI
|
|
667
670
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
668
671
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
669
672
|
# response.
|
673
|
+
#
|
674
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
675
|
+
# effort.
|
670
676
|
reasoning_effort: nil,
|
671
677
|
# An object specifying the format that the model must output.
|
672
678
|
#
|
@@ -12,7 +12,8 @@ module OpenAI
|
|
12
12
|
sig { returns(String) }
|
13
13
|
attr_accessor :key
|
14
14
|
|
15
|
-
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte
|
15
|
+
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
|
16
|
+
# `nin`.
|
16
17
|
#
|
17
18
|
# - `eq`: equals
|
18
19
|
# - `ne`: not equal
|
@@ -20,6 +21,8 @@ module OpenAI
|
|
20
21
|
# - `gte`: greater than or equal
|
21
22
|
# - `lt`: less than
|
22
23
|
# - `lte`: less than or equal
|
24
|
+
# - `in`: in
|
25
|
+
# - `nin`: not in
|
23
26
|
sig { returns(OpenAI::ComparisonFilter::Type::OrSymbol) }
|
24
27
|
attr_accessor :type
|
25
28
|
|
@@ -40,7 +43,8 @@ module OpenAI
|
|
40
43
|
def self.new(
|
41
44
|
# The key to compare against the value.
|
42
45
|
key:,
|
43
|
-
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte
|
46
|
+
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
|
47
|
+
# `nin`.
|
44
48
|
#
|
45
49
|
# - `eq`: equals
|
46
50
|
# - `ne`: not equal
|
@@ -48,6 +52,8 @@ module OpenAI
|
|
48
52
|
# - `gte`: greater than or equal
|
49
53
|
# - `lt`: less than
|
50
54
|
# - `lte`: less than or equal
|
55
|
+
# - `in`: in
|
56
|
+
# - `nin`: not in
|
51
57
|
type:,
|
52
58
|
# The value to compare against the attribute key; supports string, number, or
|
53
59
|
# boolean types.
|
@@ -67,7 +73,8 @@ module OpenAI
|
|
67
73
|
def to_hash
|
68
74
|
end
|
69
75
|
|
70
|
-
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte
|
76
|
+
# Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
|
77
|
+
# `nin`.
|
71
78
|
#
|
72
79
|
# - `eq`: equals
|
73
80
|
# - `ne`: not equal
|
@@ -75,6 +82,8 @@ module OpenAI
|
|
75
82
|
# - `gte`: greater than or equal
|
76
83
|
# - `lt`: less than
|
77
84
|
# - `lte`: less than or equal
|
85
|
+
# - `in`: in
|
86
|
+
# - `nin`: not in
|
78
87
|
module Type
|
79
88
|
extend OpenAI::Internal::Type::Enum
|
80
89
|
|
@@ -103,13 +112,43 @@ module OpenAI
|
|
103
112
|
module Value
|
104
113
|
extend OpenAI::Internal::Type::Union
|
105
114
|
|
106
|
-
Variants =
|
115
|
+
Variants =
|
116
|
+
T.type_alias do
|
117
|
+
T.any(
|
118
|
+
String,
|
119
|
+
Float,
|
120
|
+
T::Boolean,
|
121
|
+
T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants]
|
122
|
+
)
|
123
|
+
end
|
124
|
+
|
125
|
+
module UnionMember3
|
126
|
+
extend OpenAI::Internal::Type::Union
|
127
|
+
|
128
|
+
Variants = T.type_alias { T.any(String, Float) }
|
129
|
+
|
130
|
+
sig do
|
131
|
+
override.returns(
|
132
|
+
T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants]
|
133
|
+
)
|
134
|
+
end
|
135
|
+
def self.variants
|
136
|
+
end
|
137
|
+
end
|
107
138
|
|
108
139
|
sig do
|
109
140
|
override.returns(T::Array[OpenAI::ComparisonFilter::Value::Variants])
|
110
141
|
end
|
111
142
|
def self.variants
|
112
143
|
end
|
144
|
+
|
145
|
+
UnionMember3Array =
|
146
|
+
T.let(
|
147
|
+
OpenAI::Internal::Type::ArrayOf[
|
148
|
+
union: OpenAI::ComparisonFilter::Value::UnionMember3
|
149
|
+
],
|
150
|
+
OpenAI::Internal::Type::Converter
|
151
|
+
)
|
113
152
|
end
|
114
153
|
end
|
115
154
|
end
|
@@ -70,8 +70,8 @@ module OpenAI
|
|
70
70
|
testing_criteria:
|
71
71
|
T::Array[
|
72
72
|
T.any(
|
73
|
-
OpenAI::
|
74
|
-
OpenAI::
|
73
|
+
OpenAI::Graders::LabelModelGrader::OrHash,
|
74
|
+
OpenAI::Graders::StringCheckGrader::OrHash,
|
75
75
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
|
76
76
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython::OrHash,
|
77
77
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel::OrHash
|
@@ -223,8 +223,8 @@ module OpenAI
|
|
223
223
|
Variants =
|
224
224
|
T.type_alias do
|
225
225
|
T.any(
|
226
|
-
OpenAI::
|
227
|
-
OpenAI::
|
226
|
+
OpenAI::Graders::LabelModelGrader,
|
227
|
+
OpenAI::Graders::StringCheckGrader,
|
228
228
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity,
|
229
229
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython,
|
230
230
|
OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel
|
@@ -68,8 +68,8 @@ module OpenAI
|
|
68
68
|
testing_criteria:
|
69
69
|
T::Array[
|
70
70
|
T.any(
|
71
|
-
OpenAI::
|
72
|
-
OpenAI::
|
71
|
+
OpenAI::Graders::LabelModelGrader::OrHash,
|
72
|
+
OpenAI::Graders::StringCheckGrader::OrHash,
|
73
73
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
|
74
74
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython::OrHash,
|
75
75
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel::OrHash
|
@@ -221,8 +221,8 @@ module OpenAI
|
|
221
221
|
Variants =
|
222
222
|
T.type_alias do
|
223
223
|
T.any(
|
224
|
-
OpenAI::
|
225
|
-
OpenAI::
|
224
|
+
OpenAI::Graders::LabelModelGrader,
|
225
|
+
OpenAI::Graders::StringCheckGrader,
|
226
226
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity,
|
227
227
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython,
|
228
228
|
OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel
|
@@ -72,8 +72,8 @@ module OpenAI
|
|
72
72
|
testing_criteria:
|
73
73
|
T::Array[
|
74
74
|
T.any(
|
75
|
-
OpenAI::
|
76
|
-
OpenAI::
|
75
|
+
OpenAI::Graders::LabelModelGrader::OrHash,
|
76
|
+
OpenAI::Graders::StringCheckGrader::OrHash,
|
77
77
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
|
78
78
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython::OrHash,
|
79
79
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel::OrHash
|
@@ -225,8 +225,8 @@ module OpenAI
|
|
225
225
|
Variants =
|
226
226
|
T.type_alias do
|
227
227
|
T.any(
|
228
|
-
OpenAI::
|
229
|
-
OpenAI::
|
228
|
+
OpenAI::Graders::LabelModelGrader,
|
229
|
+
OpenAI::Graders::StringCheckGrader,
|
230
230
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity,
|
231
231
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython,
|
232
232
|
OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel
|
@@ -70,8 +70,8 @@ module OpenAI
|
|
70
70
|
testing_criteria:
|
71
71
|
T::Array[
|
72
72
|
T.any(
|
73
|
-
OpenAI::
|
74
|
-
OpenAI::
|
73
|
+
OpenAI::Graders::LabelModelGrader::OrHash,
|
74
|
+
OpenAI::Graders::StringCheckGrader::OrHash,
|
75
75
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
|
76
76
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython::OrHash,
|
77
77
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel::OrHash
|
@@ -223,8 +223,8 @@ module OpenAI
|
|
223
223
|
Variants =
|
224
224
|
T.type_alias do
|
225
225
|
T.any(
|
226
|
-
OpenAI::
|
227
|
-
OpenAI::
|
226
|
+
OpenAI::Graders::LabelModelGrader,
|
227
|
+
OpenAI::Graders::StringCheckGrader,
|
228
228
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity,
|
229
229
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython,
|
230
230
|
OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel
|
@@ -892,6 +892,9 @@ module OpenAI
|
|
892
892
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
893
893
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
894
894
|
# response.
|
895
|
+
#
|
896
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
897
|
+
# effort.
|
895
898
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
896
899
|
attr_accessor :reasoning_effort
|
897
900
|
|
@@ -992,6 +995,9 @@ module OpenAI
|
|
992
995
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
993
996
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
994
997
|
# response.
|
998
|
+
#
|
999
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1000
|
+
# effort.
|
995
1001
|
reasoning_effort: nil,
|
996
1002
|
# An object specifying the format that the model must output.
|
997
1003
|
#
|
@@ -515,6 +515,9 @@ module OpenAI
|
|
515
515
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
516
516
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
517
517
|
# response.
|
518
|
+
#
|
519
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
520
|
+
# effort.
|
518
521
|
sig do
|
519
522
|
returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
|
520
523
|
end
|
@@ -574,6 +577,9 @@ module OpenAI
|
|
574
577
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
575
578
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
576
579
|
# response.
|
580
|
+
#
|
581
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
582
|
+
# effort.
|
577
583
|
reasoning_effort: nil,
|
578
584
|
# Sampling temperature. This is a query parameter used to select responses.
|
579
585
|
temperature: nil,
|
@@ -1120,6 +1126,9 @@ module OpenAI
|
|
1120
1126
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1121
1127
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1122
1128
|
# response.
|
1129
|
+
#
|
1130
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1131
|
+
# effort.
|
1123
1132
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
|
1124
1133
|
attr_accessor :reasoning_effort
|
1125
1134
|
|
@@ -1241,6 +1250,9 @@ module OpenAI
|
|
1241
1250
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1242
1251
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1243
1252
|
# response.
|
1253
|
+
#
|
1254
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1255
|
+
# effort.
|
1244
1256
|
reasoning_effort: nil,
|
1245
1257
|
# A seed value to initialize the randomness, during sampling.
|
1246
1258
|
seed: nil,
|
@@ -425,6 +425,9 @@ module OpenAI
|
|
425
425
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
426
426
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
427
427
|
# response.
|
428
|
+
#
|
429
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
430
|
+
# effort.
|
428
431
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
429
432
|
attr_accessor :reasoning_effort
|
430
433
|
|
@@ -482,6 +485,9 @@ module OpenAI
|
|
482
485
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
483
486
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
484
487
|
# response.
|
488
|
+
#
|
489
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
490
|
+
# effort.
|
485
491
|
reasoning_effort: nil,
|
486
492
|
# Sampling temperature. This is a query parameter used to select responses.
|
487
493
|
temperature: nil,
|
@@ -1078,6 +1084,9 @@ module OpenAI
|
|
1078
1084
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1079
1085
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1080
1086
|
# response.
|
1087
|
+
#
|
1088
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1089
|
+
# effort.
|
1081
1090
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
1082
1091
|
attr_accessor :reasoning_effort
|
1083
1092
|
|
@@ -1216,6 +1225,9 @@ module OpenAI
|
|
1216
1225
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1217
1226
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1218
1227
|
# response.
|
1228
|
+
#
|
1229
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1230
|
+
# effort.
|
1219
1231
|
reasoning_effort: nil,
|
1220
1232
|
# A seed value to initialize the randomness, during sampling.
|
1221
1233
|
seed: nil,
|
@@ -515,6 +515,9 @@ module OpenAI
|
|
515
515
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
516
516
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
517
517
|
# response.
|
518
|
+
#
|
519
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
520
|
+
# effort.
|
518
521
|
sig do
|
519
522
|
returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
|
520
523
|
end
|
@@ -574,6 +577,9 @@ module OpenAI
|
|
574
577
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
575
578
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
576
579
|
# response.
|
580
|
+
#
|
581
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
582
|
+
# effort.
|
577
583
|
reasoning_effort: nil,
|
578
584
|
# Sampling temperature. This is a query parameter used to select responses.
|
579
585
|
temperature: nil,
|
@@ -1120,6 +1126,9 @@ module OpenAI
|
|
1120
1126
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1121
1127
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1122
1128
|
# response.
|
1129
|
+
#
|
1130
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1131
|
+
# effort.
|
1123
1132
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
|
1124
1133
|
attr_accessor :reasoning_effort
|
1125
1134
|
|
@@ -1241,6 +1250,9 @@ module OpenAI
|
|
1241
1250
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1242
1251
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1243
1252
|
# response.
|
1253
|
+
#
|
1254
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1255
|
+
# effort.
|
1244
1256
|
reasoning_effort: nil,
|
1245
1257
|
# A seed value to initialize the randomness, during sampling.
|
1246
1258
|
seed: nil,
|
@@ -511,6 +511,9 @@ module OpenAI
|
|
511
511
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
512
512
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
513
513
|
# response.
|
514
|
+
#
|
515
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
516
|
+
# effort.
|
514
517
|
sig do
|
515
518
|
returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
|
516
519
|
end
|
@@ -570,6 +573,9 @@ module OpenAI
|
|
570
573
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
571
574
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
572
575
|
# response.
|
576
|
+
#
|
577
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
578
|
+
# effort.
|
573
579
|
reasoning_effort: nil,
|
574
580
|
# Sampling temperature. This is a query parameter used to select responses.
|
575
581
|
temperature: nil,
|
@@ -1116,6 +1122,9 @@ module OpenAI
|
|
1116
1122
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1117
1123
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1118
1124
|
# response.
|
1125
|
+
#
|
1126
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1127
|
+
# effort.
|
1119
1128
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
|
1120
1129
|
attr_accessor :reasoning_effort
|
1121
1130
|
|
@@ -1237,6 +1246,9 @@ module OpenAI
|
|
1237
1246
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1238
1247
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1239
1248
|
# response.
|
1249
|
+
#
|
1250
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1251
|
+
# effort.
|
1240
1252
|
reasoning_effort: nil,
|
1241
1253
|
# A seed value to initialize the randomness, during sampling.
|
1242
1254
|
seed: nil,
|
@@ -517,6 +517,9 @@ module OpenAI
|
|
517
517
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
518
518
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
519
519
|
# response.
|
520
|
+
#
|
521
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
522
|
+
# effort.
|
520
523
|
sig do
|
521
524
|
returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
|
522
525
|
end
|
@@ -576,6 +579,9 @@ module OpenAI
|
|
576
579
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
577
580
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
578
581
|
# response.
|
582
|
+
#
|
583
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
584
|
+
# effort.
|
579
585
|
reasoning_effort: nil,
|
580
586
|
# Sampling temperature. This is a query parameter used to select responses.
|
581
587
|
temperature: nil,
|
@@ -1122,6 +1128,9 @@ module OpenAI
|
|
1122
1128
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1123
1129
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1124
1130
|
# response.
|
1131
|
+
#
|
1132
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1133
|
+
# effort.
|
1125
1134
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
|
1126
1135
|
attr_accessor :reasoning_effort
|
1127
1136
|
|
@@ -1243,6 +1252,9 @@ module OpenAI
|
|
1243
1252
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
1244
1253
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
1245
1254
|
# response.
|
1255
|
+
#
|
1256
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
1257
|
+
# effort.
|
1246
1258
|
reasoning_effort: nil,
|
1247
1259
|
# A seed value to initialize the randomness, during sampling.
|
1248
1260
|
seed: nil,
|
@@ -399,6 +399,9 @@ module OpenAI
|
|
399
399
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
400
400
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
401
401
|
# response.
|
402
|
+
#
|
403
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
404
|
+
# effort.
|
402
405
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
403
406
|
attr_accessor :reasoning_effort
|
404
407
|
|
@@ -432,6 +435,9 @@ module OpenAI
|
|
432
435
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
433
436
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
434
437
|
# response.
|
438
|
+
#
|
439
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
440
|
+
# effort.
|
435
441
|
reasoning_effort: nil,
|
436
442
|
# A seed value to initialize the randomness, during sampling.
|
437
443
|
seed: nil,
|
@@ -38,7 +38,8 @@ module OpenAI
|
|
38
38
|
|
39
39
|
# Control how much effort the model will exert to match the style and features,
|
40
40
|
# especially facial features, of input images. This parameter is only supported
|
41
|
-
# for `gpt-image-1`. Supports `high` and
|
41
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
42
|
+
# `low`. Defaults to `low`.
|
42
43
|
sig do
|
43
44
|
returns(T.nilable(OpenAI::ImageEditParams::InputFidelity::OrSymbol))
|
44
45
|
end
|
@@ -161,7 +162,8 @@ module OpenAI
|
|
161
162
|
background: nil,
|
162
163
|
# Control how much effort the model will exert to match the style and features,
|
163
164
|
# especially facial features, of input images. This parameter is only supported
|
164
|
-
# for `gpt-image-1`. Supports `high` and
|
165
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
166
|
+
# `low`. Defaults to `low`.
|
165
167
|
input_fidelity: nil,
|
166
168
|
# An additional image whose fully transparent areas (e.g. where alpha is zero)
|
167
169
|
# indicate where `image` should be edited. If there are multiple images provided,
|
@@ -294,7 +296,8 @@ module OpenAI
|
|
294
296
|
|
295
297
|
# Control how much effort the model will exert to match the style and features,
|
296
298
|
# especially facial features, of input images. This parameter is only supported
|
297
|
-
# for `gpt-image-1`. Supports `high` and
|
299
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
300
|
+
# `low`. Defaults to `low`.
|
298
301
|
module InputFidelity
|
299
302
|
extend OpenAI::Internal::Type::Enum
|
300
303
|
|
@@ -11,6 +11,8 @@ module OpenAI
|
|
11
11
|
DALL_E_2 = T.let(:"dall-e-2", OpenAI::ImageModel::TaggedSymbol)
|
12
12
|
DALL_E_3 = T.let(:"dall-e-3", OpenAI::ImageModel::TaggedSymbol)
|
13
13
|
GPT_IMAGE_1 = T.let(:"gpt-image-1", OpenAI::ImageModel::TaggedSymbol)
|
14
|
+
GPT_IMAGE_1_MINI =
|
15
|
+
T.let(:"gpt-image-1-mini", OpenAI::ImageModel::TaggedSymbol)
|
14
16
|
|
15
17
|
sig { override.returns(T::Array[OpenAI::ImageModel::TaggedSymbol]) }
|
16
18
|
def self.values
|
@@ -707,6 +707,26 @@ module OpenAI
|
|
707
707
|
:"gpt-4o-mini-realtime-preview-2024-12-17",
|
708
708
|
OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
|
709
709
|
)
|
710
|
+
GPT_REALTIME_MINI =
|
711
|
+
T.let(
|
712
|
+
:"gpt-realtime-mini",
|
713
|
+
OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
|
714
|
+
)
|
715
|
+
GPT_REALTIME_MINI_2025_10_06 =
|
716
|
+
T.let(
|
717
|
+
:"gpt-realtime-mini-2025-10-06",
|
718
|
+
OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
|
719
|
+
)
|
720
|
+
GPT_AUDIO_MINI =
|
721
|
+
T.let(
|
722
|
+
:"gpt-audio-mini",
|
723
|
+
OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
|
724
|
+
)
|
725
|
+
GPT_AUDIO_MINI_2025_10_06 =
|
726
|
+
T.let(
|
727
|
+
:"gpt-audio-mini-2025-10-06",
|
728
|
+
OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
|
729
|
+
)
|
710
730
|
|
711
731
|
sig do
|
712
732
|
override.returns(
|
@@ -508,6 +508,26 @@ module OpenAI
|
|
508
508
|
:"gpt-4o-mini-realtime-preview-2024-12-17",
|
509
509
|
OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
|
510
510
|
)
|
511
|
+
GPT_REALTIME_MINI =
|
512
|
+
T.let(
|
513
|
+
:"gpt-realtime-mini",
|
514
|
+
OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
|
515
|
+
)
|
516
|
+
GPT_REALTIME_MINI_2025_10_06 =
|
517
|
+
T.let(
|
518
|
+
:"gpt-realtime-mini-2025-10-06",
|
519
|
+
OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
|
520
|
+
)
|
521
|
+
GPT_AUDIO_MINI =
|
522
|
+
T.let(
|
523
|
+
:"gpt-audio-mini",
|
524
|
+
OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
|
525
|
+
)
|
526
|
+
GPT_AUDIO_MINI_2025_10_06 =
|
527
|
+
T.let(
|
528
|
+
:"gpt-audio-mini-2025-10-06",
|
529
|
+
OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
|
530
|
+
)
|
511
531
|
end
|
512
532
|
|
513
533
|
module OutputModality
|
@@ -1308,6 +1308,26 @@ module OpenAI
|
|
1308
1308
|
:"gpt-4o-mini-realtime-preview-2024-12-17",
|
1309
1309
|
OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
|
1310
1310
|
)
|
1311
|
+
GPT_REALTIME_MINI =
|
1312
|
+
T.let(
|
1313
|
+
:"gpt-realtime-mini",
|
1314
|
+
OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
|
1315
|
+
)
|
1316
|
+
GPT_REALTIME_MINI_2025_10_06 =
|
1317
|
+
T.let(
|
1318
|
+
:"gpt-realtime-mini-2025-10-06",
|
1319
|
+
OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
|
1320
|
+
)
|
1321
|
+
GPT_AUDIO_MINI =
|
1322
|
+
T.let(
|
1323
|
+
:"gpt-audio-mini",
|
1324
|
+
OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
|
1325
|
+
)
|
1326
|
+
GPT_AUDIO_MINI_2025_10_06 =
|
1327
|
+
T.let(
|
1328
|
+
:"gpt-audio-mini-2025-10-06",
|
1329
|
+
OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
|
1330
|
+
)
|
1311
1331
|
end
|
1312
1332
|
|
1313
1333
|
module OutputModality
|
@@ -11,6 +11,9 @@ module OpenAI
|
|
11
11
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
12
12
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
13
13
|
# response.
|
14
|
+
#
|
15
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
16
|
+
# effort.
|
14
17
|
sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
|
15
18
|
attr_accessor :effort
|
16
19
|
|
@@ -46,6 +49,9 @@ module OpenAI
|
|
46
49
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
47
50
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
48
51
|
# response.
|
52
|
+
#
|
53
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
54
|
+
# effort.
|
49
55
|
effort: nil,
|
50
56
|
# **Deprecated:** use `summary` instead.
|
51
57
|
#
|
@@ -7,6 +7,9 @@ module OpenAI
|
|
7
7
|
# supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
|
8
8
|
# effort can result in faster responses and fewer tokens used on reasoning in a
|
9
9
|
# response.
|
10
|
+
#
|
11
|
+
# Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
|
12
|
+
# effort.
|
10
13
|
module ReasoningEffort
|
11
14
|
extend OpenAI::Internal::Type::Enum
|
12
15
|
|
@@ -781,7 +781,8 @@ module OpenAI
|
|
781
781
|
|
782
782
|
# Control how much effort the model will exert to match the style and features,
|
783
783
|
# especially facial features, of input images. This parameter is only supported
|
784
|
-
# for `gpt-image-1`. Supports `high` and
|
784
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
785
|
+
# `low`. Defaults to `low`.
|
785
786
|
sig do
|
786
787
|
returns(
|
787
788
|
T.nilable(
|
@@ -946,7 +947,8 @@ module OpenAI
|
|
946
947
|
background: nil,
|
947
948
|
# Control how much effort the model will exert to match the style and features,
|
948
949
|
# especially facial features, of input images. This parameter is only supported
|
949
|
-
# for `gpt-image-1`. Supports `high` and
|
950
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
951
|
+
# `low`. Defaults to `low`.
|
950
952
|
input_fidelity: nil,
|
951
953
|
# Optional mask for inpainting. Contains `image_url` (string, optional) and
|
952
954
|
# `file_id` (string, optional).
|
@@ -1046,7 +1048,8 @@ module OpenAI
|
|
1046
1048
|
|
1047
1049
|
# Control how much effort the model will exert to match the style and features,
|
1048
1050
|
# especially facial features, of input images. This parameter is only supported
|
1049
|
-
# for `gpt-image-1`. Supports `high` and
|
1051
|
+
# for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
|
1052
|
+
# `low`. Defaults to `low`.
|
1050
1053
|
module InputFidelity
|
1051
1054
|
extend OpenAI::Internal::Type::Enum
|
1052
1055
|
|
@@ -1139,6 +1142,11 @@ module OpenAI
|
|
1139
1142
|
:"gpt-image-1",
|
1140
1143
|
OpenAI::Responses::Tool::ImageGeneration::Model::TaggedSymbol
|
1141
1144
|
)
|
1145
|
+
GPT_IMAGE_1_MINI =
|
1146
|
+
T.let(
|
1147
|
+
:"gpt-image-1-mini",
|
1148
|
+
OpenAI::Responses::Tool::ImageGeneration::Model::TaggedSymbol
|
1149
|
+
)
|
1142
1150
|
|
1143
1151
|
sig do
|
1144
1152
|
override.returns(
|