openai 0.29.0 → 0.31.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (222) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +22 -0
  3. data/README.md +1 -1
  4. data/lib/openai/client.rb +4 -0
  5. data/lib/openai/internal/conversation_cursor_page.rb +2 -2
  6. data/lib/openai/models/all_models.rb +2 -0
  7. data/lib/openai/models/beta/assistant_create_params.rb +3 -0
  8. data/lib/openai/models/beta/assistant_update_params.rb +3 -0
  9. data/lib/openai/models/beta/chatkit/chat_session.rb +94 -0
  10. data/lib/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rb +22 -0
  11. data/lib/openai/models/beta/chatkit/chat_session_chatkit_configuration.rb +38 -0
  12. data/lib/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rb +125 -0
  13. data/lib/openai/models/beta/chatkit/chat_session_expires_after_param.rb +30 -0
  14. data/lib/openai/models/beta/chatkit/chat_session_file_upload.rb +38 -0
  15. data/lib/openai/models/beta/chatkit/chat_session_history.rb +34 -0
  16. data/lib/openai/models/beta/chatkit/chat_session_rate_limits.rb +22 -0
  17. data/lib/openai/models/beta/chatkit/chat_session_rate_limits_param.rb +22 -0
  18. data/lib/openai/models/beta/chatkit/chat_session_status.rb +20 -0
  19. data/lib/openai/models/beta/chatkit/chat_session_workflow_param.rb +80 -0
  20. data/lib/openai/models/beta/chatkit/chatkit_attachment.rb +69 -0
  21. data/lib/openai/models/beta/chatkit/chatkit_response_output_text.rb +143 -0
  22. data/lib/openai/models/beta/chatkit/chatkit_thread.rb +145 -0
  23. data/lib/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rb +65 -0
  24. data/lib/openai/models/beta/chatkit/chatkit_thread_item_list.rb +374 -0
  25. data/lib/openai/models/beta/chatkit/chatkit_thread_user_message_item.rb +183 -0
  26. data/lib/openai/models/beta/chatkit/chatkit_widget_item.rb +64 -0
  27. data/lib/openai/models/beta/chatkit/session_cancel_params.rb +18 -0
  28. data/lib/openai/models/beta/chatkit/session_create_params.rb +63 -0
  29. data/lib/openai/models/beta/chatkit/thread_delete_params.rb +18 -0
  30. data/lib/openai/models/beta/chatkit/thread_delete_response.rb +39 -0
  31. data/lib/openai/models/beta/chatkit/thread_list_items_params.rb +66 -0
  32. data/lib/openai/models/beta/chatkit/thread_list_params.rb +75 -0
  33. data/lib/openai/models/beta/chatkit/thread_retrieve_params.rb +18 -0
  34. data/lib/openai/models/beta/chatkit_upload_file_params.rb +28 -0
  35. data/lib/openai/models/beta/chatkit_upload_file_response.rb +25 -0
  36. data/lib/openai/models/beta/chatkit_workflow.rb +78 -0
  37. data/lib/openai/models/beta/file_part.rb +56 -0
  38. data/lib/openai/models/beta/image_part.rb +64 -0
  39. data/lib/openai/models/beta/threads/run_create_params.rb +3 -0
  40. data/lib/openai/models/chat/completion_create_params.rb +3 -0
  41. data/lib/openai/models/comparison_filter.rb +29 -6
  42. data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +3 -0
  43. data/lib/openai/models/evals/run_cancel_response.rb +6 -0
  44. data/lib/openai/models/evals/run_create_params.rb +6 -0
  45. data/lib/openai/models/evals/run_create_response.rb +6 -0
  46. data/lib/openai/models/evals/run_list_response.rb +6 -0
  47. data/lib/openai/models/evals/run_retrieve_response.rb +6 -0
  48. data/lib/openai/models/graders/score_model_grader.rb +3 -0
  49. data/lib/openai/models/image_edit_params.rb +4 -2
  50. data/lib/openai/models/image_model.rb +1 -0
  51. data/lib/openai/models/realtime/realtime_session.rb +4 -0
  52. data/lib/openai/models/realtime/realtime_session_create_request.rb +12 -0
  53. data/lib/openai/models/realtime/realtime_session_create_response.rb +12 -0
  54. data/lib/openai/models/reasoning.rb +3 -0
  55. data/lib/openai/models/reasoning_effort.rb +3 -0
  56. data/lib/openai/models/responses/tool.rb +5 -2
  57. data/lib/openai/models/responses_model.rb +2 -0
  58. data/lib/openai/models/vector_stores/vector_store_file.rb +3 -3
  59. data/lib/openai/models/video.rb +122 -0
  60. data/lib/openai/models/video_create_error.rb +21 -0
  61. data/lib/openai/models/video_create_params.rb +54 -0
  62. data/lib/openai/models/video_delete_params.rb +14 -0
  63. data/lib/openai/models/video_delete_response.rb +35 -0
  64. data/lib/openai/models/video_download_content_params.rb +34 -0
  65. data/lib/openai/models/video_list_params.rb +54 -0
  66. data/lib/openai/models/video_model.rb +15 -0
  67. data/lib/openai/models/video_remix_params.rb +22 -0
  68. data/lib/openai/models/video_retrieve_params.rb +14 -0
  69. data/lib/openai/models/video_seconds.rb +16 -0
  70. data/lib/openai/models/video_size.rb +17 -0
  71. data/lib/openai/models.rb +22 -0
  72. data/lib/openai/resources/beta/chatkit/sessions.rb +71 -0
  73. data/lib/openai/resources/beta/chatkit/threads.rb +126 -0
  74. data/lib/openai/resources/beta/chatkit.rb +50 -0
  75. data/lib/openai/resources/beta.rb +4 -0
  76. data/lib/openai/resources/files.rb +1 -1
  77. data/lib/openai/resources/videos.rb +165 -0
  78. data/lib/openai/version.rb +1 -1
  79. data/lib/openai.rb +46 -0
  80. data/rbi/openai/client.rbi +3 -0
  81. data/rbi/openai/models/all_models.rbi +10 -0
  82. data/rbi/openai/models/beta/assistant_create_params.rbi +6 -0
  83. data/rbi/openai/models/beta/assistant_update_params.rbi +6 -0
  84. data/rbi/openai/models/beta/chatkit/chat_session.rbi +141 -0
  85. data/rbi/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rbi +35 -0
  86. data/rbi/openai/models/beta/chatkit/chat_session_chatkit_configuration.rbi +87 -0
  87. data/rbi/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rbi +256 -0
  88. data/rbi/openai/models/beta/chatkit/chat_session_expires_after_param.rbi +43 -0
  89. data/rbi/openai/models/beta/chatkit/chat_session_file_upload.rbi +61 -0
  90. data/rbi/openai/models/beta/chatkit/chat_session_history.rbi +52 -0
  91. data/rbi/openai/models/beta/chatkit/chat_session_rate_limits.rbi +37 -0
  92. data/rbi/openai/models/beta/chatkit/chat_session_rate_limits_param.rbi +40 -0
  93. data/rbi/openai/models/beta/chatkit/chat_session_status.rbi +43 -0
  94. data/rbi/openai/models/beta/chatkit/chat_session_workflow_param.rbi +166 -0
  95. data/rbi/openai/models/beta/chatkit/chatkit_attachment.rbi +116 -0
  96. data/rbi/openai/models/beta/chatkit/chatkit_response_output_text.rbi +287 -0
  97. data/rbi/openai/models/beta/chatkit/chatkit_thread.rbi +220 -0
  98. data/rbi/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rbi +94 -0
  99. data/rbi/openai/models/beta/chatkit/chatkit_thread_item_list.rbi +590 -0
  100. data/rbi/openai/models/beta/chatkit/chatkit_thread_user_message_item.rbi +324 -0
  101. data/rbi/openai/models/beta/chatkit/chatkit_widget_item.rbi +87 -0
  102. data/rbi/openai/models/beta/chatkit/session_cancel_params.rbi +34 -0
  103. data/rbi/openai/models/beta/chatkit/session_create_params.rbi +136 -0
  104. data/rbi/openai/models/beta/chatkit/thread_delete_params.rbi +34 -0
  105. data/rbi/openai/models/beta/chatkit/thread_delete_response.rbi +55 -0
  106. data/rbi/openai/models/beta/chatkit/thread_list_items_params.rbi +138 -0
  107. data/rbi/openai/models/beta/chatkit/thread_list_params.rbi +145 -0
  108. data/rbi/openai/models/beta/chatkit/thread_retrieve_params.rbi +34 -0
  109. data/rbi/openai/models/beta/chatkit_upload_file_params.rbi +50 -0
  110. data/rbi/openai/models/beta/chatkit_upload_file_response.rbi +25 -0
  111. data/rbi/openai/models/beta/chatkit_workflow.rbi +134 -0
  112. data/rbi/openai/models/beta/file_part.rbi +74 -0
  113. data/rbi/openai/models/beta/image_part.rbi +82 -0
  114. data/rbi/openai/models/beta/threads/run_create_params.rbi +6 -0
  115. data/rbi/openai/models/chat/completion_create_params.rbi +6 -0
  116. data/rbi/openai/models/comparison_filter.rbi +43 -4
  117. data/rbi/openai/models/eval_create_response.rbi +4 -4
  118. data/rbi/openai/models/eval_list_response.rbi +4 -4
  119. data/rbi/openai/models/eval_retrieve_response.rbi +4 -4
  120. data/rbi/openai/models/eval_update_response.rbi +4 -4
  121. data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +6 -0
  122. data/rbi/openai/models/evals/run_cancel_response.rbi +12 -0
  123. data/rbi/openai/models/evals/run_create_params.rbi +12 -0
  124. data/rbi/openai/models/evals/run_create_response.rbi +12 -0
  125. data/rbi/openai/models/evals/run_list_response.rbi +12 -0
  126. data/rbi/openai/models/evals/run_retrieve_response.rbi +12 -0
  127. data/rbi/openai/models/graders/score_model_grader.rbi +6 -0
  128. data/rbi/openai/models/image_edit_params.rbi +6 -3
  129. data/rbi/openai/models/image_model.rbi +2 -0
  130. data/rbi/openai/models/realtime/realtime_session.rbi +20 -0
  131. data/rbi/openai/models/realtime/realtime_session_create_request.rbi +20 -0
  132. data/rbi/openai/models/realtime/realtime_session_create_response.rbi +20 -0
  133. data/rbi/openai/models/reasoning.rbi +6 -0
  134. data/rbi/openai/models/reasoning_effort.rbi +3 -0
  135. data/rbi/openai/models/responses/tool.rbi +11 -3
  136. data/rbi/openai/models/responses_model.rbi +10 -0
  137. data/rbi/openai/models/vector_stores/vector_store_file.rbi +3 -3
  138. data/rbi/openai/models/video.rbi +143 -0
  139. data/rbi/openai/models/video_create_error.rbi +26 -0
  140. data/rbi/openai/models/video_create_params.rbi +87 -0
  141. data/rbi/openai/models/video_delete_params.rbi +27 -0
  142. data/rbi/openai/models/video_delete_response.rbi +46 -0
  143. data/rbi/openai/models/video_download_content_params.rbi +89 -0
  144. data/rbi/openai/models/video_list_params.rbi +91 -0
  145. data/rbi/openai/models/video_model.rbi +19 -0
  146. data/rbi/openai/models/video_remix_params.rbi +40 -0
  147. data/rbi/openai/models/video_retrieve_params.rbi +27 -0
  148. data/rbi/openai/models/video_seconds.rbi +20 -0
  149. data/rbi/openai/models/video_size.rbi +23 -0
  150. data/rbi/openai/models.rbi +22 -0
  151. data/rbi/openai/resources/beta/assistants.rbi +6 -0
  152. data/rbi/openai/resources/beta/chatkit/sessions.rbi +61 -0
  153. data/rbi/openai/resources/beta/chatkit/threads.rbi +110 -0
  154. data/rbi/openai/resources/beta/chatkit.rbi +35 -0
  155. data/rbi/openai/resources/beta/threads/runs.rbi +6 -0
  156. data/rbi/openai/resources/beta.rbi +3 -0
  157. data/rbi/openai/resources/chat/completions.rbi +6 -0
  158. data/rbi/openai/resources/files.rbi +1 -1
  159. data/rbi/openai/resources/images.rbi +4 -2
  160. data/rbi/openai/resources/videos.rbi +121 -0
  161. data/sig/openai/client.rbs +2 -0
  162. data/sig/openai/models/all_models.rbs +4 -0
  163. data/sig/openai/models/beta/chatkit/chat_session.rbs +69 -0
  164. data/sig/openai/models/beta/chatkit/chat_session_automatic_thread_titling.rbs +17 -0
  165. data/sig/openai/models/beta/chatkit/chat_session_chatkit_configuration.rbs +34 -0
  166. data/sig/openai/models/beta/chatkit/chat_session_chatkit_configuration_param.rbs +103 -0
  167. data/sig/openai/models/beta/chatkit/chat_session_expires_after_param.rbs +20 -0
  168. data/sig/openai/models/beta/chatkit/chat_session_file_upload.rbs +30 -0
  169. data/sig/openai/models/beta/chatkit/chat_session_history.rbs +19 -0
  170. data/sig/openai/models/beta/chatkit/chat_session_rate_limits.rbs +18 -0
  171. data/sig/openai/models/beta/chatkit/chat_session_rate_limits_param.rbs +20 -0
  172. data/sig/openai/models/beta/chatkit/chat_session_status.rbs +19 -0
  173. data/sig/openai/models/beta/chatkit/chat_session_workflow_param.rbs +69 -0
  174. data/sig/openai/models/beta/chatkit/chatkit_attachment.rbs +57 -0
  175. data/sig/openai/models/beta/chatkit/chatkit_response_output_text.rbs +114 -0
  176. data/sig/openai/models/beta/chatkit/chatkit_thread.rbs +96 -0
  177. data/sig/openai/models/beta/chatkit/chatkit_thread_assistant_message_item.rbs +51 -0
  178. data/sig/openai/models/beta/chatkit/chatkit_thread_item_list.rbs +276 -0
  179. data/sig/openai/models/beta/chatkit/chatkit_thread_user_message_item.rbs +127 -0
  180. data/sig/openai/models/beta/chatkit/chatkit_widget_item.rbs +51 -0
  181. data/sig/openai/models/beta/chatkit/session_cancel_params.rbs +19 -0
  182. data/sig/openai/models/beta/chatkit/session_create_params.rbs +62 -0
  183. data/sig/openai/models/beta/chatkit/thread_delete_params.rbs +19 -0
  184. data/sig/openai/models/beta/chatkit/thread_delete_response.rbs +30 -0
  185. data/sig/openai/models/beta/chatkit/thread_list_items_params.rbs +66 -0
  186. data/sig/openai/models/beta/chatkit/thread_list_params.rbs +73 -0
  187. data/sig/openai/models/beta/chatkit/thread_retrieve_params.rbs +19 -0
  188. data/sig/openai/models/beta/chatkit_upload_file_params.rbs +26 -0
  189. data/sig/openai/models/beta/chatkit_upload_file_response.rbs +14 -0
  190. data/sig/openai/models/beta/chatkit_workflow.rbs +55 -0
  191. data/sig/openai/models/beta/file_part.rbs +42 -0
  192. data/sig/openai/models/beta/image_part.rbs +47 -0
  193. data/sig/openai/models/comparison_filter.rbs +15 -1
  194. data/sig/openai/models/eval_create_response.rbs +2 -2
  195. data/sig/openai/models/eval_list_response.rbs +2 -2
  196. data/sig/openai/models/eval_retrieve_response.rbs +2 -2
  197. data/sig/openai/models/eval_update_response.rbs +2 -2
  198. data/sig/openai/models/image_model.rbs +3 -1
  199. data/sig/openai/models/realtime/realtime_session.rbs +8 -0
  200. data/sig/openai/models/realtime/realtime_session_create_request.rbs +8 -0
  201. data/sig/openai/models/realtime/realtime_session_create_response.rbs +8 -0
  202. data/sig/openai/models/responses/tool.rbs +2 -1
  203. data/sig/openai/models/responses_model.rbs +4 -0
  204. data/sig/openai/models/video.rbs +88 -0
  205. data/sig/openai/models/video_create_error.rbs +15 -0
  206. data/sig/openai/models/video_create_params.rbs +58 -0
  207. data/sig/openai/models/video_delete_params.rbs +14 -0
  208. data/sig/openai/models/video_delete_response.rbs +22 -0
  209. data/sig/openai/models/video_download_content_params.rbs +40 -0
  210. data/sig/openai/models/video_list_params.rbs +55 -0
  211. data/sig/openai/models/video_model.rbs +14 -0
  212. data/sig/openai/models/video_remix_params.rbs +23 -0
  213. data/sig/openai/models/video_retrieve_params.rbs +15 -0
  214. data/sig/openai/models/video_seconds.rbs +15 -0
  215. data/sig/openai/models/video_size.rbs +16 -0
  216. data/sig/openai/models.rbs +22 -0
  217. data/sig/openai/resources/beta/chatkit/sessions.rbs +25 -0
  218. data/sig/openai/resources/beta/chatkit/threads.rbs +39 -0
  219. data/sig/openai/resources/beta/chatkit.rbs +18 -0
  220. data/sig/openai/resources/beta.rbs +2 -0
  221. data/sig/openai/resources/videos.rbs +45 -0
  222. metadata +140 -2
@@ -230,6 +230,9 @@ module OpenAI
230
230
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
231
231
  # effort can result in faster responses and fewer tokens used on reasoning in a
232
232
  # response.
233
+ #
234
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
235
+ # effort.
233
236
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
234
237
  attr_accessor :reasoning_effort
235
238
 
@@ -667,6 +670,9 @@ module OpenAI
667
670
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
668
671
  # effort can result in faster responses and fewer tokens used on reasoning in a
669
672
  # response.
673
+ #
674
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
675
+ # effort.
670
676
  reasoning_effort: nil,
671
677
  # An object specifying the format that the model must output.
672
678
  #
@@ -12,7 +12,8 @@ module OpenAI
12
12
  sig { returns(String) }
13
13
  attr_accessor :key
14
14
 
15
- # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`.
15
+ # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
16
+ # `nin`.
16
17
  #
17
18
  # - `eq`: equals
18
19
  # - `ne`: not equal
@@ -20,6 +21,8 @@ module OpenAI
20
21
  # - `gte`: greater than or equal
21
22
  # - `lt`: less than
22
23
  # - `lte`: less than or equal
24
+ # - `in`: in
25
+ # - `nin`: not in
23
26
  sig { returns(OpenAI::ComparisonFilter::Type::OrSymbol) }
24
27
  attr_accessor :type
25
28
 
@@ -40,7 +43,8 @@ module OpenAI
40
43
  def self.new(
41
44
  # The key to compare against the value.
42
45
  key:,
43
- # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`.
46
+ # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
47
+ # `nin`.
44
48
  #
45
49
  # - `eq`: equals
46
50
  # - `ne`: not equal
@@ -48,6 +52,8 @@ module OpenAI
48
52
  # - `gte`: greater than or equal
49
53
  # - `lt`: less than
50
54
  # - `lte`: less than or equal
55
+ # - `in`: in
56
+ # - `nin`: not in
51
57
  type:,
52
58
  # The value to compare against the attribute key; supports string, number, or
53
59
  # boolean types.
@@ -67,7 +73,8 @@ module OpenAI
67
73
  def to_hash
68
74
  end
69
75
 
70
- # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`.
76
+ # Specifies the comparison operator: `eq`, `ne`, `gt`, `gte`, `lt`, `lte`, `in`,
77
+ # `nin`.
71
78
  #
72
79
  # - `eq`: equals
73
80
  # - `ne`: not equal
@@ -75,6 +82,8 @@ module OpenAI
75
82
  # - `gte`: greater than or equal
76
83
  # - `lt`: less than
77
84
  # - `lte`: less than or equal
85
+ # - `in`: in
86
+ # - `nin`: not in
78
87
  module Type
79
88
  extend OpenAI::Internal::Type::Enum
80
89
 
@@ -103,13 +112,43 @@ module OpenAI
103
112
  module Value
104
113
  extend OpenAI::Internal::Type::Union
105
114
 
106
- Variants = T.type_alias { T.any(String, Float, T::Boolean) }
115
+ Variants =
116
+ T.type_alias do
117
+ T.any(
118
+ String,
119
+ Float,
120
+ T::Boolean,
121
+ T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants]
122
+ )
123
+ end
124
+
125
+ module UnionMember3
126
+ extend OpenAI::Internal::Type::Union
127
+
128
+ Variants = T.type_alias { T.any(String, Float) }
129
+
130
+ sig do
131
+ override.returns(
132
+ T::Array[OpenAI::ComparisonFilter::Value::UnionMember3::Variants]
133
+ )
134
+ end
135
+ def self.variants
136
+ end
137
+ end
107
138
 
108
139
  sig do
109
140
  override.returns(T::Array[OpenAI::ComparisonFilter::Value::Variants])
110
141
  end
111
142
  def self.variants
112
143
  end
144
+
145
+ UnionMember3Array =
146
+ T.let(
147
+ OpenAI::Internal::Type::ArrayOf[
148
+ union: OpenAI::ComparisonFilter::Value::UnionMember3
149
+ ],
150
+ OpenAI::Internal::Type::Converter
151
+ )
113
152
  end
114
153
  end
115
154
  end
@@ -70,8 +70,8 @@ module OpenAI
70
70
  testing_criteria:
71
71
  T::Array[
72
72
  T.any(
73
- OpenAI::Models::Graders::LabelModelGrader::OrHash,
74
- OpenAI::Models::Graders::StringCheckGrader::OrHash,
73
+ OpenAI::Graders::LabelModelGrader::OrHash,
74
+ OpenAI::Graders::StringCheckGrader::OrHash,
75
75
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
76
76
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython::OrHash,
77
77
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel::OrHash
@@ -223,8 +223,8 @@ module OpenAI
223
223
  Variants =
224
224
  T.type_alias do
225
225
  T.any(
226
- OpenAI::Models::Graders::LabelModelGrader,
227
- OpenAI::Models::Graders::StringCheckGrader,
226
+ OpenAI::Graders::LabelModelGrader,
227
+ OpenAI::Graders::StringCheckGrader,
228
228
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderTextSimilarity,
229
229
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderPython,
230
230
  OpenAI::Models::EvalCreateResponse::TestingCriterion::EvalGraderScoreModel
@@ -68,8 +68,8 @@ module OpenAI
68
68
  testing_criteria:
69
69
  T::Array[
70
70
  T.any(
71
- OpenAI::Models::Graders::LabelModelGrader::OrHash,
72
- OpenAI::Models::Graders::StringCheckGrader::OrHash,
71
+ OpenAI::Graders::LabelModelGrader::OrHash,
72
+ OpenAI::Graders::StringCheckGrader::OrHash,
73
73
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
74
74
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython::OrHash,
75
75
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel::OrHash
@@ -221,8 +221,8 @@ module OpenAI
221
221
  Variants =
222
222
  T.type_alias do
223
223
  T.any(
224
- OpenAI::Models::Graders::LabelModelGrader,
225
- OpenAI::Models::Graders::StringCheckGrader,
224
+ OpenAI::Graders::LabelModelGrader,
225
+ OpenAI::Graders::StringCheckGrader,
226
226
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderTextSimilarity,
227
227
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderPython,
228
228
  OpenAI::Models::EvalListResponse::TestingCriterion::EvalGraderScoreModel
@@ -72,8 +72,8 @@ module OpenAI
72
72
  testing_criteria:
73
73
  T::Array[
74
74
  T.any(
75
- OpenAI::Models::Graders::LabelModelGrader::OrHash,
76
- OpenAI::Models::Graders::StringCheckGrader::OrHash,
75
+ OpenAI::Graders::LabelModelGrader::OrHash,
76
+ OpenAI::Graders::StringCheckGrader::OrHash,
77
77
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
78
78
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython::OrHash,
79
79
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel::OrHash
@@ -225,8 +225,8 @@ module OpenAI
225
225
  Variants =
226
226
  T.type_alias do
227
227
  T.any(
228
- OpenAI::Models::Graders::LabelModelGrader,
229
- OpenAI::Models::Graders::StringCheckGrader,
228
+ OpenAI::Graders::LabelModelGrader,
229
+ OpenAI::Graders::StringCheckGrader,
230
230
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderTextSimilarity,
231
231
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderPython,
232
232
  OpenAI::Models::EvalRetrieveResponse::TestingCriterion::EvalGraderScoreModel
@@ -70,8 +70,8 @@ module OpenAI
70
70
  testing_criteria:
71
71
  T::Array[
72
72
  T.any(
73
- OpenAI::Models::Graders::LabelModelGrader::OrHash,
74
- OpenAI::Models::Graders::StringCheckGrader::OrHash,
73
+ OpenAI::Graders::LabelModelGrader::OrHash,
74
+ OpenAI::Graders::StringCheckGrader::OrHash,
75
75
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity::OrHash,
76
76
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython::OrHash,
77
77
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel::OrHash
@@ -223,8 +223,8 @@ module OpenAI
223
223
  Variants =
224
224
  T.type_alias do
225
225
  T.any(
226
- OpenAI::Models::Graders::LabelModelGrader,
227
- OpenAI::Models::Graders::StringCheckGrader,
226
+ OpenAI::Graders::LabelModelGrader,
227
+ OpenAI::Graders::StringCheckGrader,
228
228
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderTextSimilarity,
229
229
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderPython,
230
230
  OpenAI::Models::EvalUpdateResponse::TestingCriterion::EvalGraderScoreModel
@@ -892,6 +892,9 @@ module OpenAI
892
892
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
893
893
  # effort can result in faster responses and fewer tokens used on reasoning in a
894
894
  # response.
895
+ #
896
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
897
+ # effort.
895
898
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
896
899
  attr_accessor :reasoning_effort
897
900
 
@@ -992,6 +995,9 @@ module OpenAI
992
995
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
993
996
  # effort can result in faster responses and fewer tokens used on reasoning in a
994
997
  # response.
998
+ #
999
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1000
+ # effort.
995
1001
  reasoning_effort: nil,
996
1002
  # An object specifying the format that the model must output.
997
1003
  #
@@ -515,6 +515,9 @@ module OpenAI
515
515
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
516
516
  # effort can result in faster responses and fewer tokens used on reasoning in a
517
517
  # response.
518
+ #
519
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
520
+ # effort.
518
521
  sig do
519
522
  returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
520
523
  end
@@ -574,6 +577,9 @@ module OpenAI
574
577
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
575
578
  # effort can result in faster responses and fewer tokens used on reasoning in a
576
579
  # response.
580
+ #
581
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
582
+ # effort.
577
583
  reasoning_effort: nil,
578
584
  # Sampling temperature. This is a query parameter used to select responses.
579
585
  temperature: nil,
@@ -1120,6 +1126,9 @@ module OpenAI
1120
1126
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1121
1127
  # effort can result in faster responses and fewer tokens used on reasoning in a
1122
1128
  # response.
1129
+ #
1130
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1131
+ # effort.
1123
1132
  sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
1124
1133
  attr_accessor :reasoning_effort
1125
1134
 
@@ -1241,6 +1250,9 @@ module OpenAI
1241
1250
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1242
1251
  # effort can result in faster responses and fewer tokens used on reasoning in a
1243
1252
  # response.
1253
+ #
1254
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1255
+ # effort.
1244
1256
  reasoning_effort: nil,
1245
1257
  # A seed value to initialize the randomness, during sampling.
1246
1258
  seed: nil,
@@ -425,6 +425,9 @@ module OpenAI
425
425
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
426
426
  # effort can result in faster responses and fewer tokens used on reasoning in a
427
427
  # response.
428
+ #
429
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
430
+ # effort.
428
431
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
429
432
  attr_accessor :reasoning_effort
430
433
 
@@ -482,6 +485,9 @@ module OpenAI
482
485
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
483
486
  # effort can result in faster responses and fewer tokens used on reasoning in a
484
487
  # response.
488
+ #
489
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
490
+ # effort.
485
491
  reasoning_effort: nil,
486
492
  # Sampling temperature. This is a query parameter used to select responses.
487
493
  temperature: nil,
@@ -1078,6 +1084,9 @@ module OpenAI
1078
1084
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1079
1085
  # effort can result in faster responses and fewer tokens used on reasoning in a
1080
1086
  # response.
1087
+ #
1088
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1089
+ # effort.
1081
1090
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
1082
1091
  attr_accessor :reasoning_effort
1083
1092
 
@@ -1216,6 +1225,9 @@ module OpenAI
1216
1225
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1217
1226
  # effort can result in faster responses and fewer tokens used on reasoning in a
1218
1227
  # response.
1228
+ #
1229
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1230
+ # effort.
1219
1231
  reasoning_effort: nil,
1220
1232
  # A seed value to initialize the randomness, during sampling.
1221
1233
  seed: nil,
@@ -515,6 +515,9 @@ module OpenAI
515
515
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
516
516
  # effort can result in faster responses and fewer tokens used on reasoning in a
517
517
  # response.
518
+ #
519
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
520
+ # effort.
518
521
  sig do
519
522
  returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
520
523
  end
@@ -574,6 +577,9 @@ module OpenAI
574
577
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
575
578
  # effort can result in faster responses and fewer tokens used on reasoning in a
576
579
  # response.
580
+ #
581
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
582
+ # effort.
577
583
  reasoning_effort: nil,
578
584
  # Sampling temperature. This is a query parameter used to select responses.
579
585
  temperature: nil,
@@ -1120,6 +1126,9 @@ module OpenAI
1120
1126
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1121
1127
  # effort can result in faster responses and fewer tokens used on reasoning in a
1122
1128
  # response.
1129
+ #
1130
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1131
+ # effort.
1123
1132
  sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
1124
1133
  attr_accessor :reasoning_effort
1125
1134
 
@@ -1241,6 +1250,9 @@ module OpenAI
1241
1250
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1242
1251
  # effort can result in faster responses and fewer tokens used on reasoning in a
1243
1252
  # response.
1253
+ #
1254
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1255
+ # effort.
1244
1256
  reasoning_effort: nil,
1245
1257
  # A seed value to initialize the randomness, during sampling.
1246
1258
  seed: nil,
@@ -511,6 +511,9 @@ module OpenAI
511
511
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
512
512
  # effort can result in faster responses and fewer tokens used on reasoning in a
513
513
  # response.
514
+ #
515
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
516
+ # effort.
514
517
  sig do
515
518
  returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
516
519
  end
@@ -570,6 +573,9 @@ module OpenAI
570
573
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
571
574
  # effort can result in faster responses and fewer tokens used on reasoning in a
572
575
  # response.
576
+ #
577
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
578
+ # effort.
573
579
  reasoning_effort: nil,
574
580
  # Sampling temperature. This is a query parameter used to select responses.
575
581
  temperature: nil,
@@ -1116,6 +1122,9 @@ module OpenAI
1116
1122
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1117
1123
  # effort can result in faster responses and fewer tokens used on reasoning in a
1118
1124
  # response.
1125
+ #
1126
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1127
+ # effort.
1119
1128
  sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
1120
1129
  attr_accessor :reasoning_effort
1121
1130
 
@@ -1237,6 +1246,9 @@ module OpenAI
1237
1246
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1238
1247
  # effort can result in faster responses and fewer tokens used on reasoning in a
1239
1248
  # response.
1249
+ #
1250
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1251
+ # effort.
1240
1252
  reasoning_effort: nil,
1241
1253
  # A seed value to initialize the randomness, during sampling.
1242
1254
  seed: nil,
@@ -517,6 +517,9 @@ module OpenAI
517
517
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
518
518
  # effort can result in faster responses and fewer tokens used on reasoning in a
519
519
  # response.
520
+ #
521
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
522
+ # effort.
520
523
  sig do
521
524
  returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol))
522
525
  end
@@ -576,6 +579,9 @@ module OpenAI
576
579
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
577
580
  # effort can result in faster responses and fewer tokens used on reasoning in a
578
581
  # response.
582
+ #
583
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
584
+ # effort.
579
585
  reasoning_effort: nil,
580
586
  # Sampling temperature. This is a query parameter used to select responses.
581
587
  temperature: nil,
@@ -1122,6 +1128,9 @@ module OpenAI
1122
1128
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1123
1129
  # effort can result in faster responses and fewer tokens used on reasoning in a
1124
1130
  # response.
1131
+ #
1132
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1133
+ # effort.
1125
1134
  sig { returns(T.nilable(OpenAI::ReasoningEffort::TaggedSymbol)) }
1126
1135
  attr_accessor :reasoning_effort
1127
1136
 
@@ -1243,6 +1252,9 @@ module OpenAI
1243
1252
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
1244
1253
  # effort can result in faster responses and fewer tokens used on reasoning in a
1245
1254
  # response.
1255
+ #
1256
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
1257
+ # effort.
1246
1258
  reasoning_effort: nil,
1247
1259
  # A seed value to initialize the randomness, during sampling.
1248
1260
  seed: nil,
@@ -399,6 +399,9 @@ module OpenAI
399
399
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
400
400
  # effort can result in faster responses and fewer tokens used on reasoning in a
401
401
  # response.
402
+ #
403
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
404
+ # effort.
402
405
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
403
406
  attr_accessor :reasoning_effort
404
407
 
@@ -432,6 +435,9 @@ module OpenAI
432
435
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
433
436
  # effort can result in faster responses and fewer tokens used on reasoning in a
434
437
  # response.
438
+ #
439
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
440
+ # effort.
435
441
  reasoning_effort: nil,
436
442
  # A seed value to initialize the randomness, during sampling.
437
443
  seed: nil,
@@ -38,7 +38,8 @@ module OpenAI
38
38
 
39
39
  # Control how much effort the model will exert to match the style and features,
40
40
  # especially facial features, of input images. This parameter is only supported
41
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
41
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
42
+ # `low`. Defaults to `low`.
42
43
  sig do
43
44
  returns(T.nilable(OpenAI::ImageEditParams::InputFidelity::OrSymbol))
44
45
  end
@@ -161,7 +162,8 @@ module OpenAI
161
162
  background: nil,
162
163
  # Control how much effort the model will exert to match the style and features,
163
164
  # especially facial features, of input images. This parameter is only supported
164
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
165
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
166
+ # `low`. Defaults to `low`.
165
167
  input_fidelity: nil,
166
168
  # An additional image whose fully transparent areas (e.g. where alpha is zero)
167
169
  # indicate where `image` should be edited. If there are multiple images provided,
@@ -294,7 +296,8 @@ module OpenAI
294
296
 
295
297
  # Control how much effort the model will exert to match the style and features,
296
298
  # especially facial features, of input images. This parameter is only supported
297
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
299
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
300
+ # `low`. Defaults to `low`.
298
301
  module InputFidelity
299
302
  extend OpenAI::Internal::Type::Enum
300
303
 
@@ -11,6 +11,8 @@ module OpenAI
11
11
  DALL_E_2 = T.let(:"dall-e-2", OpenAI::ImageModel::TaggedSymbol)
12
12
  DALL_E_3 = T.let(:"dall-e-3", OpenAI::ImageModel::TaggedSymbol)
13
13
  GPT_IMAGE_1 = T.let(:"gpt-image-1", OpenAI::ImageModel::TaggedSymbol)
14
+ GPT_IMAGE_1_MINI =
15
+ T.let(:"gpt-image-1-mini", OpenAI::ImageModel::TaggedSymbol)
14
16
 
15
17
  sig { override.returns(T::Array[OpenAI::ImageModel::TaggedSymbol]) }
16
18
  def self.values
@@ -707,6 +707,26 @@ module OpenAI
707
707
  :"gpt-4o-mini-realtime-preview-2024-12-17",
708
708
  OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
709
709
  )
710
+ GPT_REALTIME_MINI =
711
+ T.let(
712
+ :"gpt-realtime-mini",
713
+ OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
714
+ )
715
+ GPT_REALTIME_MINI_2025_10_06 =
716
+ T.let(
717
+ :"gpt-realtime-mini-2025-10-06",
718
+ OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
719
+ )
720
+ GPT_AUDIO_MINI =
721
+ T.let(
722
+ :"gpt-audio-mini",
723
+ OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
724
+ )
725
+ GPT_AUDIO_MINI_2025_10_06 =
726
+ T.let(
727
+ :"gpt-audio-mini-2025-10-06",
728
+ OpenAI::Realtime::RealtimeSession::Model::TaggedSymbol
729
+ )
710
730
 
711
731
  sig do
712
732
  override.returns(
@@ -508,6 +508,26 @@ module OpenAI
508
508
  :"gpt-4o-mini-realtime-preview-2024-12-17",
509
509
  OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
510
510
  )
511
+ GPT_REALTIME_MINI =
512
+ T.let(
513
+ :"gpt-realtime-mini",
514
+ OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
515
+ )
516
+ GPT_REALTIME_MINI_2025_10_06 =
517
+ T.let(
518
+ :"gpt-realtime-mini-2025-10-06",
519
+ OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
520
+ )
521
+ GPT_AUDIO_MINI =
522
+ T.let(
523
+ :"gpt-audio-mini",
524
+ OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
525
+ )
526
+ GPT_AUDIO_MINI_2025_10_06 =
527
+ T.let(
528
+ :"gpt-audio-mini-2025-10-06",
529
+ OpenAI::Realtime::RealtimeSessionCreateRequest::Model::TaggedSymbol
530
+ )
511
531
  end
512
532
 
513
533
  module OutputModality
@@ -1308,6 +1308,26 @@ module OpenAI
1308
1308
  :"gpt-4o-mini-realtime-preview-2024-12-17",
1309
1309
  OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
1310
1310
  )
1311
+ GPT_REALTIME_MINI =
1312
+ T.let(
1313
+ :"gpt-realtime-mini",
1314
+ OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
1315
+ )
1316
+ GPT_REALTIME_MINI_2025_10_06 =
1317
+ T.let(
1318
+ :"gpt-realtime-mini-2025-10-06",
1319
+ OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
1320
+ )
1321
+ GPT_AUDIO_MINI =
1322
+ T.let(
1323
+ :"gpt-audio-mini",
1324
+ OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
1325
+ )
1326
+ GPT_AUDIO_MINI_2025_10_06 =
1327
+ T.let(
1328
+ :"gpt-audio-mini-2025-10-06",
1329
+ OpenAI::Realtime::RealtimeSessionCreateResponse::Model::TaggedSymbol
1330
+ )
1311
1331
  end
1312
1332
 
1313
1333
  module OutputModality
@@ -11,6 +11,9 @@ module OpenAI
11
11
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
12
12
  # effort can result in faster responses and fewer tokens used on reasoning in a
13
13
  # response.
14
+ #
15
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
16
+ # effort.
14
17
  sig { returns(T.nilable(OpenAI::ReasoningEffort::OrSymbol)) }
15
18
  attr_accessor :effort
16
19
 
@@ -46,6 +49,9 @@ module OpenAI
46
49
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
47
50
  # effort can result in faster responses and fewer tokens used on reasoning in a
48
51
  # response.
52
+ #
53
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
54
+ # effort.
49
55
  effort: nil,
50
56
  # **Deprecated:** use `summary` instead.
51
57
  #
@@ -7,6 +7,9 @@ module OpenAI
7
7
  # supported values are `minimal`, `low`, `medium`, and `high`. Reducing reasoning
8
8
  # effort can result in faster responses and fewer tokens used on reasoning in a
9
9
  # response.
10
+ #
11
+ # Note: The `gpt-5-pro` model defaults to (and only supports) `high` reasoning
12
+ # effort.
10
13
  module ReasoningEffort
11
14
  extend OpenAI::Internal::Type::Enum
12
15
 
@@ -781,7 +781,8 @@ module OpenAI
781
781
 
782
782
  # Control how much effort the model will exert to match the style and features,
783
783
  # especially facial features, of input images. This parameter is only supported
784
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
784
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
785
+ # `low`. Defaults to `low`.
785
786
  sig do
786
787
  returns(
787
788
  T.nilable(
@@ -946,7 +947,8 @@ module OpenAI
946
947
  background: nil,
947
948
  # Control how much effort the model will exert to match the style and features,
948
949
  # especially facial features, of input images. This parameter is only supported
949
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
950
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
951
+ # `low`. Defaults to `low`.
950
952
  input_fidelity: nil,
951
953
  # Optional mask for inpainting. Contains `image_url` (string, optional) and
952
954
  # `file_id` (string, optional).
@@ -1046,7 +1048,8 @@ module OpenAI
1046
1048
 
1047
1049
  # Control how much effort the model will exert to match the style and features,
1048
1050
  # especially facial features, of input images. This parameter is only supported
1049
- # for `gpt-image-1`. Supports `high` and `low`. Defaults to `low`.
1051
+ # for `gpt-image-1`. Unsupported for `gpt-image-1-mini`. Supports `high` and
1052
+ # `low`. Defaults to `low`.
1050
1053
  module InputFidelity
1051
1054
  extend OpenAI::Internal::Type::Enum
1052
1055
 
@@ -1139,6 +1142,11 @@ module OpenAI
1139
1142
  :"gpt-image-1",
1140
1143
  OpenAI::Responses::Tool::ImageGeneration::Model::TaggedSymbol
1141
1144
  )
1145
+ GPT_IMAGE_1_MINI =
1146
+ T.let(
1147
+ :"gpt-image-1-mini",
1148
+ OpenAI::Responses::Tool::ImageGeneration::Model::TaggedSymbol
1149
+ )
1142
1150
 
1143
1151
  sig do
1144
1152
  override.returns(