openai 0.12.0 → 0.13.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (59) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +21 -0
  3. data/README.md +1 -1
  4. data/lib/openai/helpers/structured_output/json_schema_converter.rb +34 -10
  5. data/lib/openai/models/audio/transcription.rb +4 -4
  6. data/lib/openai/models/audio/transcription_verbose.rb +4 -4
  7. data/lib/openai/models/eval_create_params.rb +50 -5
  8. data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +50 -5
  9. data/lib/openai/models/evals/run_cancel_response.rb +48 -5
  10. data/lib/openai/models/evals/run_create_params.rb +50 -5
  11. data/lib/openai/models/evals/run_create_response.rb +48 -5
  12. data/lib/openai/models/evals/run_list_response.rb +48 -5
  13. data/lib/openai/models/evals/run_retrieve_response.rb +48 -5
  14. data/lib/openai/models/file_object.rb +5 -4
  15. data/lib/openai/models/graders/label_model_grader.rb +48 -5
  16. data/lib/openai/models/graders/score_model_grader.rb +48 -5
  17. data/lib/openai/models/responses/response_input_file.rb +9 -1
  18. data/lib/openai/models/responses/response_mcp_call_arguments_delta_event.rb +5 -5
  19. data/lib/openai/models/responses/response_mcp_call_arguments_done_event.rb +5 -5
  20. data/lib/openai/models/responses/response_output_text_annotation_added_event.rb +5 -5
  21. data/lib/openai/models/responses/response_stream_event.rb +3 -3
  22. data/lib/openai/models/responses/tool.rb +9 -1
  23. data/lib/openai/version.rb +1 -1
  24. data/rbi/openai/helpers/structured_output/json_schema_converter.rbi +4 -0
  25. data/rbi/openai/models/audio/transcription.rbi +4 -4
  26. data/rbi/openai/models/audio/transcription_verbose.rbi +4 -6
  27. data/rbi/openai/models/eval_create_params.rbi +76 -7
  28. data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +76 -7
  29. data/rbi/openai/models/evals/run_cancel_response.rbi +70 -5
  30. data/rbi/openai/models/evals/run_create_params.rbi +76 -7
  31. data/rbi/openai/models/evals/run_create_response.rbi +70 -5
  32. data/rbi/openai/models/evals/run_list_response.rbi +70 -5
  33. data/rbi/openai/models/evals/run_retrieve_response.rbi +70 -5
  34. data/rbi/openai/models/file_object.rbi +7 -6
  35. data/rbi/openai/models/graders/label_model_grader.rbi +74 -7
  36. data/rbi/openai/models/graders/score_model_grader.rbi +74 -7
  37. data/rbi/openai/models/responses/response_input_file.rbi +11 -0
  38. data/rbi/openai/models/responses/response_mcp_call_arguments_delta_event.rbi +3 -3
  39. data/rbi/openai/models/responses/response_mcp_call_arguments_done_event.rbi +3 -3
  40. data/rbi/openai/models/responses/response_output_text_annotation_added_event.rbi +3 -3
  41. data/rbi/openai/models/responses/tool.rbi +12 -1
  42. data/sig/openai/models/audio/transcription.rbs +4 -4
  43. data/sig/openai/models/audio/transcription_verbose.rbs +4 -4
  44. data/sig/openai/models/eval_create_params.rbs +29 -0
  45. data/sig/openai/models/evals/create_eval_completions_run_data_source.rbs +29 -0
  46. data/sig/openai/models/evals/run_cancel_response.rbs +33 -0
  47. data/sig/openai/models/evals/run_create_params.rbs +33 -0
  48. data/sig/openai/models/evals/run_create_response.rbs +33 -0
  49. data/sig/openai/models/evals/run_list_response.rbs +33 -0
  50. data/sig/openai/models/evals/run_retrieve_response.rbs +33 -0
  51. data/sig/openai/models/file_object.rbs +2 -0
  52. data/sig/openai/models/graders/label_model_grader.rbs +29 -0
  53. data/sig/openai/models/graders/score_model_grader.rbs +29 -0
  54. data/sig/openai/models/responses/response_input_file.rbs +7 -0
  55. data/sig/openai/models/responses/response_mcp_call_arguments_delta_event.rbs +4 -4
  56. data/sig/openai/models/responses/response_mcp_call_arguments_done_event.rbs +4 -4
  57. data/sig/openai/models/responses/response_output_text_annotation_added_event.rbs +4 -4
  58. data/sig/openai/models/responses/tool.rbs +9 -2
  59. metadata +2 -2
@@ -70,6 +70,13 @@ module OpenAI
70
70
  end
71
71
  attr_accessor :require_approval
72
72
 
73
+ # Optional description of the MCP server, used to provide more context.
74
+ sig { returns(T.nilable(String)) }
75
+ attr_reader :server_description
76
+
77
+ sig { params(server_description: String).void }
78
+ attr_writer :server_description
79
+
73
80
  # Give the model access to additional tools via remote Model Context Protocol
74
81
  # (MCP) servers.
75
82
  # [Learn more about MCP](https://platform.openai.com/docs/guides/tools-remote-mcp).
@@ -92,6 +99,7 @@ module OpenAI
92
99
  OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalSetting::OrSymbol
93
100
  )
94
101
  ),
102
+ server_description: String,
95
103
  type: Symbol
96
104
  ).returns(T.attached_class)
97
105
  end
@@ -107,6 +115,8 @@ module OpenAI
107
115
  headers: nil,
108
116
  # Specify which of the MCP server's tools require approval.
109
117
  require_approval: nil,
118
+ # Optional description of the MCP server, used to provide more context.
119
+ server_description: nil,
110
120
  # The type of the MCP tool. Always `mcp`.
111
121
  type: :mcp
112
122
  )
@@ -132,7 +142,8 @@ module OpenAI
132
142
  OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalFilter,
133
143
  OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalSetting::OrSymbol
134
144
  )
135
- )
145
+ ),
146
+ server_description: String
136
147
  }
137
148
  )
138
149
  end
@@ -131,16 +131,16 @@ module OpenAI
131
131
  end
132
132
  end
133
133
 
134
- type duration = { duration: Float, type: :duration }
134
+ type duration = { seconds: Float, type: :duration }
135
135
 
136
136
  class Duration < OpenAI::Internal::Type::BaseModel
137
- attr_accessor duration: Float
137
+ attr_accessor seconds: Float
138
138
 
139
139
  attr_accessor type: :duration
140
140
 
141
- def initialize: (duration: Float, ?type: :duration) -> void
141
+ def initialize: (seconds: Float, ?type: :duration) -> void
142
142
 
143
- def to_hash: -> { duration: Float, type: :duration }
143
+ def to_hash: -> { seconds: Float, type: :duration }
144
144
  end
145
145
 
146
146
  def self?.variants: -> ::Array[OpenAI::Models::Audio::Transcription::usage]
@@ -54,16 +54,16 @@ module OpenAI
54
54
  words: ::Array[OpenAI::Audio::TranscriptionWord]
55
55
  }
56
56
 
57
- type usage = { duration: Float, type: :duration }
57
+ type usage = { seconds: Float, type: :duration }
58
58
 
59
59
  class Usage < OpenAI::Internal::Type::BaseModel
60
- attr_accessor duration: Float
60
+ attr_accessor seconds: Float
61
61
 
62
62
  attr_accessor type: :duration
63
63
 
64
- def initialize: (duration: Float, ?type: :duration) -> void
64
+ def initialize: (seconds: Float, ?type: :duration) -> void
65
65
 
66
- def to_hash: -> { duration: Float, type: :duration }
66
+ def to_hash: -> { seconds: Float, type: :duration }
67
67
  end
68
68
  end
69
69
  end
@@ -218,6 +218,8 @@ module OpenAI
218
218
  String
219
219
  | OpenAI::Responses::ResponseInputText
220
220
  | OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText
221
+ | OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage
222
+ | ::Array[top]
221
223
 
222
224
  module Content
223
225
  extend OpenAI::Internal::Type::Union
@@ -234,7 +236,34 @@ module OpenAI
234
236
  def to_hash: -> { text: String, type: :output_text }
235
237
  end
236
238
 
239
+ type input_image =
240
+ { image_url: String, type: :input_image, detail: String }
241
+
242
+ class InputImage < OpenAI::Internal::Type::BaseModel
243
+ attr_accessor image_url: String
244
+
245
+ attr_accessor type: :input_image
246
+
247
+ attr_reader detail: String?
248
+
249
+ def detail=: (String) -> String
250
+
251
+ def initialize: (
252
+ image_url: String,
253
+ ?detail: String,
254
+ ?type: :input_image
255
+ ) -> void
256
+
257
+ def to_hash: -> {
258
+ image_url: String,
259
+ type: :input_image,
260
+ detail: String
261
+ }
262
+ end
263
+
237
264
  def self?.variants: -> ::Array[OpenAI::Models::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::content]
265
+
266
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
238
267
  end
239
268
 
240
269
  type role = :user | :assistant | :system | :developer
@@ -234,6 +234,8 @@ module OpenAI
234
234
  String
235
235
  | OpenAI::Responses::ResponseInputText
236
236
  | OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::Content::OutputText
237
+ | OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::Content::InputImage
238
+ | ::Array[top]
237
239
 
238
240
  module Content
239
241
  extend OpenAI::Internal::Type::Union
@@ -250,7 +252,34 @@ module OpenAI
250
252
  def to_hash: -> { text: String, type: :output_text }
251
253
  end
252
254
 
255
+ type input_image =
256
+ { image_url: String, type: :input_image, detail: String }
257
+
258
+ class InputImage < OpenAI::Internal::Type::BaseModel
259
+ attr_accessor image_url: String
260
+
261
+ attr_accessor type: :input_image
262
+
263
+ attr_reader detail: String?
264
+
265
+ def detail=: (String) -> String
266
+
267
+ def initialize: (
268
+ image_url: String,
269
+ ?detail: String,
270
+ ?type: :input_image
271
+ ) -> void
272
+
273
+ def to_hash: -> {
274
+ image_url: String,
275
+ type: :input_image,
276
+ detail: String
277
+ }
278
+ end
279
+
253
280
  def self?.variants: -> ::Array[OpenAI::Models::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::content]
281
+
282
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
254
283
  end
255
284
 
256
285
  type role = :user | :assistant | :system | :developer
@@ -350,6 +350,8 @@ module OpenAI
350
350
  String
351
351
  | OpenAI::Responses::ResponseInputText
352
352
  | OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
353
+ | OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
354
+ | ::Array[top]
353
355
 
354
356
  module Content
355
357
  extend OpenAI::Internal::Type::Union
@@ -369,7 +371,38 @@ module OpenAI
369
371
  def to_hash: -> { text: String, type: :output_text }
370
372
  end
371
373
 
374
+ type input_image =
375
+ {
376
+ image_url: String,
377
+ type: :input_image,
378
+ detail: String
379
+ }
380
+
381
+ class InputImage < OpenAI::Internal::Type::BaseModel
382
+ attr_accessor image_url: String
383
+
384
+ attr_accessor type: :input_image
385
+
386
+ attr_reader detail: String?
387
+
388
+ def detail=: (String) -> String
389
+
390
+ def initialize: (
391
+ image_url: String,
392
+ ?detail: String,
393
+ ?type: :input_image
394
+ ) -> void
395
+
396
+ def to_hash: -> {
397
+ image_url: String,
398
+ type: :input_image,
399
+ detail: String
400
+ }
401
+ end
402
+
372
403
  def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
404
+
405
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
373
406
  end
374
407
 
375
408
  type role = :user | :assistant | :system | :developer
@@ -313,6 +313,8 @@ module OpenAI
313
313
  String
314
314
  | OpenAI::Responses::ResponseInputText
315
315
  | OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText
316
+ | OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage
317
+ | ::Array[top]
316
318
 
317
319
  module Content
318
320
  extend OpenAI::Internal::Type::Union
@@ -332,7 +334,38 @@ module OpenAI
332
334
  def to_hash: -> { text: String, type: :output_text }
333
335
  end
334
336
 
337
+ type input_image =
338
+ {
339
+ image_url: String,
340
+ type: :input_image,
341
+ detail: String
342
+ }
343
+
344
+ class InputImage < OpenAI::Internal::Type::BaseModel
345
+ attr_accessor image_url: String
346
+
347
+ attr_accessor type: :input_image
348
+
349
+ attr_reader detail: String?
350
+
351
+ def detail=: (String) -> String
352
+
353
+ def initialize: (
354
+ image_url: String,
355
+ ?detail: String,
356
+ ?type: :input_image
357
+ ) -> void
358
+
359
+ def to_hash: -> {
360
+ image_url: String,
361
+ type: :input_image,
362
+ detail: String
363
+ }
364
+ end
365
+
335
366
  def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::content]
367
+
368
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
336
369
  end
337
370
 
338
371
  type role = :user | :assistant | :system | :developer
@@ -350,6 +350,8 @@ module OpenAI
350
350
  String
351
351
  | OpenAI::Responses::ResponseInputText
352
352
  | OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
353
+ | OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
354
+ | ::Array[top]
353
355
 
354
356
  module Content
355
357
  extend OpenAI::Internal::Type::Union
@@ -369,7 +371,38 @@ module OpenAI
369
371
  def to_hash: -> { text: String, type: :output_text }
370
372
  end
371
373
 
374
+ type input_image =
375
+ {
376
+ image_url: String,
377
+ type: :input_image,
378
+ detail: String
379
+ }
380
+
381
+ class InputImage < OpenAI::Internal::Type::BaseModel
382
+ attr_accessor image_url: String
383
+
384
+ attr_accessor type: :input_image
385
+
386
+ attr_reader detail: String?
387
+
388
+ def detail=: (String) -> String
389
+
390
+ def initialize: (
391
+ image_url: String,
392
+ ?detail: String,
393
+ ?type: :input_image
394
+ ) -> void
395
+
396
+ def to_hash: -> {
397
+ image_url: String,
398
+ type: :input_image,
399
+ detail: String
400
+ }
401
+ end
402
+
372
403
  def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
404
+
405
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
373
406
  end
374
407
 
375
408
  type role = :user | :assistant | :system | :developer
@@ -350,6 +350,8 @@ module OpenAI
350
350
  String
351
351
  | OpenAI::Responses::ResponseInputText
352
352
  | OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
353
+ | OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
354
+ | ::Array[top]
353
355
 
354
356
  module Content
355
357
  extend OpenAI::Internal::Type::Union
@@ -369,7 +371,38 @@ module OpenAI
369
371
  def to_hash: -> { text: String, type: :output_text }
370
372
  end
371
373
 
374
+ type input_image =
375
+ {
376
+ image_url: String,
377
+ type: :input_image,
378
+ detail: String
379
+ }
380
+
381
+ class InputImage < OpenAI::Internal::Type::BaseModel
382
+ attr_accessor image_url: String
383
+
384
+ attr_accessor type: :input_image
385
+
386
+ attr_reader detail: String?
387
+
388
+ def detail=: (String) -> String
389
+
390
+ def initialize: (
391
+ image_url: String,
392
+ ?detail: String,
393
+ ?type: :input_image
394
+ ) -> void
395
+
396
+ def to_hash: -> {
397
+ image_url: String,
398
+ type: :input_image,
399
+ detail: String
400
+ }
401
+ end
402
+
372
403
  def self?.variants: -> ::Array[OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
404
+
405
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
373
406
  end
374
407
 
375
408
  type role = :user | :assistant | :system | :developer
@@ -350,6 +350,8 @@ module OpenAI
350
350
  String
351
351
  | OpenAI::Responses::ResponseInputText
352
352
  | OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
353
+ | OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
354
+ | ::Array[top]
353
355
 
354
356
  module Content
355
357
  extend OpenAI::Internal::Type::Union
@@ -369,7 +371,38 @@ module OpenAI
369
371
  def to_hash: -> { text: String, type: :output_text }
370
372
  end
371
373
 
374
+ type input_image =
375
+ {
376
+ image_url: String,
377
+ type: :input_image,
378
+ detail: String
379
+ }
380
+
381
+ class InputImage < OpenAI::Internal::Type::BaseModel
382
+ attr_accessor image_url: String
383
+
384
+ attr_accessor type: :input_image
385
+
386
+ attr_reader detail: String?
387
+
388
+ def detail=: (String) -> String
389
+
390
+ def initialize: (
391
+ image_url: String,
392
+ ?detail: String,
393
+ ?type: :input_image
394
+ ) -> void
395
+
396
+ def to_hash: -> {
397
+ image_url: String,
398
+ type: :input_image,
399
+ detail: String
400
+ }
401
+ end
402
+
372
403
  def self?.variants: -> ::Array[OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
404
+
405
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
373
406
  end
374
407
 
375
408
  type role = :user | :assistant | :system | :developer
@@ -68,6 +68,7 @@ module OpenAI
68
68
  | :"fine-tune"
69
69
  | :"fine-tune-results"
70
70
  | :vision
71
+ | :user_data
71
72
 
72
73
  module Purpose
73
74
  extend OpenAI::Internal::Type::Enum
@@ -79,6 +80,7 @@ module OpenAI
79
80
  FINE_TUNE: :"fine-tune"
80
81
  FINE_TUNE_RESULTS: :"fine-tune-results"
81
82
  VISION: :vision
83
+ USER_DATA: :user_data
82
84
 
83
85
  def self?.values: -> ::Array[OpenAI::Models::FileObject::purpose]
84
86
  end
@@ -78,6 +78,8 @@ module OpenAI
78
78
  String
79
79
  | OpenAI::Responses::ResponseInputText
80
80
  | OpenAI::Graders::LabelModelGrader::Input::Content::OutputText
81
+ | OpenAI::Graders::LabelModelGrader::Input::Content::InputImage
82
+ | ::Array[top]
81
83
 
82
84
  module Content
83
85
  extend OpenAI::Internal::Type::Union
@@ -94,7 +96,34 @@ module OpenAI
94
96
  def to_hash: -> { text: String, type: :output_text }
95
97
  end
96
98
 
99
+ type input_image =
100
+ { image_url: String, type: :input_image, detail: String }
101
+
102
+ class InputImage < OpenAI::Internal::Type::BaseModel
103
+ attr_accessor image_url: String
104
+
105
+ attr_accessor type: :input_image
106
+
107
+ attr_reader detail: String?
108
+
109
+ def detail=: (String) -> String
110
+
111
+ def initialize: (
112
+ image_url: String,
113
+ ?detail: String,
114
+ ?type: :input_image
115
+ ) -> void
116
+
117
+ def to_hash: -> {
118
+ image_url: String,
119
+ type: :input_image,
120
+ detail: String
121
+ }
122
+ end
123
+
97
124
  def self?.variants: -> ::Array[OpenAI::Models::Graders::LabelModelGrader::Input::content]
125
+
126
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
98
127
  end
99
128
 
100
129
  type role = :user | :assistant | :system | :developer
@@ -82,6 +82,8 @@ module OpenAI
82
82
  String
83
83
  | OpenAI::Responses::ResponseInputText
84
84
  | OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText
85
+ | OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage
86
+ | ::Array[top]
85
87
 
86
88
  module Content
87
89
  extend OpenAI::Internal::Type::Union
@@ -98,7 +100,34 @@ module OpenAI
98
100
  def to_hash: -> { text: String, type: :output_text }
99
101
  end
100
102
 
103
+ type input_image =
104
+ { image_url: String, type: :input_image, detail: String }
105
+
106
+ class InputImage < OpenAI::Internal::Type::BaseModel
107
+ attr_accessor image_url: String
108
+
109
+ attr_accessor type: :input_image
110
+
111
+ attr_reader detail: String?
112
+
113
+ def detail=: (String) -> String
114
+
115
+ def initialize: (
116
+ image_url: String,
117
+ ?detail: String,
118
+ ?type: :input_image
119
+ ) -> void
120
+
121
+ def to_hash: -> {
122
+ image_url: String,
123
+ type: :input_image,
124
+ detail: String
125
+ }
126
+ end
127
+
101
128
  def self?.variants: -> ::Array[OpenAI::Models::Graders::ScoreModelGrader::Input::content]
129
+
130
+ AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
102
131
  end
103
132
 
104
133
  type role = :user | :assistant | :system | :developer
@@ -6,6 +6,7 @@ module OpenAI
6
6
  type: :input_file,
7
7
  file_data: String,
8
8
  file_id: String?,
9
+ file_url: String,
9
10
  filename: String
10
11
  }
11
12
 
@@ -18,6 +19,10 @@ module OpenAI
18
19
 
19
20
  attr_accessor file_id: String?
20
21
 
22
+ attr_reader file_url: String?
23
+
24
+ def file_url=: (String) -> String
25
+
21
26
  attr_reader filename: String?
22
27
 
23
28
  def filename=: (String) -> String
@@ -25,6 +30,7 @@ module OpenAI
25
30
  def initialize: (
26
31
  ?file_data: String,
27
32
  ?file_id: String?,
33
+ ?file_url: String,
28
34
  ?filename: String,
29
35
  ?type: :input_file
30
36
  ) -> void
@@ -33,6 +39,7 @@ module OpenAI
33
39
  type: :input_file,
34
40
  file_data: String,
35
41
  file_id: String?,
42
+ file_url: String,
36
43
  filename: String
37
44
  }
38
45
  end
@@ -7,7 +7,7 @@ module OpenAI
7
7
  item_id: String,
8
8
  output_index: Integer,
9
9
  sequence_number: Integer,
10
- type: :"response.mcp_call.arguments_delta"
10
+ type: :"response.mcp_call_arguments.delta"
11
11
  }
12
12
 
13
13
  class ResponseMcpCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel
@@ -19,14 +19,14 @@ module OpenAI
19
19
 
20
20
  attr_accessor sequence_number: Integer
21
21
 
22
- attr_accessor type: :"response.mcp_call.arguments_delta"
22
+ attr_accessor type: :"response.mcp_call_arguments.delta"
23
23
 
24
24
  def initialize: (
25
25
  delta: top,
26
26
  item_id: String,
27
27
  output_index: Integer,
28
28
  sequence_number: Integer,
29
- ?type: :"response.mcp_call.arguments_delta"
29
+ ?type: :"response.mcp_call_arguments.delta"
30
30
  ) -> void
31
31
 
32
32
  def to_hash: -> {
@@ -34,7 +34,7 @@ module OpenAI
34
34
  item_id: String,
35
35
  output_index: Integer,
36
36
  sequence_number: Integer,
37
- type: :"response.mcp_call.arguments_delta"
37
+ type: :"response.mcp_call_arguments.delta"
38
38
  }
39
39
  end
40
40
  end
@@ -7,7 +7,7 @@ module OpenAI
7
7
  item_id: String,
8
8
  output_index: Integer,
9
9
  sequence_number: Integer,
10
- type: :"response.mcp_call.arguments_done"
10
+ type: :"response.mcp_call_arguments.done"
11
11
  }
12
12
 
13
13
  class ResponseMcpCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel
@@ -19,14 +19,14 @@ module OpenAI
19
19
 
20
20
  attr_accessor sequence_number: Integer
21
21
 
22
- attr_accessor type: :"response.mcp_call.arguments_done"
22
+ attr_accessor type: :"response.mcp_call_arguments.done"
23
23
 
24
24
  def initialize: (
25
25
  arguments: top,
26
26
  item_id: String,
27
27
  output_index: Integer,
28
28
  sequence_number: Integer,
29
- ?type: :"response.mcp_call.arguments_done"
29
+ ?type: :"response.mcp_call_arguments.done"
30
30
  ) -> void
31
31
 
32
32
  def to_hash: -> {
@@ -34,7 +34,7 @@ module OpenAI
34
34
  item_id: String,
35
35
  output_index: Integer,
36
36
  sequence_number: Integer,
37
- type: :"response.mcp_call.arguments_done"
37
+ type: :"response.mcp_call_arguments.done"
38
38
  }
39
39
  end
40
40
  end
@@ -9,7 +9,7 @@ module OpenAI
9
9
  item_id: String,
10
10
  output_index: Integer,
11
11
  sequence_number: Integer,
12
- type: :"response.output_text_annotation.added"
12
+ type: :"response.output_text.annotation.added"
13
13
  }
14
14
 
15
15
  class ResponseOutputTextAnnotationAddedEvent < OpenAI::Internal::Type::BaseModel
@@ -25,7 +25,7 @@ module OpenAI
25
25
 
26
26
  attr_accessor sequence_number: Integer
27
27
 
28
- attr_accessor type: :"response.output_text_annotation.added"
28
+ attr_accessor type: :"response.output_text.annotation.added"
29
29
 
30
30
  def initialize: (
31
31
  annotation: top,
@@ -34,7 +34,7 @@ module OpenAI
34
34
  item_id: String,
35
35
  output_index: Integer,
36
36
  sequence_number: Integer,
37
- ?type: :"response.output_text_annotation.added"
37
+ ?type: :"response.output_text.annotation.added"
38
38
  ) -> void
39
39
 
40
40
  def to_hash: -> {
@@ -44,7 +44,7 @@ module OpenAI
44
44
  item_id: String,
45
45
  output_index: Integer,
46
46
  sequence_number: Integer,
47
- type: :"response.output_text_annotation.added"
47
+ type: :"response.output_text.annotation.added"
48
48
  }
49
49
  end
50
50
  end