openai 0.12.0 → 0.13.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +21 -0
- data/README.md +1 -1
- data/lib/openai/helpers/structured_output/json_schema_converter.rb +34 -10
- data/lib/openai/models/audio/transcription.rb +4 -4
- data/lib/openai/models/audio/transcription_verbose.rb +4 -4
- data/lib/openai/models/eval_create_params.rb +50 -5
- data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +50 -5
- data/lib/openai/models/evals/run_cancel_response.rb +48 -5
- data/lib/openai/models/evals/run_create_params.rb +50 -5
- data/lib/openai/models/evals/run_create_response.rb +48 -5
- data/lib/openai/models/evals/run_list_response.rb +48 -5
- data/lib/openai/models/evals/run_retrieve_response.rb +48 -5
- data/lib/openai/models/file_object.rb +5 -4
- data/lib/openai/models/graders/label_model_grader.rb +48 -5
- data/lib/openai/models/graders/score_model_grader.rb +48 -5
- data/lib/openai/models/responses/response_input_file.rb +9 -1
- data/lib/openai/models/responses/response_mcp_call_arguments_delta_event.rb +5 -5
- data/lib/openai/models/responses/response_mcp_call_arguments_done_event.rb +5 -5
- data/lib/openai/models/responses/response_output_text_annotation_added_event.rb +5 -5
- data/lib/openai/models/responses/response_stream_event.rb +3 -3
- data/lib/openai/models/responses/tool.rb +9 -1
- data/lib/openai/version.rb +1 -1
- data/rbi/openai/helpers/structured_output/json_schema_converter.rbi +4 -0
- data/rbi/openai/models/audio/transcription.rbi +4 -4
- data/rbi/openai/models/audio/transcription_verbose.rbi +4 -6
- data/rbi/openai/models/eval_create_params.rbi +76 -7
- data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +76 -7
- data/rbi/openai/models/evals/run_cancel_response.rbi +70 -5
- data/rbi/openai/models/evals/run_create_params.rbi +76 -7
- data/rbi/openai/models/evals/run_create_response.rbi +70 -5
- data/rbi/openai/models/evals/run_list_response.rbi +70 -5
- data/rbi/openai/models/evals/run_retrieve_response.rbi +70 -5
- data/rbi/openai/models/file_object.rbi +7 -6
- data/rbi/openai/models/graders/label_model_grader.rbi +74 -7
- data/rbi/openai/models/graders/score_model_grader.rbi +74 -7
- data/rbi/openai/models/responses/response_input_file.rbi +11 -0
- data/rbi/openai/models/responses/response_mcp_call_arguments_delta_event.rbi +3 -3
- data/rbi/openai/models/responses/response_mcp_call_arguments_done_event.rbi +3 -3
- data/rbi/openai/models/responses/response_output_text_annotation_added_event.rbi +3 -3
- data/rbi/openai/models/responses/tool.rbi +12 -1
- data/sig/openai/models/audio/transcription.rbs +4 -4
- data/sig/openai/models/audio/transcription_verbose.rbs +4 -4
- data/sig/openai/models/eval_create_params.rbs +29 -0
- data/sig/openai/models/evals/create_eval_completions_run_data_source.rbs +29 -0
- data/sig/openai/models/evals/run_cancel_response.rbs +33 -0
- data/sig/openai/models/evals/run_create_params.rbs +33 -0
- data/sig/openai/models/evals/run_create_response.rbs +33 -0
- data/sig/openai/models/evals/run_list_response.rbs +33 -0
- data/sig/openai/models/evals/run_retrieve_response.rbs +33 -0
- data/sig/openai/models/file_object.rbs +2 -0
- data/sig/openai/models/graders/label_model_grader.rbs +29 -0
- data/sig/openai/models/graders/score_model_grader.rbs +29 -0
- data/sig/openai/models/responses/response_input_file.rbs +7 -0
- data/sig/openai/models/responses/response_mcp_call_arguments_delta_event.rbs +4 -4
- data/sig/openai/models/responses/response_mcp_call_arguments_done_event.rbs +4 -4
- data/sig/openai/models/responses/response_output_text_annotation_added_event.rbs +4 -4
- data/sig/openai/models/responses/tool.rbs +9 -2
- metadata +2 -2
@@ -70,6 +70,13 @@ module OpenAI
|
|
70
70
|
end
|
71
71
|
attr_accessor :require_approval
|
72
72
|
|
73
|
+
# Optional description of the MCP server, used to provide more context.
|
74
|
+
sig { returns(T.nilable(String)) }
|
75
|
+
attr_reader :server_description
|
76
|
+
|
77
|
+
sig { params(server_description: String).void }
|
78
|
+
attr_writer :server_description
|
79
|
+
|
73
80
|
# Give the model access to additional tools via remote Model Context Protocol
|
74
81
|
# (MCP) servers.
|
75
82
|
# [Learn more about MCP](https://platform.openai.com/docs/guides/tools-remote-mcp).
|
@@ -92,6 +99,7 @@ module OpenAI
|
|
92
99
|
OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalSetting::OrSymbol
|
93
100
|
)
|
94
101
|
),
|
102
|
+
server_description: String,
|
95
103
|
type: Symbol
|
96
104
|
).returns(T.attached_class)
|
97
105
|
end
|
@@ -107,6 +115,8 @@ module OpenAI
|
|
107
115
|
headers: nil,
|
108
116
|
# Specify which of the MCP server's tools require approval.
|
109
117
|
require_approval: nil,
|
118
|
+
# Optional description of the MCP server, used to provide more context.
|
119
|
+
server_description: nil,
|
110
120
|
# The type of the MCP tool. Always `mcp`.
|
111
121
|
type: :mcp
|
112
122
|
)
|
@@ -132,7 +142,8 @@ module OpenAI
|
|
132
142
|
OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalFilter,
|
133
143
|
OpenAI::Responses::Tool::Mcp::RequireApproval::McpToolApprovalSetting::OrSymbol
|
134
144
|
)
|
135
|
-
)
|
145
|
+
),
|
146
|
+
server_description: String
|
136
147
|
}
|
137
148
|
)
|
138
149
|
end
|
@@ -131,16 +131,16 @@ module OpenAI
|
|
131
131
|
end
|
132
132
|
end
|
133
133
|
|
134
|
-
type duration = {
|
134
|
+
type duration = { seconds: Float, type: :duration }
|
135
135
|
|
136
136
|
class Duration < OpenAI::Internal::Type::BaseModel
|
137
|
-
attr_accessor
|
137
|
+
attr_accessor seconds: Float
|
138
138
|
|
139
139
|
attr_accessor type: :duration
|
140
140
|
|
141
|
-
def initialize: (
|
141
|
+
def initialize: (seconds: Float, ?type: :duration) -> void
|
142
142
|
|
143
|
-
def to_hash: -> {
|
143
|
+
def to_hash: -> { seconds: Float, type: :duration }
|
144
144
|
end
|
145
145
|
|
146
146
|
def self?.variants: -> ::Array[OpenAI::Models::Audio::Transcription::usage]
|
@@ -54,16 +54,16 @@ module OpenAI
|
|
54
54
|
words: ::Array[OpenAI::Audio::TranscriptionWord]
|
55
55
|
}
|
56
56
|
|
57
|
-
type usage = {
|
57
|
+
type usage = { seconds: Float, type: :duration }
|
58
58
|
|
59
59
|
class Usage < OpenAI::Internal::Type::BaseModel
|
60
|
-
attr_accessor
|
60
|
+
attr_accessor seconds: Float
|
61
61
|
|
62
62
|
attr_accessor type: :duration
|
63
63
|
|
64
|
-
def initialize: (
|
64
|
+
def initialize: (seconds: Float, ?type: :duration) -> void
|
65
65
|
|
66
|
-
def to_hash: -> {
|
66
|
+
def to_hash: -> { seconds: Float, type: :duration }
|
67
67
|
end
|
68
68
|
end
|
69
69
|
end
|
@@ -218,6 +218,8 @@ module OpenAI
|
|
218
218
|
String
|
219
219
|
| OpenAI::Responses::ResponseInputText
|
220
220
|
| OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText
|
221
|
+
| OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage
|
222
|
+
| ::Array[top]
|
221
223
|
|
222
224
|
module Content
|
223
225
|
extend OpenAI::Internal::Type::Union
|
@@ -234,7 +236,34 @@ module OpenAI
|
|
234
236
|
def to_hash: -> { text: String, type: :output_text }
|
235
237
|
end
|
236
238
|
|
239
|
+
type input_image =
|
240
|
+
{ image_url: String, type: :input_image, detail: String }
|
241
|
+
|
242
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
243
|
+
attr_accessor image_url: String
|
244
|
+
|
245
|
+
attr_accessor type: :input_image
|
246
|
+
|
247
|
+
attr_reader detail: String?
|
248
|
+
|
249
|
+
def detail=: (String) -> String
|
250
|
+
|
251
|
+
def initialize: (
|
252
|
+
image_url: String,
|
253
|
+
?detail: String,
|
254
|
+
?type: :input_image
|
255
|
+
) -> void
|
256
|
+
|
257
|
+
def to_hash: -> {
|
258
|
+
image_url: String,
|
259
|
+
type: :input_image,
|
260
|
+
detail: String
|
261
|
+
}
|
262
|
+
end
|
263
|
+
|
237
264
|
def self?.variants: -> ::Array[OpenAI::Models::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::content]
|
265
|
+
|
266
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
238
267
|
end
|
239
268
|
|
240
269
|
type role = :user | :assistant | :system | :developer
|
@@ -234,6 +234,8 @@ module OpenAI
|
|
234
234
|
String
|
235
235
|
| OpenAI::Responses::ResponseInputText
|
236
236
|
| OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::Content::OutputText
|
237
|
+
| OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::Content::InputImage
|
238
|
+
| ::Array[top]
|
237
239
|
|
238
240
|
module Content
|
239
241
|
extend OpenAI::Internal::Type::Union
|
@@ -250,7 +252,34 @@ module OpenAI
|
|
250
252
|
def to_hash: -> { text: String, type: :output_text }
|
251
253
|
end
|
252
254
|
|
255
|
+
type input_image =
|
256
|
+
{ image_url: String, type: :input_image, detail: String }
|
257
|
+
|
258
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
259
|
+
attr_accessor image_url: String
|
260
|
+
|
261
|
+
attr_accessor type: :input_image
|
262
|
+
|
263
|
+
attr_reader detail: String?
|
264
|
+
|
265
|
+
def detail=: (String) -> String
|
266
|
+
|
267
|
+
def initialize: (
|
268
|
+
image_url: String,
|
269
|
+
?detail: String,
|
270
|
+
?type: :input_image
|
271
|
+
) -> void
|
272
|
+
|
273
|
+
def to_hash: -> {
|
274
|
+
image_url: String,
|
275
|
+
type: :input_image,
|
276
|
+
detail: String
|
277
|
+
}
|
278
|
+
end
|
279
|
+
|
253
280
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::Message::content]
|
281
|
+
|
282
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
254
283
|
end
|
255
284
|
|
256
285
|
type role = :user | :assistant | :system | :developer
|
@@ -350,6 +350,8 @@ module OpenAI
|
|
350
350
|
String
|
351
351
|
| OpenAI::Responses::ResponseInputText
|
352
352
|
| OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
|
353
|
+
| OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
|
354
|
+
| ::Array[top]
|
353
355
|
|
354
356
|
module Content
|
355
357
|
extend OpenAI::Internal::Type::Union
|
@@ -369,7 +371,38 @@ module OpenAI
|
|
369
371
|
def to_hash: -> { text: String, type: :output_text }
|
370
372
|
end
|
371
373
|
|
374
|
+
type input_image =
|
375
|
+
{
|
376
|
+
image_url: String,
|
377
|
+
type: :input_image,
|
378
|
+
detail: String
|
379
|
+
}
|
380
|
+
|
381
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
382
|
+
attr_accessor image_url: String
|
383
|
+
|
384
|
+
attr_accessor type: :input_image
|
385
|
+
|
386
|
+
attr_reader detail: String?
|
387
|
+
|
388
|
+
def detail=: (String) -> String
|
389
|
+
|
390
|
+
def initialize: (
|
391
|
+
image_url: String,
|
392
|
+
?detail: String,
|
393
|
+
?type: :input_image
|
394
|
+
) -> void
|
395
|
+
|
396
|
+
def to_hash: -> {
|
397
|
+
image_url: String,
|
398
|
+
type: :input_image,
|
399
|
+
detail: String
|
400
|
+
}
|
401
|
+
end
|
402
|
+
|
372
403
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
|
404
|
+
|
405
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
373
406
|
end
|
374
407
|
|
375
408
|
type role = :user | :assistant | :system | :developer
|
@@ -313,6 +313,8 @@ module OpenAI
|
|
313
313
|
String
|
314
314
|
| OpenAI::Responses::ResponseInputText
|
315
315
|
| OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText
|
316
|
+
| OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage
|
317
|
+
| ::Array[top]
|
316
318
|
|
317
319
|
module Content
|
318
320
|
extend OpenAI::Internal::Type::Union
|
@@ -332,7 +334,38 @@ module OpenAI
|
|
332
334
|
def to_hash: -> { text: String, type: :output_text }
|
333
335
|
end
|
334
336
|
|
337
|
+
type input_image =
|
338
|
+
{
|
339
|
+
image_url: String,
|
340
|
+
type: :input_image,
|
341
|
+
detail: String
|
342
|
+
}
|
343
|
+
|
344
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
345
|
+
attr_accessor image_url: String
|
346
|
+
|
347
|
+
attr_accessor type: :input_image
|
348
|
+
|
349
|
+
attr_reader detail: String?
|
350
|
+
|
351
|
+
def detail=: (String) -> String
|
352
|
+
|
353
|
+
def initialize: (
|
354
|
+
image_url: String,
|
355
|
+
?detail: String,
|
356
|
+
?type: :input_image
|
357
|
+
) -> void
|
358
|
+
|
359
|
+
def to_hash: -> {
|
360
|
+
image_url: String,
|
361
|
+
type: :input_image,
|
362
|
+
detail: String
|
363
|
+
}
|
364
|
+
end
|
365
|
+
|
335
366
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::content]
|
367
|
+
|
368
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
336
369
|
end
|
337
370
|
|
338
371
|
type role = :user | :assistant | :system | :developer
|
@@ -350,6 +350,8 @@ module OpenAI
|
|
350
350
|
String
|
351
351
|
| OpenAI::Responses::ResponseInputText
|
352
352
|
| OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
|
353
|
+
| OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
|
354
|
+
| ::Array[top]
|
353
355
|
|
354
356
|
module Content
|
355
357
|
extend OpenAI::Internal::Type::Union
|
@@ -369,7 +371,38 @@ module OpenAI
|
|
369
371
|
def to_hash: -> { text: String, type: :output_text }
|
370
372
|
end
|
371
373
|
|
374
|
+
type input_image =
|
375
|
+
{
|
376
|
+
image_url: String,
|
377
|
+
type: :input_image,
|
378
|
+
detail: String
|
379
|
+
}
|
380
|
+
|
381
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
382
|
+
attr_accessor image_url: String
|
383
|
+
|
384
|
+
attr_accessor type: :input_image
|
385
|
+
|
386
|
+
attr_reader detail: String?
|
387
|
+
|
388
|
+
def detail=: (String) -> String
|
389
|
+
|
390
|
+
def initialize: (
|
391
|
+
image_url: String,
|
392
|
+
?detail: String,
|
393
|
+
?type: :input_image
|
394
|
+
) -> void
|
395
|
+
|
396
|
+
def to_hash: -> {
|
397
|
+
image_url: String,
|
398
|
+
type: :input_image,
|
399
|
+
detail: String
|
400
|
+
}
|
401
|
+
end
|
402
|
+
|
372
403
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
|
404
|
+
|
405
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
373
406
|
end
|
374
407
|
|
375
408
|
type role = :user | :assistant | :system | :developer
|
@@ -350,6 +350,8 @@ module OpenAI
|
|
350
350
|
String
|
351
351
|
| OpenAI::Responses::ResponseInputText
|
352
352
|
| OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
|
353
|
+
| OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
|
354
|
+
| ::Array[top]
|
353
355
|
|
354
356
|
module Content
|
355
357
|
extend OpenAI::Internal::Type::Union
|
@@ -369,7 +371,38 @@ module OpenAI
|
|
369
371
|
def to_hash: -> { text: String, type: :output_text }
|
370
372
|
end
|
371
373
|
|
374
|
+
type input_image =
|
375
|
+
{
|
376
|
+
image_url: String,
|
377
|
+
type: :input_image,
|
378
|
+
detail: String
|
379
|
+
}
|
380
|
+
|
381
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
382
|
+
attr_accessor image_url: String
|
383
|
+
|
384
|
+
attr_accessor type: :input_image
|
385
|
+
|
386
|
+
attr_reader detail: String?
|
387
|
+
|
388
|
+
def detail=: (String) -> String
|
389
|
+
|
390
|
+
def initialize: (
|
391
|
+
image_url: String,
|
392
|
+
?detail: String,
|
393
|
+
?type: :input_image
|
394
|
+
) -> void
|
395
|
+
|
396
|
+
def to_hash: -> {
|
397
|
+
image_url: String,
|
398
|
+
type: :input_image,
|
399
|
+
detail: String
|
400
|
+
}
|
401
|
+
end
|
402
|
+
|
372
403
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
|
404
|
+
|
405
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
373
406
|
end
|
374
407
|
|
375
408
|
type role = :user | :assistant | :system | :developer
|
@@ -350,6 +350,8 @@ module OpenAI
|
|
350
350
|
String
|
351
351
|
| OpenAI::Responses::ResponseInputText
|
352
352
|
| OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText
|
353
|
+
| OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage
|
354
|
+
| ::Array[top]
|
353
355
|
|
354
356
|
module Content
|
355
357
|
extend OpenAI::Internal::Type::Union
|
@@ -369,7 +371,38 @@ module OpenAI
|
|
369
371
|
def to_hash: -> { text: String, type: :output_text }
|
370
372
|
end
|
371
373
|
|
374
|
+
type input_image =
|
375
|
+
{
|
376
|
+
image_url: String,
|
377
|
+
type: :input_image,
|
378
|
+
detail: String
|
379
|
+
}
|
380
|
+
|
381
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
382
|
+
attr_accessor image_url: String
|
383
|
+
|
384
|
+
attr_accessor type: :input_image
|
385
|
+
|
386
|
+
attr_reader detail: String?
|
387
|
+
|
388
|
+
def detail=: (String) -> String
|
389
|
+
|
390
|
+
def initialize: (
|
391
|
+
image_url: String,
|
392
|
+
?detail: String,
|
393
|
+
?type: :input_image
|
394
|
+
) -> void
|
395
|
+
|
396
|
+
def to_hash: -> {
|
397
|
+
image_url: String,
|
398
|
+
type: :input_image,
|
399
|
+
detail: String
|
400
|
+
}
|
401
|
+
end
|
402
|
+
|
372
403
|
def self?.variants: -> ::Array[OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::content]
|
404
|
+
|
405
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
373
406
|
end
|
374
407
|
|
375
408
|
type role = :user | :assistant | :system | :developer
|
@@ -68,6 +68,7 @@ module OpenAI
|
|
68
68
|
| :"fine-tune"
|
69
69
|
| :"fine-tune-results"
|
70
70
|
| :vision
|
71
|
+
| :user_data
|
71
72
|
|
72
73
|
module Purpose
|
73
74
|
extend OpenAI::Internal::Type::Enum
|
@@ -79,6 +80,7 @@ module OpenAI
|
|
79
80
|
FINE_TUNE: :"fine-tune"
|
80
81
|
FINE_TUNE_RESULTS: :"fine-tune-results"
|
81
82
|
VISION: :vision
|
83
|
+
USER_DATA: :user_data
|
82
84
|
|
83
85
|
def self?.values: -> ::Array[OpenAI::Models::FileObject::purpose]
|
84
86
|
end
|
@@ -78,6 +78,8 @@ module OpenAI
|
|
78
78
|
String
|
79
79
|
| OpenAI::Responses::ResponseInputText
|
80
80
|
| OpenAI::Graders::LabelModelGrader::Input::Content::OutputText
|
81
|
+
| OpenAI::Graders::LabelModelGrader::Input::Content::InputImage
|
82
|
+
| ::Array[top]
|
81
83
|
|
82
84
|
module Content
|
83
85
|
extend OpenAI::Internal::Type::Union
|
@@ -94,7 +96,34 @@ module OpenAI
|
|
94
96
|
def to_hash: -> { text: String, type: :output_text }
|
95
97
|
end
|
96
98
|
|
99
|
+
type input_image =
|
100
|
+
{ image_url: String, type: :input_image, detail: String }
|
101
|
+
|
102
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
103
|
+
attr_accessor image_url: String
|
104
|
+
|
105
|
+
attr_accessor type: :input_image
|
106
|
+
|
107
|
+
attr_reader detail: String?
|
108
|
+
|
109
|
+
def detail=: (String) -> String
|
110
|
+
|
111
|
+
def initialize: (
|
112
|
+
image_url: String,
|
113
|
+
?detail: String,
|
114
|
+
?type: :input_image
|
115
|
+
) -> void
|
116
|
+
|
117
|
+
def to_hash: -> {
|
118
|
+
image_url: String,
|
119
|
+
type: :input_image,
|
120
|
+
detail: String
|
121
|
+
}
|
122
|
+
end
|
123
|
+
|
97
124
|
def self?.variants: -> ::Array[OpenAI::Models::Graders::LabelModelGrader::Input::content]
|
125
|
+
|
126
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
98
127
|
end
|
99
128
|
|
100
129
|
type role = :user | :assistant | :system | :developer
|
@@ -82,6 +82,8 @@ module OpenAI
|
|
82
82
|
String
|
83
83
|
| OpenAI::Responses::ResponseInputText
|
84
84
|
| OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText
|
85
|
+
| OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage
|
86
|
+
| ::Array[top]
|
85
87
|
|
86
88
|
module Content
|
87
89
|
extend OpenAI::Internal::Type::Union
|
@@ -98,7 +100,34 @@ module OpenAI
|
|
98
100
|
def to_hash: -> { text: String, type: :output_text }
|
99
101
|
end
|
100
102
|
|
103
|
+
type input_image =
|
104
|
+
{ image_url: String, type: :input_image, detail: String }
|
105
|
+
|
106
|
+
class InputImage < OpenAI::Internal::Type::BaseModel
|
107
|
+
attr_accessor image_url: String
|
108
|
+
|
109
|
+
attr_accessor type: :input_image
|
110
|
+
|
111
|
+
attr_reader detail: String?
|
112
|
+
|
113
|
+
def detail=: (String) -> String
|
114
|
+
|
115
|
+
def initialize: (
|
116
|
+
image_url: String,
|
117
|
+
?detail: String,
|
118
|
+
?type: :input_image
|
119
|
+
) -> void
|
120
|
+
|
121
|
+
def to_hash: -> {
|
122
|
+
image_url: String,
|
123
|
+
type: :input_image,
|
124
|
+
detail: String
|
125
|
+
}
|
126
|
+
end
|
127
|
+
|
101
128
|
def self?.variants: -> ::Array[OpenAI::Models::Graders::ScoreModelGrader::Input::content]
|
129
|
+
|
130
|
+
AnArrayOfInputTextAndInputImageArray: OpenAI::Internal::Type::Converter
|
102
131
|
end
|
103
132
|
|
104
133
|
type role = :user | :assistant | :system | :developer
|
@@ -6,6 +6,7 @@ module OpenAI
|
|
6
6
|
type: :input_file,
|
7
7
|
file_data: String,
|
8
8
|
file_id: String?,
|
9
|
+
file_url: String,
|
9
10
|
filename: String
|
10
11
|
}
|
11
12
|
|
@@ -18,6 +19,10 @@ module OpenAI
|
|
18
19
|
|
19
20
|
attr_accessor file_id: String?
|
20
21
|
|
22
|
+
attr_reader file_url: String?
|
23
|
+
|
24
|
+
def file_url=: (String) -> String
|
25
|
+
|
21
26
|
attr_reader filename: String?
|
22
27
|
|
23
28
|
def filename=: (String) -> String
|
@@ -25,6 +30,7 @@ module OpenAI
|
|
25
30
|
def initialize: (
|
26
31
|
?file_data: String,
|
27
32
|
?file_id: String?,
|
33
|
+
?file_url: String,
|
28
34
|
?filename: String,
|
29
35
|
?type: :input_file
|
30
36
|
) -> void
|
@@ -33,6 +39,7 @@ module OpenAI
|
|
33
39
|
type: :input_file,
|
34
40
|
file_data: String,
|
35
41
|
file_id: String?,
|
42
|
+
file_url: String,
|
36
43
|
filename: String
|
37
44
|
}
|
38
45
|
end
|
@@ -7,7 +7,7 @@ module OpenAI
|
|
7
7
|
item_id: String,
|
8
8
|
output_index: Integer,
|
9
9
|
sequence_number: Integer,
|
10
|
-
type: :"response.
|
10
|
+
type: :"response.mcp_call_arguments.delta"
|
11
11
|
}
|
12
12
|
|
13
13
|
class ResponseMcpCallArgumentsDeltaEvent < OpenAI::Internal::Type::BaseModel
|
@@ -19,14 +19,14 @@ module OpenAI
|
|
19
19
|
|
20
20
|
attr_accessor sequence_number: Integer
|
21
21
|
|
22
|
-
attr_accessor type: :"response.
|
22
|
+
attr_accessor type: :"response.mcp_call_arguments.delta"
|
23
23
|
|
24
24
|
def initialize: (
|
25
25
|
delta: top,
|
26
26
|
item_id: String,
|
27
27
|
output_index: Integer,
|
28
28
|
sequence_number: Integer,
|
29
|
-
?type: :"response.
|
29
|
+
?type: :"response.mcp_call_arguments.delta"
|
30
30
|
) -> void
|
31
31
|
|
32
32
|
def to_hash: -> {
|
@@ -34,7 +34,7 @@ module OpenAI
|
|
34
34
|
item_id: String,
|
35
35
|
output_index: Integer,
|
36
36
|
sequence_number: Integer,
|
37
|
-
type: :"response.
|
37
|
+
type: :"response.mcp_call_arguments.delta"
|
38
38
|
}
|
39
39
|
end
|
40
40
|
end
|
@@ -7,7 +7,7 @@ module OpenAI
|
|
7
7
|
item_id: String,
|
8
8
|
output_index: Integer,
|
9
9
|
sequence_number: Integer,
|
10
|
-
type: :"response.
|
10
|
+
type: :"response.mcp_call_arguments.done"
|
11
11
|
}
|
12
12
|
|
13
13
|
class ResponseMcpCallArgumentsDoneEvent < OpenAI::Internal::Type::BaseModel
|
@@ -19,14 +19,14 @@ module OpenAI
|
|
19
19
|
|
20
20
|
attr_accessor sequence_number: Integer
|
21
21
|
|
22
|
-
attr_accessor type: :"response.
|
22
|
+
attr_accessor type: :"response.mcp_call_arguments.done"
|
23
23
|
|
24
24
|
def initialize: (
|
25
25
|
arguments: top,
|
26
26
|
item_id: String,
|
27
27
|
output_index: Integer,
|
28
28
|
sequence_number: Integer,
|
29
|
-
?type: :"response.
|
29
|
+
?type: :"response.mcp_call_arguments.done"
|
30
30
|
) -> void
|
31
31
|
|
32
32
|
def to_hash: -> {
|
@@ -34,7 +34,7 @@ module OpenAI
|
|
34
34
|
item_id: String,
|
35
35
|
output_index: Integer,
|
36
36
|
sequence_number: Integer,
|
37
|
-
type: :"response.
|
37
|
+
type: :"response.mcp_call_arguments.done"
|
38
38
|
}
|
39
39
|
end
|
40
40
|
end
|
@@ -9,7 +9,7 @@ module OpenAI
|
|
9
9
|
item_id: String,
|
10
10
|
output_index: Integer,
|
11
11
|
sequence_number: Integer,
|
12
|
-
type: :"response.
|
12
|
+
type: :"response.output_text.annotation.added"
|
13
13
|
}
|
14
14
|
|
15
15
|
class ResponseOutputTextAnnotationAddedEvent < OpenAI::Internal::Type::BaseModel
|
@@ -25,7 +25,7 @@ module OpenAI
|
|
25
25
|
|
26
26
|
attr_accessor sequence_number: Integer
|
27
27
|
|
28
|
-
attr_accessor type: :"response.
|
28
|
+
attr_accessor type: :"response.output_text.annotation.added"
|
29
29
|
|
30
30
|
def initialize: (
|
31
31
|
annotation: top,
|
@@ -34,7 +34,7 @@ module OpenAI
|
|
34
34
|
item_id: String,
|
35
35
|
output_index: Integer,
|
36
36
|
sequence_number: Integer,
|
37
|
-
?type: :"response.
|
37
|
+
?type: :"response.output_text.annotation.added"
|
38
38
|
) -> void
|
39
39
|
|
40
40
|
def to_hash: -> {
|
@@ -44,7 +44,7 @@ module OpenAI
|
|
44
44
|
item_id: String,
|
45
45
|
output_index: Integer,
|
46
46
|
sequence_number: Integer,
|
47
|
-
type: :"response.
|
47
|
+
type: :"response.output_text.annotation.added"
|
48
48
|
}
|
49
49
|
end
|
50
50
|
end
|