openai 0.21.0 → 0.21.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +8 -0
- data/README.md +1 -1
- data/lib/openai/models/eval_create_params.rb +10 -6
- data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +10 -6
- data/lib/openai/models/evals/run_cancel_response.rb +10 -6
- data/lib/openai/models/evals/run_create_params.rb +10 -6
- data/lib/openai/models/evals/run_create_response.rb +10 -6
- data/lib/openai/models/evals/run_list_response.rb +10 -6
- data/lib/openai/models/evals/run_retrieve_response.rb +10 -6
- data/lib/openai/models/graders/label_model_grader.rb +10 -6
- data/lib/openai/models/graders/score_model_grader.rb +10 -6
- data/lib/openai/models/responses/easy_input_message.rb +3 -3
- data/lib/openai/models/responses/response.rb +1 -1
- data/lib/openai/models/responses/response_content.rb +4 -1
- data/lib/openai/models/responses/response_create_params.rb +1 -1
- data/lib/openai/models/responses/response_input_audio.rb +39 -23
- data/lib/openai/models/responses/response_input_content.rb +4 -1
- data/lib/openai/models/responses/response_input_item.rb +2 -2
- data/lib/openai/models/responses/response_input_message_item.rb +2 -2
- data/lib/openai/version.rb +1 -1
- data/rbi/openai/models/eval_create_params.rbi +5 -1
- data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +5 -1
- data/rbi/openai/models/evals/run_cancel_response.rbi +3 -1
- data/rbi/openai/models/evals/run_create_params.rbi +5 -1
- data/rbi/openai/models/evals/run_create_response.rbi +3 -1
- data/rbi/openai/models/evals/run_list_response.rbi +3 -1
- data/rbi/openai/models/evals/run_retrieve_response.rbi +3 -1
- data/rbi/openai/models/graders/label_model_grader.rbi +5 -1
- data/rbi/openai/models/graders/score_model_grader.rbi +5 -1
- data/rbi/openai/models/responses/response.rbi +2 -2
- data/rbi/openai/models/responses/response_content.rbi +1 -0
- data/rbi/openai/models/responses/response_create_params.rbi +2 -2
- data/rbi/openai/models/responses/response_input_audio.rbi +85 -34
- data/rbi/openai/models/responses/response_input_content.rbi +2 -1
- data/rbi/openai/models/responses/response_input_item.rbi +6 -3
- data/rbi/openai/models/responses/response_input_message_item.rbi +2 -1
- data/rbi/openai/resources/responses.rbi +2 -2
- data/sig/openai/models/eval_create_params.rbs +2 -1
- data/sig/openai/models/evals/create_eval_completions_run_data_source.rbs +2 -1
- data/sig/openai/models/evals/run_cancel_response.rbs +2 -1
- data/sig/openai/models/evals/run_create_params.rbs +2 -1
- data/sig/openai/models/evals/run_create_response.rbs +2 -1
- data/sig/openai/models/evals/run_list_response.rbs +2 -1
- data/sig/openai/models/evals/run_retrieve_response.rbs +2 -1
- data/sig/openai/models/graders/label_model_grader.rbs +2 -1
- data/sig/openai/models/graders/score_model_grader.rbs +2 -1
- data/sig/openai/models/responses/response_content.rbs +1 -0
- data/sig/openai/models/responses/response_input_audio.rbs +32 -15
- data/sig/openai/models/responses/response_input_content.rbs +1 -0
- metadata +1 -1
@@ -4,17 +4,10 @@ module OpenAI
|
|
4
4
|
module Models
|
5
5
|
module Responses
|
6
6
|
class ResponseInputAudio < OpenAI::Internal::Type::BaseModel
|
7
|
-
# @!attribute
|
8
|
-
# Base64-encoded audio data.
|
7
|
+
# @!attribute input_audio
|
9
8
|
#
|
10
|
-
# @return [
|
11
|
-
required :
|
12
|
-
|
13
|
-
# @!attribute format_
|
14
|
-
# The format of the audio data. Currently supported formats are `mp3` and `wav`.
|
15
|
-
#
|
16
|
-
# @return [Symbol, OpenAI::Models::Responses::ResponseInputAudio::Format]
|
17
|
-
required :format_, enum: -> { OpenAI::Responses::ResponseInputAudio::Format }, api_name: :format
|
9
|
+
# @return [OpenAI::Models::Responses::ResponseInputAudio::InputAudio]
|
10
|
+
required :input_audio, -> { OpenAI::Responses::ResponseInputAudio::InputAudio }
|
18
11
|
|
19
12
|
# @!attribute type
|
20
13
|
# The type of the input item. Always `input_audio`.
|
@@ -22,29 +15,52 @@ module OpenAI
|
|
22
15
|
# @return [Symbol, :input_audio]
|
23
16
|
required :type, const: :input_audio
|
24
17
|
|
25
|
-
# @!method initialize(
|
18
|
+
# @!method initialize(input_audio:, type: :input_audio)
|
26
19
|
# Some parameter documentations has been truncated, see
|
27
20
|
# {OpenAI::Models::Responses::ResponseInputAudio} for more details.
|
28
21
|
#
|
29
22
|
# An audio input to the model.
|
30
23
|
#
|
31
|
-
# @param
|
32
|
-
#
|
33
|
-
# @param format_ [Symbol, OpenAI::Models::Responses::ResponseInputAudio::Format] The format of the audio data. Currently supported formats are `mp3` and
|
24
|
+
# @param input_audio [OpenAI::Models::Responses::ResponseInputAudio::InputAudio]
|
34
25
|
#
|
35
26
|
# @param type [Symbol, :input_audio] The type of the input item. Always `input_audio`.
|
36
27
|
|
37
|
-
#
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
28
|
+
# @see OpenAI::Models::Responses::ResponseInputAudio#input_audio
|
29
|
+
class InputAudio < OpenAI::Internal::Type::BaseModel
|
30
|
+
# @!attribute data
|
31
|
+
# Base64-encoded audio data.
|
32
|
+
#
|
33
|
+
# @return [String]
|
34
|
+
required :data, String
|
35
|
+
|
36
|
+
# @!attribute format_
|
37
|
+
# The format of the audio data. Currently supported formats are `mp3` and `wav`.
|
38
|
+
#
|
39
|
+
# @return [Symbol, OpenAI::Models::Responses::ResponseInputAudio::InputAudio::Format]
|
40
|
+
required :format_,
|
41
|
+
enum: -> { OpenAI::Responses::ResponseInputAudio::InputAudio::Format },
|
42
|
+
api_name: :format
|
43
|
+
|
44
|
+
# @!method initialize(data:, format_:)
|
45
|
+
# Some parameter documentations has been truncated, see
|
46
|
+
# {OpenAI::Models::Responses::ResponseInputAudio::InputAudio} for more details.
|
47
|
+
#
|
48
|
+
# @param data [String] Base64-encoded audio data.
|
49
|
+
#
|
50
|
+
# @param format_ [Symbol, OpenAI::Models::Responses::ResponseInputAudio::InputAudio::Format] The format of the audio data. Currently supported formats are `mp3` and
|
51
|
+
|
52
|
+
# The format of the audio data. Currently supported formats are `mp3` and `wav`.
|
53
|
+
#
|
54
|
+
# @see OpenAI::Models::Responses::ResponseInputAudio::InputAudio#format_
|
55
|
+
module Format
|
56
|
+
extend OpenAI::Internal::Type::Enum
|
42
57
|
|
43
|
-
|
44
|
-
|
58
|
+
MP3 = :mp3
|
59
|
+
WAV = :wav
|
45
60
|
|
46
|
-
|
47
|
-
|
61
|
+
# @!method self.values
|
62
|
+
# @return [Array<Symbol>]
|
63
|
+
end
|
48
64
|
end
|
49
65
|
end
|
50
66
|
end
|
@@ -18,8 +18,11 @@ module OpenAI
|
|
18
18
|
# A file input to the model.
|
19
19
|
variant :input_file, -> { OpenAI::Responses::ResponseInputFile }
|
20
20
|
|
21
|
+
# An audio input to the model.
|
22
|
+
variant :input_audio, -> { OpenAI::Responses::ResponseInputAudio }
|
23
|
+
|
21
24
|
# @!method self.variants
|
22
|
-
# @return [Array(OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile)]
|
25
|
+
# @return [Array(OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseInputAudio)]
|
23
26
|
end
|
24
27
|
end
|
25
28
|
end
|
@@ -94,7 +94,7 @@ module OpenAI
|
|
94
94
|
# A list of one or many input items to the model, containing different content
|
95
95
|
# types.
|
96
96
|
#
|
97
|
-
# @return [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile>]
|
97
|
+
# @return [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseInputAudio>]
|
98
98
|
required :content, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Responses::ResponseInputContent] }
|
99
99
|
|
100
100
|
# @!attribute role
|
@@ -124,7 +124,7 @@ module OpenAI
|
|
124
124
|
# hierarchy. Instructions given with the `developer` or `system` role take
|
125
125
|
# precedence over instructions given with the `user` role.
|
126
126
|
#
|
127
|
-
# @param content [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile>] A list of one or many input items to the model, containing different content
|
127
|
+
# @param content [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseInputAudio>] A list of one or many input items to the model, containing different content
|
128
128
|
#
|
129
129
|
# @param role [Symbol, OpenAI::Models::Responses::ResponseInputItem::Message::Role] The role of the message input. One of `user`, `system`, or `developer`.
|
130
130
|
#
|
@@ -14,7 +14,7 @@ module OpenAI
|
|
14
14
|
# A list of one or many input items to the model, containing different content
|
15
15
|
# types.
|
16
16
|
#
|
17
|
-
# @return [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile>]
|
17
|
+
# @return [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseInputAudio>]
|
18
18
|
required :content, -> { OpenAI::Internal::Type::ArrayOf[union: OpenAI::Responses::ResponseInputContent] }
|
19
19
|
|
20
20
|
# @!attribute role
|
@@ -42,7 +42,7 @@ module OpenAI
|
|
42
42
|
#
|
43
43
|
# @param id [String] The unique ID of the message input.
|
44
44
|
#
|
45
|
-
# @param content [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile>] A list of one or many input items to the model, containing different content
|
45
|
+
# @param content [Array<OpenAI::Models::Responses::ResponseInputText, OpenAI::Models::Responses::ResponseInputImage, OpenAI::Models::Responses::ResponseInputFile, OpenAI::Models::Responses::ResponseInputAudio>] A list of one or many input items to the model, containing different content
|
46
46
|
#
|
47
47
|
# @param role [Symbol, OpenAI::Models::Responses::ResponseInputMessageItem::Role] The role of the message input. One of `user`, `system`, or `developer`.
|
48
48
|
#
|
data/lib/openai/version.rb
CHANGED
@@ -482,6 +482,7 @@ module OpenAI
|
|
482
482
|
OpenAI::Responses::ResponseInputText,
|
483
483
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText,
|
484
484
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage,
|
485
|
+
OpenAI::Responses::ResponseInputAudio,
|
485
486
|
T::Array[T.anything]
|
486
487
|
)
|
487
488
|
)
|
@@ -528,6 +529,7 @@ module OpenAI
|
|
528
529
|
OpenAI::Responses::ResponseInputText::OrHash,
|
529
530
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText::OrHash,
|
530
531
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage::OrHash,
|
532
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
531
533
|
T::Array[T.anything]
|
532
534
|
),
|
533
535
|
role:
|
@@ -556,6 +558,7 @@ module OpenAI
|
|
556
558
|
OpenAI::Responses::ResponseInputText,
|
557
559
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText,
|
558
560
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage,
|
561
|
+
OpenAI::Responses::ResponseInputAudio,
|
559
562
|
T::Array[T.anything]
|
560
563
|
),
|
561
564
|
role:
|
@@ -579,6 +582,7 @@ module OpenAI
|
|
579
582
|
OpenAI::Responses::ResponseInputText,
|
580
583
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::OutputText,
|
581
584
|
OpenAI::EvalCreateParams::TestingCriterion::LabelModel::Input::EvalItem::Content::InputImage,
|
585
|
+
OpenAI::Responses::ResponseInputAudio,
|
582
586
|
T::Array[T.anything]
|
583
587
|
)
|
584
588
|
end
|
@@ -680,7 +684,7 @@ module OpenAI
|
|
680
684
|
def self.variants
|
681
685
|
end
|
682
686
|
|
683
|
-
|
687
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
684
688
|
T.let(
|
685
689
|
OpenAI::Internal::Type::ArrayOf[
|
686
690
|
OpenAI::Internal::Type::Unknown
|
@@ -523,6 +523,7 @@ module OpenAI
|
|
523
523
|
OpenAI::Responses::ResponseInputText,
|
524
524
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
525
525
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
526
|
+
OpenAI::Responses::ResponseInputAudio,
|
526
527
|
T::Array[T.anything]
|
527
528
|
)
|
528
529
|
)
|
@@ -569,6 +570,7 @@ module OpenAI
|
|
569
570
|
OpenAI::Responses::ResponseInputText::OrHash,
|
570
571
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
571
572
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
573
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
572
574
|
T::Array[T.anything]
|
573
575
|
),
|
574
576
|
role:
|
@@ -597,6 +599,7 @@ module OpenAI
|
|
597
599
|
OpenAI::Responses::ResponseInputText,
|
598
600
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
599
601
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
602
|
+
OpenAI::Responses::ResponseInputAudio,
|
600
603
|
T::Array[T.anything]
|
601
604
|
),
|
602
605
|
role:
|
@@ -620,6 +623,7 @@ module OpenAI
|
|
620
623
|
OpenAI::Responses::ResponseInputText,
|
621
624
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
622
625
|
OpenAI::Evals::CreateEvalCompletionsRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
626
|
+
OpenAI::Responses::ResponseInputAudio,
|
623
627
|
T::Array[T.anything]
|
624
628
|
)
|
625
629
|
end
|
@@ -723,7 +727,7 @@ module OpenAI
|
|
723
727
|
def self.variants
|
724
728
|
end
|
725
729
|
|
726
|
-
|
730
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
727
731
|
T.let(
|
728
732
|
OpenAI::Internal::Type::ArrayOf[
|
729
733
|
OpenAI::Internal::Type::Unknown
|
@@ -797,6 +797,7 @@ module OpenAI
|
|
797
797
|
OpenAI::Responses::ResponseInputText::OrHash,
|
798
798
|
OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
799
799
|
OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
800
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
800
801
|
T::Array[T.anything]
|
801
802
|
),
|
802
803
|
role:
|
@@ -842,6 +843,7 @@ module OpenAI
|
|
842
843
|
OpenAI::Responses::ResponseInputText,
|
843
844
|
OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
844
845
|
OpenAI::Models::Evals::RunCancelResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
846
|
+
OpenAI::Responses::ResponseInputAudio,
|
845
847
|
T::Array[T.anything]
|
846
848
|
)
|
847
849
|
end
|
@@ -945,7 +947,7 @@ module OpenAI
|
|
945
947
|
def self.variants
|
946
948
|
end
|
947
949
|
|
948
|
-
|
950
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
949
951
|
T.let(
|
950
952
|
OpenAI::Internal::Type::ArrayOf[
|
951
953
|
OpenAI::Internal::Type::Unknown
|
@@ -701,6 +701,7 @@ module OpenAI
|
|
701
701
|
OpenAI::Responses::ResponseInputText,
|
702
702
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
703
703
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
704
|
+
OpenAI::Responses::ResponseInputAudio,
|
704
705
|
T::Array[T.anything]
|
705
706
|
)
|
706
707
|
)
|
@@ -747,6 +748,7 @@ module OpenAI
|
|
747
748
|
OpenAI::Responses::ResponseInputText::OrHash,
|
748
749
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
749
750
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
751
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
750
752
|
T::Array[T.anything]
|
751
753
|
),
|
752
754
|
role:
|
@@ -775,6 +777,7 @@ module OpenAI
|
|
775
777
|
OpenAI::Responses::ResponseInputText,
|
776
778
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
777
779
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
780
|
+
OpenAI::Responses::ResponseInputAudio,
|
778
781
|
T::Array[T.anything]
|
779
782
|
),
|
780
783
|
role:
|
@@ -798,6 +801,7 @@ module OpenAI
|
|
798
801
|
OpenAI::Responses::ResponseInputText,
|
799
802
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
800
803
|
OpenAI::Evals::RunCreateParams::DataSource::CreateEvalResponsesRunDataSource::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
804
|
+
OpenAI::Responses::ResponseInputAudio,
|
801
805
|
T::Array[T.anything]
|
802
806
|
)
|
803
807
|
end
|
@@ -901,7 +905,7 @@ module OpenAI
|
|
901
905
|
def self.variants
|
902
906
|
end
|
903
907
|
|
904
|
-
|
908
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
905
909
|
T.let(
|
906
910
|
OpenAI::Internal::Type::ArrayOf[
|
907
911
|
OpenAI::Internal::Type::Unknown
|
@@ -797,6 +797,7 @@ module OpenAI
|
|
797
797
|
OpenAI::Responses::ResponseInputText::OrHash,
|
798
798
|
OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
799
799
|
OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
800
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
800
801
|
T::Array[T.anything]
|
801
802
|
),
|
802
803
|
role:
|
@@ -842,6 +843,7 @@ module OpenAI
|
|
842
843
|
OpenAI::Responses::ResponseInputText,
|
843
844
|
OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
844
845
|
OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
846
|
+
OpenAI::Responses::ResponseInputAudio,
|
845
847
|
T::Array[T.anything]
|
846
848
|
)
|
847
849
|
end
|
@@ -945,7 +947,7 @@ module OpenAI
|
|
945
947
|
def self.variants
|
946
948
|
end
|
947
949
|
|
948
|
-
|
950
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
949
951
|
T.let(
|
950
952
|
OpenAI::Internal::Type::ArrayOf[
|
951
953
|
OpenAI::Internal::Type::Unknown
|
@@ -793,6 +793,7 @@ module OpenAI
|
|
793
793
|
OpenAI::Responses::ResponseInputText::OrHash,
|
794
794
|
OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
795
795
|
OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
796
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
796
797
|
T::Array[T.anything]
|
797
798
|
),
|
798
799
|
role:
|
@@ -838,6 +839,7 @@ module OpenAI
|
|
838
839
|
OpenAI::Responses::ResponseInputText,
|
839
840
|
OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
840
841
|
OpenAI::Models::Evals::RunListResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
842
|
+
OpenAI::Responses::ResponseInputAudio,
|
841
843
|
T::Array[T.anything]
|
842
844
|
)
|
843
845
|
end
|
@@ -941,7 +943,7 @@ module OpenAI
|
|
941
943
|
def self.variants
|
942
944
|
end
|
943
945
|
|
944
|
-
|
946
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
945
947
|
T.let(
|
946
948
|
OpenAI::Internal::Type::ArrayOf[
|
947
949
|
OpenAI::Internal::Type::Unknown
|
@@ -799,6 +799,7 @@ module OpenAI
|
|
799
799
|
OpenAI::Responses::ResponseInputText::OrHash,
|
800
800
|
OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText::OrHash,
|
801
801
|
OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage::OrHash,
|
802
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
802
803
|
T::Array[T.anything]
|
803
804
|
),
|
804
805
|
role:
|
@@ -844,6 +845,7 @@ module OpenAI
|
|
844
845
|
OpenAI::Responses::ResponseInputText,
|
845
846
|
OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::OutputText,
|
846
847
|
OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::InputMessages::Template::Template::EvalItem::Content::InputImage,
|
848
|
+
OpenAI::Responses::ResponseInputAudio,
|
847
849
|
T::Array[T.anything]
|
848
850
|
)
|
849
851
|
end
|
@@ -947,7 +949,7 @@ module OpenAI
|
|
947
949
|
def self.variants
|
948
950
|
end
|
949
951
|
|
950
|
-
|
952
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
951
953
|
T.let(
|
952
954
|
OpenAI::Internal::Type::ArrayOf[
|
953
955
|
OpenAI::Internal::Type::Unknown
|
@@ -93,6 +93,7 @@ module OpenAI
|
|
93
93
|
OpenAI::Responses::ResponseInputText,
|
94
94
|
OpenAI::Graders::LabelModelGrader::Input::Content::OutputText,
|
95
95
|
OpenAI::Graders::LabelModelGrader::Input::Content::InputImage,
|
96
|
+
OpenAI::Responses::ResponseInputAudio,
|
96
97
|
T::Array[T.anything]
|
97
98
|
)
|
98
99
|
)
|
@@ -136,6 +137,7 @@ module OpenAI
|
|
136
137
|
OpenAI::Responses::ResponseInputText::OrHash,
|
137
138
|
OpenAI::Graders::LabelModelGrader::Input::Content::OutputText::OrHash,
|
138
139
|
OpenAI::Graders::LabelModelGrader::Input::Content::InputImage::OrHash,
|
140
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
139
141
|
T::Array[T.anything]
|
140
142
|
),
|
141
143
|
role: OpenAI::Graders::LabelModelGrader::Input::Role::OrSymbol,
|
@@ -162,6 +164,7 @@ module OpenAI
|
|
162
164
|
OpenAI::Responses::ResponseInputText,
|
163
165
|
OpenAI::Graders::LabelModelGrader::Input::Content::OutputText,
|
164
166
|
OpenAI::Graders::LabelModelGrader::Input::Content::InputImage,
|
167
|
+
OpenAI::Responses::ResponseInputAudio,
|
165
168
|
T::Array[T.anything]
|
166
169
|
),
|
167
170
|
role: OpenAI::Graders::LabelModelGrader::Input::Role::OrSymbol,
|
@@ -183,6 +186,7 @@ module OpenAI
|
|
183
186
|
OpenAI::Responses::ResponseInputText,
|
184
187
|
OpenAI::Graders::LabelModelGrader::Input::Content::OutputText,
|
185
188
|
OpenAI::Graders::LabelModelGrader::Input::Content::InputImage,
|
189
|
+
OpenAI::Responses::ResponseInputAudio,
|
186
190
|
T::Array[T.anything]
|
187
191
|
)
|
188
192
|
end
|
@@ -282,7 +286,7 @@ module OpenAI
|
|
282
286
|
def self.variants
|
283
287
|
end
|
284
288
|
|
285
|
-
|
289
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
286
290
|
T.let(
|
287
291
|
OpenAI::Internal::Type::ArrayOf[
|
288
292
|
OpenAI::Internal::Type::Unknown
|
@@ -100,6 +100,7 @@ module OpenAI
|
|
100
100
|
OpenAI::Responses::ResponseInputText,
|
101
101
|
OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText,
|
102
102
|
OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage,
|
103
|
+
OpenAI::Responses::ResponseInputAudio,
|
103
104
|
T::Array[T.anything]
|
104
105
|
)
|
105
106
|
)
|
@@ -143,6 +144,7 @@ module OpenAI
|
|
143
144
|
OpenAI::Responses::ResponseInputText::OrHash,
|
144
145
|
OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText::OrHash,
|
145
146
|
OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage::OrHash,
|
147
|
+
OpenAI::Responses::ResponseInputAudio::OrHash,
|
146
148
|
T::Array[T.anything]
|
147
149
|
),
|
148
150
|
role: OpenAI::Graders::ScoreModelGrader::Input::Role::OrSymbol,
|
@@ -169,6 +171,7 @@ module OpenAI
|
|
169
171
|
OpenAI::Responses::ResponseInputText,
|
170
172
|
OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText,
|
171
173
|
OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage,
|
174
|
+
OpenAI::Responses::ResponseInputAudio,
|
172
175
|
T::Array[T.anything]
|
173
176
|
),
|
174
177
|
role: OpenAI::Graders::ScoreModelGrader::Input::Role::OrSymbol,
|
@@ -190,6 +193,7 @@ module OpenAI
|
|
190
193
|
OpenAI::Responses::ResponseInputText,
|
191
194
|
OpenAI::Graders::ScoreModelGrader::Input::Content::OutputText,
|
192
195
|
OpenAI::Graders::ScoreModelGrader::Input::Content::InputImage,
|
196
|
+
OpenAI::Responses::ResponseInputAudio,
|
193
197
|
T::Array[T.anything]
|
194
198
|
)
|
195
199
|
end
|
@@ -289,7 +293,7 @@ module OpenAI
|
|
289
293
|
def self.variants
|
290
294
|
end
|
291
295
|
|
292
|
-
|
296
|
+
AnArrayOfInputTextInputImageAndInputAudioArray =
|
293
297
|
T.let(
|
294
298
|
OpenAI::Internal::Type::ArrayOf[
|
295
299
|
OpenAI::Internal::Type::Unknown
|
@@ -116,7 +116,7 @@ module OpenAI
|
|
116
116
|
# Learn more about
|
117
117
|
# [built-in tools](https://platform.openai.com/docs/guides/tools).
|
118
118
|
# - **MCP Tools**: Integrations with third-party systems via custom MCP servers or
|
119
|
-
# predefined connectors such as Google Drive and
|
119
|
+
# predefined connectors such as Google Drive and SharePoint. Learn more about
|
120
120
|
# [MCP Tools](https://platform.openai.com/docs/guides/tools-connectors-mcp).
|
121
121
|
# - **Function calls (custom tools)**: Functions that are defined by you, enabling
|
122
122
|
# the model to call your own code with strongly typed arguments and outputs.
|
@@ -438,7 +438,7 @@ module OpenAI
|
|
438
438
|
# Learn more about
|
439
439
|
# [built-in tools](https://platform.openai.com/docs/guides/tools).
|
440
440
|
# - **MCP Tools**: Integrations with third-party systems via custom MCP servers or
|
441
|
-
# predefined connectors such as Google Drive and
|
441
|
+
# predefined connectors such as Google Drive and SharePoint. Learn more about
|
442
442
|
# [MCP Tools](https://platform.openai.com/docs/guides/tools-connectors-mcp).
|
443
443
|
# - **Function calls (custom tools)**: Functions that are defined by you, enabling
|
444
444
|
# the model to call your own code with strongly typed arguments and outputs.
|
@@ -13,6 +13,7 @@ module OpenAI
|
|
13
13
|
OpenAI::Responses::ResponseInputText,
|
14
14
|
OpenAI::Responses::ResponseInputImage,
|
15
15
|
OpenAI::Responses::ResponseInputFile,
|
16
|
+
OpenAI::Responses::ResponseInputAudio,
|
16
17
|
OpenAI::Responses::ResponseOutputText,
|
17
18
|
OpenAI::Responses::ResponseOutputRefusal
|
18
19
|
)
|
@@ -313,7 +313,7 @@ module OpenAI
|
|
313
313
|
# Learn more about
|
314
314
|
# [built-in tools](https://platform.openai.com/docs/guides/tools).
|
315
315
|
# - **MCP Tools**: Integrations with third-party systems via custom MCP servers or
|
316
|
-
# predefined connectors such as Google Drive and
|
316
|
+
# predefined connectors such as Google Drive and SharePoint. Learn more about
|
317
317
|
# [MCP Tools](https://platform.openai.com/docs/guides/tools-connectors-mcp).
|
318
318
|
# - **Function calls (custom tools)**: Functions that are defined by you, enabling
|
319
319
|
# the model to call your own code with strongly typed arguments and outputs.
|
@@ -618,7 +618,7 @@ module OpenAI
|
|
618
618
|
# Learn more about
|
619
619
|
# [built-in tools](https://platform.openai.com/docs/guides/tools).
|
620
620
|
# - **MCP Tools**: Integrations with third-party systems via custom MCP servers or
|
621
|
-
# predefined connectors such as Google Drive and
|
621
|
+
# predefined connectors such as Google Drive and SharePoint. Learn more about
|
622
622
|
# [MCP Tools](https://platform.openai.com/docs/guides/tools-connectors-mcp).
|
623
623
|
# - **Function calls (custom tools)**: Functions that are defined by you, enabling
|
624
624
|
# the model to call your own code with strongly typed arguments and outputs.
|