openai 0.4.1 → 0.5.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +27 -0
- data/README.md +1 -1
- data/lib/openai/internal/util.rb +5 -1
- data/lib/openai/models/audio/transcription_text_delta_event.rb +3 -3
- data/lib/openai/models/audio/transcription_text_done_event.rb +3 -3
- data/lib/openai/models/chat/chat_completion.rb +4 -4
- data/lib/openai/models/chat/chat_completion_chunk.rb +4 -4
- data/lib/openai/models/chat/completion_create_params.rb +4 -4
- data/lib/openai/models/fine_tuning/alpha/grader_run_params.rb +17 -30
- data/lib/openai/models/fine_tuning/fine_tuning_job.rb +3 -5
- data/lib/openai/models/graders/multi_grader.rb +11 -4
- data/lib/openai/models/image_edit_params.rb +2 -2
- data/lib/openai/models/responses/response.rb +4 -4
- data/lib/openai/models/responses/response_code_interpreter_call_code_delta_event.rb +5 -5
- data/lib/openai/models/responses/response_code_interpreter_call_code_done_event.rb +5 -5
- data/lib/openai/models/responses/response_create_params.rb +6 -4
- data/lib/openai/models/responses/response_includable.rb +3 -0
- data/lib/openai/models/responses/response_output_text.rb +120 -4
- data/lib/openai/models/responses/response_retrieve_params.rb +11 -1
- data/lib/openai/models/responses/response_stream_event.rb +2 -2
- data/lib/openai/resources/beta/threads/messages.rb +11 -0
- data/lib/openai/resources/beta/threads/runs/steps.rb +5 -0
- data/lib/openai/resources/beta/threads/runs.rb +17 -0
- data/lib/openai/resources/beta/threads.rb +15 -2
- data/lib/openai/resources/containers/files/content.rb +3 -2
- data/lib/openai/resources/fine_tuning/alpha/graders.rb +6 -3
- data/lib/openai/resources/responses.rb +49 -3
- data/lib/openai/version.rb +1 -1
- data/lib/openai.rb +3 -1
- data/rbi/openai/models/audio/transcription_text_delta_event.rbi +4 -4
- data/rbi/openai/models/audio/transcription_text_done_event.rbi +4 -4
- data/rbi/openai/models/chat/chat_completion.rbi +6 -6
- data/rbi/openai/models/chat/chat_completion_chunk.rbi +6 -6
- data/rbi/openai/models/chat/completion_create_params.rbi +6 -6
- data/rbi/openai/models/fine_tuning/alpha/grader_run_params.rbi +24 -43
- data/rbi/openai/models/fine_tuning/fine_tuning_job.rbi +2 -3
- data/rbi/openai/models/graders/multi_grader.rbi +27 -32
- data/rbi/openai/models/image_edit_params.rbi +3 -3
- data/rbi/openai/models/responses/response.rbi +6 -6
- data/rbi/openai/models/responses/response_code_interpreter_call_code_delta_event.rbi +3 -3
- data/rbi/openai/models/responses/response_code_interpreter_call_code_done_event.rbi +3 -3
- data/rbi/openai/models/responses/response_create_params.rbi +10 -6
- data/rbi/openai/models/responses/response_includable.rbi +7 -0
- data/rbi/openai/models/responses/response_output_text.rbi +189 -1
- data/rbi/openai/models/responses/response_retrieve_params.rbi +11 -0
- data/rbi/openai/resources/chat/completions.rbi +4 -4
- data/rbi/openai/resources/containers/files/content.rbi +1 -1
- data/rbi/openai/resources/fine_tuning/alpha/graders.rbi +10 -5
- data/rbi/openai/resources/images.rbi +1 -1
- data/rbi/openai/resources/responses.rbi +49 -5
- data/sig/openai/models/audio/transcription_text_delta_event.rbs +10 -5
- data/sig/openai/models/audio/transcription_text_done_event.rbs +10 -5
- data/sig/openai/models/fine_tuning/alpha/grader_run_params.rbs +6 -14
- data/sig/openai/models/fine_tuning/fine_tuning_job.rbs +1 -1
- data/sig/openai/models/graders/multi_grader.rbs +7 -7
- data/sig/openai/models/responses/response_code_interpreter_call_code_delta_event.rbs +4 -4
- data/sig/openai/models/responses/response_code_interpreter_call_code_done_event.rbs +4 -4
- data/sig/openai/models/responses/response_includable.rbs +2 -0
- data/sig/openai/models/responses/response_output_text.rbs +104 -2
- data/sig/openai/models/responses/response_retrieve_params.rbs +10 -1
- data/sig/openai/resources/containers/files/content.rbs +1 -1
- data/sig/openai/resources/fine_tuning/alpha/graders.rbs +1 -1
- data/sig/openai/resources/responses.rbs +9 -1
- metadata +2 -2
@@ -30,17 +30,22 @@ module OpenAI
|
|
30
30
|
end
|
31
31
|
attr_accessor :grader
|
32
32
|
|
33
|
-
# The model sample to be evaluated.
|
33
|
+
# The model sample to be evaluated. This value will be used to populate the
|
34
|
+
# `sample` namespace. See
|
35
|
+
# [the guide](https://platform.openai.com/docs/guides/graders) for more details.
|
36
|
+
# The `output_json` variable will be populated if the model sample is a valid JSON
|
37
|
+
# string.
|
34
38
|
sig { returns(String) }
|
35
39
|
attr_accessor :model_sample
|
36
40
|
|
37
|
-
# The
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
42
|
-
|
43
|
-
|
41
|
+
# The dataset item provided to the grader. This will be used to populate the
|
42
|
+
# `item` namespace. See
|
43
|
+
# [the guide](https://platform.openai.com/docs/guides/graders) for more details.
|
44
|
+
sig { returns(T.nilable(T.anything)) }
|
45
|
+
attr_reader :item
|
46
|
+
|
47
|
+
sig { params(item: T.anything).void }
|
48
|
+
attr_writer :item
|
44
49
|
|
45
50
|
sig do
|
46
51
|
params(
|
@@ -53,18 +58,23 @@ module OpenAI
|
|
53
58
|
OpenAI::Graders::MultiGrader::OrHash
|
54
59
|
),
|
55
60
|
model_sample: String,
|
56
|
-
|
57
|
-
OpenAI::FineTuning::Alpha::GraderRunParams::ReferenceAnswer::Variants,
|
61
|
+
item: T.anything,
|
58
62
|
request_options: OpenAI::RequestOptions::OrHash
|
59
63
|
).returns(T.attached_class)
|
60
64
|
end
|
61
65
|
def self.new(
|
62
66
|
# The grader used for the fine-tuning job.
|
63
67
|
grader:,
|
64
|
-
# The model sample to be evaluated.
|
68
|
+
# The model sample to be evaluated. This value will be used to populate the
|
69
|
+
# `sample` namespace. See
|
70
|
+
# [the guide](https://platform.openai.com/docs/guides/graders) for more details.
|
71
|
+
# The `output_json` variable will be populated if the model sample is a valid JSON
|
72
|
+
# string.
|
65
73
|
model_sample:,
|
66
|
-
# The
|
67
|
-
|
74
|
+
# The dataset item provided to the grader. This will be used to populate the
|
75
|
+
# `item` namespace. See
|
76
|
+
# [the guide](https://platform.openai.com/docs/guides/graders) for more details.
|
77
|
+
item: nil,
|
68
78
|
request_options: {}
|
69
79
|
)
|
70
80
|
end
|
@@ -81,8 +91,7 @@ module OpenAI
|
|
81
91
|
OpenAI::Graders::MultiGrader
|
82
92
|
),
|
83
93
|
model_sample: String,
|
84
|
-
|
85
|
-
OpenAI::FineTuning::Alpha::GraderRunParams::ReferenceAnswer::Variants,
|
94
|
+
item: T.anything,
|
86
95
|
request_options: OpenAI::RequestOptions
|
87
96
|
}
|
88
97
|
)
|
@@ -115,34 +124,6 @@ module OpenAI
|
|
115
124
|
def self.variants
|
116
125
|
end
|
117
126
|
end
|
118
|
-
|
119
|
-
# The reference answer for the evaluation.
|
120
|
-
module ReferenceAnswer
|
121
|
-
extend OpenAI::Internal::Type::Union
|
122
|
-
|
123
|
-
Variants =
|
124
|
-
T.type_alias do
|
125
|
-
T.any(String, T.anything, T::Array[T.anything], Float)
|
126
|
-
end
|
127
|
-
|
128
|
-
sig do
|
129
|
-
override.returns(
|
130
|
-
T::Array[
|
131
|
-
OpenAI::FineTuning::Alpha::GraderRunParams::ReferenceAnswer::Variants
|
132
|
-
]
|
133
|
-
)
|
134
|
-
end
|
135
|
-
def self.variants
|
136
|
-
end
|
137
|
-
|
138
|
-
UnionMember2Array =
|
139
|
-
T.let(
|
140
|
-
OpenAI::Internal::Type::ArrayOf[
|
141
|
-
OpenAI::Internal::Type::Unknown
|
142
|
-
],
|
143
|
-
OpenAI::Internal::Type::Converter
|
144
|
-
)
|
145
|
-
end
|
146
127
|
end
|
147
128
|
end
|
148
129
|
end
|
@@ -359,7 +359,7 @@ module OpenAI
|
|
359
359
|
# returned when running `supervised` jobs.
|
360
360
|
sig do
|
361
361
|
params(
|
362
|
-
batch_size: T.nilable(T.any(
|
362
|
+
batch_size: T.nilable(T.any(Symbol, Integer)),
|
363
363
|
learning_rate_multiplier: T.any(Symbol, Float),
|
364
364
|
n_epochs: T.any(Symbol, Integer)
|
365
365
|
).returns(T.attached_class)
|
@@ -399,8 +399,7 @@ module OpenAI
|
|
399
399
|
module BatchSize
|
400
400
|
extend OpenAI::Internal::Type::Union
|
401
401
|
|
402
|
-
Variants =
|
403
|
-
T.type_alias { T.nilable(T.any(T.anything, Symbol, Integer)) }
|
402
|
+
Variants = T.type_alias { T.any(Symbol, Integer) }
|
404
403
|
|
405
404
|
sig do
|
406
405
|
override.returns(
|
@@ -15,18 +15,17 @@ module OpenAI
|
|
15
15
|
sig { returns(String) }
|
16
16
|
attr_accessor :calculate_output
|
17
17
|
|
18
|
+
# A StringCheckGrader object that performs a string comparison between input and
|
19
|
+
# reference using a specified operation.
|
18
20
|
sig do
|
19
21
|
returns(
|
20
|
-
T
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
27
|
-
OpenAI::Graders::LabelModelGrader
|
28
|
-
)
|
29
|
-
]
|
22
|
+
T.any(
|
23
|
+
OpenAI::Graders::StringCheckGrader,
|
24
|
+
OpenAI::Graders::TextSimilarityGrader,
|
25
|
+
OpenAI::Graders::PythonGrader,
|
26
|
+
OpenAI::Graders::ScoreModelGrader,
|
27
|
+
OpenAI::Graders::LabelModelGrader
|
28
|
+
)
|
30
29
|
)
|
31
30
|
end
|
32
31
|
attr_accessor :graders
|
@@ -45,16 +44,13 @@ module OpenAI
|
|
45
44
|
params(
|
46
45
|
calculate_output: String,
|
47
46
|
graders:
|
48
|
-
T
|
49
|
-
|
50
|
-
|
51
|
-
|
52
|
-
|
53
|
-
|
54
|
-
|
55
|
-
OpenAI::Graders::LabelModelGrader::OrHash
|
56
|
-
)
|
57
|
-
],
|
47
|
+
T.any(
|
48
|
+
OpenAI::Graders::StringCheckGrader::OrHash,
|
49
|
+
OpenAI::Graders::TextSimilarityGrader::OrHash,
|
50
|
+
OpenAI::Graders::PythonGrader::OrHash,
|
51
|
+
OpenAI::Graders::ScoreModelGrader::OrHash,
|
52
|
+
OpenAI::Graders::LabelModelGrader::OrHash
|
53
|
+
),
|
58
54
|
name: String,
|
59
55
|
type: Symbol
|
60
56
|
).returns(T.attached_class)
|
@@ -62,6 +58,8 @@ module OpenAI
|
|
62
58
|
def self.new(
|
63
59
|
# A formula to calculate the output based on grader results.
|
64
60
|
calculate_output:,
|
61
|
+
# A StringCheckGrader object that performs a string comparison between input and
|
62
|
+
# reference using a specified operation.
|
65
63
|
graders:,
|
66
64
|
# The name of the grader.
|
67
65
|
name:,
|
@@ -75,16 +73,13 @@ module OpenAI
|
|
75
73
|
{
|
76
74
|
calculate_output: String,
|
77
75
|
graders:
|
78
|
-
T
|
79
|
-
|
80
|
-
|
81
|
-
|
82
|
-
|
83
|
-
|
84
|
-
|
85
|
-
OpenAI::Graders::LabelModelGrader
|
86
|
-
)
|
87
|
-
],
|
76
|
+
T.any(
|
77
|
+
OpenAI::Graders::StringCheckGrader,
|
78
|
+
OpenAI::Graders::TextSimilarityGrader,
|
79
|
+
OpenAI::Graders::PythonGrader,
|
80
|
+
OpenAI::Graders::ScoreModelGrader,
|
81
|
+
OpenAI::Graders::LabelModelGrader
|
82
|
+
),
|
88
83
|
name: String,
|
89
84
|
type: Symbol
|
90
85
|
}
|
@@ -95,7 +90,7 @@ module OpenAI
|
|
95
90
|
|
96
91
|
# A StringCheckGrader object that performs a string comparison between input and
|
97
92
|
# reference using a specified operation.
|
98
|
-
module
|
93
|
+
module Graders
|
99
94
|
extend OpenAI::Internal::Type::Union
|
100
95
|
|
101
96
|
Variants =
|
@@ -111,7 +106,7 @@ module OpenAI
|
|
111
106
|
|
112
107
|
sig do
|
113
108
|
override.returns(
|
114
|
-
T::Array[OpenAI::Graders::MultiGrader::
|
109
|
+
T::Array[OpenAI::Graders::MultiGrader::Graders::Variants]
|
115
110
|
)
|
116
111
|
end
|
117
112
|
def self.variants
|
@@ -14,7 +14,7 @@ module OpenAI
|
|
14
14
|
# The image(s) to edit. Must be a supported image file or an array of images.
|
15
15
|
#
|
16
16
|
# For `gpt-image-1`, each image should be a `png`, `webp`, or `jpg` file less than
|
17
|
-
#
|
17
|
+
# 50MB. You can provide up to 16 images.
|
18
18
|
#
|
19
19
|
# For `dall-e-2`, you can only provide one image, and it should be a square `png`
|
20
20
|
# file less than 4MB.
|
@@ -106,7 +106,7 @@ module OpenAI
|
|
106
106
|
# The image(s) to edit. Must be a supported image file or an array of images.
|
107
107
|
#
|
108
108
|
# For `gpt-image-1`, each image should be a `png`, `webp`, or `jpg` file less than
|
109
|
-
#
|
109
|
+
# 50MB. You can provide up to 16 images.
|
110
110
|
#
|
111
111
|
# For `dall-e-2`, you can only provide one image, and it should be a square `png`
|
112
112
|
# file less than 4MB.
|
@@ -179,7 +179,7 @@ module OpenAI
|
|
179
179
|
# The image(s) to edit. Must be a supported image file or an array of images.
|
180
180
|
#
|
181
181
|
# For `gpt-image-1`, each image should be a `png`, `webp`, or `jpg` file less than
|
182
|
-
#
|
182
|
+
# 50MB. You can provide up to 16 images.
|
183
183
|
#
|
184
184
|
# For `dall-e-2`, you can only provide one image, and it should be a square `png`
|
185
185
|
# file less than 4MB.
|
@@ -160,9 +160,9 @@ module OpenAI
|
|
160
160
|
# utilize scale tier credits until they are exhausted.
|
161
161
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
162
162
|
# be processed using the default service tier with a lower uptime SLA and no
|
163
|
-
# latency
|
163
|
+
# latency guarantee.
|
164
164
|
# - If set to 'default', the request will be processed using the default service
|
165
|
-
# tier with a lower uptime SLA and no latency
|
165
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
166
166
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
167
167
|
# service tier.
|
168
168
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -390,9 +390,9 @@ module OpenAI
|
|
390
390
|
# utilize scale tier credits until they are exhausted.
|
391
391
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
392
392
|
# be processed using the default service tier with a lower uptime SLA and no
|
393
|
-
# latency
|
393
|
+
# latency guarantee.
|
394
394
|
# - If set to 'default', the request will be processed using the default service
|
395
|
-
# tier with a lower uptime SLA and no latency
|
395
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
396
396
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
397
397
|
# service tier.
|
398
398
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -588,9 +588,9 @@ module OpenAI
|
|
588
588
|
# utilize scale tier credits until they are exhausted.
|
589
589
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
590
590
|
# be processed using the default service tier with a lower uptime SLA and no
|
591
|
-
# latency
|
591
|
+
# latency guarantee.
|
592
592
|
# - If set to 'default', the request will be processed using the default service
|
593
|
-
# tier with a lower uptime SLA and no latency
|
593
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
594
594
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
595
595
|
# service tier.
|
596
596
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -24,7 +24,7 @@ module OpenAI
|
|
24
24
|
sig { returns(Integer) }
|
25
25
|
attr_accessor :sequence_number
|
26
26
|
|
27
|
-
# The type of the event. Always `response.
|
27
|
+
# The type of the event. Always `response.code_interpreter_call_code.delta`.
|
28
28
|
sig { returns(Symbol) }
|
29
29
|
attr_accessor :type
|
30
30
|
|
@@ -44,8 +44,8 @@ module OpenAI
|
|
44
44
|
output_index:,
|
45
45
|
# The sequence number of this event.
|
46
46
|
sequence_number:,
|
47
|
-
# The type of the event. Always `response.
|
48
|
-
type: :"response.
|
47
|
+
# The type of the event. Always `response.code_interpreter_call_code.delta`.
|
48
|
+
type: :"response.code_interpreter_call_code.delta"
|
49
49
|
)
|
50
50
|
end
|
51
51
|
|
@@ -24,7 +24,7 @@ module OpenAI
|
|
24
24
|
sig { returns(Integer) }
|
25
25
|
attr_accessor :sequence_number
|
26
26
|
|
27
|
-
# The type of the event. Always `response.
|
27
|
+
# The type of the event. Always `response.code_interpreter_call_code.done`.
|
28
28
|
sig { returns(Symbol) }
|
29
29
|
attr_accessor :type
|
30
30
|
|
@@ -44,8 +44,8 @@ module OpenAI
|
|
44
44
|
output_index:,
|
45
45
|
# The sequence number of this event.
|
46
46
|
sequence_number:,
|
47
|
-
# The type of the event. Always `response.
|
48
|
-
type: :"response.
|
47
|
+
# The type of the event. Always `response.code_interpreter_call_code.done`.
|
48
|
+
type: :"response.code_interpreter_call_code.done"
|
49
49
|
)
|
50
50
|
end
|
51
51
|
|
@@ -63,6 +63,8 @@ module OpenAI
|
|
63
63
|
# multi-turn conversations when using the Responses API statelessly (like when
|
64
64
|
# the `store` parameter is set to `false`, or when an organization is enrolled
|
65
65
|
# in the zero data retention program).
|
66
|
+
# - `code_interpreter_call.outputs`: Includes the outputs of python code execution
|
67
|
+
# in code interpreter tool call items.
|
66
68
|
sig do
|
67
69
|
returns(
|
68
70
|
T.nilable(T::Array[OpenAI::Responses::ResponseIncludable::OrSymbol])
|
@@ -121,9 +123,9 @@ module OpenAI
|
|
121
123
|
# utilize scale tier credits until they are exhausted.
|
122
124
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
123
125
|
# be processed using the default service tier with a lower uptime SLA and no
|
124
|
-
# latency
|
126
|
+
# latency guarantee.
|
125
127
|
# - If set to 'default', the request will be processed using the default service
|
126
|
-
# tier with a lower uptime SLA and no latency
|
128
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
127
129
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
128
130
|
# service tier.
|
129
131
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -372,6 +374,8 @@ module OpenAI
|
|
372
374
|
# multi-turn conversations when using the Responses API statelessly (like when
|
373
375
|
# the `store` parameter is set to `false`, or when an organization is enrolled
|
374
376
|
# in the zero data retention program).
|
377
|
+
# - `code_interpreter_call.outputs`: Includes the outputs of python code execution
|
378
|
+
# in code interpreter tool call items.
|
375
379
|
include: nil,
|
376
380
|
# Inserts a system (or developer) message as the first item in the model's
|
377
381
|
# context.
|
@@ -409,9 +413,9 @@ module OpenAI
|
|
409
413
|
# utilize scale tier credits until they are exhausted.
|
410
414
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
411
415
|
# be processed using the default service tier with a lower uptime SLA and no
|
412
|
-
# latency
|
416
|
+
# latency guarantee.
|
413
417
|
# - If set to 'default', the request will be processed using the default service
|
414
|
-
# tier with a lower uptime SLA and no latency
|
418
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
415
419
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
416
420
|
# service tier.
|
417
421
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -570,9 +574,9 @@ module OpenAI
|
|
570
574
|
# utilize scale tier credits until they are exhausted.
|
571
575
|
# - If set to 'auto', and the Project is not Scale tier enabled, the request will
|
572
576
|
# be processed using the default service tier with a lower uptime SLA and no
|
573
|
-
# latency
|
577
|
+
# latency guarantee.
|
574
578
|
# - If set to 'default', the request will be processed using the default service
|
575
|
-
# tier with a lower uptime SLA and no latency
|
579
|
+
# tier with a lower uptime SLA and no latency guarantee.
|
576
580
|
# - If set to 'flex', the request will be processed with the Flex Processing
|
577
581
|
# service tier.
|
578
582
|
# [Learn more](https://platform.openai.com/docs/guides/flex-processing).
|
@@ -16,6 +16,8 @@ module OpenAI
|
|
16
16
|
# multi-turn conversations when using the Responses API statelessly (like when
|
17
17
|
# the `store` parameter is set to `false`, or when an organization is enrolled
|
18
18
|
# in the zero data retention program).
|
19
|
+
# - `code_interpreter_call.outputs`: Includes the outputs of python code execution
|
20
|
+
# in code interpreter tool call items.
|
19
21
|
module ResponseIncludable
|
20
22
|
extend OpenAI::Internal::Type::Enum
|
21
23
|
|
@@ -43,6 +45,11 @@ module OpenAI
|
|
43
45
|
:"reasoning.encrypted_content",
|
44
46
|
OpenAI::Responses::ResponseIncludable::TaggedSymbol
|
45
47
|
)
|
48
|
+
CODE_INTERPRETER_CALL_OUTPUTS =
|
49
|
+
T.let(
|
50
|
+
:"code_interpreter_call.outputs",
|
51
|
+
OpenAI::Responses::ResponseIncludable::TaggedSymbol
|
52
|
+
)
|
46
53
|
|
47
54
|
sig do
|
48
55
|
override.returns(
|
@@ -19,6 +19,7 @@ module OpenAI
|
|
19
19
|
T.any(
|
20
20
|
OpenAI::Responses::ResponseOutputText::Annotation::FileCitation,
|
21
21
|
OpenAI::Responses::ResponseOutputText::Annotation::URLCitation,
|
22
|
+
OpenAI::Responses::ResponseOutputText::Annotation::ContainerFileCitation,
|
22
23
|
OpenAI::Responses::ResponseOutputText::Annotation::FilePath
|
23
24
|
)
|
24
25
|
]
|
@@ -38,6 +39,21 @@ module OpenAI
|
|
38
39
|
sig { returns(Symbol) }
|
39
40
|
attr_accessor :type
|
40
41
|
|
42
|
+
sig do
|
43
|
+
returns(
|
44
|
+
T.nilable(T::Array[OpenAI::Responses::ResponseOutputText::Logprob])
|
45
|
+
)
|
46
|
+
end
|
47
|
+
attr_reader :logprobs
|
48
|
+
|
49
|
+
sig do
|
50
|
+
params(
|
51
|
+
logprobs:
|
52
|
+
T::Array[OpenAI::Responses::ResponseOutputText::Logprob::OrHash]
|
53
|
+
).void
|
54
|
+
end
|
55
|
+
attr_writer :logprobs
|
56
|
+
|
41
57
|
# A text output from the model.
|
42
58
|
sig do
|
43
59
|
params(
|
@@ -46,10 +62,13 @@ module OpenAI
|
|
46
62
|
T.any(
|
47
63
|
OpenAI::Responses::ResponseOutputText::Annotation::FileCitation::OrHash,
|
48
64
|
OpenAI::Responses::ResponseOutputText::Annotation::URLCitation::OrHash,
|
65
|
+
OpenAI::Responses::ResponseOutputText::Annotation::ContainerFileCitation::OrHash,
|
49
66
|
OpenAI::Responses::ResponseOutputText::Annotation::FilePath::OrHash
|
50
67
|
)
|
51
68
|
],
|
52
69
|
text: String,
|
70
|
+
logprobs:
|
71
|
+
T::Array[OpenAI::Responses::ResponseOutputText::Logprob::OrHash],
|
53
72
|
type: Symbol
|
54
73
|
).returns(T.attached_class)
|
55
74
|
end
|
@@ -58,6 +77,7 @@ module OpenAI
|
|
58
77
|
annotations:,
|
59
78
|
# The text output from the model.
|
60
79
|
text:,
|
80
|
+
logprobs: nil,
|
61
81
|
# The type of the output text. Always `output_text`.
|
62
82
|
type: :output_text
|
63
83
|
)
|
@@ -71,11 +91,13 @@ module OpenAI
|
|
71
91
|
T.any(
|
72
92
|
OpenAI::Responses::ResponseOutputText::Annotation::FileCitation,
|
73
93
|
OpenAI::Responses::ResponseOutputText::Annotation::URLCitation,
|
94
|
+
OpenAI::Responses::ResponseOutputText::Annotation::ContainerFileCitation,
|
74
95
|
OpenAI::Responses::ResponseOutputText::Annotation::FilePath
|
75
96
|
)
|
76
97
|
],
|
77
98
|
text: String,
|
78
|
-
type: Symbol
|
99
|
+
type: Symbol,
|
100
|
+
logprobs: T::Array[OpenAI::Responses::ResponseOutputText::Logprob]
|
79
101
|
}
|
80
102
|
)
|
81
103
|
end
|
@@ -91,6 +113,7 @@ module OpenAI
|
|
91
113
|
T.any(
|
92
114
|
OpenAI::Responses::ResponseOutputText::Annotation::FileCitation,
|
93
115
|
OpenAI::Responses::ResponseOutputText::Annotation::URLCitation,
|
116
|
+
OpenAI::Responses::ResponseOutputText::Annotation::ContainerFileCitation,
|
94
117
|
OpenAI::Responses::ResponseOutputText::Annotation::FilePath
|
95
118
|
)
|
96
119
|
end
|
@@ -209,6 +232,74 @@ module OpenAI
|
|
209
232
|
end
|
210
233
|
end
|
211
234
|
|
235
|
+
class ContainerFileCitation < OpenAI::Internal::Type::BaseModel
|
236
|
+
OrHash =
|
237
|
+
T.type_alias do
|
238
|
+
T.any(
|
239
|
+
OpenAI::Responses::ResponseOutputText::Annotation::ContainerFileCitation,
|
240
|
+
OpenAI::Internal::AnyHash
|
241
|
+
)
|
242
|
+
end
|
243
|
+
|
244
|
+
# The ID of the container file.
|
245
|
+
sig { returns(String) }
|
246
|
+
attr_accessor :container_id
|
247
|
+
|
248
|
+
# The index of the last character of the container file citation in the message.
|
249
|
+
sig { returns(Integer) }
|
250
|
+
attr_accessor :end_index
|
251
|
+
|
252
|
+
# The ID of the file.
|
253
|
+
sig { returns(String) }
|
254
|
+
attr_accessor :file_id
|
255
|
+
|
256
|
+
# The index of the first character of the container file citation in the message.
|
257
|
+
sig { returns(Integer) }
|
258
|
+
attr_accessor :start_index
|
259
|
+
|
260
|
+
# The type of the container file citation. Always `container_file_citation`.
|
261
|
+
sig { returns(Symbol) }
|
262
|
+
attr_accessor :type
|
263
|
+
|
264
|
+
# A citation for a container file used to generate a model response.
|
265
|
+
sig do
|
266
|
+
params(
|
267
|
+
container_id: String,
|
268
|
+
end_index: Integer,
|
269
|
+
file_id: String,
|
270
|
+
start_index: Integer,
|
271
|
+
type: Symbol
|
272
|
+
).returns(T.attached_class)
|
273
|
+
end
|
274
|
+
def self.new(
|
275
|
+
# The ID of the container file.
|
276
|
+
container_id:,
|
277
|
+
# The index of the last character of the container file citation in the message.
|
278
|
+
end_index:,
|
279
|
+
# The ID of the file.
|
280
|
+
file_id:,
|
281
|
+
# The index of the first character of the container file citation in the message.
|
282
|
+
start_index:,
|
283
|
+
# The type of the container file citation. Always `container_file_citation`.
|
284
|
+
type: :container_file_citation
|
285
|
+
)
|
286
|
+
end
|
287
|
+
|
288
|
+
sig do
|
289
|
+
override.returns(
|
290
|
+
{
|
291
|
+
container_id: String,
|
292
|
+
end_index: Integer,
|
293
|
+
file_id: String,
|
294
|
+
start_index: Integer,
|
295
|
+
type: Symbol
|
296
|
+
}
|
297
|
+
)
|
298
|
+
end
|
299
|
+
def to_hash
|
300
|
+
end
|
301
|
+
end
|
302
|
+
|
212
303
|
class FilePath < OpenAI::Internal::Type::BaseModel
|
213
304
|
OrHash =
|
214
305
|
T.type_alias do
|
@@ -265,6 +356,103 @@ module OpenAI
|
|
265
356
|
def self.variants
|
266
357
|
end
|
267
358
|
end
|
359
|
+
|
360
|
+
class Logprob < OpenAI::Internal::Type::BaseModel
|
361
|
+
OrHash =
|
362
|
+
T.type_alias do
|
363
|
+
T.any(
|
364
|
+
OpenAI::Responses::ResponseOutputText::Logprob,
|
365
|
+
OpenAI::Internal::AnyHash
|
366
|
+
)
|
367
|
+
end
|
368
|
+
|
369
|
+
sig { returns(String) }
|
370
|
+
attr_accessor :token
|
371
|
+
|
372
|
+
sig { returns(T::Array[Integer]) }
|
373
|
+
attr_accessor :bytes
|
374
|
+
|
375
|
+
sig { returns(Float) }
|
376
|
+
attr_accessor :logprob
|
377
|
+
|
378
|
+
sig do
|
379
|
+
returns(
|
380
|
+
T::Array[
|
381
|
+
OpenAI::Responses::ResponseOutputText::Logprob::TopLogprob
|
382
|
+
]
|
383
|
+
)
|
384
|
+
end
|
385
|
+
attr_accessor :top_logprobs
|
386
|
+
|
387
|
+
# The log probability of a token.
|
388
|
+
sig do
|
389
|
+
params(
|
390
|
+
token: String,
|
391
|
+
bytes: T::Array[Integer],
|
392
|
+
logprob: Float,
|
393
|
+
top_logprobs:
|
394
|
+
T::Array[
|
395
|
+
OpenAI::Responses::ResponseOutputText::Logprob::TopLogprob::OrHash
|
396
|
+
]
|
397
|
+
).returns(T.attached_class)
|
398
|
+
end
|
399
|
+
def self.new(token:, bytes:, logprob:, top_logprobs:)
|
400
|
+
end
|
401
|
+
|
402
|
+
sig do
|
403
|
+
override.returns(
|
404
|
+
{
|
405
|
+
token: String,
|
406
|
+
bytes: T::Array[Integer],
|
407
|
+
logprob: Float,
|
408
|
+
top_logprobs:
|
409
|
+
T::Array[
|
410
|
+
OpenAI::Responses::ResponseOutputText::Logprob::TopLogprob
|
411
|
+
]
|
412
|
+
}
|
413
|
+
)
|
414
|
+
end
|
415
|
+
def to_hash
|
416
|
+
end
|
417
|
+
|
418
|
+
class TopLogprob < OpenAI::Internal::Type::BaseModel
|
419
|
+
OrHash =
|
420
|
+
T.type_alias do
|
421
|
+
T.any(
|
422
|
+
OpenAI::Responses::ResponseOutputText::Logprob::TopLogprob,
|
423
|
+
OpenAI::Internal::AnyHash
|
424
|
+
)
|
425
|
+
end
|
426
|
+
|
427
|
+
sig { returns(String) }
|
428
|
+
attr_accessor :token
|
429
|
+
|
430
|
+
sig { returns(T::Array[Integer]) }
|
431
|
+
attr_accessor :bytes
|
432
|
+
|
433
|
+
sig { returns(Float) }
|
434
|
+
attr_accessor :logprob
|
435
|
+
|
436
|
+
# The top log probability of a token.
|
437
|
+
sig do
|
438
|
+
params(
|
439
|
+
token: String,
|
440
|
+
bytes: T::Array[Integer],
|
441
|
+
logprob: Float
|
442
|
+
).returns(T.attached_class)
|
443
|
+
end
|
444
|
+
def self.new(token:, bytes:, logprob:)
|
445
|
+
end
|
446
|
+
|
447
|
+
sig do
|
448
|
+
override.returns(
|
449
|
+
{ token: String, bytes: T::Array[Integer], logprob: Float }
|
450
|
+
)
|
451
|
+
end
|
452
|
+
def to_hash
|
453
|
+
end
|
454
|
+
end
|
455
|
+
end
|
268
456
|
end
|
269
457
|
end
|
270
458
|
end
|