openai 0.31.0 → 0.33.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +22 -0
- data/README.md +1 -1
- data/lib/openai/internal/util.rb +5 -5
- data/lib/openai/models/audio/transcription_create_params.rb +42 -11
- data/lib/openai/models/audio/transcription_create_response.rb +4 -1
- data/lib/openai/models/audio/transcription_diarized.rb +160 -0
- data/lib/openai/models/audio/transcription_diarized_segment.rb +65 -0
- data/lib/openai/models/audio/transcription_stream_event.rb +7 -4
- data/lib/openai/models/audio/transcription_text_delta_event.rb +10 -1
- data/lib/openai/models/audio/transcription_text_segment_event.rb +63 -0
- data/lib/openai/models/audio_model.rb +1 -0
- data/lib/openai/models/audio_response_format.rb +5 -2
- data/lib/openai/models/realtime/audio_transcription.rb +8 -6
- data/lib/openai/models/vector_store_create_params.rb +10 -1
- data/lib/openai/resources/audio/transcriptions.rb +12 -4
- data/lib/openai/resources/beta/chatkit.rb +0 -26
- data/lib/openai/resources/vector_stores.rb +3 -1
- data/lib/openai/version.rb +1 -1
- data/lib/openai.rb +3 -4
- data/rbi/openai/models/audio/transcription_create_params.rbi +66 -16
- data/rbi/openai/models/audio/transcription_create_response.rbi +1 -0
- data/rbi/openai/models/audio/transcription_diarized.rbi +281 -0
- data/rbi/openai/models/audio/transcription_diarized_segment.rbi +87 -0
- data/rbi/openai/models/audio/transcription_stream_event.rbi +4 -3
- data/rbi/openai/models/audio/transcription_text_delta_event.rbi +14 -1
- data/rbi/openai/models/audio/transcription_text_segment_event.rbi +86 -0
- data/rbi/openai/models/audio_model.rbi +2 -0
- data/rbi/openai/models/audio_response_format.rbi +6 -2
- data/rbi/openai/models/realtime/audio_transcription.rbi +15 -12
- data/rbi/openai/models/vector_store_create_params.rbi +13 -0
- data/rbi/openai/resources/audio/transcriptions.rbi +52 -14
- data/rbi/openai/resources/beta/chatkit.rbi +0 -15
- data/rbi/openai/resources/vector_stores.rbi +4 -0
- data/sig/openai/models/audio/transcription_create_params.rbs +14 -0
- data/sig/openai/models/audio/transcription_create_response.rbs +3 -1
- data/sig/openai/models/audio/transcription_diarized.rbs +129 -0
- data/sig/openai/models/audio/transcription_diarized_segment.rbs +47 -0
- data/sig/openai/models/audio/transcription_stream_event.rbs +2 -1
- data/sig/openai/models/audio/transcription_text_delta_event.rbs +9 -2
- data/sig/openai/models/audio/transcription_text_segment_event.rbs +47 -0
- data/sig/openai/models/audio_model.rbs +5 -1
- data/sig/openai/models/audio_response_format.rbs +3 -1
- data/sig/openai/models/realtime/audio_transcription.rbs +2 -2
- data/sig/openai/models/vector_store_create_params.rbs +7 -0
- data/sig/openai/resources/audio/transcriptions.rbs +4 -0
- data/sig/openai/resources/beta/chatkit.rbs +0 -5
- data/sig/openai/resources/vector_stores.rbs +1 -0
- metadata +11 -14
- data/lib/openai/models/beta/chatkit_upload_file_params.rb +0 -28
- data/lib/openai/models/beta/chatkit_upload_file_response.rb +0 -25
- data/lib/openai/models/beta/file_part.rb +0 -56
- data/lib/openai/models/beta/image_part.rb +0 -64
- data/rbi/openai/models/beta/chatkit_upload_file_params.rbi +0 -50
- data/rbi/openai/models/beta/chatkit_upload_file_response.rbi +0 -25
- data/rbi/openai/models/beta/file_part.rbi +0 -74
- data/rbi/openai/models/beta/image_part.rbi +0 -82
- data/sig/openai/models/beta/chatkit_upload_file_params.rbs +0 -26
- data/sig/openai/models/beta/chatkit_upload_file_response.rbs +0 -14
- data/sig/openai/models/beta/file_part.rbs +0 -42
- data/sig/openai/models/beta/image_part.rbs +0 -47
@@ -0,0 +1,129 @@
|
|
1
|
+
module OpenAI
|
2
|
+
module Models
|
3
|
+
module Audio
|
4
|
+
type transcription_diarized =
|
5
|
+
{
|
6
|
+
duration: Float,
|
7
|
+
segments: ::Array[OpenAI::Audio::TranscriptionDiarizedSegment],
|
8
|
+
task: :transcribe,
|
9
|
+
text: String,
|
10
|
+
usage: OpenAI::Models::Audio::TranscriptionDiarized::usage
|
11
|
+
}
|
12
|
+
|
13
|
+
class TranscriptionDiarized < OpenAI::Internal::Type::BaseModel
|
14
|
+
attr_accessor duration: Float
|
15
|
+
|
16
|
+
attr_accessor segments: ::Array[OpenAI::Audio::TranscriptionDiarizedSegment]
|
17
|
+
|
18
|
+
attr_accessor task: :transcribe
|
19
|
+
|
20
|
+
attr_accessor text: String
|
21
|
+
|
22
|
+
attr_reader usage: OpenAI::Models::Audio::TranscriptionDiarized::usage?
|
23
|
+
|
24
|
+
def usage=: (
|
25
|
+
OpenAI::Models::Audio::TranscriptionDiarized::usage
|
26
|
+
) -> OpenAI::Models::Audio::TranscriptionDiarized::usage
|
27
|
+
|
28
|
+
def initialize: (
|
29
|
+
duration: Float,
|
30
|
+
segments: ::Array[OpenAI::Audio::TranscriptionDiarizedSegment],
|
31
|
+
text: String,
|
32
|
+
?usage: OpenAI::Models::Audio::TranscriptionDiarized::usage,
|
33
|
+
?task: :transcribe
|
34
|
+
) -> void
|
35
|
+
|
36
|
+
def to_hash: -> {
|
37
|
+
duration: Float,
|
38
|
+
segments: ::Array[OpenAI::Audio::TranscriptionDiarizedSegment],
|
39
|
+
task: :transcribe,
|
40
|
+
text: String,
|
41
|
+
usage: OpenAI::Models::Audio::TranscriptionDiarized::usage
|
42
|
+
}
|
43
|
+
|
44
|
+
type usage =
|
45
|
+
OpenAI::Audio::TranscriptionDiarized::Usage::Tokens
|
46
|
+
| OpenAI::Audio::TranscriptionDiarized::Usage::Duration
|
47
|
+
|
48
|
+
module Usage
|
49
|
+
extend OpenAI::Internal::Type::Union
|
50
|
+
|
51
|
+
type tokens =
|
52
|
+
{
|
53
|
+
input_tokens: Integer,
|
54
|
+
output_tokens: Integer,
|
55
|
+
total_tokens: Integer,
|
56
|
+
type: :tokens,
|
57
|
+
input_token_details: OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails
|
58
|
+
}
|
59
|
+
|
60
|
+
class Tokens < OpenAI::Internal::Type::BaseModel
|
61
|
+
attr_accessor input_tokens: Integer
|
62
|
+
|
63
|
+
attr_accessor output_tokens: Integer
|
64
|
+
|
65
|
+
attr_accessor total_tokens: Integer
|
66
|
+
|
67
|
+
attr_accessor type: :tokens
|
68
|
+
|
69
|
+
attr_reader input_token_details: OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails?
|
70
|
+
|
71
|
+
def input_token_details=: (
|
72
|
+
OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails
|
73
|
+
) -> OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails
|
74
|
+
|
75
|
+
def initialize: (
|
76
|
+
input_tokens: Integer,
|
77
|
+
output_tokens: Integer,
|
78
|
+
total_tokens: Integer,
|
79
|
+
?input_token_details: OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails,
|
80
|
+
?type: :tokens
|
81
|
+
) -> void
|
82
|
+
|
83
|
+
def to_hash: -> {
|
84
|
+
input_tokens: Integer,
|
85
|
+
output_tokens: Integer,
|
86
|
+
total_tokens: Integer,
|
87
|
+
type: :tokens,
|
88
|
+
input_token_details: OpenAI::Audio::TranscriptionDiarized::Usage::Tokens::InputTokenDetails
|
89
|
+
}
|
90
|
+
|
91
|
+
type input_token_details =
|
92
|
+
{ audio_tokens: Integer, text_tokens: Integer }
|
93
|
+
|
94
|
+
class InputTokenDetails < OpenAI::Internal::Type::BaseModel
|
95
|
+
attr_reader audio_tokens: Integer?
|
96
|
+
|
97
|
+
def audio_tokens=: (Integer) -> Integer
|
98
|
+
|
99
|
+
attr_reader text_tokens: Integer?
|
100
|
+
|
101
|
+
def text_tokens=: (Integer) -> Integer
|
102
|
+
|
103
|
+
def initialize: (
|
104
|
+
?audio_tokens: Integer,
|
105
|
+
?text_tokens: Integer
|
106
|
+
) -> void
|
107
|
+
|
108
|
+
def to_hash: -> { audio_tokens: Integer, text_tokens: Integer }
|
109
|
+
end
|
110
|
+
end
|
111
|
+
|
112
|
+
type duration = { seconds: Float, type: :duration }
|
113
|
+
|
114
|
+
class Duration < OpenAI::Internal::Type::BaseModel
|
115
|
+
attr_accessor seconds: Float
|
116
|
+
|
117
|
+
attr_accessor type: :duration
|
118
|
+
|
119
|
+
def initialize: (seconds: Float, ?type: :duration) -> void
|
120
|
+
|
121
|
+
def to_hash: -> { seconds: Float, type: :duration }
|
122
|
+
end
|
123
|
+
|
124
|
+
def self?.variants: -> ::Array[OpenAI::Models::Audio::TranscriptionDiarized::usage]
|
125
|
+
end
|
126
|
+
end
|
127
|
+
end
|
128
|
+
end
|
129
|
+
end
|
@@ -0,0 +1,47 @@
|
|
1
|
+
module OpenAI
|
2
|
+
module Models
|
3
|
+
module Audio
|
4
|
+
type transcription_diarized_segment =
|
5
|
+
{
|
6
|
+
id: String,
|
7
|
+
end_: Float,
|
8
|
+
speaker: String,
|
9
|
+
start: Float,
|
10
|
+
text: String,
|
11
|
+
type: :"transcript.text.segment"
|
12
|
+
}
|
13
|
+
|
14
|
+
class TranscriptionDiarizedSegment < OpenAI::Internal::Type::BaseModel
|
15
|
+
attr_accessor id: String
|
16
|
+
|
17
|
+
attr_accessor end_: Float
|
18
|
+
|
19
|
+
attr_accessor speaker: String
|
20
|
+
|
21
|
+
attr_accessor start: Float
|
22
|
+
|
23
|
+
attr_accessor text: String
|
24
|
+
|
25
|
+
attr_accessor type: :"transcript.text.segment"
|
26
|
+
|
27
|
+
def initialize: (
|
28
|
+
id: String,
|
29
|
+
end_: Float,
|
30
|
+
speaker: String,
|
31
|
+
start: Float,
|
32
|
+
text: String,
|
33
|
+
?type: :"transcript.text.segment"
|
34
|
+
) -> void
|
35
|
+
|
36
|
+
def to_hash: -> {
|
37
|
+
id: String,
|
38
|
+
end_: Float,
|
39
|
+
speaker: String,
|
40
|
+
start: Float,
|
41
|
+
text: String,
|
42
|
+
type: :"transcript.text.segment"
|
43
|
+
}
|
44
|
+
end
|
45
|
+
end
|
46
|
+
end
|
47
|
+
end
|
@@ -2,7 +2,8 @@ module OpenAI
|
|
2
2
|
module Models
|
3
3
|
module Audio
|
4
4
|
type transcription_stream_event =
|
5
|
-
OpenAI::Audio::
|
5
|
+
OpenAI::Audio::TranscriptionTextSegmentEvent
|
6
|
+
| OpenAI::Audio::TranscriptionTextDeltaEvent
|
6
7
|
| OpenAI::Audio::TranscriptionTextDoneEvent
|
7
8
|
|
8
9
|
module TranscriptionStreamEvent
|
@@ -5,7 +5,8 @@ module OpenAI
|
|
5
5
|
{
|
6
6
|
delta: String,
|
7
7
|
type: :"transcript.text.delta",
|
8
|
-
logprobs: ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob]
|
8
|
+
logprobs: ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob],
|
9
|
+
segment_id: String
|
9
10
|
}
|
10
11
|
|
11
12
|
class TranscriptionTextDeltaEvent < OpenAI::Internal::Type::BaseModel
|
@@ -19,16 +20,22 @@ module OpenAI
|
|
19
20
|
::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob]
|
20
21
|
) -> ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob]
|
21
22
|
|
23
|
+
attr_reader segment_id: String?
|
24
|
+
|
25
|
+
def segment_id=: (String) -> String
|
26
|
+
|
22
27
|
def initialize: (
|
23
28
|
delta: String,
|
24
29
|
?logprobs: ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob],
|
30
|
+
?segment_id: String,
|
25
31
|
?type: :"transcript.text.delta"
|
26
32
|
) -> void
|
27
33
|
|
28
34
|
def to_hash: -> {
|
29
35
|
delta: String,
|
30
36
|
type: :"transcript.text.delta",
|
31
|
-
logprobs: ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob]
|
37
|
+
logprobs: ::Array[OpenAI::Audio::TranscriptionTextDeltaEvent::Logprob],
|
38
|
+
segment_id: String
|
32
39
|
}
|
33
40
|
|
34
41
|
type logprob =
|
@@ -0,0 +1,47 @@
|
|
1
|
+
module OpenAI
|
2
|
+
module Models
|
3
|
+
module Audio
|
4
|
+
type transcription_text_segment_event =
|
5
|
+
{
|
6
|
+
id: String,
|
7
|
+
end_: Float,
|
8
|
+
speaker: String,
|
9
|
+
start: Float,
|
10
|
+
text: String,
|
11
|
+
type: :"transcript.text.segment"
|
12
|
+
}
|
13
|
+
|
14
|
+
class TranscriptionTextSegmentEvent < OpenAI::Internal::Type::BaseModel
|
15
|
+
attr_accessor id: String
|
16
|
+
|
17
|
+
attr_accessor end_: Float
|
18
|
+
|
19
|
+
attr_accessor speaker: String
|
20
|
+
|
21
|
+
attr_accessor start: Float
|
22
|
+
|
23
|
+
attr_accessor text: String
|
24
|
+
|
25
|
+
attr_accessor type: :"transcript.text.segment"
|
26
|
+
|
27
|
+
def initialize: (
|
28
|
+
id: String,
|
29
|
+
end_: Float,
|
30
|
+
speaker: String,
|
31
|
+
start: Float,
|
32
|
+
text: String,
|
33
|
+
?type: :"transcript.text.segment"
|
34
|
+
) -> void
|
35
|
+
|
36
|
+
def to_hash: -> {
|
37
|
+
id: String,
|
38
|
+
end_: Float,
|
39
|
+
speaker: String,
|
40
|
+
start: Float,
|
41
|
+
text: String,
|
42
|
+
type: :"transcript.text.segment"
|
43
|
+
}
|
44
|
+
end
|
45
|
+
end
|
46
|
+
end
|
47
|
+
end
|
@@ -1,7 +1,10 @@
|
|
1
1
|
module OpenAI
|
2
2
|
module Models
|
3
3
|
type audio_model =
|
4
|
-
:"whisper-1"
|
4
|
+
:"whisper-1"
|
5
|
+
| :"gpt-4o-transcribe"
|
6
|
+
| :"gpt-4o-mini-transcribe"
|
7
|
+
| :"gpt-4o-transcribe-diarize"
|
5
8
|
|
6
9
|
module AudioModel
|
7
10
|
extend OpenAI::Internal::Type::Enum
|
@@ -9,6 +12,7 @@ module OpenAI
|
|
9
12
|
WHISPER_1: :"whisper-1"
|
10
13
|
GPT_4O_TRANSCRIBE: :"gpt-4o-transcribe"
|
11
14
|
GPT_4O_MINI_TRANSCRIBE: :"gpt-4o-mini-transcribe"
|
15
|
+
GPT_4O_TRANSCRIBE_DIARIZE: :"gpt-4o-transcribe-diarize"
|
12
16
|
|
13
17
|
def self?.values: -> ::Array[OpenAI::Models::audio_model]
|
14
18
|
end
|
@@ -1,6 +1,7 @@
|
|
1
1
|
module OpenAI
|
2
2
|
module Models
|
3
|
-
type audio_response_format =
|
3
|
+
type audio_response_format =
|
4
|
+
:json | :text | :srt | :verbose_json | :vtt | :diarized_json
|
4
5
|
|
5
6
|
module AudioResponseFormat
|
6
7
|
extend OpenAI::Internal::Type::Enum
|
@@ -10,6 +11,7 @@ module OpenAI
|
|
10
11
|
SRT: :srt
|
11
12
|
VERBOSE_JSON: :verbose_json
|
12
13
|
VTT: :vtt
|
14
|
+
DIARIZED_JSON: :diarized_json
|
13
15
|
|
14
16
|
def self?.values: -> ::Array[OpenAI::Models::audio_response_format]
|
15
17
|
end
|
@@ -37,17 +37,17 @@ module OpenAI
|
|
37
37
|
|
38
38
|
type model =
|
39
39
|
:"whisper-1"
|
40
|
-
| :"gpt-4o-transcribe-latest"
|
41
40
|
| :"gpt-4o-mini-transcribe"
|
42
41
|
| :"gpt-4o-transcribe"
|
42
|
+
| :"gpt-4o-transcribe-diarize"
|
43
43
|
|
44
44
|
module Model
|
45
45
|
extend OpenAI::Internal::Type::Enum
|
46
46
|
|
47
47
|
WHISPER_1: :"whisper-1"
|
48
|
-
GPT_4O_TRANSCRIBE_LATEST: :"gpt-4o-transcribe-latest"
|
49
48
|
GPT_4O_MINI_TRANSCRIBE: :"gpt-4o-mini-transcribe"
|
50
49
|
GPT_4O_TRANSCRIBE: :"gpt-4o-transcribe"
|
50
|
+
GPT_4O_TRANSCRIBE_DIARIZE: :"gpt-4o-transcribe-diarize"
|
51
51
|
|
52
52
|
def self?.values: -> ::Array[OpenAI::Models::Realtime::AudioTranscription::model]
|
53
53
|
end
|
@@ -3,6 +3,7 @@ module OpenAI
|
|
3
3
|
type vector_store_create_params =
|
4
4
|
{
|
5
5
|
chunking_strategy: OpenAI::Models::file_chunking_strategy_param,
|
6
|
+
description: String,
|
6
7
|
expires_after: OpenAI::VectorStoreCreateParams::ExpiresAfter,
|
7
8
|
file_ids: ::Array[String],
|
8
9
|
metadata: OpenAI::Models::metadata?,
|
@@ -20,6 +21,10 @@ module OpenAI
|
|
20
21
|
OpenAI::Models::file_chunking_strategy_param
|
21
22
|
) -> OpenAI::Models::file_chunking_strategy_param
|
22
23
|
|
24
|
+
attr_reader description: String?
|
25
|
+
|
26
|
+
def description=: (String) -> String
|
27
|
+
|
23
28
|
attr_reader expires_after: OpenAI::VectorStoreCreateParams::ExpiresAfter?
|
24
29
|
|
25
30
|
def expires_after=: (
|
@@ -38,6 +43,7 @@ module OpenAI
|
|
38
43
|
|
39
44
|
def initialize: (
|
40
45
|
?chunking_strategy: OpenAI::Models::file_chunking_strategy_param,
|
46
|
+
?description: String,
|
41
47
|
?expires_after: OpenAI::VectorStoreCreateParams::ExpiresAfter,
|
42
48
|
?file_ids: ::Array[String],
|
43
49
|
?metadata: OpenAI::Models::metadata?,
|
@@ -47,6 +53,7 @@ module OpenAI
|
|
47
53
|
|
48
54
|
def to_hash: -> {
|
49
55
|
chunking_strategy: OpenAI::Models::file_chunking_strategy_param,
|
56
|
+
description: String,
|
50
57
|
expires_after: OpenAI::VectorStoreCreateParams::ExpiresAfter,
|
51
58
|
file_ids: ::Array[String],
|
52
59
|
metadata: OpenAI::Models::metadata?,
|
@@ -7,6 +7,8 @@ module OpenAI
|
|
7
7
|
model: OpenAI::Models::Audio::TranscriptionCreateParams::model,
|
8
8
|
?chunking_strategy: OpenAI::Models::Audio::TranscriptionCreateParams::chunking_strategy?,
|
9
9
|
?include: ::Array[OpenAI::Models::Audio::transcription_include],
|
10
|
+
?known_speaker_names: ::Array[String],
|
11
|
+
?known_speaker_references: ::Array[String],
|
10
12
|
?language: String,
|
11
13
|
?prompt: String,
|
12
14
|
?response_format: OpenAI::Models::audio_response_format,
|
@@ -20,6 +22,8 @@ module OpenAI
|
|
20
22
|
model: OpenAI::Models::Audio::TranscriptionCreateParams::model,
|
21
23
|
?chunking_strategy: OpenAI::Models::Audio::TranscriptionCreateParams::chunking_strategy?,
|
22
24
|
?include: ::Array[OpenAI::Models::Audio::transcription_include],
|
25
|
+
?known_speaker_names: ::Array[String],
|
26
|
+
?known_speaker_references: ::Array[String],
|
23
27
|
?language: String,
|
24
28
|
?prompt: String,
|
25
29
|
?response_format: OpenAI::Models::audio_response_format,
|
@@ -6,11 +6,6 @@ module OpenAI
|
|
6
6
|
|
7
7
|
attr_reader threads: OpenAI::Resources::Beta::ChatKit::Threads
|
8
8
|
|
9
|
-
def upload_file: (
|
10
|
-
file: OpenAI::Internal::file_input,
|
11
|
-
?request_options: OpenAI::request_opts
|
12
|
-
) -> OpenAI::Models::Beta::chatkit_upload_file_response
|
13
|
-
|
14
9
|
def initialize: (client: OpenAI::Client) -> void
|
15
10
|
end
|
16
11
|
end
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: openai
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.33.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- OpenAI
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2025-10-
|
11
|
+
date: 2025-10-17 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: connection_pool
|
@@ -79,11 +79,14 @@ files:
|
|
79
79
|
- lib/openai/models/audio/transcription.rb
|
80
80
|
- lib/openai/models/audio/transcription_create_params.rb
|
81
81
|
- lib/openai/models/audio/transcription_create_response.rb
|
82
|
+
- lib/openai/models/audio/transcription_diarized.rb
|
83
|
+
- lib/openai/models/audio/transcription_diarized_segment.rb
|
82
84
|
- lib/openai/models/audio/transcription_include.rb
|
83
85
|
- lib/openai/models/audio/transcription_segment.rb
|
84
86
|
- lib/openai/models/audio/transcription_stream_event.rb
|
85
87
|
- lib/openai/models/audio/transcription_text_delta_event.rb
|
86
88
|
- lib/openai/models/audio/transcription_text_done_event.rb
|
89
|
+
- lib/openai/models/audio/transcription_text_segment_event.rb
|
87
90
|
- lib/openai/models/audio/transcription_verbose.rb
|
88
91
|
- lib/openai/models/audio/transcription_word.rb
|
89
92
|
- lib/openai/models/audio/translation.rb
|
@@ -139,14 +142,10 @@ files:
|
|
139
142
|
- lib/openai/models/beta/chatkit/thread_list_items_params.rb
|
140
143
|
- lib/openai/models/beta/chatkit/thread_list_params.rb
|
141
144
|
- lib/openai/models/beta/chatkit/thread_retrieve_params.rb
|
142
|
-
- lib/openai/models/beta/chatkit_upload_file_params.rb
|
143
|
-
- lib/openai/models/beta/chatkit_upload_file_response.rb
|
144
145
|
- lib/openai/models/beta/chatkit_workflow.rb
|
145
146
|
- lib/openai/models/beta/code_interpreter_tool.rb
|
146
|
-
- lib/openai/models/beta/file_part.rb
|
147
147
|
- lib/openai/models/beta/file_search_tool.rb
|
148
148
|
- lib/openai/models/beta/function_tool.rb
|
149
|
-
- lib/openai/models/beta/image_part.rb
|
150
149
|
- lib/openai/models/beta/message_stream_event.rb
|
151
150
|
- lib/openai/models/beta/run_step_stream_event.rb
|
152
151
|
- lib/openai/models/beta/run_stream_event.rb
|
@@ -811,11 +810,14 @@ files:
|
|
811
810
|
- rbi/openai/models/audio/transcription.rbi
|
812
811
|
- rbi/openai/models/audio/transcription_create_params.rbi
|
813
812
|
- rbi/openai/models/audio/transcription_create_response.rbi
|
813
|
+
- rbi/openai/models/audio/transcription_diarized.rbi
|
814
|
+
- rbi/openai/models/audio/transcription_diarized_segment.rbi
|
814
815
|
- rbi/openai/models/audio/transcription_include.rbi
|
815
816
|
- rbi/openai/models/audio/transcription_segment.rbi
|
816
817
|
- rbi/openai/models/audio/transcription_stream_event.rbi
|
817
818
|
- rbi/openai/models/audio/transcription_text_delta_event.rbi
|
818
819
|
- rbi/openai/models/audio/transcription_text_done_event.rbi
|
820
|
+
- rbi/openai/models/audio/transcription_text_segment_event.rbi
|
819
821
|
- rbi/openai/models/audio/transcription_verbose.rbi
|
820
822
|
- rbi/openai/models/audio/transcription_word.rbi
|
821
823
|
- rbi/openai/models/audio/translation.rbi
|
@@ -871,14 +873,10 @@ files:
|
|
871
873
|
- rbi/openai/models/beta/chatkit/thread_list_items_params.rbi
|
872
874
|
- rbi/openai/models/beta/chatkit/thread_list_params.rbi
|
873
875
|
- rbi/openai/models/beta/chatkit/thread_retrieve_params.rbi
|
874
|
-
- rbi/openai/models/beta/chatkit_upload_file_params.rbi
|
875
|
-
- rbi/openai/models/beta/chatkit_upload_file_response.rbi
|
876
876
|
- rbi/openai/models/beta/chatkit_workflow.rbi
|
877
877
|
- rbi/openai/models/beta/code_interpreter_tool.rbi
|
878
|
-
- rbi/openai/models/beta/file_part.rbi
|
879
878
|
- rbi/openai/models/beta/file_search_tool.rbi
|
880
879
|
- rbi/openai/models/beta/function_tool.rbi
|
881
|
-
- rbi/openai/models/beta/image_part.rbi
|
882
880
|
- rbi/openai/models/beta/message_stream_event.rbi
|
883
881
|
- rbi/openai/models/beta/run_step_stream_event.rbi
|
884
882
|
- rbi/openai/models/beta/run_stream_event.rbi
|
@@ -1532,11 +1530,14 @@ files:
|
|
1532
1530
|
- sig/openai/models/audio/transcription.rbs
|
1533
1531
|
- sig/openai/models/audio/transcription_create_params.rbs
|
1534
1532
|
- sig/openai/models/audio/transcription_create_response.rbs
|
1533
|
+
- sig/openai/models/audio/transcription_diarized.rbs
|
1534
|
+
- sig/openai/models/audio/transcription_diarized_segment.rbs
|
1535
1535
|
- sig/openai/models/audio/transcription_include.rbs
|
1536
1536
|
- sig/openai/models/audio/transcription_segment.rbs
|
1537
1537
|
- sig/openai/models/audio/transcription_stream_event.rbs
|
1538
1538
|
- sig/openai/models/audio/transcription_text_delta_event.rbs
|
1539
1539
|
- sig/openai/models/audio/transcription_text_done_event.rbs
|
1540
|
+
- sig/openai/models/audio/transcription_text_segment_event.rbs
|
1540
1541
|
- sig/openai/models/audio/transcription_verbose.rbs
|
1541
1542
|
- sig/openai/models/audio/transcription_word.rbs
|
1542
1543
|
- sig/openai/models/audio/translation.rbs
|
@@ -1592,14 +1593,10 @@ files:
|
|
1592
1593
|
- sig/openai/models/beta/chatkit/thread_list_items_params.rbs
|
1593
1594
|
- sig/openai/models/beta/chatkit/thread_list_params.rbs
|
1594
1595
|
- sig/openai/models/beta/chatkit/thread_retrieve_params.rbs
|
1595
|
-
- sig/openai/models/beta/chatkit_upload_file_params.rbs
|
1596
|
-
- sig/openai/models/beta/chatkit_upload_file_response.rbs
|
1597
1596
|
- sig/openai/models/beta/chatkit_workflow.rbs
|
1598
1597
|
- sig/openai/models/beta/code_interpreter_tool.rbs
|
1599
|
-
- sig/openai/models/beta/file_part.rbs
|
1600
1598
|
- sig/openai/models/beta/file_search_tool.rbs
|
1601
1599
|
- sig/openai/models/beta/function_tool.rbs
|
1602
|
-
- sig/openai/models/beta/image_part.rbs
|
1603
1600
|
- sig/openai/models/beta/message_stream_event.rbs
|
1604
1601
|
- sig/openai/models/beta/run_step_stream_event.rbs
|
1605
1602
|
- sig/openai/models/beta/run_stream_event.rbs
|
@@ -1,28 +0,0 @@
|
|
1
|
-
# frozen_string_literal: true
|
2
|
-
|
3
|
-
module OpenAI
|
4
|
-
module Models
|
5
|
-
module Beta
|
6
|
-
# @see OpenAI::Resources::Beta::ChatKit#upload_file
|
7
|
-
class ChatKitUploadFileParams < OpenAI::Internal::Type::BaseModel
|
8
|
-
extend OpenAI::Internal::Type::RequestParameters::Converter
|
9
|
-
include OpenAI::Internal::Type::RequestParameters
|
10
|
-
|
11
|
-
# @!attribute file
|
12
|
-
# Binary file contents to store with the ChatKit session. Supports PDFs and PNG,
|
13
|
-
# JPG, JPEG, GIF, or WEBP images.
|
14
|
-
#
|
15
|
-
# @return [Pathname, StringIO, IO, String, OpenAI::FilePart]
|
16
|
-
required :file, OpenAI::Internal::Type::FileInput
|
17
|
-
|
18
|
-
# @!method initialize(file:, request_options: {})
|
19
|
-
# Some parameter documentations has been truncated, see
|
20
|
-
# {OpenAI::Models::Beta::ChatKitUploadFileParams} for more details.
|
21
|
-
#
|
22
|
-
# @param file [Pathname, StringIO, IO, String, OpenAI::FilePart] Binary file contents to store with the ChatKit session. Supports PDFs and PNG, J
|
23
|
-
#
|
24
|
-
# @param request_options [OpenAI::RequestOptions, Hash{Symbol=>Object}]
|
25
|
-
end
|
26
|
-
end
|
27
|
-
end
|
28
|
-
end
|
@@ -1,25 +0,0 @@
|
|
1
|
-
# frozen_string_literal: true
|
2
|
-
|
3
|
-
module OpenAI
|
4
|
-
module Models
|
5
|
-
module Beta
|
6
|
-
# Represents either a file or image attachment.
|
7
|
-
#
|
8
|
-
# @see OpenAI::Resources::Beta::ChatKit#upload_file
|
9
|
-
module ChatKitUploadFileResponse
|
10
|
-
extend OpenAI::Internal::Type::Union
|
11
|
-
|
12
|
-
discriminator :type
|
13
|
-
|
14
|
-
# Metadata for a non-image file uploaded through ChatKit.
|
15
|
-
variant :file, -> { OpenAI::Beta::FilePart }
|
16
|
-
|
17
|
-
# Metadata for an image uploaded through ChatKit.
|
18
|
-
variant :image, -> { OpenAI::Beta::ImagePart }
|
19
|
-
|
20
|
-
# @!method self.variants
|
21
|
-
# @return [Array(OpenAI::Models::Beta::FilePart, OpenAI::Models::Beta::ImagePart)]
|
22
|
-
end
|
23
|
-
end
|
24
|
-
end
|
25
|
-
end
|
@@ -1,56 +0,0 @@
|
|
1
|
-
# frozen_string_literal: true
|
2
|
-
|
3
|
-
module OpenAI
|
4
|
-
module Models
|
5
|
-
module Beta
|
6
|
-
class FilePart < OpenAI::Internal::Type::BaseModel
|
7
|
-
# @!attribute id
|
8
|
-
# Unique identifier for the uploaded file.
|
9
|
-
#
|
10
|
-
# @return [String]
|
11
|
-
required :id, String
|
12
|
-
|
13
|
-
# @!attribute mime_type
|
14
|
-
# MIME type reported for the uploaded file. Defaults to null when unknown.
|
15
|
-
#
|
16
|
-
# @return [String, nil]
|
17
|
-
required :mime_type, String, nil?: true
|
18
|
-
|
19
|
-
# @!attribute name
|
20
|
-
# Original filename supplied by the uploader. Defaults to null when unnamed.
|
21
|
-
#
|
22
|
-
# @return [String, nil]
|
23
|
-
required :name, String, nil?: true
|
24
|
-
|
25
|
-
# @!attribute type
|
26
|
-
# Type discriminator that is always `file`.
|
27
|
-
#
|
28
|
-
# @return [Symbol, :file]
|
29
|
-
required :type, const: :file
|
30
|
-
|
31
|
-
# @!attribute upload_url
|
32
|
-
# Signed URL for downloading the uploaded file. Defaults to null when no download
|
33
|
-
# link is available.
|
34
|
-
#
|
35
|
-
# @return [String, nil]
|
36
|
-
required :upload_url, String, nil?: true
|
37
|
-
|
38
|
-
# @!method initialize(id:, mime_type:, name:, upload_url:, type: :file)
|
39
|
-
# Some parameter documentations has been truncated, see
|
40
|
-
# {OpenAI::Models::Beta::FilePart} for more details.
|
41
|
-
#
|
42
|
-
# Metadata for a non-image file uploaded through ChatKit.
|
43
|
-
#
|
44
|
-
# @param id [String] Unique identifier for the uploaded file.
|
45
|
-
#
|
46
|
-
# @param mime_type [String, nil] MIME type reported for the uploaded file. Defaults to null when unknown.
|
47
|
-
#
|
48
|
-
# @param name [String, nil] Original filename supplied by the uploader. Defaults to null when unnamed.
|
49
|
-
#
|
50
|
-
# @param upload_url [String, nil] Signed URL for downloading the uploaded file. Defaults to null when no download
|
51
|
-
#
|
52
|
-
# @param type [Symbol, :file] Type discriminator that is always `file`.
|
53
|
-
end
|
54
|
-
end
|
55
|
-
end
|
56
|
-
end
|