openai 0.6.0 → 0.7.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +15 -0
- data/README.md +1 -1
- data/lib/openai/internal/util.rb +8 -9
- data/lib/openai/models/evals/create_eval_completions_run_data_source.rb +67 -1
- data/lib/openai/models/evals/run_cancel_response.rb +77 -1
- data/lib/openai/models/evals/run_create_params.rb +79 -1
- data/lib/openai/models/evals/run_create_response.rb +77 -1
- data/lib/openai/models/evals/run_list_response.rb +76 -1
- data/lib/openai/models/evals/run_retrieve_response.rb +77 -1
- data/lib/openai/resources/chat/completions.rb +2 -0
- data/lib/openai/resources/responses.rb +2 -0
- data/lib/openai/version.rb +1 -1
- data/rbi/openai/internal/util.rbi +2 -0
- data/rbi/openai/models/evals/create_eval_completions_run_data_source.rbi +110 -0
- data/rbi/openai/models/evals/run_cancel_response.rbi +187 -0
- data/rbi/openai/models/evals/run_create_params.rbi +222 -0
- data/rbi/openai/models/evals/run_create_response.rbi +187 -0
- data/rbi/openai/models/evals/run_list_response.rbi +187 -0
- data/rbi/openai/models/evals/run_retrieve_response.rbi +187 -0
- data/sig/openai/models/evals/create_eval_completions_run_data_source.rbs +29 -0
- data/sig/openai/models/evals/run_cancel_response.rbs +39 -0
- data/sig/openai/models/evals/run_create_params.rbs +39 -0
- data/sig/openai/models/evals/run_create_response.rbs +39 -0
- data/sig/openai/models/evals/run_list_response.rbs +39 -0
- data/sig/openai/models/evals/run_retrieve_response.rbs +39 -0
- metadata +2 -2
@@ -427,6 +427,8 @@ module OpenAI
|
|
427
427
|
max_completion_tokens: Integer,
|
428
428
|
seed: Integer,
|
429
429
|
temperature: Float,
|
430
|
+
text: OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text,
|
431
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
430
432
|
top_p: Float
|
431
433
|
}
|
432
434
|
|
@@ -443,6 +445,18 @@ module OpenAI
|
|
443
445
|
|
444
446
|
def temperature=: (Float) -> Float
|
445
447
|
|
448
|
+
attr_reader text: OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text?
|
449
|
+
|
450
|
+
def text=: (
|
451
|
+
OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text
|
452
|
+
) -> OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text
|
453
|
+
|
454
|
+
attr_reader tools: ::Array[OpenAI::Models::Responses::tool]?
|
455
|
+
|
456
|
+
def tools=: (
|
457
|
+
::Array[OpenAI::Models::Responses::tool]
|
458
|
+
) -> ::Array[OpenAI::Models::Responses::tool]
|
459
|
+
|
446
460
|
attr_reader top_p: Float?
|
447
461
|
|
448
462
|
def top_p=: (Float) -> Float
|
@@ -451,6 +465,8 @@ module OpenAI
|
|
451
465
|
?max_completion_tokens: Integer,
|
452
466
|
?seed: Integer,
|
453
467
|
?temperature: Float,
|
468
|
+
?text: OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text,
|
469
|
+
?tools: ::Array[OpenAI::Models::Responses::tool],
|
454
470
|
?top_p: Float
|
455
471
|
) -> void
|
456
472
|
|
@@ -458,8 +474,31 @@ module OpenAI
|
|
458
474
|
max_completion_tokens: Integer,
|
459
475
|
seed: Integer,
|
460
476
|
temperature: Float,
|
477
|
+
text: OpenAI::Models::Evals::RunCreateResponse::DataSource::Responses::SamplingParams::Text,
|
478
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
461
479
|
top_p: Float
|
462
480
|
}
|
481
|
+
|
482
|
+
type text =
|
483
|
+
{
|
484
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
485
|
+
}
|
486
|
+
|
487
|
+
class Text < OpenAI::Internal::Type::BaseModel
|
488
|
+
attr_reader format_: OpenAI::Models::Responses::response_format_text_config?
|
489
|
+
|
490
|
+
def format_=: (
|
491
|
+
OpenAI::Models::Responses::response_format_text_config
|
492
|
+
) -> OpenAI::Models::Responses::response_format_text_config
|
493
|
+
|
494
|
+
def initialize: (
|
495
|
+
?format_: OpenAI::Models::Responses::response_format_text_config
|
496
|
+
) -> void
|
497
|
+
|
498
|
+
def to_hash: -> {
|
499
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
500
|
+
}
|
501
|
+
end
|
463
502
|
end
|
464
503
|
end
|
465
504
|
|
@@ -427,6 +427,8 @@ module OpenAI
|
|
427
427
|
max_completion_tokens: Integer,
|
428
428
|
seed: Integer,
|
429
429
|
temperature: Float,
|
430
|
+
text: OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text,
|
431
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
430
432
|
top_p: Float
|
431
433
|
}
|
432
434
|
|
@@ -443,6 +445,18 @@ module OpenAI
|
|
443
445
|
|
444
446
|
def temperature=: (Float) -> Float
|
445
447
|
|
448
|
+
attr_reader text: OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text?
|
449
|
+
|
450
|
+
def text=: (
|
451
|
+
OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text
|
452
|
+
) -> OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text
|
453
|
+
|
454
|
+
attr_reader tools: ::Array[OpenAI::Models::Responses::tool]?
|
455
|
+
|
456
|
+
def tools=: (
|
457
|
+
::Array[OpenAI::Models::Responses::tool]
|
458
|
+
) -> ::Array[OpenAI::Models::Responses::tool]
|
459
|
+
|
446
460
|
attr_reader top_p: Float?
|
447
461
|
|
448
462
|
def top_p=: (Float) -> Float
|
@@ -451,6 +465,8 @@ module OpenAI
|
|
451
465
|
?max_completion_tokens: Integer,
|
452
466
|
?seed: Integer,
|
453
467
|
?temperature: Float,
|
468
|
+
?text: OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text,
|
469
|
+
?tools: ::Array[OpenAI::Models::Responses::tool],
|
454
470
|
?top_p: Float
|
455
471
|
) -> void
|
456
472
|
|
@@ -458,8 +474,31 @@ module OpenAI
|
|
458
474
|
max_completion_tokens: Integer,
|
459
475
|
seed: Integer,
|
460
476
|
temperature: Float,
|
477
|
+
text: OpenAI::Models::Evals::RunListResponse::DataSource::Responses::SamplingParams::Text,
|
478
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
461
479
|
top_p: Float
|
462
480
|
}
|
481
|
+
|
482
|
+
type text =
|
483
|
+
{
|
484
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
485
|
+
}
|
486
|
+
|
487
|
+
class Text < OpenAI::Internal::Type::BaseModel
|
488
|
+
attr_reader format_: OpenAI::Models::Responses::response_format_text_config?
|
489
|
+
|
490
|
+
def format_=: (
|
491
|
+
OpenAI::Models::Responses::response_format_text_config
|
492
|
+
) -> OpenAI::Models::Responses::response_format_text_config
|
493
|
+
|
494
|
+
def initialize: (
|
495
|
+
?format_: OpenAI::Models::Responses::response_format_text_config
|
496
|
+
) -> void
|
497
|
+
|
498
|
+
def to_hash: -> {
|
499
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
500
|
+
}
|
501
|
+
end
|
463
502
|
end
|
464
503
|
end
|
465
504
|
|
@@ -427,6 +427,8 @@ module OpenAI
|
|
427
427
|
max_completion_tokens: Integer,
|
428
428
|
seed: Integer,
|
429
429
|
temperature: Float,
|
430
|
+
text: OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text,
|
431
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
430
432
|
top_p: Float
|
431
433
|
}
|
432
434
|
|
@@ -443,6 +445,18 @@ module OpenAI
|
|
443
445
|
|
444
446
|
def temperature=: (Float) -> Float
|
445
447
|
|
448
|
+
attr_reader text: OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text?
|
449
|
+
|
450
|
+
def text=: (
|
451
|
+
OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text
|
452
|
+
) -> OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text
|
453
|
+
|
454
|
+
attr_reader tools: ::Array[OpenAI::Models::Responses::tool]?
|
455
|
+
|
456
|
+
def tools=: (
|
457
|
+
::Array[OpenAI::Models::Responses::tool]
|
458
|
+
) -> ::Array[OpenAI::Models::Responses::tool]
|
459
|
+
|
446
460
|
attr_reader top_p: Float?
|
447
461
|
|
448
462
|
def top_p=: (Float) -> Float
|
@@ -451,6 +465,8 @@ module OpenAI
|
|
451
465
|
?max_completion_tokens: Integer,
|
452
466
|
?seed: Integer,
|
453
467
|
?temperature: Float,
|
468
|
+
?text: OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text,
|
469
|
+
?tools: ::Array[OpenAI::Models::Responses::tool],
|
454
470
|
?top_p: Float
|
455
471
|
) -> void
|
456
472
|
|
@@ -458,8 +474,31 @@ module OpenAI
|
|
458
474
|
max_completion_tokens: Integer,
|
459
475
|
seed: Integer,
|
460
476
|
temperature: Float,
|
477
|
+
text: OpenAI::Models::Evals::RunRetrieveResponse::DataSource::Responses::SamplingParams::Text,
|
478
|
+
tools: ::Array[OpenAI::Models::Responses::tool],
|
461
479
|
top_p: Float
|
462
480
|
}
|
481
|
+
|
482
|
+
type text =
|
483
|
+
{
|
484
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
485
|
+
}
|
486
|
+
|
487
|
+
class Text < OpenAI::Internal::Type::BaseModel
|
488
|
+
attr_reader format_: OpenAI::Models::Responses::response_format_text_config?
|
489
|
+
|
490
|
+
def format_=: (
|
491
|
+
OpenAI::Models::Responses::response_format_text_config
|
492
|
+
) -> OpenAI::Models::Responses::response_format_text_config
|
493
|
+
|
494
|
+
def initialize: (
|
495
|
+
?format_: OpenAI::Models::Responses::response_format_text_config
|
496
|
+
) -> void
|
497
|
+
|
498
|
+
def to_hash: -> {
|
499
|
+
format_: OpenAI::Models::Responses::response_format_text_config
|
500
|
+
}
|
501
|
+
end
|
463
502
|
end
|
464
503
|
end
|
465
504
|
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: openai
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.7.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- OpenAI
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2025-06-
|
11
|
+
date: 2025-06-09 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: connection_pool
|