omniai-openai 1.1.1 → 1.2.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/README.md +29 -9
- data/lib/omniai/openai/client.rb +9 -4
- data/lib/omniai/openai/config.rb +3 -1
- data/lib/omniai/openai/version.rb +1 -1
- metadata +3 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: e3afb339ec2a0e69d5a268fd134aae4119e3140ccad493f22d7db43b817145a4
|
4
|
+
data.tar.gz: f49c3cfd25228e49555981cc93e05e6702d3ce4edde9fa07277bc68c83312361
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 4455db3523b81301793b936528312394cb0f98b8f51ad7a49f5f6fe5811c2969dcaf679649deb0cb90bfec5fdf5838f3ba6444eea645d560ceaaaf96b9c15f5f
|
7
|
+
data.tar.gz: 804f70fd6067cba057aaa7dbbe7d65f5edaa389e5315fb4a107f154fbafa1139d33ad637439e8b838036767c02e4e2cfed0d69c43c0a36cb805bb4dfe21eb907
|
data/README.md
CHANGED
@@ -25,7 +25,7 @@ A client may also be passed the following options:
|
|
25
25
|
- `api_key` (required - default is `ENV['OPENAI_API_KEY']`)
|
26
26
|
- `organization` (optional)
|
27
27
|
- `project` (optional)
|
28
|
-
- `host` (optional)
|
28
|
+
- `host` (optional) useful for usage with Ollama or LocalAI
|
29
29
|
|
30
30
|
### Configuration
|
31
31
|
|
@@ -36,21 +36,41 @@ OmniAI::OpenAI.configure do |config|
|
|
36
36
|
config.api_key = 'sk-...' # default: ENV['OPENAI_API_KEY']
|
37
37
|
config.organization = '...' # default: ENV['OPENAI_ORGANIZATION']
|
38
38
|
config.project = '...' # default: ENV['OPENAI_PROJECT']
|
39
|
-
config.host = '...' # default: 'https://api.openai.com'
|
39
|
+
config.host = '...' # default: 'https://api.openai.com' - override for usage with LocalAI / Ollama
|
40
40
|
end
|
41
41
|
```
|
42
42
|
|
43
|
+
#### Usage with [LocalAI](https://localai.io/)
|
44
|
+
|
45
|
+
LocalAI offers [built in compatability with the OpenAI specification](https://localai.io/). To initialize a client that points to a Ollama change the host accordingly:
|
46
|
+
|
47
|
+
```ruby
|
48
|
+
client = OmniAI::OpenAI::Client.new(host: 'http://localhost:8080', api_key: nil)
|
49
|
+
```
|
50
|
+
|
51
|
+
_For details on installation or running LocalAI see the [getting started tutorial](https://localai.io/basics/getting_started/)._
|
52
|
+
|
53
|
+
#### Usage with [Ollama](https://www.ollama.com/)
|
54
|
+
|
55
|
+
Ollama offers [built in compatability with the OpenAI specification](https://ollama.com/blog/openai-compatibility). To initialize a client that points to a Ollama change the host accordingly:
|
56
|
+
|
57
|
+
```ruby
|
58
|
+
client = OmniAI::OpenAI::Client.new(host: 'http://localhost:11434', api_key: nil)
|
59
|
+
```
|
60
|
+
|
61
|
+
_For details on installation or running Ollama checkout [the project README](https://github.com/ollama/ollama)._
|
62
|
+
|
43
63
|
### Chat
|
44
64
|
|
45
65
|
A chat completion is generated by passing in prompts using any a variety of formats:
|
46
66
|
|
47
67
|
```ruby
|
48
|
-
completion = client.chat
|
68
|
+
completion = client.chat('Tell me a joke!')
|
49
69
|
completion.choice.message.content # 'Why did the chicken cross the road? To get to the other side.'
|
50
70
|
```
|
51
71
|
|
52
72
|
```ruby
|
53
|
-
completion = client.chat
|
73
|
+
completion = client.chat({
|
54
74
|
role: OmniAI::Chat::Role::USER,
|
55
75
|
content: 'Is it wise to jump off a bridge?'
|
56
76
|
})
|
@@ -58,7 +78,7 @@ completion.choice.message.content # 'No.'
|
|
58
78
|
```
|
59
79
|
|
60
80
|
```ruby
|
61
|
-
completion = client.chat
|
81
|
+
completion = client.chat([
|
62
82
|
{
|
63
83
|
role: OmniAI::Chat::Role::SYSTEM,
|
64
84
|
content: 'You are a helpful assistant.'
|
@@ -73,7 +93,7 @@ completion.choice.message.content # 'The capital of Canada is Ottawa.'
|
|
73
93
|
`model` takes an optional string (default is `gtp-4o`):
|
74
94
|
|
75
95
|
```ruby
|
76
|
-
completion = client.chat
|
96
|
+
completion = client.chat('How fast is a cheetah?', model: OmniAI::OpenAI::Chat::Model::GPT_3_5_TURBO)
|
77
97
|
completion.choice.message.content # 'A cheetah can reach speeds over 100 km/h.'
|
78
98
|
```
|
79
99
|
|
@@ -84,7 +104,7 @@ completion.choice.message.content # 'A cheetah can reach speeds over 100 km/h.'
|
|
84
104
|
`temperature` takes an optional float between `0.0` and `2.0` (defaults is `0.7`):
|
85
105
|
|
86
106
|
```ruby
|
87
|
-
completion = client.chat
|
107
|
+
completion = client.chat('Pick a number between 1 and 5', temperature: 2.0)
|
88
108
|
completion.choice.message.content # '3'
|
89
109
|
```
|
90
110
|
|
@@ -98,7 +118,7 @@ completion.choice.message.content # '3'
|
|
98
118
|
stream = proc do |chunk|
|
99
119
|
print(chunk.choice.delta.content) # 'Better', 'three', 'hours', ...
|
100
120
|
end
|
101
|
-
client.chat
|
121
|
+
client.chat('Be poetic.', stream:)
|
102
122
|
```
|
103
123
|
|
104
124
|
[OpenAI API Reference `stream`](https://platform.openai.com/docs/api-reference/chat/create#chat-create-stream)
|
@@ -108,7 +128,7 @@ client.chat.completion('Be poetic.', stream:)
|
|
108
128
|
`format` takes an optional symbol (`:json`) and that setes the `response_format` to `json_object`:
|
109
129
|
|
110
130
|
```ruby
|
111
|
-
completion = client.chat
|
131
|
+
completion = client.chat([
|
112
132
|
{ role: OmniAI::Chat::Role::SYSTEM, content: OmniAI::Chat::JSON_PROMPT },
|
113
133
|
{ role: OmniAI::Chat::Role::USER, content: 'What is the name of the drummer for the Beatles?' }
|
114
134
|
], format: :json)
|
data/lib/omniai/openai/client.rb
CHANGED
@@ -33,19 +33,24 @@ module OmniAI
|
|
33
33
|
logger: OmniAI::OpenAI.config.logger,
|
34
34
|
host: OmniAI::OpenAI.config.host
|
35
35
|
)
|
36
|
-
|
36
|
+
if api_key.nil? && host.eql?(Config::DEFAULT_HOST)
|
37
|
+
raise(
|
38
|
+
ArgumentError,
|
39
|
+
%(ENV['OPENAI_API_KEY'] must be defined or `api_key` must be passed when using #{Config::DEFAULT_HOST})
|
40
|
+
)
|
41
|
+
end
|
37
42
|
|
38
|
-
super(api_key:, logger:)
|
43
|
+
super(api_key:, host:, logger:)
|
39
44
|
|
40
45
|
@organization = organization
|
41
46
|
@project = project
|
42
|
-
@host = host
|
43
47
|
end
|
44
48
|
|
45
49
|
# @return [HTTP::Client]
|
46
50
|
def connection
|
47
51
|
@connection ||= begin
|
48
|
-
http = HTTP.
|
52
|
+
http = HTTP.persistent(@host)
|
53
|
+
http = http.auth("Bearer #{@api_key}") if @api_key
|
49
54
|
http = http.headers('OpenAI-Organization': @organization) if @organization
|
50
55
|
http = http.headers('OpenAI-Project': @project) if @project
|
51
56
|
http
|
data/lib/omniai/openai/config.rb
CHANGED
@@ -6,12 +6,14 @@ module OmniAI
|
|
6
6
|
class Config < OmniAI::Config
|
7
7
|
attr_accessor :organization, :project, :chat_options, :transcribe_options, :speak_options
|
8
8
|
|
9
|
+
DEFAULT_HOST = 'https://api.openai.com'
|
10
|
+
|
9
11
|
def initialize
|
10
12
|
super
|
11
13
|
@api_key = ENV.fetch('OPENAI_API_KEY', nil)
|
12
14
|
@organization = ENV.fetch('OPENAI_ORGANIZATION', nil)
|
13
15
|
@project = ENV.fetch('OPENAI_PROJECT', nil)
|
14
|
-
@host = ENV.fetch('OPENAI_HOST',
|
16
|
+
@host = ENV.fetch('OPENAI_HOST', DEFAULT_HOST)
|
15
17
|
@chat_options = {}
|
16
18
|
@transcribe_options = {}
|
17
19
|
@speak_options = {}
|
metadata
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: omniai-openai
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.2.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Kevin Sylvestre
|
@@ -69,7 +69,8 @@ files:
|
|
69
69
|
- lib/omniai/openai/transcribe.rb
|
70
70
|
- lib/omniai/openai/version.rb
|
71
71
|
homepage: https://github.com/ksylvest/omniai-openai
|
72
|
-
licenses:
|
72
|
+
licenses:
|
73
|
+
- MIT
|
73
74
|
metadata:
|
74
75
|
homepage_uri: https://github.com/ksylvest/omniai-openai
|
75
76
|
changelog_uri: https://github.com/ksylvest/omniai-openai/releases
|