omniai 0.2.0 → 1.0.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +4 -4
- data/lib/omniai/chat.rb +98 -18
- data/lib/omniai/client.rb +16 -7
- data/lib/omniai/config.rb +6 -0
- data/lib/omniai/version.rb +1 -1
- metadata +1 -2
- data/lib/omniai/chat/request.rb +0 -105
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: b730453eea6e211031174bde7e43485b5402b9f0590bb832924762f8c2f6345c
|
4
|
+
data.tar.gz: cad17682fc35243454f58b28a9bf0e2d48f9857415bfe3b72012f370b7e9d4ab
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 0ebd9716858486dc399504b8caefd3d4186cdc97da05715e8519fcbc2238b38672f6a760e5134b3231bfc4683662f69098edc3eb6bd22808e91658bb4207b1b4
|
7
|
+
data.tar.gz: b7623a59effdc4714cbf78f8dd950cde0d3746da0dfc39ec7c9a6caf745ffedbd841d0b72097b1fdd30dd73f3f8ba35e7736a4253b32b7d2fcd7367c732172b0
|
data/README.md
CHANGED
@@ -65,7 +65,7 @@ Clients that support chat (e.g. Anthropic w/ "Claude", Google w/ "Gemini", Mistr
|
|
65
65
|
#### w/ a Simple Prompt
|
66
66
|
|
67
67
|
```ruby
|
68
|
-
completion = client.chat
|
68
|
+
completion = client.chat('Tell me a joke.')
|
69
69
|
puts(completion.choice.message.content) # '...'
|
70
70
|
```
|
71
71
|
|
@@ -82,7 +82,7 @@ messages = [
|
|
82
82
|
content: 'What is the capital of Canada?',
|
83
83
|
},
|
84
84
|
]
|
85
|
-
completion = client.chat
|
85
|
+
completion = client.chat(messages, model: '...', temperature: 0.7, format: :json)
|
86
86
|
puts(completion.choice.message.content) # '...'
|
87
87
|
```
|
88
88
|
|
@@ -102,7 +102,7 @@ message = {
|
|
102
102
|
]
|
103
103
|
}
|
104
104
|
|
105
|
-
completion = client.chat
|
105
|
+
completion = client.chat(message)
|
106
106
|
puts(completion.choice.message.content) # '...'
|
107
107
|
```
|
108
108
|
|
@@ -112,5 +112,5 @@ puts(completion.choice.message.content) # '...'
|
|
112
112
|
stream = proc do |chunk|
|
113
113
|
print(chunk.choice.delta.content) # '...'
|
114
114
|
end
|
115
|
-
client.chat
|
115
|
+
client.chat('Tell me a joke.', stream:)
|
116
116
|
```
|
data/lib/omniai/chat.rb
CHANGED
@@ -1,43 +1,123 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
3
|
module OmniAI
|
4
|
-
# An abstract class that provides
|
4
|
+
# An abstract class that provides a consistent interface for processing chat requests.
|
5
5
|
#
|
6
6
|
# Usage:
|
7
7
|
#
|
8
8
|
# class OmniAI::OpenAI::Chat < OmniAI::Chat
|
9
|
-
#
|
10
|
-
#
|
9
|
+
# module Model
|
10
|
+
# GPT_4O = "gpt-4o"
|
11
11
|
# end
|
12
|
-
# end
|
13
12
|
#
|
14
|
-
#
|
13
|
+
# protected
|
14
|
+
#
|
15
|
+
# # @return [Hash]
|
16
|
+
# def payload
|
17
|
+
# raise NotImplementedError, "#{self.class.name}#payload undefined"
|
18
|
+
# end
|
15
19
|
#
|
16
|
-
#
|
20
|
+
# # @return [String]
|
21
|
+
# def path
|
22
|
+
# raise NotImplementedError, "#{self.class.name}#path undefined"
|
23
|
+
# end
|
24
|
+
# end
|
17
25
|
#
|
18
|
-
#
|
26
|
+
# client.chat(messages, model: "...", temperature: 0.0, format: :text)
|
19
27
|
class Chat
|
28
|
+
JSON_PROMPT = 'Respond with valid JSON. Do not include any non-JSON in the response.'
|
29
|
+
|
20
30
|
module Role
|
21
31
|
ASSISTANT = 'assistant'
|
22
32
|
USER = 'user'
|
23
33
|
SYSTEM = 'system'
|
24
34
|
end
|
25
35
|
|
26
|
-
def
|
27
|
-
|
36
|
+
def self.process!(...)
|
37
|
+
new(...).process!
|
28
38
|
end
|
29
39
|
|
30
|
-
# @
|
31
|
-
#
|
32
|
-
# @param
|
33
|
-
# @param model [String] optional
|
34
|
-
# @param format [Symbol] optional :text or :json
|
40
|
+
# @param messages [String] required
|
41
|
+
# @param client [OmniAI::Client] the client
|
42
|
+
# @param model [String] required
|
35
43
|
# @param temperature [Float, nil] optional
|
36
44
|
# @param stream [Proc, nil] optional
|
37
|
-
#
|
38
|
-
|
39
|
-
|
40
|
-
|
45
|
+
# @param format [Symbol, nil] optional - :json
|
46
|
+
def initialize(messages, client:, model:, temperature: nil, stream: nil, format: nil)
|
47
|
+
@messages = messages
|
48
|
+
@client = client
|
49
|
+
@model = model
|
50
|
+
@temperature = temperature
|
51
|
+
@stream = stream
|
52
|
+
@format = format
|
53
|
+
end
|
54
|
+
|
55
|
+
# @raise [ExecutionError]
|
56
|
+
def process!
|
57
|
+
response = request!
|
58
|
+
raise HTTPError, response unless response.status.ok?
|
59
|
+
|
60
|
+
parse!(response:)
|
61
|
+
end
|
62
|
+
|
63
|
+
protected
|
64
|
+
|
65
|
+
# @return [Hash]
|
66
|
+
def payload
|
67
|
+
raise NotImplementedError, "#{self.class.name}#payload undefined"
|
68
|
+
end
|
69
|
+
|
70
|
+
# @return [String]
|
71
|
+
def path
|
72
|
+
raise NotImplementedError, "#{self.class.name}#path undefined"
|
73
|
+
end
|
74
|
+
|
75
|
+
# @param response [HTTP::Response]
|
76
|
+
# @return [OmniAI::Chat::Completion]
|
77
|
+
def parse!(response:)
|
78
|
+
if @stream
|
79
|
+
stream!(response:)
|
80
|
+
else
|
81
|
+
complete!(response:)
|
82
|
+
end
|
83
|
+
end
|
84
|
+
|
85
|
+
# @param response [OmniAI::Chat::Completion]
|
86
|
+
def complete!(response:)
|
87
|
+
Completion.new(data: response.parse)
|
88
|
+
end
|
89
|
+
|
90
|
+
# @param response [HTTP::Response]
|
91
|
+
# @return [OmniAI::Chat::Stream]
|
92
|
+
def stream!(response:)
|
93
|
+
raise Error, "#{self.class.name}#stream! unstreamable" unless @stream
|
94
|
+
|
95
|
+
Stream.new(response:).stream! { |chunk| @stream.call(chunk) }
|
96
|
+
end
|
97
|
+
|
98
|
+
# @return [Array<Hash>]
|
99
|
+
def messages
|
100
|
+
arrayify(@messages).map do |content|
|
101
|
+
case content
|
102
|
+
when String then { role: OmniAI::Chat::Role::USER, content: }
|
103
|
+
when Hash then content
|
104
|
+
else raise Error, "Unsupported content=#{content.inspect}"
|
105
|
+
end
|
106
|
+
end
|
107
|
+
end
|
108
|
+
|
109
|
+
# @param value [Object, Array<Object>]
|
110
|
+
# @return [Array<Object>]
|
111
|
+
def arrayify(value)
|
112
|
+
value.is_a?(Array) ? value : [value]
|
113
|
+
end
|
114
|
+
|
115
|
+
# @return [HTTP::Response]
|
116
|
+
def request!
|
117
|
+
@client
|
118
|
+
.connection
|
119
|
+
.accept(:json)
|
120
|
+
.post(path, json: payload)
|
41
121
|
end
|
42
122
|
end
|
43
123
|
end
|
data/lib/omniai/client.rb
CHANGED
@@ -9,11 +9,6 @@ module OmniAI
|
|
9
9
|
# def initialize(api_key: ENV.fetch('OPENAI_API_KEY'), logger: nil)
|
10
10
|
# super
|
11
11
|
# end
|
12
|
-
#
|
13
|
-
# @return [OmniAI::OpenAI::Chat]
|
14
|
-
# def chat
|
15
|
-
# # TODO: implement
|
16
|
-
# end
|
17
12
|
# end
|
18
13
|
class Client
|
19
14
|
class Error < StandardError; end
|
@@ -27,13 +22,27 @@ module OmniAI
|
|
27
22
|
@logger = logger
|
28
23
|
end
|
29
24
|
|
25
|
+
# @return [String]
|
26
|
+
def inspect
|
27
|
+
masked_api_key = "#{api_key[..2]}***" if api_key
|
28
|
+
"#<#{self.class.name} api_key=#{masked_api_key.inspect}>"
|
29
|
+
end
|
30
|
+
|
30
31
|
# @return [HTTP::Client]
|
31
32
|
def connection
|
32
33
|
raise NotImplementedError, "#{self.class.name}#connection undefined"
|
33
34
|
end
|
34
35
|
|
35
|
-
# @
|
36
|
-
|
36
|
+
# @raise [OmniAI::Error]
|
37
|
+
#
|
38
|
+
# @param messages [String, Array, Hash]
|
39
|
+
# @param model [String] optional
|
40
|
+
# @param format [Symbol] optional :text or :json
|
41
|
+
# @param temperature [Float, nil] optional
|
42
|
+
# @param stream [Proc, nil] optional
|
43
|
+
#
|
44
|
+
# @return [OmniAI::Chat::Completion]
|
45
|
+
def chat(messages, model:, temperature: nil, format: nil, stream: nil)
|
37
46
|
raise NotImplementedError, "#{self.class.name}#chat undefined"
|
38
47
|
end
|
39
48
|
end
|
data/lib/omniai/config.rb
CHANGED
@@ -4,5 +4,11 @@ module OmniAI
|
|
4
4
|
# A configuration for each agent w/ `api_key` / `host` / `logger`.
|
5
5
|
class Config
|
6
6
|
attr_accessor :api_key, :host, :logger
|
7
|
+
|
8
|
+
# @return [String]
|
9
|
+
def inspect
|
10
|
+
masked_api_key = "#{api_key[..2]}***" if api_key
|
11
|
+
"#<#{self.class.name} api_key=#{masked_api_key.inspect} host=#{host.inspect}>"
|
12
|
+
end
|
7
13
|
end
|
8
14
|
end
|
data/lib/omniai/version.rb
CHANGED
metadata
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: omniai
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 1.0.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Kevin Sylvestre
|
@@ -72,7 +72,6 @@ files:
|
|
72
72
|
- lib/omniai/chat/content.rb
|
73
73
|
- lib/omniai/chat/delta.rb
|
74
74
|
- lib/omniai/chat/message.rb
|
75
|
-
- lib/omniai/chat/request.rb
|
76
75
|
- lib/omniai/chat/stream.rb
|
77
76
|
- lib/omniai/chat/usage.rb
|
78
77
|
- lib/omniai/client.rb
|
data/lib/omniai/chat/request.rb
DELETED
@@ -1,105 +0,0 @@
|
|
1
|
-
# frozen_string_literal: true
|
2
|
-
|
3
|
-
module OmniAI
|
4
|
-
class Chat
|
5
|
-
# An abstract class that provides a consistent interface for processing chat requests.
|
6
|
-
#
|
7
|
-
# Usage:
|
8
|
-
#
|
9
|
-
# class OmniAI::OpenAI::ChatGPT::Request < OmniAI::Chat::Request
|
10
|
-
# module Model
|
11
|
-
# CHAT = "davinci"
|
12
|
-
# end
|
13
|
-
# def completion(messages, model:, temperature: 0.0, format: :text)
|
14
|
-
# end
|
15
|
-
# end
|
16
|
-
#
|
17
|
-
# Once defined, the subclass can be used to interface with the vendor's chat API as follows:
|
18
|
-
#
|
19
|
-
# client.chat.completion(messages, model: "...", temperature: 0.0, format: :text)
|
20
|
-
class Request
|
21
|
-
# @param client [OmniAI::Client] the client
|
22
|
-
# @param messages [String] required
|
23
|
-
# @param model [String] required
|
24
|
-
# @param temperature [Float, nil] optional
|
25
|
-
# @param stream [Proc, nil] optional
|
26
|
-
# @param format [Symbol, nil] optional - :json
|
27
|
-
def initialize(client:, messages:, model:, temperature: nil, stream: nil, format: nil)
|
28
|
-
@client = client
|
29
|
-
@messages = messages
|
30
|
-
@model = model
|
31
|
-
@temperature = temperature
|
32
|
-
@stream = stream
|
33
|
-
@format = format
|
34
|
-
end
|
35
|
-
|
36
|
-
# @raise [ExecutionError]
|
37
|
-
def process!
|
38
|
-
response = request!
|
39
|
-
raise HTTPError, response unless response.status.ok?
|
40
|
-
|
41
|
-
parse!(response:)
|
42
|
-
end
|
43
|
-
|
44
|
-
protected
|
45
|
-
|
46
|
-
# @return [Hash]
|
47
|
-
def payload
|
48
|
-
raise NotImplementedError, "#{self.class.name}#payload undefined"
|
49
|
-
end
|
50
|
-
|
51
|
-
# @return [String]
|
52
|
-
def path
|
53
|
-
raise NotImplementedError, "#{self.class.name}#path undefined"
|
54
|
-
end
|
55
|
-
|
56
|
-
# @param response [HTTP::Response]
|
57
|
-
# @return [OmniAI::Chat::Completion]
|
58
|
-
def parse!(response:)
|
59
|
-
if @stream
|
60
|
-
stream!(response:)
|
61
|
-
else
|
62
|
-
complete!(response:)
|
63
|
-
end
|
64
|
-
end
|
65
|
-
|
66
|
-
# @param response [OmniAI::Chat::Completion]
|
67
|
-
def complete!(response:)
|
68
|
-
Completion.new(data: response.parse)
|
69
|
-
end
|
70
|
-
|
71
|
-
# @param response [HTTP::Response]
|
72
|
-
# @return [OmniAI::Chat::Chunk]
|
73
|
-
def stream!(response:)
|
74
|
-
raise Error, "#{self.class.name}#stream! unstreamable" unless @stream
|
75
|
-
|
76
|
-
Stream.new(response:).stream! { |chunk| @stream.call(chunk) }
|
77
|
-
end
|
78
|
-
|
79
|
-
# @return [Array<Hash>]
|
80
|
-
def messages
|
81
|
-
arrayify(@messages).map do |content|
|
82
|
-
case content
|
83
|
-
when String then { role: OmniAI::Chat::Role::USER, content: }
|
84
|
-
when Hash then content
|
85
|
-
else raise Error, "Unsupported content=#{content.inspect}"
|
86
|
-
end
|
87
|
-
end
|
88
|
-
end
|
89
|
-
|
90
|
-
# @param value [Object, Array<Object>]
|
91
|
-
# @return [Array<Object>]
|
92
|
-
def arrayify(value)
|
93
|
-
value.is_a?(Array) ? value : [value]
|
94
|
-
end
|
95
|
-
|
96
|
-
# @return [HTTP::Response]
|
97
|
-
def request!
|
98
|
-
@client
|
99
|
-
.connection
|
100
|
-
.accept(:json)
|
101
|
-
.post(path, json: payload)
|
102
|
-
end
|
103
|
-
end
|
104
|
-
end
|
105
|
-
end
|