open_ai_bot 0.3.3 → 0.3.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/Gemfile +1 -1
- data/Gemfile.lock +16 -13
- data/lib/open_ai/chat_gpt.rb +4 -1
- data/lib/open_ai/chat_thread.rb +1 -0
- data/lib/open_ai/model.rb +54 -16
- data/lib/open_ai_bot.rb +8 -0
- data/main.rb +1 -1
- data/open_ai_bot.gemspec +1 -1
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 70a3fabb10c860e1d7c45e83c74b38f7b393d4a1d31c86a73ccc4f827af6dcd8
|
4
|
+
data.tar.gz: 91436b52ead57fb638b4de1d74f573041fd541bae1352315632ae4414853fb28
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: '0908d995df133104dd4a591e7a9cff2c561f39753fc0570ac15eb3d9808ab644f7c04afef5682580c020bfca9d2e05e70573a4258427e9737bb18c2ef120ba60'
|
7
|
+
data.tar.gz: 6b54406776b00460dea2f1451313558b60bdb8fddb5dd843ec6ed64e6052a42affeae69bfe166be25ad7d641dd9d71980b25adaf0bb620462ce4b0e73d63d806
|
data/Gemfile
CHANGED
data/Gemfile.lock
CHANGED
@@ -35,26 +35,29 @@ GEM
|
|
35
35
|
dry-inflector (~> 1.0)
|
36
36
|
dry-logic (~> 1.4)
|
37
37
|
zeitwerk (~> 2.6)
|
38
|
-
event_stream_parser (0.
|
39
|
-
faraday (2.
|
40
|
-
faraday-net_http (>= 2.0, < 3.
|
41
|
-
|
42
|
-
|
43
|
-
faraday-
|
44
|
-
|
38
|
+
event_stream_parser (1.0.0)
|
39
|
+
faraday (2.13.1)
|
40
|
+
faraday-net_http (>= 2.0, < 3.5)
|
41
|
+
json
|
42
|
+
logger
|
43
|
+
faraday-multipart (1.1.0)
|
44
|
+
multipart-post (~> 2.0)
|
45
|
+
faraday-net_http (3.4.1)
|
46
|
+
net-http (>= 0.5.0)
|
45
47
|
fiber-annotation (0.2.0)
|
46
48
|
fiber-local (1.1.0)
|
47
49
|
fiber-storage
|
48
50
|
fiber-storage (0.1.1)
|
49
51
|
ice_nine (0.11.2)
|
50
52
|
io-event (1.6.0)
|
51
|
-
json (2.
|
53
|
+
json (2.12.2)
|
54
|
+
logger (1.7.0)
|
52
55
|
multipart-post (2.4.1)
|
53
|
-
net-http (0.
|
56
|
+
net-http (0.6.0)
|
54
57
|
uri
|
55
58
|
public_suffix (5.0.5)
|
56
|
-
ruby-openai (
|
57
|
-
event_stream_parser (>= 0.3.0, <
|
59
|
+
ruby-openai (8.1.0)
|
60
|
+
event_stream_parser (>= 0.3.0, < 2.0.0)
|
58
61
|
faraday (>= 1)
|
59
62
|
faraday-multipart (>= 1)
|
60
63
|
rubydium (0.4.1)
|
@@ -65,7 +68,7 @@ GEM
|
|
65
68
|
faraday (~> 2.0)
|
66
69
|
faraday-multipart (~> 1.0)
|
67
70
|
zeitwerk (~> 2.6)
|
68
|
-
uri (0.
|
71
|
+
uri (1.0.3)
|
69
72
|
zeitwerk (2.6.15)
|
70
73
|
|
71
74
|
PLATFORMS
|
@@ -74,7 +77,7 @@ PLATFORMS
|
|
74
77
|
|
75
78
|
DEPENDENCIES
|
76
79
|
down (~> 5.4)
|
77
|
-
ruby-openai (~>
|
80
|
+
ruby-openai (~> 8.1)
|
78
81
|
rubydium (>= 0.2.5)
|
79
82
|
|
80
83
|
BUNDLED WITH
|
data/lib/open_ai/chat_gpt.rb
CHANGED
@@ -145,8 +145,11 @@ module OpenAI
|
|
145
145
|
def get_tokens_info!(response)
|
146
146
|
completion_tokens = response.dig("usage", "completion_tokens")
|
147
147
|
prompt_tokens = response.dig("usage", "prompt_tokens")
|
148
|
+
cached_prompt_tokens = response.dig("usage", "prompt_tokens_details", "cached_tokens")
|
148
149
|
|
149
|
-
|
150
|
+
current_thread.model.request_cost(
|
151
|
+
completion_tokens:, prompt_tokens:, cached_prompt_tokens:, current_thread:
|
152
|
+
)
|
150
153
|
end
|
151
154
|
|
152
155
|
def send_chat_gpt_response(text, tokens_info)
|
data/lib/open_ai/chat_thread.rb
CHANGED
data/lib/open_ai/model.rb
CHANGED
@@ -1,23 +1,59 @@
|
|
1
1
|
module OpenAI
|
2
2
|
class Model
|
3
|
-
# All prices are per
|
3
|
+
# All prices are in USD per 1M tokens
|
4
4
|
MODEL_INFO = {
|
5
|
+
"gpt-4.1": {
|
6
|
+
max_context: 1_047_576,
|
7
|
+
input_price: 2.00,
|
8
|
+
cached_input_price: 0.50,
|
9
|
+
output_price: 8.00,
|
10
|
+
vision: true
|
11
|
+
},
|
12
|
+
"gpt-4.1-mini": {
|
13
|
+
max_context: 1_047_576,
|
14
|
+
input_price: 0.40,
|
15
|
+
cached_input_price: 0.10,
|
16
|
+
output_price: 1.60,
|
17
|
+
vision: true
|
18
|
+
},
|
19
|
+
"gpt-4.1-nano": {
|
20
|
+
max_context: 1_047_576,
|
21
|
+
input_price: 0.10,
|
22
|
+
cached_input_price: 0.025,
|
23
|
+
output_price: 0.40,
|
24
|
+
vision: true
|
25
|
+
},
|
5
26
|
"gpt-4o": {
|
6
27
|
max_context: 128_000,
|
7
|
-
|
8
|
-
|
28
|
+
input_price: 5.00,
|
29
|
+
cached_input_price: 2.50,
|
30
|
+
output_price: 20.00,
|
9
31
|
vision: true
|
10
32
|
},
|
11
|
-
"gpt-
|
12
|
-
max_context:
|
13
|
-
|
14
|
-
|
33
|
+
"gpt-4o-mini": {
|
34
|
+
max_context: 128_000,
|
35
|
+
input_price: 0.60,
|
36
|
+
cached_input_price: 0.30,
|
37
|
+
output_price: 2.40,
|
38
|
+
vision: true
|
39
|
+
},
|
40
|
+
"o3": {
|
41
|
+
max_context: 200_000,
|
42
|
+
input_price: 2.00,
|
43
|
+
cached_input_price: 0.50,
|
44
|
+
output_price: 8.00,
|
45
|
+
vision: true
|
46
|
+
},
|
47
|
+
"o4-mini": {
|
48
|
+
max_context: 200_000,
|
49
|
+
input_price: 1.10,
|
50
|
+
cached_input_price: 0.275,
|
51
|
+
output_price: 4.40,
|
52
|
+
vision: true
|
15
53
|
}
|
16
54
|
}
|
17
55
|
|
18
|
-
|
19
|
-
|
20
|
-
[:max_context, :prompt_price, :completion_price].each do |attr|
|
56
|
+
[:max_context, :input_price, :cached_input_price, :output_price].each do |attr|
|
21
57
|
define_method(attr) do
|
22
58
|
MODEL_INFO[@model][attr]
|
23
59
|
end
|
@@ -39,15 +75,17 @@ module OpenAI
|
|
39
75
|
MODEL_INFO[@model][:vision]
|
40
76
|
end
|
41
77
|
|
42
|
-
def request_cost(prompt_tokens:, completion_tokens:, current_thread:)
|
43
|
-
prompt_cost = prompt_tokens *
|
44
|
-
|
78
|
+
def request_cost(prompt_tokens:, cached_prompt_tokens:, completion_tokens:, current_thread:)
|
79
|
+
prompt_cost = prompt_tokens * input_price / 1_000_000
|
80
|
+
cached_prompt_cost = cached_prompt_tokens * cached_input_price / 1_000_000
|
81
|
+
completion_cost = completion_tokens * output_price / 1_000_000
|
45
82
|
|
46
|
-
total = prompt_cost + completion_cost
|
83
|
+
total = prompt_cost + cached_prompt_cost + completion_cost
|
47
84
|
thread_total = current_thread.total_cost
|
48
85
|
|
49
86
|
info = "\n\n" + {
|
50
|
-
|
87
|
+
cached_prompt: "#{cached_prompt_tokens} tokens (#{cached_prompt_cost.round(5)}$)",
|
88
|
+
uncached_prompt: "#{prompt_tokens} tokens (#{prompt_cost.round(5)}$)",
|
51
89
|
completion: "#{completion_tokens} tokens (#{completion_cost.round(5)}$)",
|
52
90
|
total: "#{total.round(5)}$",
|
53
91
|
total_for_this_conversation: "#{(thread_total + total).round(5)}$",
|
@@ -59,4 +97,4 @@ module OpenAI
|
|
59
97
|
{ info:, total: }
|
60
98
|
end
|
61
99
|
end
|
62
|
-
end
|
100
|
+
end
|
data/lib/open_ai_bot.rb
CHANGED
@@ -27,6 +27,14 @@ class OpenAIBot < Rubydium::Bot
|
|
27
27
|
on_command "/help", description: "Sends useful help info" do
|
28
28
|
reply(self.class.help_message)
|
29
29
|
end
|
30
|
+
on_command "/d" do
|
31
|
+
return unless @user.username == config.owner_username
|
32
|
+
return unless @target&.id.in? [config.bot_id, @user.id]
|
33
|
+
|
34
|
+
current_thread.delete(@replies_to.message_id)
|
35
|
+
safe_delete(@replies_to)
|
36
|
+
safe_delete(@msg)
|
37
|
+
end
|
30
38
|
|
31
39
|
def allowed_chat?
|
32
40
|
return true if @user.username == config.owner_username
|
data/main.rb
CHANGED
@@ -24,7 +24,7 @@ bot = bots[bot_name]
|
|
24
24
|
bot.config = YAML.load_file("#{__dir__}/config.yaml")
|
25
25
|
bot.configure do |config|
|
26
26
|
config.open_ai_client = OpenAI::Client.new(
|
27
|
-
access_token: config.
|
27
|
+
access_token: config.open_ai['token']
|
28
28
|
# organization_id: config.open_ai_organization_id
|
29
29
|
)
|
30
30
|
end
|
data/open_ai_bot.gemspec
CHANGED
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: open_ai_bot
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.3.
|
4
|
+
version: 0.3.4
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- bulgakke
|
8
8
|
autorequire:
|
9
9
|
bindir: exe
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2025-06-16 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: down
|