llm_lib 0.1.0 → 0.1.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/llm_lib/restclient.rb +0 -4
- data/lib/llm_lib.rb +5 -4
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 74ba99c44eaa6b120986df7ff3bd257dc1fa395c7a6aa3e83287850ba08a1275
|
4
|
+
data.tar.gz: b8d24197c782363f01ab737bb63563197926f665be9ac9d1e154201e9f241076
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: d5b95c2ae9e63591ee822f8abb7a47227ab34e2ec0abccd30422e3289728b3118dd6e7d33fcce7c644a556115b5436c0b681c132159ca2b2e5835a5996b25b1c
|
7
|
+
data.tar.gz: b76dd9437711b511a15dcd7ca0d7bf1830dc6e22dead75fafa9939dd2763bb442fe15d85585287f2a3bb45b13fa2a0dbfe041bb03fb58b118fb7e52a443bdbc2
|
data/lib/llm_lib/restclient.rb
CHANGED
@@ -1,6 +1,5 @@
|
|
1
1
|
class LlmLib::Restclient
|
2
2
|
|
3
|
-
# def self.post(url:, body:, headers: {})
|
4
3
|
def self.post(url:, body:, apikey:)
|
5
4
|
|
6
5
|
url = URI(url)
|
@@ -13,10 +12,7 @@ def self.post(url:, body:, apikey:)
|
|
13
12
|
request = Net::HTTP::Post.new(url.path)
|
14
13
|
request['Content-Type'] = 'application/json'
|
15
14
|
request["cache-control"] = 'no-cache'
|
16
|
-
# request["apikey"] = apikey
|
17
|
-
# headers.each { |key, value| request[key] = value }
|
18
15
|
request['Authorization'] = "Bearer #{apikey}"
|
19
|
-
# request.set_form_data(body)
|
20
16
|
request.body = JSON.generate(body)
|
21
17
|
|
22
18
|
begin
|
data/lib/llm_lib.rb
CHANGED
@@ -12,7 +12,7 @@ module LlmLib
|
|
12
12
|
# puts response
|
13
13
|
|
14
14
|
# remove
|
15
|
-
|
15
|
+
attr_reader :apikey
|
16
16
|
# def def initialize(apikey, query)
|
17
17
|
# @apikey, @query = apikey, query
|
18
18
|
# end
|
@@ -47,7 +47,7 @@ module LlmLib
|
|
47
47
|
|
48
48
|
def chat_gpt_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, stop = "\n")
|
49
49
|
model = "gpt-3.5-turbo"
|
50
|
-
response = OpenAI.send(@
|
50
|
+
response = OpenAI.send(@apikey,
|
51
51
|
model,
|
52
52
|
prompt,
|
53
53
|
max_tokens,
|
@@ -97,8 +97,9 @@ module LlmLib
|
|
97
97
|
|
98
98
|
def hugging_bloom_call(query, model = "bigscience/bloom")
|
99
99
|
response = HuggingFace.send(@api_key,
|
100
|
-
|
101
|
-
|
100
|
+
model,
|
101
|
+
query
|
102
|
+
)
|
102
103
|
response
|
103
104
|
end
|
104
105
|
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: llm_lib
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.1.
|
4
|
+
version: 0.1.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Chamath Attanayaka
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-07-
|
11
|
+
date: 2023-07-31 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: uri
|