llm_lib 0.1.1 → 0.1.2
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/llm_lib.rb +1 -38
- metadata +1 -1
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: a7fe9391bf23bcbe47c61ac8e8449536eb3052a19f8150813bf125861492d167
|
4
|
+
data.tar.gz: c2107e7106b73200502730c128fe0220a2a374d8f1ba6cfe5e6db48c8e78b95d
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: cc56b149e4e2e127857528a6c1754cf4e472eab9a09a4e34932b9a63d5d200eba80718b9c87dfa0f2d6d7901e9c95526b28dd348e870cc171a601b245c465830
|
7
|
+
data.tar.gz: d456b665985410472ba8248a517ab0bcc9126532cf11df46bf2e4134e8aafbb3903cdf9215164b280ff8f7b1ce55920f5af41792bb70268b54d8c71ca202300b
|
data/lib/llm_lib.rb
CHANGED
@@ -11,40 +11,10 @@ module LlmLib
|
|
11
11
|
# response = client.chat_gpt_call(prompt, max_tokens)
|
12
12
|
# puts response
|
13
13
|
|
14
|
-
# remove
|
15
|
-
attr_reader :apikey
|
16
|
-
# def def initialize(apikey, query)
|
17
|
-
# @apikey, @query = apikey, query
|
18
|
-
# end
|
19
|
-
|
20
14
|
def initialize(apikey)
|
21
15
|
@apikey = apikey
|
22
16
|
end
|
23
17
|
|
24
|
-
# def chat_gpt_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, logprobs = nil, stop = "\n")
|
25
|
-
# model = "text-davinci-003"
|
26
|
-
# response = self.class.send(@api_key, {
|
27
|
-
# "model" => model,
|
28
|
-
# "prompt" => prompt,
|
29
|
-
# "max_tokens" => max_tokens,
|
30
|
-
# "temperature" => temperature,
|
31
|
-
# "top_p" => top_p,
|
32
|
-
# "n" => n,
|
33
|
-
# "stream" => stream,
|
34
|
-
# "logprobs" => logprobs,
|
35
|
-
# "stop" => stop
|
36
|
-
# })
|
37
|
-
# response
|
38
|
-
# end
|
39
|
-
|
40
|
-
# def self.send(apikey, body)
|
41
|
-
# LlmLib::Restclient.post(
|
42
|
-
# body: body,
|
43
|
-
# url: "https://api.openai.com/v1/completions",
|
44
|
-
# apikey: apikey
|
45
|
-
# )
|
46
|
-
# end
|
47
|
-
|
48
18
|
def chat_gpt_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, stop = "\n")
|
49
19
|
model = "gpt-3.5-turbo"
|
50
20
|
response = OpenAI.send(@apikey,
|
@@ -62,7 +32,7 @@ module LlmLib
|
|
62
32
|
|
63
33
|
def gpt4_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, stop = "\n")
|
64
34
|
model = "gpt-4"
|
65
|
-
response = OpenAI.send(@
|
35
|
+
response = OpenAI.send(@apikey,
|
66
36
|
model,
|
67
37
|
prompt,
|
68
38
|
max_tokens,
|
@@ -117,13 +87,6 @@ module LlmLib
|
|
117
87
|
response
|
118
88
|
end
|
119
89
|
|
120
|
-
# def self.send(apikey, body, model)
|
121
|
-
# LlmLib::Restclient.post(
|
122
|
-
# body: body,
|
123
|
-
# url: "https://api-inference.huggingface.co/models/#{model}",
|
124
|
-
# apikey: apikey
|
125
|
-
# )
|
126
|
-
# end
|
127
90
|
end
|
128
91
|
|
129
92
|
end
|