llm_lib 0.1.1 → 0.1.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/lib/llm_lib.rb +1 -38
- metadata +1 -1
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: a7fe9391bf23bcbe47c61ac8e8449536eb3052a19f8150813bf125861492d167
|
|
4
|
+
data.tar.gz: c2107e7106b73200502730c128fe0220a2a374d8f1ba6cfe5e6db48c8e78b95d
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: cc56b149e4e2e127857528a6c1754cf4e472eab9a09a4e34932b9a63d5d200eba80718b9c87dfa0f2d6d7901e9c95526b28dd348e870cc171a601b245c465830
|
|
7
|
+
data.tar.gz: d456b665985410472ba8248a517ab0bcc9126532cf11df46bf2e4134e8aafbb3903cdf9215164b280ff8f7b1ce55920f5af41792bb70268b54d8c71ca202300b
|
data/lib/llm_lib.rb
CHANGED
|
@@ -11,40 +11,10 @@ module LlmLib
|
|
|
11
11
|
# response = client.chat_gpt_call(prompt, max_tokens)
|
|
12
12
|
# puts response
|
|
13
13
|
|
|
14
|
-
# remove
|
|
15
|
-
attr_reader :apikey
|
|
16
|
-
# def def initialize(apikey, query)
|
|
17
|
-
# @apikey, @query = apikey, query
|
|
18
|
-
# end
|
|
19
|
-
|
|
20
14
|
def initialize(apikey)
|
|
21
15
|
@apikey = apikey
|
|
22
16
|
end
|
|
23
17
|
|
|
24
|
-
# def chat_gpt_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, logprobs = nil, stop = "\n")
|
|
25
|
-
# model = "text-davinci-003"
|
|
26
|
-
# response = self.class.send(@api_key, {
|
|
27
|
-
# "model" => model,
|
|
28
|
-
# "prompt" => prompt,
|
|
29
|
-
# "max_tokens" => max_tokens,
|
|
30
|
-
# "temperature" => temperature,
|
|
31
|
-
# "top_p" => top_p,
|
|
32
|
-
# "n" => n,
|
|
33
|
-
# "stream" => stream,
|
|
34
|
-
# "logprobs" => logprobs,
|
|
35
|
-
# "stop" => stop
|
|
36
|
-
# })
|
|
37
|
-
# response
|
|
38
|
-
# end
|
|
39
|
-
|
|
40
|
-
# def self.send(apikey, body)
|
|
41
|
-
# LlmLib::Restclient.post(
|
|
42
|
-
# body: body,
|
|
43
|
-
# url: "https://api.openai.com/v1/completions",
|
|
44
|
-
# apikey: apikey
|
|
45
|
-
# )
|
|
46
|
-
# end
|
|
47
|
-
|
|
48
18
|
def chat_gpt_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, stop = "\n")
|
|
49
19
|
model = "gpt-3.5-turbo"
|
|
50
20
|
response = OpenAI.send(@apikey,
|
|
@@ -62,7 +32,7 @@ module LlmLib
|
|
|
62
32
|
|
|
63
33
|
def gpt4_call(prompt, max_tokens, temperature = 0, top_p = 1, n = 1, stream = false, stop = "\n")
|
|
64
34
|
model = "gpt-4"
|
|
65
|
-
response = OpenAI.send(@
|
|
35
|
+
response = OpenAI.send(@apikey,
|
|
66
36
|
model,
|
|
67
37
|
prompt,
|
|
68
38
|
max_tokens,
|
|
@@ -117,13 +87,6 @@ module LlmLib
|
|
|
117
87
|
response
|
|
118
88
|
end
|
|
119
89
|
|
|
120
|
-
# def self.send(apikey, body, model)
|
|
121
|
-
# LlmLib::Restclient.post(
|
|
122
|
-
# body: body,
|
|
123
|
-
# url: "https://api-inference.huggingface.co/models/#{model}",
|
|
124
|
-
# apikey: apikey
|
|
125
|
-
# )
|
|
126
|
-
# end
|
|
127
90
|
end
|
|
128
91
|
|
|
129
92
|
end
|