langchainrb 0.3.15 → 0.4.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +10 -0
- data/Gemfile.lock +14 -1
- data/README.md +22 -22
- data/lib/{agent → langchain/agent}/base.rb +1 -1
- data/lib/{agent → langchain/agent}/chain_of_thought_agent/chain_of_thought_agent.rb +13 -14
- data/lib/{llm → langchain/llm}/base.rb +3 -3
- data/lib/{llm → langchain/llm}/cohere.rb +1 -1
- data/lib/{llm → langchain/llm}/google_palm.rb +3 -3
- data/lib/{llm → langchain/llm}/hugging_face.rb +1 -1
- data/lib/{llm → langchain/llm}/openai.rb +18 -6
- data/lib/{llm → langchain/llm}/replicate.rb +3 -3
- data/lib/{prompt → langchain/prompt}/base.rb +2 -2
- data/lib/{prompt → langchain/prompt}/few_shot_prompt_template.rb +1 -1
- data/lib/{prompt → langchain/prompt}/loading.rb +3 -3
- data/lib/{prompt → langchain/prompt}/prompt_template.rb +1 -1
- data/lib/{tool → langchain/tool}/base.rb +5 -5
- data/lib/{tool → langchain/tool}/calculator.rb +4 -2
- data/lib/langchain/tool/ruby_code_interpreter.rb +27 -0
- data/lib/{tool → langchain/tool}/serp_api.rb +3 -1
- data/lib/{tool → langchain/tool}/wikipedia.rb +3 -1
- data/lib/langchain/utils/token_length_validator.rb +57 -0
- data/lib/{vectorsearch → langchain/vectorsearch}/base.rb +5 -5
- data/lib/{vectorsearch → langchain/vectorsearch}/chroma.rb +1 -1
- data/lib/{vectorsearch → langchain/vectorsearch}/milvus.rb +1 -1
- data/lib/{vectorsearch → langchain/vectorsearch}/pgvector.rb +15 -4
- data/lib/{vectorsearch → langchain/vectorsearch}/pinecone.rb +1 -1
- data/lib/{vectorsearch → langchain/vectorsearch}/qdrant.rb +1 -1
- data/lib/{vectorsearch → langchain/vectorsearch}/weaviate.rb +1 -1
- data/lib/langchain.rb +42 -35
- data/lib/version.rb +1 -1
- metadata +72 -28
- /data/lib/{agent → langchain/agent}/chain_of_thought_agent/chain_of_thought_agent_prompt.json +0 -0
- /data/lib/{llm → langchain/llm}/prompts/summarize_template.json +0 -0
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 100c9989f1a6787168064f6d810f009c6ded8129949f0040a7c1ed05ebff9d20
|
4
|
+
data.tar.gz: 0cf01f053841e9a45353d72b62df9a29c0daf7268fb69aff794aae830f7db448
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: becea5f089e7455e8dd9437cb3647c921eebb3155a9685697190e0e757d31ec8d9f4572faa3e299911b44b371c097bd9b5edf3da35772562027d1b12cd743b9e
|
7
|
+
data.tar.gz: 44cfe6f8cb2056d6a0a5c619ded7828a9cdad0e5f4e8876242adad2f13251ea1c3dbb64c9feffea25b61e91f163245954c1f272fadb9f4f3a2a06cadd64bd9a0
|
data/CHANGELOG.md
CHANGED
@@ -1,5 +1,15 @@
|
|
1
1
|
## [Unreleased]
|
2
2
|
|
3
|
+
## [0.4.1] - 2023-06-02
|
4
|
+
- Beautiful colored log messages
|
5
|
+
- 🛠️ Tools
|
6
|
+
- Introducing `Langchain::Tool::RubyCodeInterpreter`, a tool executes sandboxed Ruby code
|
7
|
+
|
8
|
+
## [0.4.0] - 2023-06-01
|
9
|
+
- [BREAKING] Everything is namespaced under `Langchain::` now
|
10
|
+
- Pgvector similarity search uses the cosine distance by default now
|
11
|
+
- OpenAI token length validation using tiktoken_ruby
|
12
|
+
|
3
13
|
## [0.3.15] - 2023-05-30
|
4
14
|
- Drop Ruby 2.7 support. It had reached EOD.
|
5
15
|
- Bump pgvector-ruby to 0.2
|
data/Gemfile.lock
CHANGED
@@ -1,7 +1,9 @@
|
|
1
1
|
PATH
|
2
2
|
remote: .
|
3
3
|
specs:
|
4
|
-
langchainrb (0.
|
4
|
+
langchainrb (0.4.1)
|
5
|
+
colorize (~> 0.8.1)
|
6
|
+
tiktoken_ruby (~> 0.0.5)
|
5
7
|
|
6
8
|
GEM
|
7
9
|
remote: https://rubygems.org/
|
@@ -31,6 +33,7 @@ GEM
|
|
31
33
|
ast (2.4.2)
|
32
34
|
builder (3.2.4)
|
33
35
|
byebug (11.1.3)
|
36
|
+
childprocess (4.1.0)
|
34
37
|
chroma-db (0.3.0)
|
35
38
|
dry-monads (~> 1.6)
|
36
39
|
ruby-next-core (>= 0.15.0)
|
@@ -38,6 +41,7 @@ GEM
|
|
38
41
|
cohere-ruby (0.9.4)
|
39
42
|
faraday (>= 1.0.0)
|
40
43
|
faraday_middleware (>= 1.0.0)
|
44
|
+
colorize (0.8.1)
|
41
45
|
concurrent-ruby (1.2.2)
|
42
46
|
crass (1.0.6)
|
43
47
|
diff-lcs (1.5.0)
|
@@ -205,6 +209,7 @@ GEM
|
|
205
209
|
zeitwerk (~> 2.5)
|
206
210
|
rainbow (3.1.1)
|
207
211
|
rake (13.0.6)
|
212
|
+
rb_sys (0.9.78)
|
208
213
|
regexp_parser (2.8.0)
|
209
214
|
replicate-ruby (0.2.2)
|
210
215
|
addressable
|
@@ -248,6 +253,8 @@ GEM
|
|
248
253
|
ruby-rc4 (0.1.5)
|
249
254
|
ruby2_keywords (0.0.5)
|
250
255
|
rubyzip (2.3.2)
|
256
|
+
safe_ruby (1.0.4)
|
257
|
+
childprocess (>= 0.3.9)
|
251
258
|
standard (1.28.2)
|
252
259
|
language_server-protocol (~> 3.17.0.2)
|
253
260
|
lint_roller (~> 1.0)
|
@@ -262,6 +269,11 @@ GEM
|
|
262
269
|
standardrb (1.0.1)
|
263
270
|
standard
|
264
271
|
thor (1.2.1)
|
272
|
+
tiktoken_ruby (0.0.5)
|
273
|
+
rb_sys (~> 0.9.68)
|
274
|
+
tiktoken_ruby (0.0.5-arm64-darwin)
|
275
|
+
tiktoken_ruby (0.0.5-x86_64-darwin)
|
276
|
+
tiktoken_ruby (0.0.5-x86_64-linux)
|
265
277
|
treetop (1.6.12)
|
266
278
|
polyglot (~> 0.3)
|
267
279
|
ttfunk (1.7.0)
|
@@ -306,6 +318,7 @@ DEPENDENCIES
|
|
306
318
|
replicate-ruby (~> 0.2.2)
|
307
319
|
rspec (~> 3.0)
|
308
320
|
ruby-openai (~> 4.0.0)
|
321
|
+
safe_ruby (~> 1.0.4)
|
309
322
|
standardrb
|
310
323
|
weaviate-ruby (~> 0.8.0)
|
311
324
|
wikipedia-client (~> 1.17.0)
|
data/README.md
CHANGED
@@ -44,7 +44,7 @@ Add `gem "weaviate-ruby", "~> 0.8.0"` to your Gemfile.
|
|
44
44
|
|
45
45
|
Pick the vector search database you'll be using and instantiate the client:
|
46
46
|
```ruby
|
47
|
-
client = Vectorsearch::Weaviate.new(
|
47
|
+
client = Langchain::Vectorsearch::Weaviate.new(
|
48
48
|
url: ENV["WEAVIATE_URL"],
|
49
49
|
api_key: ENV["WEAVIATE_API_KEY"],
|
50
50
|
llm: :openai, # or :cohere
|
@@ -52,10 +52,10 @@ client = Vectorsearch::Weaviate.new(
|
|
52
52
|
)
|
53
53
|
|
54
54
|
# You can instantiate any other supported vector search database:
|
55
|
-
client = Vectorsearch::Milvus.new(...) # `gem "milvus", "~> 0.9.0"`
|
56
|
-
client = Vectorsearch::Qdrant.new(...) # `gem"qdrant-ruby", "~> 0.9.0"`
|
57
|
-
client = Vectorsearch::Pinecone.new(...) # `gem "pinecone", "~> 0.1.6"`
|
58
|
-
client = Vectorsearch::Chroma.new(...) # `gem "chroma-db", "~> 0.3.0"`
|
55
|
+
client = Langchain::Vectorsearch::Milvus.new(...) # `gem "milvus", "~> 0.9.0"`
|
56
|
+
client = Langchain::Vectorsearch::Qdrant.new(...) # `gem"qdrant-ruby", "~> 0.9.0"`
|
57
|
+
client = Langchain::Vectorsearch::Pinecone.new(...) # `gem "pinecone", "~> 0.1.6"`
|
58
|
+
client = Langchain::Vectorsearch::Chroma.new(...) # `gem "chroma-db", "~> 0.3.0"`
|
59
59
|
```
|
60
60
|
|
61
61
|
```ruby
|
@@ -107,7 +107,7 @@ Add `gem "ruby-openai", "~> 4.0.0"` to your Gemfile.
|
|
107
107
|
|
108
108
|
#### OpenAI
|
109
109
|
```ruby
|
110
|
-
openai = LLM::OpenAI.new(api_key: ENV["OPENAI_API_KEY"])
|
110
|
+
openai = Langchain::LLM::OpenAI.new(api_key: ENV["OPENAI_API_KEY"])
|
111
111
|
```
|
112
112
|
```ruby
|
113
113
|
openai.embed(text: "foo bar")
|
@@ -120,7 +120,7 @@ openai.complete(prompt: "What is the meaning of life?")
|
|
120
120
|
Add `gem "cohere-ruby", "~> 0.9.3"` to your Gemfile.
|
121
121
|
|
122
122
|
```ruby
|
123
|
-
cohere = LLM::Cohere.new(api_key: ENV["COHERE_API_KEY"])
|
123
|
+
cohere = Langchain::LLM::Cohere.new(api_key: ENV["COHERE_API_KEY"])
|
124
124
|
```
|
125
125
|
```ruby
|
126
126
|
cohere.embed(text: "foo bar")
|
@@ -132,19 +132,19 @@ cohere.complete(prompt: "What is the meaning of life?")
|
|
132
132
|
#### HuggingFace
|
133
133
|
Add `gem "hugging-face", "~> 0.3.2"` to your Gemfile.
|
134
134
|
```ruby
|
135
|
-
cohere = LLM::HuggingFace.new(api_key: ENV["HUGGING_FACE_API_KEY"])
|
135
|
+
cohere = Langchain::LLM::HuggingFace.new(api_key: ENV["HUGGING_FACE_API_KEY"])
|
136
136
|
```
|
137
137
|
|
138
138
|
#### Replicate
|
139
139
|
Add `gem "replicate-ruby", "~> 0.2.2"` to your Gemfile.
|
140
140
|
```ruby
|
141
|
-
cohere = LLM::Replicate.new(api_key: ENV["REPLICATE_API_KEY"])
|
141
|
+
cohere = Langchain::LLM::Replicate.new(api_key: ENV["REPLICATE_API_KEY"])
|
142
142
|
```
|
143
143
|
|
144
144
|
#### Google PaLM (Pathways Language Model)
|
145
145
|
Add `"google_palm_api", "~> 0.1.0"` to your Gemfile.
|
146
146
|
```ruby
|
147
|
-
google_palm = LLM::GooglePalm.new(api_key: ENV["GOOGLE_PALM_API_KEY"])
|
147
|
+
google_palm = Langchain::LLM::GooglePalm.new(api_key: ENV["GOOGLE_PALM_API_KEY"])
|
148
148
|
```
|
149
149
|
|
150
150
|
### Using Prompts 📋
|
@@ -154,21 +154,21 @@ google_palm = LLM::GooglePalm.new(api_key: ENV["GOOGLE_PALM_API_KEY"])
|
|
154
154
|
Create a prompt with one input variable:
|
155
155
|
|
156
156
|
```ruby
|
157
|
-
prompt = Prompt::PromptTemplate.new(template: "Tell me a {adjective} joke.", input_variables: ["adjective"])
|
157
|
+
prompt = Langchain::Prompt::PromptTemplate.new(template: "Tell me a {adjective} joke.", input_variables: ["adjective"])
|
158
158
|
prompt.format(adjective: "funny") # "Tell me a funny joke."
|
159
159
|
```
|
160
160
|
|
161
161
|
Create a prompt with multiple input variables:
|
162
162
|
|
163
163
|
```ruby
|
164
|
-
prompt = Prompt::PromptTemplate.new(template: "Tell me a {adjective} joke about {content}.", input_variables: ["adjective", "content"])
|
164
|
+
prompt = Langchain::Prompt::PromptTemplate.new(template: "Tell me a {adjective} joke about {content}.", input_variables: ["adjective", "content"])
|
165
165
|
prompt.format(adjective: "funny", content: "chickens") # "Tell me a funny joke about chickens."
|
166
166
|
```
|
167
167
|
|
168
168
|
Creating a PromptTemplate using just a prompt and no input_variables:
|
169
169
|
|
170
170
|
```ruby
|
171
|
-
prompt = Prompt::PromptTemplate.from_template("Tell me a {adjective} joke about {content}.")
|
171
|
+
prompt = Langchain::Prompt::PromptTemplate.from_template("Tell me a {adjective} joke about {content}.")
|
172
172
|
prompt.input_variables # ["adjective", "content"]
|
173
173
|
prompt.format(adjective: "funny", content: "chickens") # "Tell me a funny joke about chickens."
|
174
174
|
```
|
@@ -182,7 +182,7 @@ prompt.save(file_path: "spec/fixtures/prompt/prompt_template.json")
|
|
182
182
|
Loading a new prompt template using a JSON file:
|
183
183
|
|
184
184
|
```ruby
|
185
|
-
prompt = Prompt.load_from_path(file_path: "spec/fixtures/prompt/prompt_template.json")
|
185
|
+
prompt = Langchain::Prompt.load_from_path(file_path: "spec/fixtures/prompt/prompt_template.json")
|
186
186
|
prompt.input_variables # ["adjective", "content"]
|
187
187
|
```
|
188
188
|
|
@@ -191,10 +191,10 @@ prompt.input_variables # ["adjective", "content"]
|
|
191
191
|
Create a prompt with a few shot examples:
|
192
192
|
|
193
193
|
```ruby
|
194
|
-
prompt = Prompt::FewShotPromptTemplate.new(
|
194
|
+
prompt = Langchain::Prompt::FewShotPromptTemplate.new(
|
195
195
|
prefix: "Write antonyms for the following words.",
|
196
196
|
suffix: "Input: {adjective}\nOutput:",
|
197
|
-
example_prompt: Prompt::PromptTemplate.new(
|
197
|
+
example_prompt: Langchain::Prompt::PromptTemplate.new(
|
198
198
|
input_variables: ["input", "output"],
|
199
199
|
template: "Input: {input}\nOutput: {output}"
|
200
200
|
),
|
@@ -228,7 +228,7 @@ prompt.save(file_path: "spec/fixtures/prompt/few_shot_prompt_template.json")
|
|
228
228
|
Loading a new prompt template using a JSON file:
|
229
229
|
|
230
230
|
```ruby
|
231
|
-
prompt = Prompt.load_from_path(file_path: "spec/fixtures/prompt/few_shot_prompt_template.json")
|
231
|
+
prompt = Langchain::Prompt.load_from_path(file_path: "spec/fixtures/prompt/few_shot_prompt_template.json")
|
232
232
|
prompt.prefix # "Write antonyms for the following words."
|
233
233
|
```
|
234
234
|
|
@@ -237,10 +237,10 @@ Agents are semi-autonomous bots that can respond to user questions and use avail
|
|
237
237
|
|
238
238
|
#### Chain-of-Thought Agent
|
239
239
|
|
240
|
-
Add `gem "openai
|
240
|
+
Add `gem "ruby-openai"`, `gem "eqn"`, and `gem "google_search_results"` to your Gemfile
|
241
241
|
|
242
242
|
```ruby
|
243
|
-
agent = Agent::ChainOfThoughtAgent.new(llm: :openai, llm_api_key: ENV["OPENAI_API_KEY"], tools: ['search', 'calculator'])
|
243
|
+
agent = Langchain::Agent::ChainOfThoughtAgent.new(llm: :openai, llm_api_key: ENV["OPENAI_API_KEY"], tools: ['search', 'calculator'])
|
244
244
|
|
245
245
|
agent.tools
|
246
246
|
# => ["search", "calculator"]
|
@@ -260,10 +260,10 @@ agent.run(question: "How many full soccer fields would be needed to cover the di
|
|
260
260
|
| Name | Description | ENV Requirements | Gem Requirements |
|
261
261
|
| ------------ | :------------------------------------------------: | :-----------------------------------------------------------: | :---------------------------------------: |
|
262
262
|
| "calculator" | Useful for getting the result of a math expression | | `gem "eqn", "~> 1.6.5"` |
|
263
|
-
| "
|
263
|
+
| "ruby_code_interpreter" | Interprets Ruby expressions | | `gem "safe_ruby", "~> 1.0.4"` |
|
264
|
+
| "search" | A wrapper around Google Search | `ENV["SERPAPI_API_KEY"]` (https://serpapi.com/manage-api-key) | `gem "google_search_results", "~> 2.0.0"` |
|
264
265
|
| "wikipedia" | Calls Wikipedia API to retrieve the summary | | `gem "wikipedia-client", "~> 1.17.0"` |
|
265
266
|
|
266
|
-
|
267
267
|
#### Loaders 🚚
|
268
268
|
|
269
269
|
Need to read data from various sources? Load it up.
|
@@ -273,7 +273,7 @@ Need to read data from various sources? Load it up.
|
|
273
273
|
Just call `Langchan::Loader.load` with the path to the file or a URL you want to load.
|
274
274
|
|
275
275
|
```ruby
|
276
|
-
|
276
|
+
Langchain::Loader.load('/path/to/file.pdf')
|
277
277
|
```
|
278
278
|
|
279
279
|
or
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Agent
|
3
|
+
module Langchain::Agent
|
4
4
|
class ChainOfThoughtAgent < Base
|
5
5
|
attr_reader :llm, :llm_api_key, :llm_client, :tools
|
6
6
|
|
@@ -11,14 +11,14 @@ module Agent
|
|
11
11
|
# @param tools [Array] The tools to use
|
12
12
|
# @return [ChainOfThoughtAgent] The Agent::ChainOfThoughtAgent instance
|
13
13
|
def initialize(llm:, llm_api_key:, tools: [])
|
14
|
-
LLM::Base.validate_llm!(llm: llm)
|
15
|
-
Tool::Base.validate_tools!(tools: tools)
|
14
|
+
Langchain::LLM::Base.validate_llm!(llm: llm)
|
15
|
+
Langchain::Tool::Base.validate_tools!(tools: tools)
|
16
16
|
|
17
17
|
@llm = llm
|
18
18
|
@llm_api_key = llm_api_key
|
19
19
|
@tools = tools
|
20
20
|
|
21
|
-
@llm_client = LLM.const_get(LLM::Base::LLMS.fetch(llm)).new(api_key: llm_api_key)
|
21
|
+
@llm_client = Langchain::LLM.const_get(Langchain::LLM::Base::LLMS.fetch(llm)).new(api_key: llm_api_key)
|
22
22
|
end
|
23
23
|
|
24
24
|
# Validate tools when they're re-assigned
|
@@ -26,7 +26,7 @@ module Agent
|
|
26
26
|
# @param value [Array] The tools to use
|
27
27
|
# @return [Array] The tools that will be used
|
28
28
|
def tools=(value)
|
29
|
-
Tool::Base.validate_tools!(tools: value)
|
29
|
+
Langchain::Tool::Base.validate_tools!(tools: value)
|
30
30
|
@tools = value
|
31
31
|
end
|
32
32
|
|
@@ -42,7 +42,7 @@ module Agent
|
|
42
42
|
)
|
43
43
|
|
44
44
|
loop do
|
45
|
-
Langchain.logger.info("
|
45
|
+
Langchain.logger.info("[#{self.class.name}]".red + ": Sending the prompt to the #{llm} LLM")
|
46
46
|
response = llm_client.complete(
|
47
47
|
prompt: prompt,
|
48
48
|
stop_sequences: ["Observation:"],
|
@@ -59,12 +59,11 @@ module Agent
|
|
59
59
|
# Find the input to the action in the "Action Input: [action_input]" format
|
60
60
|
action_input = response.match(/Action Input: "?(.*)"?/)&.send(:[], -1)
|
61
61
|
|
62
|
-
Langchain.logger.info("Agent: Using the \"#{action}\" Tool with \"#{action_input}\"")
|
63
|
-
|
64
62
|
# Retrieve the Tool::[ToolName] class and call `execute`` with action_input as the input
|
65
|
-
|
66
|
-
|
67
|
-
|
63
|
+
tool = Langchain::Tool.const_get(Langchain::Tool::Base::TOOLS[action.strip])
|
64
|
+
Langchain.logger.info("[#{self.class.name}]".red + ": Invoking \"#{tool}\" Tool with \"#{action_input}\"")
|
65
|
+
|
66
|
+
result = tool.execute(input: action_input)
|
68
67
|
|
69
68
|
# Append the Observation to the prompt
|
70
69
|
prompt += if prompt.end_with?("Observation:")
|
@@ -91,7 +90,7 @@ module Agent
|
|
91
90
|
question: question,
|
92
91
|
tool_names: "[#{tools.join(", ")}]",
|
93
92
|
tools: tools.map do |tool|
|
94
|
-
"#{tool}: #{Tool.const_get(Tool::Base::TOOLS[tool]).const_get(:DESCRIPTION)}"
|
93
|
+
"#{tool}: #{Langchain::Tool.const_get(Langchain::Tool::Base::TOOLS[tool]).const_get(:DESCRIPTION)}"
|
95
94
|
end.join("\n")
|
96
95
|
)
|
97
96
|
end
|
@@ -99,8 +98,8 @@ module Agent
|
|
99
98
|
# Load the PromptTemplate from the JSON file
|
100
99
|
# @return [PromptTemplate] PromptTemplate instance
|
101
100
|
def prompt_template
|
102
|
-
@template ||= Prompt.load_from_path(
|
103
|
-
file_path: Langchain.root.join("agent/chain_of_thought_agent/chain_of_thought_agent_prompt.json")
|
101
|
+
@template ||= Langchain::Prompt.load_from_path(
|
102
|
+
file_path: Langchain.root.join("langchain/agent/chain_of_thought_agent/chain_of_thought_agent_prompt.json")
|
104
103
|
)
|
105
104
|
end
|
106
105
|
end
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module LLM
|
3
|
+
module Langchain::LLM
|
4
4
|
class Base
|
5
5
|
attr_reader :client
|
6
6
|
|
@@ -42,8 +42,8 @@ module LLM
|
|
42
42
|
# @param llm [Symbol] The LLM to use
|
43
43
|
def self.validate_llm!(llm:)
|
44
44
|
# TODO: Fix so this works when `llm` value is a string instead of a symbol
|
45
|
-
unless LLM::Base::LLMS.key?(llm)
|
46
|
-
raise ArgumentError, "LLM must be one of #{LLM::Base::LLMS.keys}"
|
45
|
+
unless Langchain::LLM::Base::LLMS.key?(llm)
|
46
|
+
raise ArgumentError, "LLM must be one of #{Langchain::LLM::Base::LLMS.keys}"
|
47
47
|
end
|
48
48
|
end
|
49
49
|
end
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module LLM
|
3
|
+
module Langchain::LLM
|
4
4
|
class GooglePalm < Base
|
5
5
|
# Wrapper around the Google PaLM (Pathways Language Model) APIs.
|
6
6
|
|
@@ -89,8 +89,8 @@ module LLM
|
|
89
89
|
# @return [String] The summarization
|
90
90
|
#
|
91
91
|
def summarize(text:)
|
92
|
-
prompt_template = Prompt.load_from_path(
|
93
|
-
file_path: Langchain.root.join("llm/prompts/summarize_template.json")
|
92
|
+
prompt_template = Langchain::Prompt.load_from_path(
|
93
|
+
file_path: Langchain.root.join("langchain/llm/prompts/summarize_template.json")
|
94
94
|
)
|
95
95
|
prompt = prompt_template.format(text: text)
|
96
96
|
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module LLM
|
3
|
+
module Langchain::LLM
|
4
4
|
class OpenAI < Base
|
5
5
|
DEFAULTS = {
|
6
6
|
temperature: 0.0,
|
@@ -25,9 +25,13 @@ module LLM
|
|
25
25
|
# @return [Array] The embedding
|
26
26
|
#
|
27
27
|
def embed(text:)
|
28
|
+
model = DEFAULTS[:embeddings_model_name]
|
29
|
+
|
30
|
+
Langchain::Utils::TokenLengthValidator.validate!(text, model)
|
31
|
+
|
28
32
|
response = client.embeddings(
|
29
33
|
parameters: {
|
30
|
-
model:
|
34
|
+
model: model,
|
31
35
|
input: text
|
32
36
|
}
|
33
37
|
)
|
@@ -41,8 +45,12 @@ module LLM
|
|
41
45
|
# @return [String] The completion
|
42
46
|
#
|
43
47
|
def complete(prompt:, **params)
|
48
|
+
model = DEFAULTS[:completion_model_name]
|
49
|
+
|
50
|
+
Langchain::Utils::TokenLengthValidator.validate!(prompt, model)
|
51
|
+
|
44
52
|
default_params = {
|
45
|
-
model:
|
53
|
+
model: model,
|
46
54
|
temperature: DEFAULTS[:temperature],
|
47
55
|
prompt: prompt
|
48
56
|
}
|
@@ -64,8 +72,12 @@ module LLM
|
|
64
72
|
# @return [String] The chat completion
|
65
73
|
#
|
66
74
|
def chat(prompt:, **params)
|
75
|
+
model = DEFAULTS[:chat_completion_model_name]
|
76
|
+
|
77
|
+
Langchain::Utils::TokenLengthValidator.validate!(prompt, model)
|
78
|
+
|
67
79
|
default_params = {
|
68
|
-
model:
|
80
|
+
model: model,
|
69
81
|
temperature: DEFAULTS[:temperature],
|
70
82
|
# TODO: Figure out how to introduce persisted conversations
|
71
83
|
messages: [{role: "user", content: prompt}]
|
@@ -88,8 +100,8 @@ module LLM
|
|
88
100
|
# @return [String] The summary
|
89
101
|
#
|
90
102
|
def summarize(text:)
|
91
|
-
prompt_template = Prompt.load_from_path(
|
92
|
-
file_path: Langchain.root.join("llm/prompts/summarize_template.json")
|
103
|
+
prompt_template = Langchain::Prompt.load_from_path(
|
104
|
+
file_path: Langchain.root.join("langchain/llm/prompts/summarize_template.json")
|
93
105
|
)
|
94
106
|
prompt = prompt_template.format(text: text)
|
95
107
|
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module LLM
|
3
|
+
module Langchain::LLM
|
4
4
|
class Replicate < Base
|
5
5
|
# Wrapper around Replicate.com LLM provider
|
6
6
|
# Use it directly:
|
@@ -89,8 +89,8 @@ module LLM
|
|
89
89
|
# @return [String] The summary
|
90
90
|
#
|
91
91
|
def summarize(text:)
|
92
|
-
prompt_template = Prompt.load_from_path(
|
93
|
-
file_path: Langchain.root.join("llm/prompts/summarize_template.json")
|
92
|
+
prompt_template = Langchain::Prompt.load_from_path(
|
93
|
+
file_path: Langchain.root.join("langchain/llm/prompts/summarize_template.json")
|
94
94
|
)
|
95
95
|
prompt = prompt_template.format(text: text)
|
96
96
|
|
@@ -3,7 +3,7 @@
|
|
3
3
|
require "strscan"
|
4
4
|
require "json"
|
5
5
|
|
6
|
-
module Prompt
|
6
|
+
module Langchain::Prompt
|
7
7
|
class Base
|
8
8
|
def format(**kwargs)
|
9
9
|
raise NotImplementedError
|
@@ -29,7 +29,7 @@ module Prompt
|
|
29
29
|
#
|
30
30
|
def validate(template:, input_variables:)
|
31
31
|
input_variables_set = @input_variables.uniq
|
32
|
-
variables_from_template = Prompt::Base.extract_variables_from_template(template)
|
32
|
+
variables_from_template = Langchain::Prompt::Base.extract_variables_from_template(template)
|
33
33
|
|
34
34
|
missing_variables = variables_from_template - input_variables_set
|
35
35
|
extra_variables = input_variables_set - variables_from_template
|
@@ -3,10 +3,10 @@
|
|
3
3
|
require "strscan"
|
4
4
|
require "pathname"
|
5
5
|
|
6
|
-
module Prompt
|
6
|
+
module Langchain::Prompt
|
7
7
|
TYPE_TO_LOADER = {
|
8
|
-
"prompt" => ->(config) {
|
9
|
-
"few_shot" => ->(config) {
|
8
|
+
"prompt" => ->(config) { load_prompt(config) },
|
9
|
+
"few_shot" => ->(config) { load_few_shot_prompt(config) }
|
10
10
|
}
|
11
11
|
|
12
12
|
class << self
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Tool
|
3
|
+
module Langchain::Tool
|
4
4
|
class Base
|
5
5
|
# How to add additional Tools?
|
6
6
|
# 1. Create a new file in lib/tool/your_tool_name.rb
|
@@ -10,9 +10,9 @@ module Tool
|
|
10
10
|
# 4. Add your tool to the README.md
|
11
11
|
|
12
12
|
TOOLS = {
|
13
|
-
"calculator" => "Tool::Calculator",
|
14
|
-
"search" => "Tool::SerpApi",
|
15
|
-
"wikipedia" => "Tool::Wikipedia"
|
13
|
+
"calculator" => "Langchain::Tool::Calculator",
|
14
|
+
"search" => "Langchain::Tool::SerpApi",
|
15
|
+
"wikipedia" => "Langchain::Tool::Wikipedia"
|
16
16
|
}
|
17
17
|
|
18
18
|
def self.description(value)
|
@@ -40,7 +40,7 @@ module Tool
|
|
40
40
|
# @raise [ArgumentError] If any of the tools are not supported
|
41
41
|
#
|
42
42
|
def self.validate_tools!(tools:)
|
43
|
-
unrecognized_tools = tools - Tool::Base::TOOLS.keys
|
43
|
+
unrecognized_tools = tools - Langchain::Tool::Base::TOOLS.keys
|
44
44
|
|
45
45
|
if unrecognized_tools.any?
|
46
46
|
raise ArgumentError, "Unrecognized Tools: #{unrecognized_tools}"
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Tool
|
3
|
+
module Langchain::Tool
|
4
4
|
class Calculator < Base
|
5
5
|
description <<~DESC
|
6
6
|
Useful for getting the result of a math expression.
|
@@ -18,11 +18,13 @@ module Tool
|
|
18
18
|
# @param input [String] math expression
|
19
19
|
# @return [String] Answer
|
20
20
|
def execute(input:)
|
21
|
+
Langchain.logger.info("[#{self.class.name}]".light_blue + ": Executing \"#{input}\"")
|
22
|
+
|
21
23
|
Eqn::Calculator.calc(input)
|
22
24
|
rescue Eqn::ParseError, Eqn::NoVariableValueError
|
23
25
|
# Sometimes the input is not a pure math expression, e.g: "12F in Celsius"
|
24
26
|
# We can use the google answer box to evaluate this expression
|
25
|
-
hash_results = Tool::SerpApi.execute_search(input: input)
|
27
|
+
hash_results = Langchain::Tool::SerpApi.execute_search(input: input)
|
26
28
|
hash_results.dig(:answer_box, :to)
|
27
29
|
end
|
28
30
|
end
|
@@ -0,0 +1,27 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
module Langchain::Tool
|
4
|
+
class RubyCodeInterpreter < Base
|
5
|
+
description <<~DESC
|
6
|
+
A Ruby code interpreter. Use this to execute ruby expressions. Input should be a valid ruby expression. If you want to see the output of the tool, make sure to return a value.
|
7
|
+
DESC
|
8
|
+
|
9
|
+
def initialize(timeout: 30)
|
10
|
+
@timeout = timeout
|
11
|
+
depends_on "safe_ruby"
|
12
|
+
require "safe_ruby"
|
13
|
+
end
|
14
|
+
|
15
|
+
# @param input [String] ruby code expression
|
16
|
+
# @return [String] Answer
|
17
|
+
def execute(input:)
|
18
|
+
Langchain.logger.info("[#{self.class.name}]".light_blue + ": Executing \"#{input}\"")
|
19
|
+
|
20
|
+
safe_eval(input)
|
21
|
+
end
|
22
|
+
|
23
|
+
def safe_eval(code)
|
24
|
+
SafeRuby.eval(code, timeout: @timeout)
|
25
|
+
end
|
26
|
+
end
|
27
|
+
end
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Tool
|
3
|
+
module Langchain::Tool
|
4
4
|
class SerpApi < Base
|
5
5
|
# Wrapper around SerpAPI
|
6
6
|
# Set ENV["SERPAPI_API_KEY"] to use it
|
@@ -33,6 +33,8 @@ module Tool
|
|
33
33
|
# TODO: Glance at all of the fields that langchain Python looks through: https://github.com/hwchase17/langchain/blob/v0.0.166/langchain/utilities/serpapi.py#L128-L156
|
34
34
|
# We may need to do the same thing here.
|
35
35
|
def execute(input:)
|
36
|
+
Langchain.logger.info("[#{self.class.name}]".light_blue + ": Executing \"#{input}\"")
|
37
|
+
|
36
38
|
hash_results = execute_search(input: input)
|
37
39
|
|
38
40
|
hash_results.dig(:answer_box, :answer) ||
|
@@ -1,6 +1,6 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Tool
|
3
|
+
module Langchain::Tool
|
4
4
|
class Wikipedia < Base
|
5
5
|
# Tool that adds the capability to search using the Wikipedia API
|
6
6
|
|
@@ -22,6 +22,8 @@ module Tool
|
|
22
22
|
# @param input [String] search query
|
23
23
|
# @return [String] Answer
|
24
24
|
def execute(input:)
|
25
|
+
Langchain.logger.info("[#{self.class.name}]".light_blue + ": Executing \"#{input}\"")
|
26
|
+
|
25
27
|
page = ::Wikipedia.find(input)
|
26
28
|
# It would be nice to figure out a way to provide page.content but the LLM token limit is an issue
|
27
29
|
page.summary
|
@@ -0,0 +1,57 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
require "tiktoken_ruby"
|
4
|
+
|
5
|
+
module Langchain
|
6
|
+
module Utils
|
7
|
+
class TokenLimitExceeded < StandardError; end
|
8
|
+
|
9
|
+
class TokenLengthValidator
|
10
|
+
#
|
11
|
+
# This class is meant to validate the length of the text passed in to OpenAI's API.
|
12
|
+
# It is used to validate the token length before the API call is made
|
13
|
+
#
|
14
|
+
TOKEN_LIMITS = {
|
15
|
+
# Source:
|
16
|
+
# https://platform.openai.com/docs/api-reference/embeddings
|
17
|
+
# https://platform.openai.com/docs/models/gpt-4
|
18
|
+
"text-embedding-ada-002" => 8191,
|
19
|
+
"gpt-3.5-turbo" => 4096,
|
20
|
+
"gpt-3.5-turbo-0301" => 4096,
|
21
|
+
"text-davinci-003" => 4097,
|
22
|
+
"text-davinci-002" => 4097,
|
23
|
+
"code-davinci-002" => 8001,
|
24
|
+
"gpt-4" => 8192,
|
25
|
+
"gpt-4-0314" => 8192,
|
26
|
+
"gpt-4-32k" => 32768,
|
27
|
+
"gpt-4-32k-0314" => 32768,
|
28
|
+
"text-curie-001" => 2049,
|
29
|
+
"text-babbage-001" => 2049,
|
30
|
+
"text-ada-001" => 2049,
|
31
|
+
"davinci" => 2049,
|
32
|
+
"curie" => 2049,
|
33
|
+
"babbage" => 2049,
|
34
|
+
"ada" => 2049
|
35
|
+
}.freeze
|
36
|
+
|
37
|
+
#
|
38
|
+
# Validate the length of the text passed in to OpenAI's API
|
39
|
+
#
|
40
|
+
# @param text [String] The text to validate
|
41
|
+
# @param model_name [String] The model name to validate against
|
42
|
+
# @return [Boolean] Whether the text is valid or not
|
43
|
+
# @raise [TokenLimitExceeded] If the text is too long
|
44
|
+
#
|
45
|
+
def self.validate!(text, model_name)
|
46
|
+
encoder = Tiktoken.encoding_for_model(model_name)
|
47
|
+
token_length = encoder.encode(text).length
|
48
|
+
|
49
|
+
if token_length > TOKEN_LIMITS[model_name]
|
50
|
+
raise TokenLimitExceeded, "This model's maximum context length is #{TOKEN_LIMITS[model_name]} tokens, but the given text is #{token_length} tokens long."
|
51
|
+
end
|
52
|
+
|
53
|
+
true
|
54
|
+
end
|
55
|
+
end
|
56
|
+
end
|
57
|
+
end
|
@@ -2,7 +2,7 @@
|
|
2
2
|
|
3
3
|
require "forwardable"
|
4
4
|
|
5
|
-
module Vectorsearch
|
5
|
+
module Langchain::Vectorsearch
|
6
6
|
class Base
|
7
7
|
extend Forwardable
|
8
8
|
|
@@ -13,12 +13,12 @@ module Vectorsearch
|
|
13
13
|
# @param llm [Symbol] The LLM to use
|
14
14
|
# @param llm_api_key [String] The API key for the LLM
|
15
15
|
def initialize(llm:, llm_api_key:)
|
16
|
-
LLM::Base.validate_llm!(llm: llm)
|
16
|
+
Langchain::LLM::Base.validate_llm!(llm: llm)
|
17
17
|
|
18
18
|
@llm = llm
|
19
19
|
@llm_api_key = llm_api_key
|
20
20
|
|
21
|
-
@llm_client = LLM.const_get(LLM::Base::LLMS.fetch(llm)).new(api_key: llm_api_key)
|
21
|
+
@llm_client = Langchain::LLM.const_get(Langchain::LLM::Base::LLMS.fetch(llm)).new(api_key: llm_api_key)
|
22
22
|
end
|
23
23
|
|
24
24
|
# Method supported by Vectorsearch DB to create a default schema
|
@@ -51,10 +51,10 @@ module Vectorsearch
|
|
51
51
|
:default_dimension
|
52
52
|
|
53
53
|
def generate_prompt(question:, context:)
|
54
|
-
prompt_template = Prompt::FewShotPromptTemplate.new(
|
54
|
+
prompt_template = Langchain::Prompt::FewShotPromptTemplate.new(
|
55
55
|
prefix: "Context:",
|
56
56
|
suffix: "---\nQuestion: {question}\n---\nAnswer:",
|
57
|
-
example_prompt: Prompt::PromptTemplate.new(
|
57
|
+
example_prompt: Langchain::Prompt::PromptTemplate.new(
|
58
58
|
template: "{context}",
|
59
59
|
input_variables: ["context"]
|
60
60
|
),
|
@@ -1,8 +1,17 @@
|
|
1
1
|
# frozen_string_literal: true
|
2
2
|
|
3
|
-
module Vectorsearch
|
3
|
+
module Langchain::Vectorsearch
|
4
4
|
# The PostgreSQL vector search adapter
|
5
5
|
class Pgvector < Base
|
6
|
+
# The operators supported by the PostgreSQL vector search adapter
|
7
|
+
OPERATORS = {
|
8
|
+
"cosine_distance" => "<=>",
|
9
|
+
"euclidean_distance" => "<->"
|
10
|
+
}
|
11
|
+
DEFAULT_OPERATOR = "cosine_distance"
|
12
|
+
|
13
|
+
attr_reader :operator, :quoted_table_name
|
14
|
+
|
6
15
|
# @param url [String] The URL of the PostgreSQL database
|
7
16
|
# @param index_name [String] The name of the table to use for the index
|
8
17
|
# @param llm [String] The URL of the Language Layer API
|
@@ -18,6 +27,8 @@ module Vectorsearch
|
|
18
27
|
@client.type_map_for_results = PG::BasicTypeMapForResults.new(@client, registry: registry)
|
19
28
|
|
20
29
|
@index_name = index_name
|
30
|
+
@quoted_table_name = @client.quote_ident(index_name)
|
31
|
+
@operator = OPERATORS[DEFAULT_OPERATOR]
|
21
32
|
|
22
33
|
super(llm: llm, llm_api_key: llm_api_key)
|
23
34
|
end
|
@@ -31,7 +42,7 @@ module Vectorsearch
|
|
31
42
|
end
|
32
43
|
values = texts.length.times.map { |i| "($#{2 * i + 1}, $#{2 * i + 2})" }.join(",")
|
33
44
|
client.exec_params(
|
34
|
-
"INSERT INTO #{
|
45
|
+
"INSERT INTO #{quoted_table_name} (content, vectors) VALUES #{values};",
|
35
46
|
data
|
36
47
|
)
|
37
48
|
end
|
@@ -42,7 +53,7 @@ module Vectorsearch
|
|
42
53
|
client.exec("CREATE EXTENSION IF NOT EXISTS vector;")
|
43
54
|
client.exec(
|
44
55
|
<<~SQL
|
45
|
-
CREATE TABLE IF NOT EXISTS #{
|
56
|
+
CREATE TABLE IF NOT EXISTS #{quoted_table_name} (
|
46
57
|
id serial PRIMARY KEY,
|
47
58
|
content TEXT,
|
48
59
|
vectors VECTOR(#{default_dimension})
|
@@ -73,7 +84,7 @@ module Vectorsearch
|
|
73
84
|
result = client.transaction do |conn|
|
74
85
|
conn.exec("SET LOCAL ivfflat.probes = 10;")
|
75
86
|
query = <<~SQL
|
76
|
-
SELECT id, content FROM #{
|
87
|
+
SELECT id, content FROM #{quoted_table_name} ORDER BY vectors #{operator} $1 ASC LIMIT $2;
|
77
88
|
SQL
|
78
89
|
conn.exec_params(query, [embedding, k])
|
79
90
|
end
|
data/lib/langchain.rb
CHANGED
@@ -2,9 +2,11 @@
|
|
2
2
|
|
3
3
|
require "logger"
|
4
4
|
require "pathname"
|
5
|
+
require "colorize"
|
5
6
|
|
6
7
|
require_relative "./version"
|
7
8
|
require_relative "./dependency_helper"
|
9
|
+
|
8
10
|
module Langchain
|
9
11
|
class << self
|
10
12
|
attr_accessor :logger
|
@@ -12,13 +14,26 @@ module Langchain
|
|
12
14
|
attr_reader :root
|
13
15
|
end
|
14
16
|
|
15
|
-
@logger ||= ::Logger.new($stdout, level: :warn, formatter: ->(severity, datetime, progname, msg) { "[LangChain.rb] #{msg}\n" })
|
17
|
+
@logger ||= ::Logger.new($stdout, level: :warn, formatter: ->(severity, datetime, progname, msg) { "[LangChain.rb]".yellow + " #{msg}\n" })
|
16
18
|
|
17
19
|
@root = Pathname.new(__dir__)
|
18
20
|
|
19
21
|
autoload :Loader, "langchain/loader"
|
20
22
|
autoload :Data, "langchain/data"
|
21
23
|
|
24
|
+
module Agent
|
25
|
+
autoload :Base, "langchain/agent/base"
|
26
|
+
autoload :ChainOfThoughtAgent, "langchain/agent/chain_of_thought_agent/chain_of_thought_agent.rb"
|
27
|
+
end
|
28
|
+
|
29
|
+
module Tool
|
30
|
+
autoload :Base, "langchain/tool/base"
|
31
|
+
autoload :Calculator, "langchain/tool/calculator"
|
32
|
+
autoload :RubyCodeInterpreter, "langchain/tool/ruby_code_interpreter"
|
33
|
+
autoload :SerpApi, "langchain/tool/serp_api"
|
34
|
+
autoload :Wikipedia, "langchain/tool/wikipedia"
|
35
|
+
end
|
36
|
+
|
22
37
|
module Processors
|
23
38
|
autoload :Base, "langchain/processors/base"
|
24
39
|
autoload :CSV, "langchain/processors/csv"
|
@@ -29,43 +44,35 @@ module Langchain
|
|
29
44
|
autoload :PDF, "langchain/processors/pdf"
|
30
45
|
autoload :Text, "langchain/processors/text"
|
31
46
|
end
|
32
|
-
end
|
33
|
-
|
34
|
-
module Agent
|
35
|
-
autoload :Base, "agent/base"
|
36
|
-
autoload :ChainOfThoughtAgent, "agent/chain_of_thought_agent/chain_of_thought_agent.rb"
|
37
|
-
end
|
38
47
|
|
39
|
-
module
|
40
|
-
|
41
|
-
|
42
|
-
autoload :Milvus, "vectorsearch/milvus"
|
43
|
-
autoload :Pinecone, "vectorsearch/pinecone"
|
44
|
-
autoload :Pgvector, "vectorsearch/pgvector"
|
45
|
-
autoload :Qdrant, "vectorsearch/qdrant"
|
46
|
-
autoload :Weaviate, "vectorsearch/weaviate"
|
47
|
-
end
|
48
|
+
module Utils
|
49
|
+
autoload :TokenLengthValidator, "langchain/utils/token_length_validator"
|
50
|
+
end
|
48
51
|
|
49
|
-
module
|
50
|
-
|
51
|
-
|
52
|
-
|
53
|
-
|
54
|
-
|
55
|
-
|
56
|
-
|
52
|
+
module Vectorsearch
|
53
|
+
autoload :Base, "langchain/vectorsearch/base"
|
54
|
+
autoload :Chroma, "langchain/vectorsearch/chroma"
|
55
|
+
autoload :Milvus, "langchain/vectorsearch/milvus"
|
56
|
+
autoload :Pinecone, "langchain/vectorsearch/pinecone"
|
57
|
+
autoload :Pgvector, "langchain/vectorsearch/pgvector"
|
58
|
+
autoload :Qdrant, "langchain/vectorsearch/qdrant"
|
59
|
+
autoload :Weaviate, "langchain/vectorsearch/weaviate"
|
60
|
+
end
|
57
61
|
|
58
|
-
module
|
59
|
-
|
62
|
+
module LLM
|
63
|
+
autoload :Base, "langchain/llm/base"
|
64
|
+
autoload :Cohere, "langchain/llm/cohere"
|
65
|
+
autoload :GooglePalm, "langchain/llm/google_palm"
|
66
|
+
autoload :HuggingFace, "langchain/llm/hugging_face"
|
67
|
+
autoload :OpenAI, "langchain/llm/openai"
|
68
|
+
autoload :Replicate, "langchain/llm/replicate"
|
69
|
+
end
|
60
70
|
|
61
|
-
|
62
|
-
|
63
|
-
autoload :FewShotPromptTemplate, "prompt/few_shot_prompt_template"
|
64
|
-
end
|
71
|
+
module Prompt
|
72
|
+
require_relative "langchain/prompt/loading"
|
65
73
|
|
66
|
-
|
67
|
-
|
68
|
-
|
69
|
-
|
70
|
-
autoload :Wikipedia, "tool/wikipedia"
|
74
|
+
autoload :Base, "langchain/prompt/base"
|
75
|
+
autoload :PromptTemplate, "langchain/prompt/prompt_template"
|
76
|
+
autoload :FewShotPromptTemplate, "langchain/prompt/few_shot_prompt_template"
|
77
|
+
end
|
71
78
|
end
|
data/lib/version.rb
CHANGED
metadata
CHANGED
@@ -1,15 +1,43 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: langchainrb
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.4.1
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Andrei Bondarev
|
8
8
|
autorequire:
|
9
9
|
bindir: exe
|
10
10
|
cert_chain: []
|
11
|
-
date: 2023-
|
11
|
+
date: 2023-06-02 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
|
+
- !ruby/object:Gem::Dependency
|
14
|
+
name: tiktoken_ruby
|
15
|
+
requirement: !ruby/object:Gem::Requirement
|
16
|
+
requirements:
|
17
|
+
- - "~>"
|
18
|
+
- !ruby/object:Gem::Version
|
19
|
+
version: 0.0.5
|
20
|
+
type: :runtime
|
21
|
+
prerelease: false
|
22
|
+
version_requirements: !ruby/object:Gem::Requirement
|
23
|
+
requirements:
|
24
|
+
- - "~>"
|
25
|
+
- !ruby/object:Gem::Version
|
26
|
+
version: 0.0.5
|
27
|
+
- !ruby/object:Gem::Dependency
|
28
|
+
name: colorize
|
29
|
+
requirement: !ruby/object:Gem::Requirement
|
30
|
+
requirements:
|
31
|
+
- - "~>"
|
32
|
+
- !ruby/object:Gem::Version
|
33
|
+
version: 0.8.1
|
34
|
+
type: :runtime
|
35
|
+
prerelease: false
|
36
|
+
version_requirements: !ruby/object:Gem::Requirement
|
37
|
+
requirements:
|
38
|
+
- - "~>"
|
39
|
+
- !ruby/object:Gem::Version
|
40
|
+
version: 0.8.1
|
13
41
|
- !ruby/object:Gem::Dependency
|
14
42
|
name: dotenv-rails
|
15
43
|
requirement: !ruby/object:Gem::Requirement
|
@@ -262,6 +290,20 @@ dependencies:
|
|
262
290
|
- - "~>"
|
263
291
|
- !ruby/object:Gem::Version
|
264
292
|
version: 4.0.0
|
293
|
+
- !ruby/object:Gem::Dependency
|
294
|
+
name: safe_ruby
|
295
|
+
requirement: !ruby/object:Gem::Requirement
|
296
|
+
requirements:
|
297
|
+
- - "~>"
|
298
|
+
- !ruby/object:Gem::Version
|
299
|
+
version: 1.0.4
|
300
|
+
type: :development
|
301
|
+
prerelease: false
|
302
|
+
version_requirements: !ruby/object:Gem::Requirement
|
303
|
+
requirements:
|
304
|
+
- - "~>"
|
305
|
+
- !ruby/object:Gem::Version
|
306
|
+
version: 1.0.4
|
265
307
|
- !ruby/object:Gem::Dependency
|
266
308
|
name: weaviate-ruby
|
267
309
|
requirement: !ruby/object:Gem::Requirement
|
@@ -311,12 +353,19 @@ files:
|
|
311
353
|
- examples/store_and_query_with_pinecone.rb
|
312
354
|
- examples/store_and_query_with_qdrant.rb
|
313
355
|
- examples/store_and_query_with_weaviate.rb
|
314
|
-
- lib/agent/base.rb
|
315
|
-
- lib/agent/chain_of_thought_agent/chain_of_thought_agent.rb
|
316
|
-
- lib/agent/chain_of_thought_agent/chain_of_thought_agent_prompt.json
|
317
356
|
- lib/dependency_helper.rb
|
318
357
|
- lib/langchain.rb
|
358
|
+
- lib/langchain/agent/base.rb
|
359
|
+
- lib/langchain/agent/chain_of_thought_agent/chain_of_thought_agent.rb
|
360
|
+
- lib/langchain/agent/chain_of_thought_agent/chain_of_thought_agent_prompt.json
|
319
361
|
- lib/langchain/data.rb
|
362
|
+
- lib/langchain/llm/base.rb
|
363
|
+
- lib/langchain/llm/cohere.rb
|
364
|
+
- lib/langchain/llm/google_palm.rb
|
365
|
+
- lib/langchain/llm/hugging_face.rb
|
366
|
+
- lib/langchain/llm/openai.rb
|
367
|
+
- lib/langchain/llm/prompts/summarize_template.json
|
368
|
+
- lib/langchain/llm/replicate.rb
|
320
369
|
- lib/langchain/loader.rb
|
321
370
|
- lib/langchain/processors/base.rb
|
322
371
|
- lib/langchain/processors/csv.rb
|
@@ -326,29 +375,24 @@ files:
|
|
326
375
|
- lib/langchain/processors/jsonl.rb
|
327
376
|
- lib/langchain/processors/pdf.rb
|
328
377
|
- lib/langchain/processors/text.rb
|
378
|
+
- lib/langchain/prompt/base.rb
|
379
|
+
- lib/langchain/prompt/few_shot_prompt_template.rb
|
380
|
+
- lib/langchain/prompt/loading.rb
|
381
|
+
- lib/langchain/prompt/prompt_template.rb
|
382
|
+
- lib/langchain/tool/base.rb
|
383
|
+
- lib/langchain/tool/calculator.rb
|
384
|
+
- lib/langchain/tool/ruby_code_interpreter.rb
|
385
|
+
- lib/langchain/tool/serp_api.rb
|
386
|
+
- lib/langchain/tool/wikipedia.rb
|
387
|
+
- lib/langchain/utils/token_length_validator.rb
|
388
|
+
- lib/langchain/vectorsearch/base.rb
|
389
|
+
- lib/langchain/vectorsearch/chroma.rb
|
390
|
+
- lib/langchain/vectorsearch/milvus.rb
|
391
|
+
- lib/langchain/vectorsearch/pgvector.rb
|
392
|
+
- lib/langchain/vectorsearch/pinecone.rb
|
393
|
+
- lib/langchain/vectorsearch/qdrant.rb
|
394
|
+
- lib/langchain/vectorsearch/weaviate.rb
|
329
395
|
- lib/langchainrb.rb
|
330
|
-
- lib/llm/base.rb
|
331
|
-
- lib/llm/cohere.rb
|
332
|
-
- lib/llm/google_palm.rb
|
333
|
-
- lib/llm/hugging_face.rb
|
334
|
-
- lib/llm/openai.rb
|
335
|
-
- lib/llm/prompts/summarize_template.json
|
336
|
-
- lib/llm/replicate.rb
|
337
|
-
- lib/prompt/base.rb
|
338
|
-
- lib/prompt/few_shot_prompt_template.rb
|
339
|
-
- lib/prompt/loading.rb
|
340
|
-
- lib/prompt/prompt_template.rb
|
341
|
-
- lib/tool/base.rb
|
342
|
-
- lib/tool/calculator.rb
|
343
|
-
- lib/tool/serp_api.rb
|
344
|
-
- lib/tool/wikipedia.rb
|
345
|
-
- lib/vectorsearch/base.rb
|
346
|
-
- lib/vectorsearch/chroma.rb
|
347
|
-
- lib/vectorsearch/milvus.rb
|
348
|
-
- lib/vectorsearch/pgvector.rb
|
349
|
-
- lib/vectorsearch/pinecone.rb
|
350
|
-
- lib/vectorsearch/qdrant.rb
|
351
|
-
- lib/vectorsearch/weaviate.rb
|
352
396
|
- lib/version.rb
|
353
397
|
- sig/langchain.rbs
|
354
398
|
homepage: https://rubygems.org/gems/langchainrb
|
@@ -373,7 +417,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
373
417
|
- !ruby/object:Gem::Version
|
374
418
|
version: '0'
|
375
419
|
requirements: []
|
376
|
-
rubygems_version: 3.3
|
420
|
+
rubygems_version: 3.2.3
|
377
421
|
signing_key:
|
378
422
|
specification_version: 4
|
379
423
|
summary: Build LLM-backed Ruby applications with Ruby's LangChain
|
/data/lib/{agent → langchain/agent}/chain_of_thought_agent/chain_of_thought_agent_prompt.json
RENAMED
File without changes
|
File without changes
|