langchainrb 0.12.1 → 0.13.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -9,8 +9,8 @@ module Langchain::Tool
9
9
  # gem "wikipedia-client", "~> 1.17.0"
10
10
  #
11
11
  # Usage:
12
- # weather = Langchain::Tool::Wikipedia.new
13
- # weather.execute(input: "The Roman Empire")
12
+ # wikipedia = Langchain::Tool::Wikipedia.new
13
+ # wikipedia.execute(input: "The Roman Empire")
14
14
  #
15
15
  NAME = "wikipedia"
16
16
  ANNOTATIONS_PATH = Langchain.root.join("./langchain/tool/#{NAME}/#{NAME}.json").to_path
@@ -1,5 +1,5 @@
1
1
  # frozen_string_literal: true
2
2
 
3
3
  module Langchain
4
- VERSION = "0.12.1"
4
+ VERSION = "0.13.0"
5
5
  end
data/lib/langchain.rb CHANGED
@@ -12,6 +12,7 @@ loader.inflector.inflect(
12
12
  "ai21_response" => "AI21Response",
13
13
  "ai21_validator" => "AI21Validator",
14
14
  "csv" => "CSV",
15
+ "google_vertex_ai" => "GoogleVertexAI",
15
16
  "html" => "HTML",
16
17
  "json" => "JSON",
17
18
  "jsonl" => "JSONL",
@@ -21,6 +22,7 @@ loader.inflector.inflect(
21
22
  "openai" => "OpenAI",
22
23
  "openai_validator" => "OpenAIValidator",
23
24
  "openai_response" => "OpenAIResponse",
25
+ "openai_message" => "OpenAIMessage",
24
26
  "pdf" => "PDF"
25
27
  )
26
28
  loader.collapse("#{__dir__}/langchain/llm/response")
@@ -31,6 +33,7 @@ loader.collapse("#{__dir__}/langchain/tool/database")
31
33
  loader.collapse("#{__dir__}/langchain/tool/file_system")
32
34
  loader.collapse("#{__dir__}/langchain/tool/google_search")
33
35
  loader.collapse("#{__dir__}/langchain/tool/ruby_code_interpreter")
36
+ loader.collapse("#{__dir__}/langchain/tool/news_retriever")
34
37
  loader.collapse("#{__dir__}/langchain/tool/vectorsearch")
35
38
  loader.collapse("#{__dir__}/langchain/tool/weather")
36
39
  loader.collapse("#{__dir__}/langchain/tool/wikipedia")
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: langchainrb
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.12.1
4
+ version: 0.13.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Andrei Bondarev
8
8
  autorequire:
9
9
  bindir: exe
10
10
  cert_chain: []
11
- date: 2024-05-13 00:00:00.000000000 Z
11
+ date: 2024-05-14 00:00:00.000000000 Z
12
12
  dependencies:
13
13
  - !ruby/object:Gem::Dependency
14
14
  name: activesupport
@@ -347,19 +347,19 @@ dependencies:
347
347
  - !ruby/object:Gem::Version
348
348
  version: 1.6.5
349
349
  - !ruby/object:Gem::Dependency
350
- name: google-apis-aiplatform_v1
350
+ name: googleauth
351
351
  requirement: !ruby/object:Gem::Requirement
352
352
  requirements:
353
- - - "~>"
353
+ - - ">="
354
354
  - !ruby/object:Gem::Version
355
- version: '0.7'
355
+ version: '0'
356
356
  type: :development
357
357
  prerelease: false
358
358
  version_requirements: !ruby/object:Gem::Requirement
359
359
  requirements:
360
- - - "~>"
360
+ - - ">="
361
361
  - !ruby/object:Gem::Version
362
- version: '0.7'
362
+ version: '0'
363
363
  - !ruby/object:Gem::Dependency
364
364
  name: google_palm_api
365
365
  requirement: !ruby/object:Gem::Requirement
@@ -708,7 +708,9 @@ files:
708
708
  - README.md
709
709
  - lib/langchain.rb
710
710
  - lib/langchain/assistants/assistant.rb
711
- - lib/langchain/assistants/message.rb
711
+ - lib/langchain/assistants/messages/base.rb
712
+ - lib/langchain/assistants/messages/google_gemini_message.rb
713
+ - lib/langchain/assistants/messages/openai_message.rb
712
714
  - lib/langchain/assistants/thread.rb
713
715
  - lib/langchain/chunk.rb
714
716
  - lib/langchain/chunker/base.rb
@@ -735,6 +737,7 @@ files:
735
737
  - lib/langchain/llm/azure.rb
736
738
  - lib/langchain/llm/base.rb
737
739
  - lib/langchain/llm/cohere.rb
740
+ - lib/langchain/llm/google_gemini.rb
738
741
  - lib/langchain/llm/google_palm.rb
739
742
  - lib/langchain/llm/google_vertex_ai.rb
740
743
  - lib/langchain/llm/hugging_face.rb
@@ -751,8 +754,8 @@ files:
751
754
  - lib/langchain/llm/response/aws_titan_response.rb
752
755
  - lib/langchain/llm/response/base_response.rb
753
756
  - lib/langchain/llm/response/cohere_response.rb
757
+ - lib/langchain/llm/response/google_gemini_response.rb
754
758
  - lib/langchain/llm/response/google_palm_response.rb
755
- - lib/langchain/llm/response/google_vertex_ai_response.rb
756
759
  - lib/langchain/llm/response/hugging_face_response.rb
757
760
  - lib/langchain/llm/response/llama_cpp_response.rb
758
761
  - lib/langchain/llm/response/mistral_ai_response.rb
@@ -791,6 +794,8 @@ files:
791
794
  - lib/langchain/tool/file_system/file_system.rb
792
795
  - lib/langchain/tool/google_search/google_search.json
793
796
  - lib/langchain/tool/google_search/google_search.rb
797
+ - lib/langchain/tool/news_retriever/news_retriever.json
798
+ - lib/langchain/tool/news_retriever/news_retriever.rb
794
799
  - lib/langchain/tool/ruby_code_interpreter/ruby_code_interpreter.json
795
800
  - lib/langchain/tool/ruby_code_interpreter/ruby_code_interpreter.rb
796
801
  - lib/langchain/tool/vectorsearch/vectorsearch.json
@@ -1,58 +0,0 @@
1
- # frozen_string_literal: true
2
-
3
- module Langchain
4
- # Langchain::Message are the messages that are sent to LLM chat methods
5
- class Message
6
- attr_reader :role, :content, :tool_calls, :tool_call_id
7
-
8
- ROLES = %w[
9
- system
10
- assistant
11
- user
12
- tool
13
- ].freeze
14
-
15
- # @param role [String] The role of the message
16
- # @param content [String] The content of the message
17
- # @param tool_calls [Array<Hash>] Tool calls to be made
18
- # @param tool_call_id [String] The ID of the tool call to be made
19
- def initialize(role:, content: nil, tool_calls: [], tool_call_id: nil) # TODO: Implement image_file: reference (https://platform.openai.com/docs/api-reference/messages/object#messages/object-content)
20
- raise ArgumentError, "Role must be one of #{ROLES.join(", ")}" unless ROLES.include?(role)
21
- raise ArgumentError, "Tool calls must be an array of hashes" unless tool_calls.is_a?(Array) && tool_calls.all? { |tool_call| tool_call.is_a?(Hash) }
22
-
23
- @role = role
24
- # Some Tools return content as a JSON hence `.to_s`
25
- @content = content.to_s
26
- @tool_calls = tool_calls
27
- @tool_call_id = tool_call_id
28
- end
29
-
30
- # Convert the message to an OpenAI API-compatible hash
31
- #
32
- # @return [Hash] The message as an OpenAI API-compatible hash
33
- def to_openai_format
34
- {}.tap do |h|
35
- h[:role] = role
36
- h[:content] = content if content # Content is nil for tool calls
37
- h[:tool_calls] = tool_calls if tool_calls.any?
38
- h[:tool_call_id] = tool_call_id if tool_call_id
39
- end
40
- end
41
-
42
- def assistant?
43
- role == "assistant"
44
- end
45
-
46
- def system?
47
- role == "system"
48
- end
49
-
50
- def user?
51
- role == "user"
52
- end
53
-
54
- def tool?
55
- role == "tool"
56
- end
57
- end
58
- end
@@ -1,33 +0,0 @@
1
- # frozen_string_literal: true
2
-
3
- module Langchain::LLM
4
- class GoogleVertexAiResponse < BaseResponse
5
- attr_reader :prompt_tokens
6
-
7
- def initialize(raw_response, model: nil)
8
- @prompt_tokens = prompt_tokens
9
- super(raw_response, model: model)
10
- end
11
-
12
- def completion
13
- # completions&.dig(0, "output")
14
- raw_response.predictions[0]["content"]
15
- end
16
-
17
- def embedding
18
- embeddings.first
19
- end
20
-
21
- def completions
22
- raw_response.predictions.map { |p| p["content"] }
23
- end
24
-
25
- def total_tokens
26
- raw_response.dig(:predictions, 0, :embeddings, :statistics, :token_count)
27
- end
28
-
29
- def embeddings
30
- [raw_response.dig(:predictions, 0, :embeddings, :values)]
31
- end
32
- end
33
- end