langchain-ocr-lib 0.3.0__tar.gz → 0.3.2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/PKG-INFO +12 -8
  2. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/README.md +9 -5
  3. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/pyproject.toml +3 -3
  4. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/di_config.py +15 -11
  5. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/langfuse_manager/langfuse_manager.py +12 -21
  6. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/langfuse_settings.py +4 -0
  7. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/ollama_chat_settings.py +1 -1
  8. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/openai_chat_settings.py +3 -3
  9. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/vllm_chat_settings.py +3 -3
  10. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/__init__.py +0 -0
  11. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/chains/__init__.py +0 -0
  12. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/chains/chain.py +0 -0
  13. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/converter/__init__.py +0 -0
  14. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/converter/converter.py +0 -0
  15. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/di_binding_keys/__init__.py +0 -0
  16. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/di_binding_keys/binding_keys.py +0 -0
  17. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/__init__.py +0 -0
  18. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/chains/__init__.py +0 -0
  19. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/chains/ocr_chain.py +0 -0
  20. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/converter/__init__.py +0 -0
  21. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/converter/image_converter.py +0 -0
  22. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/converter/pdf_converter.py +0 -0
  23. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/langfuse_manager/__init__.py +0 -0
  24. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/llms/__init__.py +0 -0
  25. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/llms/llm_factory.py +0 -0
  26. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/llms/llm_type.py +0 -0
  27. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/__init__.py +0 -0
  28. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/language_settings.py +0 -0
  29. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/settings/llm_class_type_settings.py +0 -0
  30. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/tracers/__init__.py +0 -0
  31. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/impl/tracers/langfuse_traced_chain.py +0 -0
  32. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/language_mapping/language_mapping.py +0 -0
  33. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/main.py +0 -0
  34. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/prompt_templates/__init__.py +0 -0
  35. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/prompt_templates/ocr_prompt.py +0 -0
  36. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/tracers/__init__.py +0 -0
  37. {langchain_ocr_lib-0.3.0 → langchain_ocr_lib-0.3.2}/src/langchain_ocr_lib/tracers/traced_chain.py +0 -0
@@ -1,7 +1,7 @@
1
- Metadata-Version: 2.1
1
+ Metadata-Version: 2.3
2
2
  Name: langchain-ocr-lib
3
- Version: 0.3.0
4
- Summary:
3
+ Version: 0.3.2
4
+ Summary: Modular, vision-LLM-powered chain to convert image and PDF documents into clean Markdown.
5
5
  License: MIT
6
6
  Author: Andreas Klos
7
7
  Author-email: aklos@outlook.de
@@ -25,9 +25,13 @@ Requires-Dist: pytest-asyncio (>=0.25.0,<0.26.0)
25
25
  Requires-Dist: pyyaml (>=6.0.2,<7.0.0)
26
26
  Description-Content-Type: text/markdown
27
27
 
28
- # langchain_ocr_lib
28
+ # langchain-ocr-lib
29
29
 
30
- **langchain_ocr_lib** is the OCR processing engine behind LangChain-OCR. It provides a modular, vision-LLM-powered Chain to convert image and PDF documents into clean Markdown. Designed for direct CLI usage or integration into larger applications.
30
+ **langchain-ocr-lib** is the OCR processing engine behind LangChain-OCR. It provides a modular, vision-LLM-powered Chain to convert image and PDF documents into clean Markdown. Designed for direct CLI usage or integration into larger applications.
31
+
32
+ <div align="center">
33
+ <img src="./images/logo.png" alt="Logo" style="width:30%;">
34
+ </div>
31
35
 
32
36
  ## Table of Contents
33
37
 
@@ -137,7 +141,7 @@ class Converter:
137
141
  return self._converter.convert2markdown(filename=filename)
138
142
 
139
143
  converter = Converter()
140
- markdown = converter.convert("../docs/invoice.pdf") # Adjust the file path as needed
144
+ markdown = converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
141
145
  print(markdown)
142
146
  ```
143
147
 
@@ -174,7 +178,7 @@ class Converter:
174
178
  self._converter.convert(filename=filename)
175
179
 
176
180
  converter = Converter()
177
- converter.convert("../docs/invoice.pdf") # Adjust the file path as needed
181
+ converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
178
182
  ```
179
183
 
180
184
  ### 4.3 Docker
@@ -183,6 +187,6 @@ Run OCR via Docker without local Python setup:
183
187
 
184
188
  ```bash
185
189
  docker build -t ocr -f langchain_ocr_lib/Dockerfile .
186
- docker run --net=host -it --rm -v ./docs:/app/docs:ro ocr docs/invoice.png
190
+ docker run --net=host -it --rm -v ./examples:/app/examples:ro ocr examples/invoice.png
187
191
  ```
188
192
 
@@ -1,6 +1,10 @@
1
- # langchain_ocr_lib
1
+ # langchain-ocr-lib
2
2
 
3
- **langchain_ocr_lib** is the OCR processing engine behind LangChain-OCR. It provides a modular, vision-LLM-powered Chain to convert image and PDF documents into clean Markdown. Designed for direct CLI usage or integration into larger applications.
3
+ **langchain-ocr-lib** is the OCR processing engine behind LangChain-OCR. It provides a modular, vision-LLM-powered Chain to convert image and PDF documents into clean Markdown. Designed for direct CLI usage or integration into larger applications.
4
+
5
+ <div align="center">
6
+ <img src="./images/logo.png" alt="Logo" style="width:30%;">
7
+ </div>
4
8
 
5
9
  ## Table of Contents
6
10
 
@@ -110,7 +114,7 @@ class Converter:
110
114
  return self._converter.convert2markdown(filename=filename)
111
115
 
112
116
  converter = Converter()
113
- markdown = converter.convert("../docs/invoice.pdf") # Adjust the file path as needed
117
+ markdown = converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
114
118
  print(markdown)
115
119
  ```
116
120
 
@@ -147,7 +151,7 @@ class Converter:
147
151
  self._converter.convert(filename=filename)
148
152
 
149
153
  converter = Converter()
150
- converter.convert("../docs/invoice.pdf") # Adjust the file path as needed
154
+ converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
151
155
  ```
152
156
 
153
157
  ### 4.3 Docker
@@ -156,5 +160,5 @@ Run OCR via Docker without local Python setup:
156
160
 
157
161
  ```bash
158
162
  docker build -t ocr -f langchain_ocr_lib/Dockerfile .
159
- docker run --net=host -it --rm -v ./docs:/app/docs:ro ocr docs/invoice.png
163
+ docker run --net=host -it --rm -v ./examples:/app/examples:ro ocr examples/invoice.png
160
164
  ```
@@ -1,5 +1,5 @@
1
1
  [build-system]
2
- requires = ["poetry-core==1.8.5"]
2
+ requires = ["poetry-core==2.1.2"]
3
3
  build-backend = "poetry.core.masonry.api"
4
4
 
5
5
  [tool.poetry.scripts]
@@ -7,8 +7,8 @@ langchain-ocr = "langchain_ocr_lib.main:main"
7
7
 
8
8
  [tool.poetry]
9
9
  name = "langchain-ocr-lib"
10
- version = "0.3.0"
11
- description = ""
10
+ version = "0.3.2"
11
+ description = "Modular, vision-LLM-powered chain to convert image and PDF documents into clean Markdown."
12
12
  authors = ["Andreas Klos <aklos@outlook.de>"]
13
13
  readme = "README.md"
14
14
  packages = [{ include = "langchain_ocr_lib", from = "src" }]
@@ -14,6 +14,7 @@ from langchain_ocr_lib.di_binding_keys.binding_keys import (
14
14
  from langchain_ollama import ChatOllama
15
15
  from langchain_openai import ChatOpenAI
16
16
  from langfuse import Langfuse
17
+ from functools import partial
17
18
 
18
19
  from langchain_ocr_lib.impl.chains.ocr_chain import OcrChain
19
20
  from langchain_ocr_lib.impl.settings.ollama_chat_settings import OllamaSettings
@@ -50,16 +51,17 @@ def lib_di_config(binder: Binder):
50
51
 
51
52
  if llm_class_type_settings.llm_type == "ollama":
52
53
  settings = OllamaSettings()
53
- llm_instance = llm_provider(settings, ChatOllama)
54
+ partial_llm_provider = partial(llm_provider,settings, ChatOllama)
54
55
  elif llm_class_type_settings.llm_type == "openai":
55
56
  settings = OpenAISettings()
56
- llm_instance = llm_provider(settings, ChatOpenAI)
57
+ partial_llm_provider = partial(llm_provider,settings, ChatOpenAI)
57
58
  elif llm_class_type_settings.llm_type == "vllm":
58
59
  settings = VllmSettings()
59
- llm_instance = llm_provider(settings, ChatOpenAI)
60
+ partial_llm_provider = partial(llm_provider,settings, ChatOpenAI)
60
61
  else:
61
62
  raise NotImplementedError("Configured LLM is not implemented")
62
- binder.bind(LargeLanguageModelKey, llm_instance)
63
+
64
+ binder.bind_to_provider(LargeLanguageModelKey, partial_llm_provider)
63
65
 
64
66
  prompt = ocr_prompt_template_builder(language=language_settings.language, model_name=settings.model)
65
67
 
@@ -78,17 +80,19 @@ def lib_di_config(binder: Binder):
78
80
  managed_prompts={
79
81
  OcrChain.__name__: prompt,
80
82
  },
83
+ enabled=langfuse_settings.enabled,
81
84
  ),
82
85
  )
83
86
 
84
- binder.bind(OcrChainKey, OcrChain())
87
+ binder.bind(OcrChainKey if langfuse_settings.enabled else LangfuseTracedChainKey, OcrChain())
85
88
 
86
- binder.bind(
87
- LangfuseTracedChainKey,
88
- LangfuseTracedChain(
89
- settings=langfuse_settings,
90
- ),
91
- )
89
+ if langfuse_settings.enabled:
90
+ binder.bind(
91
+ LangfuseTracedChainKey,
92
+ LangfuseTracedChain(
93
+ settings=langfuse_settings,
94
+ ),
95
+ )
92
96
 
93
97
  binder.bind(PdfConverterKey, Pdf2MarkdownConverter())
94
98
  binder.bind(ImageConverterKey, Image2MarkdownConverter())
@@ -32,22 +32,10 @@ class LangfuseManager:
32
32
  def __init__(
33
33
  self,
34
34
  managed_prompts: dict[str, str],
35
+ enabled: bool = True,
35
36
  ):
36
37
  self._managed_prompts = managed_prompts
37
-
38
- def init_prompts(self) -> None:
39
- """
40
- Initialize the prompts managed by the LangfuseManager.
41
-
42
- This method iterates over the keys of the managed prompts and retrieves
43
- each prompt using the `get_langfuse_prompt` method.
44
-
45
- Returns
46
- -------
47
- None
48
- """
49
- for key in list(self._managed_prompts.keys()):
50
- self.get_langfuse_prompt(key)
38
+ self._enabled = enabled
51
39
 
52
40
  def get_langfuse_prompt(self, base_prompt_name: str) -> Optional[ChatPromptClient]:
53
41
  """
@@ -70,6 +58,9 @@ class LangfuseManager:
70
58
  Exception
71
59
  If an error occurs while retrieving the prompt template from Langfuse.
72
60
  """
61
+ if not self._enabled:
62
+ logger.info("Langfuse is not enabled. Using fallback prompt.")
63
+ return None
73
64
  try:
74
65
  langfuse_prompt = self._langfuse.get_prompt(base_prompt_name)
75
66
  except NotFoundError:
@@ -86,11 +77,7 @@ class LangfuseManager:
86
77
  )
87
78
  langfuse_prompt = self._langfuse.get_prompt(base_prompt_name)
88
79
  except Exception as error:
89
- logger.error(
90
- "Error occured while getting prompt template from langfuse. Error:\n{error}",
91
- extra={error: error},
92
- )
93
- return None
80
+ logger.error(f"Error occurred while getting prompt template from langfuse. Error:\n{error}")
94
81
  return langfuse_prompt
95
82
 
96
83
  def get_base_llm(self, name: str) -> LLM:
@@ -108,9 +95,12 @@ class LangfuseManager:
108
95
  The base Large Language Model. If the Langfuse prompt is not found,
109
96
  returns the LLM with a fallback configuration.
110
97
  """
98
+ if not self._enabled:
99
+ logger.info("Langfuse is not enabled. Using fallback LLM.")
100
+ return self._llm
111
101
  langfuse_prompt = self.get_langfuse_prompt(name)
112
102
  if not langfuse_prompt:
113
- logger.error("Using fallback for llm")
103
+ logger.warning("Could not retrieve prompt template from langfuse. Using fallback LLM.")
114
104
  return self._llm
115
105
 
116
106
  return self._llm.with_config({"configurable": langfuse_prompt.config})
@@ -135,7 +125,8 @@ class LangfuseManager:
135
125
  """
136
126
  langfuse_prompt = self.get_langfuse_prompt(name)
137
127
  if not langfuse_prompt:
138
- logger.error("Could not retrieve prompt template from langfuse. Using fallback value.")
128
+ if self._enabled:
129
+ logger.warning("Could not retrieve prompt template from langfuse. Using fallback value.")
139
130
  fallback = self._managed_prompts[name]
140
131
  if isinstance(fallback, ChatPromptTemplate):
141
132
  return fallback
@@ -27,3 +27,7 @@ class LangfuseSettings(BaseSettings):
27
27
  secret_key: str = Field(default="", description="The secret key for Langfuse.")
28
28
  public_key: str = Field(default="", description="The public key for Langfuse.")
29
29
  host: str = Field(default="https://api.langchain.com", description="The host for Langfuse.")
30
+ enabled: bool = Field(
31
+ default=True,
32
+ description="Whether to enable Langfuse. If set to False, Langfuse will not be used.",
33
+ )
@@ -35,7 +35,7 @@ class OllamaSettings(BaseSettings):
35
35
  env_prefix = "OLLAMA_"
36
36
  case_sensitive = False
37
37
 
38
- model: str = Field(default="gemma3:4b-it-q4_K_M")
38
+ model: str = Field(default="gemma3:4b-it-q4_K_M", title="LLM Model")
39
39
  base_url: str = Field(default="http://localhost:11434")
40
40
  top_k: int = Field(default=0, title="LLM Top K")
41
41
  top_p: float = Field(default=0, title="LLM Top P")
@@ -28,10 +28,10 @@ class OpenAISettings(BaseSettings):
28
28
  env_prefix = "OPENAI_"
29
29
  case_sensitive = False
30
30
 
31
- model: str = Field(default="gpt-4o-mini-search-preview-2025-03-11", description="The model identifier")
31
+ model: str = Field(default="gpt-4o-mini-search-preview-2025-03-11", description="The model identifier", title="LLM Model")
32
32
  api_key: str = Field(default="", description="The API key for authentication")
33
- top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step")
34
- temperature: float = Field(default=0, description="What sampling temperature to use")
33
+ top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step", title="Top P")
34
+ temperature: float = Field(default=0, description="What sampling temperature to use", title="Temperature")
35
35
  base_url: str = Field(
36
36
  default="https://api.openai.com/v1",
37
37
  description="The base URL for the OpenAI API endpoint",
@@ -28,10 +28,10 @@ class VllmSettings(BaseSettings):
28
28
  env_prefix = "VLLM_"
29
29
  case_sensitive = False
30
30
 
31
- model: str = Field(default="", description="The model identifier")
31
+ model: str = Field(default="", description="The model identifier", title="LLM Model")
32
32
  api_key: str = Field(default="", description="The API key for authentication")
33
- top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step")
34
- temperature: float = Field(default=0, description="What sampling temperature to use")
33
+ top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step", title="Top P")
34
+ temperature: float = Field(default=0, description="What sampling temperature to use", title="Temperature")
35
35
  base_url: str = Field(
36
36
  default="http://localhost:8000/v1",
37
37
  description="The base URL for the Vllm API endpoint",