langchain-ocr-lib 0.3.0__py3-none-any.whl → 0.3.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- langchain_ocr_lib/di_config.py +15 -11
- langchain_ocr_lib/impl/langfuse_manager/langfuse_manager.py +12 -21
- langchain_ocr_lib/impl/settings/langfuse_settings.py +4 -0
- langchain_ocr_lib/impl/settings/ollama_chat_settings.py +1 -1
- langchain_ocr_lib/impl/settings/openai_chat_settings.py +3 -3
- langchain_ocr_lib/impl/settings/vllm_chat_settings.py +3 -3
- {langchain_ocr_lib-0.3.0.dist-info → langchain_ocr_lib-0.3.2.dist-info}/METADATA +12 -8
- {langchain_ocr_lib-0.3.0.dist-info → langchain_ocr_lib-0.3.2.dist-info}/RECORD +10 -10
- {langchain_ocr_lib-0.3.0.dist-info → langchain_ocr_lib-0.3.2.dist-info}/WHEEL +1 -1
- {langchain_ocr_lib-0.3.0.dist-info → langchain_ocr_lib-0.3.2.dist-info}/entry_points.txt +0 -0
langchain_ocr_lib/di_config.py
CHANGED
@@ -14,6 +14,7 @@ from langchain_ocr_lib.di_binding_keys.binding_keys import (
|
|
14
14
|
from langchain_ollama import ChatOllama
|
15
15
|
from langchain_openai import ChatOpenAI
|
16
16
|
from langfuse import Langfuse
|
17
|
+
from functools import partial
|
17
18
|
|
18
19
|
from langchain_ocr_lib.impl.chains.ocr_chain import OcrChain
|
19
20
|
from langchain_ocr_lib.impl.settings.ollama_chat_settings import OllamaSettings
|
@@ -50,16 +51,17 @@ def lib_di_config(binder: Binder):
|
|
50
51
|
|
51
52
|
if llm_class_type_settings.llm_type == "ollama":
|
52
53
|
settings = OllamaSettings()
|
53
|
-
|
54
|
+
partial_llm_provider = partial(llm_provider,settings, ChatOllama)
|
54
55
|
elif llm_class_type_settings.llm_type == "openai":
|
55
56
|
settings = OpenAISettings()
|
56
|
-
|
57
|
+
partial_llm_provider = partial(llm_provider,settings, ChatOpenAI)
|
57
58
|
elif llm_class_type_settings.llm_type == "vllm":
|
58
59
|
settings = VllmSettings()
|
59
|
-
|
60
|
+
partial_llm_provider = partial(llm_provider,settings, ChatOpenAI)
|
60
61
|
else:
|
61
62
|
raise NotImplementedError("Configured LLM is not implemented")
|
62
|
-
|
63
|
+
|
64
|
+
binder.bind_to_provider(LargeLanguageModelKey, partial_llm_provider)
|
63
65
|
|
64
66
|
prompt = ocr_prompt_template_builder(language=language_settings.language, model_name=settings.model)
|
65
67
|
|
@@ -78,17 +80,19 @@ def lib_di_config(binder: Binder):
|
|
78
80
|
managed_prompts={
|
79
81
|
OcrChain.__name__: prompt,
|
80
82
|
},
|
83
|
+
enabled=langfuse_settings.enabled,
|
81
84
|
),
|
82
85
|
)
|
83
86
|
|
84
|
-
binder.bind(OcrChainKey, OcrChain())
|
87
|
+
binder.bind(OcrChainKey if langfuse_settings.enabled else LangfuseTracedChainKey, OcrChain())
|
85
88
|
|
86
|
-
|
87
|
-
|
88
|
-
|
89
|
-
|
90
|
-
|
91
|
-
|
89
|
+
if langfuse_settings.enabled:
|
90
|
+
binder.bind(
|
91
|
+
LangfuseTracedChainKey,
|
92
|
+
LangfuseTracedChain(
|
93
|
+
settings=langfuse_settings,
|
94
|
+
),
|
95
|
+
)
|
92
96
|
|
93
97
|
binder.bind(PdfConverterKey, Pdf2MarkdownConverter())
|
94
98
|
binder.bind(ImageConverterKey, Image2MarkdownConverter())
|
@@ -32,22 +32,10 @@ class LangfuseManager:
|
|
32
32
|
def __init__(
|
33
33
|
self,
|
34
34
|
managed_prompts: dict[str, str],
|
35
|
+
enabled: bool = True,
|
35
36
|
):
|
36
37
|
self._managed_prompts = managed_prompts
|
37
|
-
|
38
|
-
def init_prompts(self) -> None:
|
39
|
-
"""
|
40
|
-
Initialize the prompts managed by the LangfuseManager.
|
41
|
-
|
42
|
-
This method iterates over the keys of the managed prompts and retrieves
|
43
|
-
each prompt using the `get_langfuse_prompt` method.
|
44
|
-
|
45
|
-
Returns
|
46
|
-
-------
|
47
|
-
None
|
48
|
-
"""
|
49
|
-
for key in list(self._managed_prompts.keys()):
|
50
|
-
self.get_langfuse_prompt(key)
|
38
|
+
self._enabled = enabled
|
51
39
|
|
52
40
|
def get_langfuse_prompt(self, base_prompt_name: str) -> Optional[ChatPromptClient]:
|
53
41
|
"""
|
@@ -70,6 +58,9 @@ class LangfuseManager:
|
|
70
58
|
Exception
|
71
59
|
If an error occurs while retrieving the prompt template from Langfuse.
|
72
60
|
"""
|
61
|
+
if not self._enabled:
|
62
|
+
logger.info("Langfuse is not enabled. Using fallback prompt.")
|
63
|
+
return None
|
73
64
|
try:
|
74
65
|
langfuse_prompt = self._langfuse.get_prompt(base_prompt_name)
|
75
66
|
except NotFoundError:
|
@@ -86,11 +77,7 @@ class LangfuseManager:
|
|
86
77
|
)
|
87
78
|
langfuse_prompt = self._langfuse.get_prompt(base_prompt_name)
|
88
79
|
except Exception as error:
|
89
|
-
logger.error(
|
90
|
-
"Error occured while getting prompt template from langfuse. Error:\n{error}",
|
91
|
-
extra={error: error},
|
92
|
-
)
|
93
|
-
return None
|
80
|
+
logger.error(f"Error occurred while getting prompt template from langfuse. Error:\n{error}")
|
94
81
|
return langfuse_prompt
|
95
82
|
|
96
83
|
def get_base_llm(self, name: str) -> LLM:
|
@@ -108,9 +95,12 @@ class LangfuseManager:
|
|
108
95
|
The base Large Language Model. If the Langfuse prompt is not found,
|
109
96
|
returns the LLM with a fallback configuration.
|
110
97
|
"""
|
98
|
+
if not self._enabled:
|
99
|
+
logger.info("Langfuse is not enabled. Using fallback LLM.")
|
100
|
+
return self._llm
|
111
101
|
langfuse_prompt = self.get_langfuse_prompt(name)
|
112
102
|
if not langfuse_prompt:
|
113
|
-
logger.
|
103
|
+
logger.warning("Could not retrieve prompt template from langfuse. Using fallback LLM.")
|
114
104
|
return self._llm
|
115
105
|
|
116
106
|
return self._llm.with_config({"configurable": langfuse_prompt.config})
|
@@ -135,7 +125,8 @@ class LangfuseManager:
|
|
135
125
|
"""
|
136
126
|
langfuse_prompt = self.get_langfuse_prompt(name)
|
137
127
|
if not langfuse_prompt:
|
138
|
-
|
128
|
+
if self._enabled:
|
129
|
+
logger.warning("Could not retrieve prompt template from langfuse. Using fallback value.")
|
139
130
|
fallback = self._managed_prompts[name]
|
140
131
|
if isinstance(fallback, ChatPromptTemplate):
|
141
132
|
return fallback
|
@@ -27,3 +27,7 @@ class LangfuseSettings(BaseSettings):
|
|
27
27
|
secret_key: str = Field(default="", description="The secret key for Langfuse.")
|
28
28
|
public_key: str = Field(default="", description="The public key for Langfuse.")
|
29
29
|
host: str = Field(default="https://api.langchain.com", description="The host for Langfuse.")
|
30
|
+
enabled: bool = Field(
|
31
|
+
default=True,
|
32
|
+
description="Whether to enable Langfuse. If set to False, Langfuse will not be used.",
|
33
|
+
)
|
@@ -35,7 +35,7 @@ class OllamaSettings(BaseSettings):
|
|
35
35
|
env_prefix = "OLLAMA_"
|
36
36
|
case_sensitive = False
|
37
37
|
|
38
|
-
model: str = Field(default="gemma3:4b-it-q4_K_M")
|
38
|
+
model: str = Field(default="gemma3:4b-it-q4_K_M", title="LLM Model")
|
39
39
|
base_url: str = Field(default="http://localhost:11434")
|
40
40
|
top_k: int = Field(default=0, title="LLM Top K")
|
41
41
|
top_p: float = Field(default=0, title="LLM Top P")
|
@@ -28,10 +28,10 @@ class OpenAISettings(BaseSettings):
|
|
28
28
|
env_prefix = "OPENAI_"
|
29
29
|
case_sensitive = False
|
30
30
|
|
31
|
-
model: str = Field(default="gpt-4o-mini-search-preview-2025-03-11", description="The model identifier")
|
31
|
+
model: str = Field(default="gpt-4o-mini-search-preview-2025-03-11", description="The model identifier", title="LLM Model")
|
32
32
|
api_key: str = Field(default="", description="The API key for authentication")
|
33
|
-
top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step")
|
34
|
-
temperature: float = Field(default=0, description="What sampling temperature to use")
|
33
|
+
top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step", title="Top P")
|
34
|
+
temperature: float = Field(default=0, description="What sampling temperature to use", title="Temperature")
|
35
35
|
base_url: str = Field(
|
36
36
|
default="https://api.openai.com/v1",
|
37
37
|
description="The base URL for the OpenAI API endpoint",
|
@@ -28,10 +28,10 @@ class VllmSettings(BaseSettings):
|
|
28
28
|
env_prefix = "VLLM_"
|
29
29
|
case_sensitive = False
|
30
30
|
|
31
|
-
model: str = Field(default="", description="The model identifier")
|
31
|
+
model: str = Field(default="", description="The model identifier", title="LLM Model")
|
32
32
|
api_key: str = Field(default="", description="The API key for authentication")
|
33
|
-
top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step")
|
34
|
-
temperature: float = Field(default=0, description="What sampling temperature to use")
|
33
|
+
top_p: float = Field(default=1.0, description="Total probability mass of tokens to consider at each step", title="Top P")
|
34
|
+
temperature: float = Field(default=0, description="What sampling temperature to use", title="Temperature")
|
35
35
|
base_url: str = Field(
|
36
36
|
default="http://localhost:8000/v1",
|
37
37
|
description="The base URL for the Vllm API endpoint",
|
@@ -1,7 +1,7 @@
|
|
1
|
-
Metadata-Version: 2.
|
1
|
+
Metadata-Version: 2.3
|
2
2
|
Name: langchain-ocr-lib
|
3
|
-
Version: 0.3.
|
4
|
-
Summary:
|
3
|
+
Version: 0.3.2
|
4
|
+
Summary: Modular, vision-LLM-powered chain to convert image and PDF documents into clean Markdown.
|
5
5
|
License: MIT
|
6
6
|
Author: Andreas Klos
|
7
7
|
Author-email: aklos@outlook.de
|
@@ -25,9 +25,13 @@ Requires-Dist: pytest-asyncio (>=0.25.0,<0.26.0)
|
|
25
25
|
Requires-Dist: pyyaml (>=6.0.2,<7.0.0)
|
26
26
|
Description-Content-Type: text/markdown
|
27
27
|
|
28
|
-
#
|
28
|
+
# langchain-ocr-lib
|
29
29
|
|
30
|
-
**
|
30
|
+
**langchain-ocr-lib** is the OCR processing engine behind LangChain-OCR. It provides a modular, vision-LLM-powered Chain to convert image and PDF documents into clean Markdown. Designed for direct CLI usage or integration into larger applications.
|
31
|
+
|
32
|
+
<div align="center">
|
33
|
+
<img src="./images/logo.png" alt="Logo" style="width:30%;">
|
34
|
+
</div>
|
31
35
|
|
32
36
|
## Table of Contents
|
33
37
|
|
@@ -137,7 +141,7 @@ class Converter:
|
|
137
141
|
return self._converter.convert2markdown(filename=filename)
|
138
142
|
|
139
143
|
converter = Converter()
|
140
|
-
markdown = converter.convert("../
|
144
|
+
markdown = converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
|
141
145
|
print(markdown)
|
142
146
|
```
|
143
147
|
|
@@ -174,7 +178,7 @@ class Converter:
|
|
174
178
|
self._converter.convert(filename=filename)
|
175
179
|
|
176
180
|
converter = Converter()
|
177
|
-
converter.convert("../
|
181
|
+
converter.convert("../examples/invoice.pdf") # Adjust the file path as needed
|
178
182
|
```
|
179
183
|
|
180
184
|
### 4.3 Docker
|
@@ -183,6 +187,6 @@ Run OCR via Docker without local Python setup:
|
|
183
187
|
|
184
188
|
```bash
|
185
189
|
docker build -t ocr -f langchain_ocr_lib/Dockerfile .
|
186
|
-
docker run --net=host -it --rm -v ./
|
190
|
+
docker run --net=host -it --rm -v ./examples:/app/examples:ro ocr examples/invoice.png
|
187
191
|
```
|
188
192
|
|
@@ -5,7 +5,7 @@ langchain_ocr_lib/converter/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMp
|
|
5
5
|
langchain_ocr_lib/converter/converter.py,sha256=oDUNzVWD743RgqIal7T4OVv-Z1RKE9uQYzAIPpgY3o8,1280
|
6
6
|
langchain_ocr_lib/di_binding_keys/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
7
7
|
langchain_ocr_lib/di_binding_keys/binding_keys.py,sha256=jE8rwNcLaI0NflIMkK0vu0LVy5o4y0pYgdjbpDNTGyk,338
|
8
|
-
langchain_ocr_lib/di_config.py,sha256=
|
8
|
+
langchain_ocr_lib/di_config.py,sha256=LvRnptts1VCDa3HVP7jtBtQTyG-QEErbyplGX86uaUA,3802
|
9
9
|
langchain_ocr_lib/impl/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
10
10
|
langchain_ocr_lib/impl/chains/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
11
11
|
langchain_ocr_lib/impl/chains/ocr_chain.py,sha256=stE8RLE1ieRHf6XHreKCRfhNfXzw9fNLTake7xQBGL8,2673
|
@@ -13,17 +13,17 @@ langchain_ocr_lib/impl/converter/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRk
|
|
13
13
|
langchain_ocr_lib/impl/converter/image_converter.py,sha256=G1rDOCbudWNL4sDvSGJ7CeeFrWUblfWPGaZf5JsnpiM,2871
|
14
14
|
langchain_ocr_lib/impl/converter/pdf_converter.py,sha256=pTHPojuNLCSWJp4FzXBHshXva2sBGyOs6Y7jnKJrnNo,3760
|
15
15
|
langchain_ocr_lib/impl/langfuse_manager/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
16
|
-
langchain_ocr_lib/impl/langfuse_manager/langfuse_manager.py,sha256=
|
16
|
+
langchain_ocr_lib/impl/langfuse_manager/langfuse_manager.py,sha256=C2waQ1Mvqz6bECXAcaMiBsE8TLc2kLr5QUIgXqgH_cE,5311
|
17
17
|
langchain_ocr_lib/impl/llms/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
18
18
|
langchain_ocr_lib/impl/llms/llm_factory.py,sha256=9DsUdoYNrjeWLGA9ISDdHN2cxcQ7DquNQ5it6zSxHlg,2199
|
19
19
|
langchain_ocr_lib/impl/llms/llm_type.py,sha256=_LKtdVuTRYX6gupkxJtEtIwrbtiMvZmG8WOxfzlm42M,286
|
20
20
|
langchain_ocr_lib/impl/settings/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
21
|
-
langchain_ocr_lib/impl/settings/langfuse_settings.py,sha256=
|
21
|
+
langchain_ocr_lib/impl/settings/langfuse_settings.py,sha256=QT4_VwYj0msFbgL3qIQ-oer3Lt0qny0FFAyfssGu-q0,962
|
22
22
|
langchain_ocr_lib/impl/settings/language_settings.py,sha256=tdAC1t5wGu1MoH1jhjkDnxnX4Ui7giwxt7Qm8_LPkP8,627
|
23
23
|
langchain_ocr_lib/impl/settings/llm_class_type_settings.py,sha256=4KC6zxby13wn38rB8055J8LNVTsmUfrOiyLtLuToHaM,598
|
24
|
-
langchain_ocr_lib/impl/settings/ollama_chat_settings.py,sha256=
|
25
|
-
langchain_ocr_lib/impl/settings/openai_chat_settings.py,sha256=
|
26
|
-
langchain_ocr_lib/impl/settings/vllm_chat_settings.py,sha256=
|
24
|
+
langchain_ocr_lib/impl/settings/ollama_chat_settings.py,sha256=YQkgD7CfOjHN5wkpJakO0GfM7-D2GqoJLP1gB2932ms,1525
|
25
|
+
langchain_ocr_lib/impl/settings/openai_chat_settings.py,sha256=NqVfkcI8OoD8TVxyv4l0G9ycUC6LIs6Qs4kQRL24doA,1331
|
26
|
+
langchain_ocr_lib/impl/settings/vllm_chat_settings.py,sha256=Zr4L6Urp-f1JZu7Q1dwL6671EQbrIIYL0ubJSQlod3c,1281
|
27
27
|
langchain_ocr_lib/impl/tracers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
28
28
|
langchain_ocr_lib/impl/tracers/langfuse_traced_chain.py,sha256=syjwNt8HfVmaWXZ-ElFYsc-KwpnKQz2LE3K5jV7c3GE,1599
|
29
29
|
langchain_ocr_lib/language_mapping/language_mapping.py,sha256=VY7WkkZauoHNxkvgUYbig0rDmlKqDkz24cXMd6A7txM,700
|
@@ -32,7 +32,7 @@ langchain_ocr_lib/prompt_templates/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQe
|
|
32
32
|
langchain_ocr_lib/prompt_templates/ocr_prompt.py,sha256=3Be1AL-HJkxPnAP0DNH1MqvAxFWTCeM5UOKP63xkHsY,3543
|
33
33
|
langchain_ocr_lib/tracers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
34
34
|
langchain_ocr_lib/tracers/traced_chain.py,sha256=uxRkdLNn_G6dAsti_gUuF7muhIj10xrOUL7HUga40oc,3056
|
35
|
-
langchain_ocr_lib-0.3.
|
36
|
-
langchain_ocr_lib-0.3.
|
37
|
-
langchain_ocr_lib-0.3.
|
38
|
-
langchain_ocr_lib-0.3.
|
35
|
+
langchain_ocr_lib-0.3.2.dist-info/METADATA,sha256=MYG6NqcxYMkVY2UQ0SbSCdNmpSJJ98NUb5KLBSDciyg,6440
|
36
|
+
langchain_ocr_lib-0.3.2.dist-info/WHEEL,sha256=fGIA9gx4Qxk2KDKeNJCbOEwSrmLtjWCwzBz351GyrPQ,88
|
37
|
+
langchain_ocr_lib-0.3.2.dist-info/entry_points.txt,sha256=l4mIs0tnIgbJYuVveZySQKVBnqNMHS-8ZZtLwz8ag5k,61
|
38
|
+
langchain_ocr_lib-0.3.2.dist-info/RECORD,,
|
File without changes
|