langchain 0.2.15__py3-none-any.whl → 0.3.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of langchain might be problematic. Click here for more details.
- langchain/agents/agent.py +23 -19
- langchain/agents/agent_toolkits/vectorstore/toolkit.py +10 -7
- langchain/agents/chat/base.py +1 -1
- langchain/agents/conversational/base.py +1 -1
- langchain/agents/conversational_chat/base.py +1 -1
- langchain/agents/mrkl/base.py +1 -1
- langchain/agents/openai_assistant/base.py +8 -7
- langchain/agents/openai_functions_agent/base.py +6 -5
- langchain/agents/openai_functions_multi_agent/base.py +6 -5
- langchain/agents/openai_tools/base.py +8 -3
- langchain/agents/react/base.py +1 -1
- langchain/agents/self_ask_with_search/base.py +1 -1
- langchain/agents/structured_chat/base.py +1 -1
- langchain/agents/structured_chat/output_parser.py +1 -1
- langchain/chains/api/base.py +14 -12
- langchain/chains/base.py +17 -9
- langchain/chains/combine_documents/base.py +1 -1
- langchain/chains/combine_documents/map_reduce.py +14 -10
- langchain/chains/combine_documents/map_rerank.py +17 -14
- langchain/chains/combine_documents/reduce.py +5 -3
- langchain/chains/combine_documents/refine.py +11 -8
- langchain/chains/combine_documents/stuff.py +8 -6
- langchain/chains/constitutional_ai/models.py +1 -1
- langchain/chains/conversation/base.py +13 -11
- langchain/chains/conversational_retrieval/base.py +10 -8
- langchain/chains/elasticsearch_database/base.py +11 -9
- langchain/chains/flare/base.py +5 -2
- langchain/chains/hyde/base.py +6 -4
- langchain/chains/llm.py +7 -7
- langchain/chains/llm_checker/base.py +8 -6
- langchain/chains/llm_math/base.py +8 -6
- langchain/chains/llm_summarization_checker/base.py +8 -6
- langchain/chains/mapreduce.py +5 -3
- langchain/chains/moderation.py +14 -12
- langchain/chains/natbot/base.py +8 -6
- langchain/chains/openai_functions/base.py +3 -3
- langchain/chains/openai_functions/citation_fuzzy_match.py +1 -1
- langchain/chains/openai_functions/extraction.py +8 -4
- langchain/chains/openai_functions/qa_with_structure.py +5 -2
- langchain/chains/openai_functions/tagging.py +5 -2
- langchain/chains/openai_tools/extraction.py +2 -2
- langchain/chains/prompt_selector.py +1 -1
- langchain/chains/qa_generation/base.py +1 -1
- langchain/chains/qa_with_sources/base.py +8 -6
- langchain/chains/qa_with_sources/retrieval.py +1 -1
- langchain/chains/qa_with_sources/vector_db.py +4 -3
- langchain/chains/query_constructor/schema.py +5 -4
- langchain/chains/retrieval_qa/base.py +12 -9
- langchain/chains/router/base.py +5 -3
- langchain/chains/router/embedding_router.py +5 -3
- langchain/chains/router/llm_router.py +6 -5
- langchain/chains/sequential.py +17 -13
- langchain/chains/structured_output/base.py +8 -8
- langchain/chains/transform.py +1 -1
- langchain/chat_models/base.py +2 -2
- langchain/evaluation/agents/trajectory_eval_chain.py +4 -3
- langchain/evaluation/comparison/eval_chain.py +4 -3
- langchain/evaluation/criteria/eval_chain.py +4 -3
- langchain/evaluation/embedding_distance/base.py +4 -3
- langchain/evaluation/qa/eval_chain.py +7 -4
- langchain/evaluation/qa/generate_chain.py +1 -1
- langchain/evaluation/scoring/eval_chain.py +4 -3
- langchain/evaluation/string_distance/base.py +1 -1
- langchain/indexes/vectorstore.py +9 -7
- langchain/memory/chat_memory.py +1 -1
- langchain/memory/combined.py +5 -3
- langchain/memory/entity.py +4 -3
- langchain/memory/summary.py +1 -1
- langchain/memory/vectorstore.py +1 -1
- langchain/memory/vectorstore_token_buffer_memory.py +1 -1
- langchain/output_parsers/fix.py +3 -2
- langchain/output_parsers/pandas_dataframe.py +3 -2
- langchain/output_parsers/retry.py +4 -3
- langchain/output_parsers/structured.py +1 -1
- langchain/output_parsers/yaml.py +5 -2
- langchain/pydantic_v1/__init__.py +20 -0
- langchain/pydantic_v1/dataclasses.py +20 -0
- langchain/pydantic_v1/main.py +20 -0
- langchain/retrievers/contextual_compression.py +4 -2
- langchain/retrievers/document_compressors/base.py +4 -2
- langchain/retrievers/document_compressors/chain_extract.py +4 -2
- langchain/retrievers/document_compressors/chain_filter.py +4 -2
- langchain/retrievers/document_compressors/cohere_rerank.py +8 -6
- langchain/retrievers/document_compressors/cross_encoder_rerank.py +5 -3
- langchain/retrievers/document_compressors/embeddings_filter.py +5 -4
- langchain/retrievers/document_compressors/listwise_rerank.py +4 -3
- langchain/retrievers/ensemble.py +18 -14
- langchain/retrievers/multi_vector.py +5 -4
- langchain/retrievers/self_query/base.py +19 -8
- langchain/retrievers/time_weighted_retriever.py +4 -3
- langchain/smith/evaluation/config.py +7 -5
- {langchain-0.2.15.dist-info → langchain-0.3.0.dist-info}/METADATA +5 -5
- {langchain-0.2.15.dist-info → langchain-0.3.0.dist-info}/RECORD +96 -96
- {langchain-0.2.15.dist-info → langchain-0.3.0.dist-info}/LICENSE +0 -0
- {langchain-0.2.15.dist-info → langchain-0.3.0.dist-info}/WHEEL +0 -0
- {langchain-0.2.15.dist-info → langchain-0.3.0.dist-info}/entry_points.txt +0 -0
langchain/retrievers/ensemble.py
CHANGED
|
@@ -24,8 +24,6 @@ from langchain_core.callbacks import (
|
|
|
24
24
|
CallbackManagerForRetrieverRun,
|
|
25
25
|
)
|
|
26
26
|
from langchain_core.documents import Document
|
|
27
|
-
from langchain_core.load.dump import dumpd
|
|
28
|
-
from langchain_core.pydantic_v1 import root_validator
|
|
29
27
|
from langchain_core.retrievers import BaseRetriever, RetrieverLike
|
|
30
28
|
from langchain_core.runnables import RunnableConfig
|
|
31
29
|
from langchain_core.runnables.config import ensure_config, patch_config
|
|
@@ -33,6 +31,7 @@ from langchain_core.runnables.utils import (
|
|
|
33
31
|
ConfigurableFieldSpec,
|
|
34
32
|
get_unique_config_specs,
|
|
35
33
|
)
|
|
34
|
+
from pydantic import model_validator
|
|
36
35
|
|
|
37
36
|
T = TypeVar("T")
|
|
38
37
|
H = TypeVar("H", bound=Hashable)
|
|
@@ -83,8 +82,9 @@ class EnsembleRetriever(BaseRetriever):
|
|
|
83
82
|
spec for retriever in self.retrievers for spec in retriever.config_specs
|
|
84
83
|
)
|
|
85
84
|
|
|
86
|
-
@
|
|
87
|
-
|
|
85
|
+
@model_validator(mode="before")
|
|
86
|
+
@classmethod
|
|
87
|
+
def set_weights(cls, values: Dict[str, Any]) -> Any:
|
|
88
88
|
if not values.get("weights"):
|
|
89
89
|
n_retrievers = len(values["retrievers"])
|
|
90
90
|
values["weights"] = [1 / n_retrievers] * n_retrievers
|
|
@@ -106,9 +106,9 @@ class EnsembleRetriever(BaseRetriever):
|
|
|
106
106
|
local_metadata=self.metadata,
|
|
107
107
|
)
|
|
108
108
|
run_manager = callback_manager.on_retriever_start(
|
|
109
|
-
|
|
109
|
+
None,
|
|
110
110
|
input,
|
|
111
|
-
name=config.get("run_name"),
|
|
111
|
+
name=config.get("run_name") or self.get_name(),
|
|
112
112
|
**kwargs,
|
|
113
113
|
)
|
|
114
114
|
try:
|
|
@@ -139,9 +139,9 @@ class EnsembleRetriever(BaseRetriever):
|
|
|
139
139
|
local_metadata=self.metadata,
|
|
140
140
|
)
|
|
141
141
|
run_manager = await callback_manager.on_retriever_start(
|
|
142
|
-
|
|
142
|
+
None,
|
|
143
143
|
input,
|
|
144
|
-
name=config.get("run_name"),
|
|
144
|
+
name=config.get("run_name") or self.get_name(),
|
|
145
145
|
**kwargs,
|
|
146
146
|
)
|
|
147
147
|
try:
|
|
@@ -310,9 +310,11 @@ class EnsembleRetriever(BaseRetriever):
|
|
|
310
310
|
for doc_list, weight in zip(doc_lists, self.weights):
|
|
311
311
|
for rank, doc in enumerate(doc_list, start=1):
|
|
312
312
|
rrf_score[
|
|
313
|
-
|
|
314
|
-
|
|
315
|
-
|
|
313
|
+
(
|
|
314
|
+
doc.page_content
|
|
315
|
+
if self.id_key is None
|
|
316
|
+
else doc.metadata[self.id_key]
|
|
317
|
+
)
|
|
316
318
|
] += weight / (rank + self.c)
|
|
317
319
|
|
|
318
320
|
# Docs are deduplicated by their contents then sorted by their scores
|
|
@@ -320,9 +322,11 @@ class EnsembleRetriever(BaseRetriever):
|
|
|
320
322
|
sorted_docs = sorted(
|
|
321
323
|
unique_by_key(
|
|
322
324
|
all_docs,
|
|
323
|
-
lambda doc:
|
|
324
|
-
|
|
325
|
-
|
|
325
|
+
lambda doc: (
|
|
326
|
+
doc.page_content
|
|
327
|
+
if self.id_key is None
|
|
328
|
+
else doc.metadata[self.id_key]
|
|
329
|
+
),
|
|
326
330
|
),
|
|
327
331
|
reverse=True,
|
|
328
332
|
key=lambda doc: rrf_score[
|
|
@@ -1,15 +1,15 @@
|
|
|
1
1
|
from enum import Enum
|
|
2
|
-
from typing import Dict, List, Optional
|
|
2
|
+
from typing import Any, Dict, List, Optional
|
|
3
3
|
|
|
4
4
|
from langchain_core.callbacks import (
|
|
5
5
|
AsyncCallbackManagerForRetrieverRun,
|
|
6
6
|
CallbackManagerForRetrieverRun,
|
|
7
7
|
)
|
|
8
8
|
from langchain_core.documents import Document
|
|
9
|
-
from langchain_core.pydantic_v1 import Field, root_validator
|
|
10
9
|
from langchain_core.retrievers import BaseRetriever
|
|
11
10
|
from langchain_core.stores import BaseStore, ByteStore
|
|
12
11
|
from langchain_core.vectorstores import VectorStore
|
|
12
|
+
from pydantic import Field, model_validator
|
|
13
13
|
|
|
14
14
|
from langchain.storage._lc_store import create_kv_docstore
|
|
15
15
|
|
|
@@ -41,8 +41,9 @@ class MultiVectorRetriever(BaseRetriever):
|
|
|
41
41
|
search_type: SearchType = SearchType.similarity
|
|
42
42
|
"""Type of search to perform (similarity / mmr)"""
|
|
43
43
|
|
|
44
|
-
@
|
|
45
|
-
|
|
44
|
+
@model_validator(mode="before")
|
|
45
|
+
@classmethod
|
|
46
|
+
def shim_docstore(cls, values: Dict) -> Any:
|
|
46
47
|
byte_store = values.get("byte_store")
|
|
47
48
|
docstore = values.get("docstore")
|
|
48
49
|
if byte_store is not None:
|
|
@@ -9,11 +9,11 @@ from langchain_core.callbacks.manager import (
|
|
|
9
9
|
)
|
|
10
10
|
from langchain_core.documents import Document
|
|
11
11
|
from langchain_core.language_models import BaseLanguageModel
|
|
12
|
-
from langchain_core.pydantic_v1 import Field, root_validator
|
|
13
12
|
from langchain_core.retrievers import BaseRetriever
|
|
14
13
|
from langchain_core.runnables import Runnable
|
|
15
14
|
from langchain_core.structured_query import StructuredQuery, Visitor
|
|
16
15
|
from langchain_core.vectorstores import VectorStore
|
|
16
|
+
from pydantic import ConfigDict, Field, model_validator
|
|
17
17
|
|
|
18
18
|
from langchain.chains.query_constructor.base import load_query_constructor_runnable
|
|
19
19
|
from langchain.chains.query_constructor.schema import AttributeInfo
|
|
@@ -48,6 +48,7 @@ def _get_builtin_translator(vectorstore: VectorStore) -> Visitor:
|
|
|
48
48
|
MongoDBAtlasTranslator,
|
|
49
49
|
)
|
|
50
50
|
from langchain_community.query_constructors.myscale import MyScaleTranslator
|
|
51
|
+
from langchain_community.query_constructors.neo4j import Neo4jTranslator
|
|
51
52
|
from langchain_community.query_constructors.opensearch import OpenSearchTranslator
|
|
52
53
|
from langchain_community.query_constructors.pgvector import PGVectorTranslator
|
|
53
54
|
from langchain_community.query_constructors.pinecone import PineconeTranslator
|
|
@@ -70,6 +71,7 @@ def _get_builtin_translator(vectorstore: VectorStore) -> Visitor:
|
|
|
70
71
|
Dingo,
|
|
71
72
|
Milvus,
|
|
72
73
|
MyScale,
|
|
74
|
+
Neo4jVector,
|
|
73
75
|
OpenSearchVectorSearch,
|
|
74
76
|
PGVector,
|
|
75
77
|
Qdrant,
|
|
@@ -111,11 +113,10 @@ def _get_builtin_translator(vectorstore: VectorStore) -> Visitor:
|
|
|
111
113
|
TimescaleVector: TimescaleVectorTranslator,
|
|
112
114
|
OpenSearchVectorSearch: OpenSearchTranslator,
|
|
113
115
|
CommunityMongoDBAtlasVectorSearch: MongoDBAtlasTranslator,
|
|
116
|
+
Neo4jVector: Neo4jTranslator,
|
|
114
117
|
}
|
|
115
118
|
if isinstance(vectorstore, DatabricksVectorSearch):
|
|
116
119
|
return DatabricksVectorSearchTranslator()
|
|
117
|
-
if isinstance(vectorstore, Qdrant):
|
|
118
|
-
return QdrantTranslator(metadata_key=vectorstore.metadata_payload_key)
|
|
119
120
|
elif isinstance(vectorstore, MyScale):
|
|
120
121
|
return MyScaleTranslator(metadata_key=vectorstore.metadata_column)
|
|
121
122
|
elif isinstance(vectorstore, Redis):
|
|
@@ -177,6 +178,14 @@ def _get_builtin_translator(vectorstore: VectorStore) -> Visitor:
|
|
|
177
178
|
if isinstance(vectorstore, PGVector):
|
|
178
179
|
return NewPGVectorTranslator()
|
|
179
180
|
|
|
181
|
+
try:
|
|
182
|
+
from langchain_qdrant import QdrantVectorStore
|
|
183
|
+
except ImportError:
|
|
184
|
+
pass
|
|
185
|
+
else:
|
|
186
|
+
if isinstance(vectorstore, QdrantVectorStore):
|
|
187
|
+
return QdrantTranslator(metadata_key=vectorstore.metadata_payload_key)
|
|
188
|
+
|
|
180
189
|
try:
|
|
181
190
|
# Added in langchain-community==0.2.11
|
|
182
191
|
from langchain_community.query_constructors.hanavector import HanaTranslator
|
|
@@ -214,12 +223,14 @@ class SelfQueryRetriever(BaseRetriever):
|
|
|
214
223
|
use_original_query: bool = False
|
|
215
224
|
"""Use original query instead of the revised new query from LLM"""
|
|
216
225
|
|
|
217
|
-
|
|
218
|
-
|
|
219
|
-
arbitrary_types_allowed
|
|
226
|
+
model_config = ConfigDict(
|
|
227
|
+
populate_by_name=True,
|
|
228
|
+
arbitrary_types_allowed=True,
|
|
229
|
+
)
|
|
220
230
|
|
|
221
|
-
@
|
|
222
|
-
|
|
231
|
+
@model_validator(mode="before")
|
|
232
|
+
@classmethod
|
|
233
|
+
def validate_translator(cls, values: Dict) -> Any:
|
|
223
234
|
"""Validate translator."""
|
|
224
235
|
if "structured_query_translator" not in values:
|
|
225
236
|
values["structured_query_translator"] = _get_builtin_translator(
|
|
@@ -7,9 +7,9 @@ from langchain_core.callbacks import (
|
|
|
7
7
|
CallbackManagerForRetrieverRun,
|
|
8
8
|
)
|
|
9
9
|
from langchain_core.documents import Document
|
|
10
|
-
from langchain_core.pydantic_v1 import Field
|
|
11
10
|
from langchain_core.retrievers import BaseRetriever
|
|
12
11
|
from langchain_core.vectorstores import VectorStore
|
|
12
|
+
from pydantic import ConfigDict, Field
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
def _get_hours_passed(time: datetime.datetime, ref_time: datetime.datetime) -> float:
|
|
@@ -46,8 +46,9 @@ class TimeWeightedVectorStoreRetriever(BaseRetriever):
|
|
|
46
46
|
None assigns no salience to documents not fetched from the vector store.
|
|
47
47
|
"""
|
|
48
48
|
|
|
49
|
-
|
|
50
|
-
arbitrary_types_allowed
|
|
49
|
+
model_config = ConfigDict(
|
|
50
|
+
arbitrary_types_allowed=True,
|
|
51
|
+
)
|
|
51
52
|
|
|
52
53
|
def _document_get_date(self, field: str, document: Document) -> datetime.datetime:
|
|
53
54
|
"""Return the value of the date field of a document."""
|
|
@@ -5,10 +5,10 @@ from typing import Any, Callable, Dict, List, Optional, Sequence, Union
|
|
|
5
5
|
from langchain_core.embeddings import Embeddings
|
|
6
6
|
from langchain_core.language_models import BaseLanguageModel
|
|
7
7
|
from langchain_core.prompts import BasePromptTemplate
|
|
8
|
-
from langchain_core.pydantic_v1 import BaseModel, Field
|
|
9
8
|
from langsmith import RunEvaluator
|
|
10
9
|
from langsmith.evaluation.evaluator import EvaluationResult, EvaluationResults
|
|
11
10
|
from langsmith.schemas import Example, Run
|
|
11
|
+
from pydantic import BaseModel, ConfigDict, Field
|
|
12
12
|
|
|
13
13
|
from langchain.evaluation.criteria.eval_chain import CRITERIA_TYPE
|
|
14
14
|
from langchain.evaluation.embedding_distance.base import (
|
|
@@ -156,8 +156,9 @@ class RunEvalConfig(BaseModel):
|
|
|
156
156
|
eval_llm: Optional[BaseLanguageModel] = None
|
|
157
157
|
"""The language model to pass to any evaluators that require one."""
|
|
158
158
|
|
|
159
|
-
|
|
160
|
-
arbitrary_types_allowed
|
|
159
|
+
model_config = ConfigDict(
|
|
160
|
+
arbitrary_types_allowed=True,
|
|
161
|
+
)
|
|
161
162
|
|
|
162
163
|
class Criteria(SingleKeyEvalConfig):
|
|
163
164
|
"""Configuration for a reference-free criteria evaluator.
|
|
@@ -217,8 +218,9 @@ class RunEvalConfig(BaseModel):
|
|
|
217
218
|
embeddings: Optional[Embeddings] = None
|
|
218
219
|
distance_metric: Optional[EmbeddingDistanceEnum] = None
|
|
219
220
|
|
|
220
|
-
|
|
221
|
-
arbitrary_types_allowed
|
|
221
|
+
model_config = ConfigDict(
|
|
222
|
+
arbitrary_types_allowed=True,
|
|
223
|
+
)
|
|
222
224
|
|
|
223
225
|
class StringDistance(SingleKeyEvalConfig):
|
|
224
226
|
"""Configuration for a string distance evaluator.
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: langchain
|
|
3
|
-
Version: 0.
|
|
3
|
+
Version: 0.3.0
|
|
4
4
|
Summary: Building applications with LLMs through composability
|
|
5
5
|
Home-page: https://github.com/langchain-ai/langchain
|
|
6
6
|
License: MIT
|
|
7
|
-
Requires-Python: >=3.
|
|
7
|
+
Requires-Python: >=3.9,<4.0
|
|
8
8
|
Classifier: License :: OSI Approved :: MIT License
|
|
9
9
|
Classifier: Programming Language :: Python :: 3
|
|
10
10
|
Classifier: Programming Language :: Python :: 3.9
|
|
@@ -15,12 +15,12 @@ Requires-Dist: PyYAML (>=5.3)
|
|
|
15
15
|
Requires-Dist: SQLAlchemy (>=1.4,<3)
|
|
16
16
|
Requires-Dist: aiohttp (>=3.8.3,<4.0.0)
|
|
17
17
|
Requires-Dist: async-timeout (>=4.0.0,<5.0.0) ; python_version < "3.11"
|
|
18
|
-
Requires-Dist: langchain-core (>=0.
|
|
19
|
-
Requires-Dist: langchain-text-splitters (>=0.
|
|
18
|
+
Requires-Dist: langchain-core (>=0.3.0,<0.4.0)
|
|
19
|
+
Requires-Dist: langchain-text-splitters (>=0.3.0,<0.4.0)
|
|
20
20
|
Requires-Dist: langsmith (>=0.1.17,<0.2.0)
|
|
21
21
|
Requires-Dist: numpy (>=1,<2) ; python_version < "3.12"
|
|
22
22
|
Requires-Dist: numpy (>=1.26.0,<2.0.0) ; python_version >= "3.12"
|
|
23
|
-
Requires-Dist: pydantic (>=
|
|
23
|
+
Requires-Dist: pydantic (>=2.7.4,<3.0.0)
|
|
24
24
|
Requires-Dist: requests (>=2,<3)
|
|
25
25
|
Requires-Dist: tenacity (>=8.1.0,<9.0.0,!=8.4.0)
|
|
26
26
|
Project-URL: Repository, https://github.com/langchain-ai/langchain
|