langchain 0.3.35 → 0.3.37
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agents/react/index.d.ts +1 -1
- package/dist/hub/base.cjs +3 -0
- package/dist/hub/base.js +3 -0
- package/dist/hub/node.cjs +3 -0
- package/dist/hub/node.js +3 -0
- package/dist/load/import_map.cjs +6 -1
- package/dist/load/import_map.d.ts +5 -0
- package/dist/load/import_map.js +5 -0
- package/package.json +1 -1
|
@@ -4,7 +4,7 @@ import type { BaseLanguageModelInterface } from "@langchain/core/language_models
|
|
|
4
4
|
import { AgentStep } from "@langchain/core/agents";
|
|
5
5
|
import { AgentRunnableSequence } from "../agent.js";
|
|
6
6
|
/**
|
|
7
|
-
* Params used by the
|
|
7
|
+
* Params used by the createReactAgent function.
|
|
8
8
|
*/
|
|
9
9
|
export type CreateReactAgentParams = {
|
|
10
10
|
/** LLM to use for the agent. */
|
package/dist/hub/base.cjs
CHANGED
|
@@ -94,6 +94,9 @@ modelClass) {
|
|
|
94
94
|
else if (modelLcName === "ChatBedrockConverse") {
|
|
95
95
|
importMapKey = "chat_models__chat_bedrock_converse";
|
|
96
96
|
}
|
|
97
|
+
else if (modelLcName === "ChatMistralAI") {
|
|
98
|
+
importMapKey = "chat_models__mistralai";
|
|
99
|
+
}
|
|
97
100
|
else if (modelLcName === "ChatMistral") {
|
|
98
101
|
importMapKey = "chat_models__mistralai";
|
|
99
102
|
}
|
package/dist/hub/base.js
CHANGED
|
@@ -87,6 +87,9 @@ modelClass) {
|
|
|
87
87
|
else if (modelLcName === "ChatBedrockConverse") {
|
|
88
88
|
importMapKey = "chat_models__chat_bedrock_converse";
|
|
89
89
|
}
|
|
90
|
+
else if (modelLcName === "ChatMistralAI") {
|
|
91
|
+
importMapKey = "chat_models__mistralai";
|
|
92
|
+
}
|
|
90
93
|
else if (modelLcName === "ChatMistral") {
|
|
91
94
|
importMapKey = "chat_models__mistralai";
|
|
92
95
|
}
|
package/dist/hub/node.cjs
CHANGED
|
@@ -97,6 +97,9 @@ async function pull(ownerRepoCommit, options) {
|
|
|
97
97
|
else if (modelName === "ChatBedrockConverse") {
|
|
98
98
|
modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/aws")))).ChatBedrockConverse;
|
|
99
99
|
}
|
|
100
|
+
else if (modelName === "ChatMistralAI") {
|
|
101
|
+
modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/mistralai")))).ChatMistralAI;
|
|
102
|
+
}
|
|
100
103
|
else if (modelName === "ChatMistral") {
|
|
101
104
|
modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/mistralai")))).ChatMistralAI;
|
|
102
105
|
}
|
package/dist/hub/node.js
CHANGED
|
@@ -61,6 +61,9 @@ export async function pull(ownerRepoCommit, options) {
|
|
|
61
61
|
else if (modelName === "ChatBedrockConverse") {
|
|
62
62
|
modelClass = (await import("@langchain/aws")).ChatBedrockConverse;
|
|
63
63
|
}
|
|
64
|
+
else if (modelName === "ChatMistralAI") {
|
|
65
|
+
modelClass = (await import("@langchain/mistralai")).ChatMistralAI;
|
|
66
|
+
}
|
|
64
67
|
else if (modelName === "ChatMistral") {
|
|
65
68
|
modelClass = (await import("@langchain/mistralai")).ChatMistralAI;
|
|
66
69
|
}
|
package/dist/load/import_map.cjs
CHANGED
|
@@ -35,7 +35,7 @@ var __importStar = (this && this.__importStar) || (function () {
|
|
|
35
35
|
})();
|
|
36
36
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
37
37
|
exports.util__math = exports.util__document = exports.storage__in_memory = exports.storage__encoder_backed = exports.stores__message__in_memory = exports.stores__file__in_memory = exports.stores__doc__in_memory = exports.stores__doc__base = exports.retrievers__matryoshka_retriever = exports.retrievers__score_threshold = exports.retrievers__hyde = exports.retrievers__document_compressors__embeddings_filter = exports.retrievers__document_compressors__chain_extract = exports.retrievers__time_weighted = exports.retrievers__parent_document = exports.retrievers__multi_vector = exports.retrievers__multi_query = exports.retrievers__ensemble = exports.retrievers__document_compressors = exports.retrievers__contextual_compression = exports.output_parsers = exports.callbacks = exports.document_transformers__openai_functions = exports.document_loaders__base = exports.memory__chat_memory = exports.memory = exports.text_splitter = exports.vectorstores__memory = exports.embeddings__fake = exports.embeddings__cache_backed = exports.chains__retrieval = exports.chains__openai_functions = exports.chains__history_aware_retriever = exports.chains__combine_documents__reduce = exports.chains__combine_documents = exports.chains = exports.tools__retriever = exports.tools__render = exports.tools__chain = exports.tools = exports.agents__openai__output_parser = exports.agents__xml__output_parser = exports.agents__react__output_parser = exports.agents__format_scratchpad__log_to_message = exports.agents__format_scratchpad__xml = exports.agents__format_scratchpad__log = exports.agents__format_scratchpad__openai_tools = exports.agents__format_scratchpad = exports.agents__toolkits = exports.agents = void 0;
|
|
38
|
-
exports.schema__output = exports.schema__output_parser = exports.schema__runnable = exports.prompts__base = exports.prompts__pipeline = exports.prompts__image = exports.prompts__chat = exports.schema = exports.schema__messages = exports.prompts__prompt = exports.embeddings__azure_openai = exports.embeddings__openai = exports.llms__azure_openai = exports.llms__openai = exports.chat_models__azure_openai = exports.chat_models__openai = exports.schema__prompt_template = exports.schema__query_constructor = exports.indexes = exports.runnables__remote = exports.smith = exports.evaluation = exports.experimental__prompts__custom_format = exports.experimental__masking = exports.experimental__chains__violation_of_expectations = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__openai_files = exports.experimental__openai_assistant = exports.experimental__autogpt = exports.util__time = void 0;
|
|
38
|
+
exports.langsmith = exports.schema__output = exports.schema__output_parser = exports.schema__runnable = exports.prompts__base = exports.prompts__pipeline = exports.prompts__image = exports.prompts__chat = exports.schema = exports.schema__messages = exports.prompts__prompt = exports.embeddings__azure_openai = exports.embeddings__openai = exports.llms__azure_openai = exports.llms__openai = exports.chat_models__azure_openai = exports.chat_models__openai = exports.schema__prompt_template = exports.schema__query_constructor = exports.indexes = exports.runnables__remote = exports.smith = exports.evaluation = exports.experimental__prompts__custom_format = exports.experimental__masking = exports.experimental__chains__violation_of_expectations = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__openai_files = exports.experimental__openai_assistant = exports.experimental__autogpt = exports.util__time = void 0;
|
|
39
39
|
exports.agents = __importStar(require("../agents/index.cjs"));
|
|
40
40
|
exports.agents__toolkits = __importStar(require("../agents/toolkits/index.cjs"));
|
|
41
41
|
exports.agents__format_scratchpad = __importStar(require("../agents/format_scratchpad/openai_functions.cjs"));
|
|
@@ -109,6 +109,7 @@ const prompt_values_1 = require("@langchain/core/prompt_values");
|
|
|
109
109
|
const runnables_1 = require("@langchain/core/runnables");
|
|
110
110
|
const output_parsers_1 = require("@langchain/core/output_parsers");
|
|
111
111
|
const outputs_1 = require("@langchain/core/outputs");
|
|
112
|
+
const langsmith_1 = require("langsmith");
|
|
112
113
|
const chat_models__openai = {
|
|
113
114
|
ChatOpenAI: openai_1.ChatOpenAI
|
|
114
115
|
};
|
|
@@ -217,3 +218,7 @@ const schema__output = {
|
|
|
217
218
|
GenerationChunk: outputs_1.GenerationChunk
|
|
218
219
|
};
|
|
219
220
|
exports.schema__output = schema__output;
|
|
221
|
+
const langsmith = {
|
|
222
|
+
Client: langsmith_1.Client
|
|
223
|
+
};
|
|
224
|
+
exports.langsmith = langsmith;
|
|
@@ -71,6 +71,7 @@ import { StringPromptValue } from "@langchain/core/prompt_values";
|
|
|
71
71
|
import { RouterRunnable, RunnableAssign, RunnableBinding, RunnableBranch, RunnableEach, RunnableMap, RunnableParallel, RunnablePassthrough, RunnablePick, RunnableRetry, RunnableSequence, RunnableWithFallbacks, RunnableWithMessageHistory } from "@langchain/core/runnables";
|
|
72
72
|
import { StringOutputParser } from "@langchain/core/output_parsers";
|
|
73
73
|
import { ChatGenerationChunk, GenerationChunk } from "@langchain/core/outputs";
|
|
74
|
+
import { Client } from "langsmith";
|
|
74
75
|
declare const chat_models__openai: {
|
|
75
76
|
ChatOpenAI: typeof ChatOpenAI;
|
|
76
77
|
};
|
|
@@ -179,3 +180,7 @@ declare const schema__output: {
|
|
|
179
180
|
GenerationChunk: typeof GenerationChunk;
|
|
180
181
|
};
|
|
181
182
|
export { schema__output };
|
|
183
|
+
declare const langsmith: {
|
|
184
|
+
Client: typeof Client;
|
|
185
|
+
};
|
|
186
|
+
export { langsmith };
|
package/dist/load/import_map.js
CHANGED
|
@@ -72,6 +72,7 @@ import { StringPromptValue } from "@langchain/core/prompt_values";
|
|
|
72
72
|
import { RouterRunnable, RunnableAssign, RunnableBinding, RunnableBranch, RunnableEach, RunnableMap, RunnableParallel, RunnablePassthrough, RunnablePick, RunnableRetry, RunnableSequence, RunnableWithFallbacks, RunnableWithMessageHistory } from "@langchain/core/runnables";
|
|
73
73
|
import { StringOutputParser } from "@langchain/core/output_parsers";
|
|
74
74
|
import { ChatGenerationChunk, GenerationChunk } from "@langchain/core/outputs";
|
|
75
|
+
import { Client } from "langsmith";
|
|
75
76
|
const chat_models__openai = {
|
|
76
77
|
ChatOpenAI
|
|
77
78
|
};
|
|
@@ -180,3 +181,7 @@ const schema__output = {
|
|
|
180
181
|
GenerationChunk
|
|
181
182
|
};
|
|
182
183
|
export { schema__output };
|
|
184
|
+
const langsmith = {
|
|
185
|
+
Client
|
|
186
|
+
};
|
|
187
|
+
export { langsmith };
|