langchain 0.3.35 → 0.3.36

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4,7 +4,7 @@ import type { BaseLanguageModelInterface } from "@langchain/core/language_models
4
4
  import { AgentStep } from "@langchain/core/agents";
5
5
  import { AgentRunnableSequence } from "../agent.js";
6
6
  /**
7
- * Params used by the createXmlAgent function.
7
+ * Params used by the createReactAgent function.
8
8
  */
9
9
  export type CreateReactAgentParams = {
10
10
  /** LLM to use for the agent. */
package/dist/hub/base.cjs CHANGED
@@ -94,6 +94,9 @@ modelClass) {
94
94
  else if (modelLcName === "ChatBedrockConverse") {
95
95
  importMapKey = "chat_models__chat_bedrock_converse";
96
96
  }
97
+ else if (modelLcName === "ChatMistralAI") {
98
+ importMapKey = "chat_models__mistralai";
99
+ }
97
100
  else if (modelLcName === "ChatMistral") {
98
101
  importMapKey = "chat_models__mistralai";
99
102
  }
package/dist/hub/base.js CHANGED
@@ -87,6 +87,9 @@ modelClass) {
87
87
  else if (modelLcName === "ChatBedrockConverse") {
88
88
  importMapKey = "chat_models__chat_bedrock_converse";
89
89
  }
90
+ else if (modelLcName === "ChatMistralAI") {
91
+ importMapKey = "chat_models__mistralai";
92
+ }
90
93
  else if (modelLcName === "ChatMistral") {
91
94
  importMapKey = "chat_models__mistralai";
92
95
  }
package/dist/hub/node.cjs CHANGED
@@ -97,6 +97,9 @@ async function pull(ownerRepoCommit, options) {
97
97
  else if (modelName === "ChatBedrockConverse") {
98
98
  modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/aws")))).ChatBedrockConverse;
99
99
  }
100
+ else if (modelName === "ChatMistralAI") {
101
+ modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/mistralai")))).ChatMistralAI;
102
+ }
100
103
  else if (modelName === "ChatMistral") {
101
104
  modelClass = (await Promise.resolve().then(() => __importStar(require("@langchain/mistralai")))).ChatMistralAI;
102
105
  }
package/dist/hub/node.js CHANGED
@@ -61,6 +61,9 @@ export async function pull(ownerRepoCommit, options) {
61
61
  else if (modelName === "ChatBedrockConverse") {
62
62
  modelClass = (await import("@langchain/aws")).ChatBedrockConverse;
63
63
  }
64
+ else if (modelName === "ChatMistralAI") {
65
+ modelClass = (await import("@langchain/mistralai")).ChatMistralAI;
66
+ }
64
67
  else if (modelName === "ChatMistral") {
65
68
  modelClass = (await import("@langchain/mistralai")).ChatMistralAI;
66
69
  }
@@ -35,7 +35,7 @@ var __importStar = (this && this.__importStar) || (function () {
35
35
  })();
36
36
  Object.defineProperty(exports, "__esModule", { value: true });
37
37
  exports.util__math = exports.util__document = exports.storage__in_memory = exports.storage__encoder_backed = exports.stores__message__in_memory = exports.stores__file__in_memory = exports.stores__doc__in_memory = exports.stores__doc__base = exports.retrievers__matryoshka_retriever = exports.retrievers__score_threshold = exports.retrievers__hyde = exports.retrievers__document_compressors__embeddings_filter = exports.retrievers__document_compressors__chain_extract = exports.retrievers__time_weighted = exports.retrievers__parent_document = exports.retrievers__multi_vector = exports.retrievers__multi_query = exports.retrievers__ensemble = exports.retrievers__document_compressors = exports.retrievers__contextual_compression = exports.output_parsers = exports.callbacks = exports.document_transformers__openai_functions = exports.document_loaders__base = exports.memory__chat_memory = exports.memory = exports.text_splitter = exports.vectorstores__memory = exports.embeddings__fake = exports.embeddings__cache_backed = exports.chains__retrieval = exports.chains__openai_functions = exports.chains__history_aware_retriever = exports.chains__combine_documents__reduce = exports.chains__combine_documents = exports.chains = exports.tools__retriever = exports.tools__render = exports.tools__chain = exports.tools = exports.agents__openai__output_parser = exports.agents__xml__output_parser = exports.agents__react__output_parser = exports.agents__format_scratchpad__log_to_message = exports.agents__format_scratchpad__xml = exports.agents__format_scratchpad__log = exports.agents__format_scratchpad__openai_tools = exports.agents__format_scratchpad = exports.agents__toolkits = exports.agents = void 0;
38
- exports.schema__output = exports.schema__output_parser = exports.schema__runnable = exports.prompts__base = exports.prompts__pipeline = exports.prompts__image = exports.prompts__chat = exports.schema = exports.schema__messages = exports.prompts__prompt = exports.embeddings__azure_openai = exports.embeddings__openai = exports.llms__azure_openai = exports.llms__openai = exports.chat_models__azure_openai = exports.chat_models__openai = exports.schema__prompt_template = exports.schema__query_constructor = exports.indexes = exports.runnables__remote = exports.smith = exports.evaluation = exports.experimental__prompts__custom_format = exports.experimental__masking = exports.experimental__chains__violation_of_expectations = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__openai_files = exports.experimental__openai_assistant = exports.experimental__autogpt = exports.util__time = void 0;
38
+ exports.langsmith = exports.schema__output = exports.schema__output_parser = exports.schema__runnable = exports.prompts__base = exports.prompts__pipeline = exports.prompts__image = exports.prompts__chat = exports.schema = exports.schema__messages = exports.prompts__prompt = exports.embeddings__azure_openai = exports.embeddings__openai = exports.llms__azure_openai = exports.llms__openai = exports.chat_models__azure_openai = exports.chat_models__openai = exports.schema__prompt_template = exports.schema__query_constructor = exports.indexes = exports.runnables__remote = exports.smith = exports.evaluation = exports.experimental__prompts__custom_format = exports.experimental__masking = exports.experimental__chains__violation_of_expectations = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__openai_files = exports.experimental__openai_assistant = exports.experimental__autogpt = exports.util__time = void 0;
39
39
  exports.agents = __importStar(require("../agents/index.cjs"));
40
40
  exports.agents__toolkits = __importStar(require("../agents/toolkits/index.cjs"));
41
41
  exports.agents__format_scratchpad = __importStar(require("../agents/format_scratchpad/openai_functions.cjs"));
@@ -109,6 +109,7 @@ const prompt_values_1 = require("@langchain/core/prompt_values");
109
109
  const runnables_1 = require("@langchain/core/runnables");
110
110
  const output_parsers_1 = require("@langchain/core/output_parsers");
111
111
  const outputs_1 = require("@langchain/core/outputs");
112
+ const langsmith_1 = require("langsmith");
112
113
  const chat_models__openai = {
113
114
  ChatOpenAI: openai_1.ChatOpenAI
114
115
  };
@@ -217,3 +218,7 @@ const schema__output = {
217
218
  GenerationChunk: outputs_1.GenerationChunk
218
219
  };
219
220
  exports.schema__output = schema__output;
221
+ const langsmith = {
222
+ Client: langsmith_1.Client
223
+ };
224
+ exports.langsmith = langsmith;
@@ -71,6 +71,7 @@ import { StringPromptValue } from "@langchain/core/prompt_values";
71
71
  import { RouterRunnable, RunnableAssign, RunnableBinding, RunnableBranch, RunnableEach, RunnableMap, RunnableParallel, RunnablePassthrough, RunnablePick, RunnableRetry, RunnableSequence, RunnableWithFallbacks, RunnableWithMessageHistory } from "@langchain/core/runnables";
72
72
  import { StringOutputParser } from "@langchain/core/output_parsers";
73
73
  import { ChatGenerationChunk, GenerationChunk } from "@langchain/core/outputs";
74
+ import { Client } from "langsmith";
74
75
  declare const chat_models__openai: {
75
76
  ChatOpenAI: typeof ChatOpenAI;
76
77
  };
@@ -179,3 +180,7 @@ declare const schema__output: {
179
180
  GenerationChunk: typeof GenerationChunk;
180
181
  };
181
182
  export { schema__output };
183
+ declare const langsmith: {
184
+ Client: typeof Client;
185
+ };
186
+ export { langsmith };
@@ -72,6 +72,7 @@ import { StringPromptValue } from "@langchain/core/prompt_values";
72
72
  import { RouterRunnable, RunnableAssign, RunnableBinding, RunnableBranch, RunnableEach, RunnableMap, RunnableParallel, RunnablePassthrough, RunnablePick, RunnableRetry, RunnableSequence, RunnableWithFallbacks, RunnableWithMessageHistory } from "@langchain/core/runnables";
73
73
  import { StringOutputParser } from "@langchain/core/output_parsers";
74
74
  import { ChatGenerationChunk, GenerationChunk } from "@langchain/core/outputs";
75
+ import { Client } from "langsmith";
75
76
  const chat_models__openai = {
76
77
  ChatOpenAI
77
78
  };
@@ -180,3 +181,7 @@ const schema__output = {
180
181
  GenerationChunk
181
182
  };
182
183
  export { schema__output };
184
+ const langsmith = {
185
+ Client
186
+ };
187
+ export { langsmith };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "langchain",
3
- "version": "0.3.35",
3
+ "version": "0.3.36",
4
4
  "description": "Typescript bindings for langchain",
5
5
  "type": "module",
6
6
  "engines": {