@langchain/core 0.3.3 → 0.3.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -88,6 +88,8 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
88
88
  };
89
89
  const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [messages], runnableConfig.runId, undefined, extra, undefined, undefined, runnableConfig.runName);
90
90
  let generationChunk;
91
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
92
+ let llmOutput;
91
93
  try {
92
94
  for await (const chunk of this._streamResponseChunks(messages, callOptions, runManagers?.[0])) {
93
95
  if (chunk.message.id == null) {
@@ -106,6 +108,16 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
106
108
  else {
107
109
  generationChunk = generationChunk.concat(chunk);
108
110
  }
111
+ if ((0, index_js_1.isAIMessageChunk)(chunk.message) &&
112
+ chunk.message.usage_metadata !== undefined) {
113
+ llmOutput = {
114
+ tokenUsage: {
115
+ promptTokens: chunk.message.usage_metadata.input_tokens,
116
+ completionTokens: chunk.message.usage_metadata.output_tokens,
117
+ totalTokens: chunk.message.usage_metadata.total_tokens,
118
+ },
119
+ };
120
+ }
109
121
  }
110
122
  }
111
123
  catch (err) {
@@ -115,6 +127,7 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
115
127
  await Promise.all((runManagers ?? []).map((runManager) => runManager?.handleLLMEnd({
116
128
  // TODO: Remove cast after figuring out inheritance
117
129
  generations: [[generationChunk]],
130
+ llmOutput,
118
131
  })));
119
132
  }
120
133
  }
@@ -154,6 +167,8 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
154
167
  try {
155
168
  const stream = await this._streamResponseChunks(baseMessages[0], parsedOptions, runManagers?.[0]);
156
169
  let aggregated;
170
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
171
+ let llmOutput;
157
172
  for await (const chunk of stream) {
158
173
  if (chunk.message.id == null) {
159
174
  const runId = runManagers?.at(0)?.runId;
@@ -166,6 +181,16 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
166
181
  else {
167
182
  aggregated = (0, stream_js_1.concat)(aggregated, chunk);
168
183
  }
184
+ if ((0, index_js_1.isAIMessageChunk)(chunk.message) &&
185
+ chunk.message.usage_metadata !== undefined) {
186
+ llmOutput = {
187
+ tokenUsage: {
188
+ promptTokens: chunk.message.usage_metadata.input_tokens,
189
+ completionTokens: chunk.message.usage_metadata.output_tokens,
190
+ totalTokens: chunk.message.usage_metadata.total_tokens,
191
+ },
192
+ };
193
+ }
169
194
  }
170
195
  if (aggregated === undefined) {
171
196
  throw new Error("Received empty response from chat model call.");
@@ -173,7 +198,7 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
173
198
  generations.push([aggregated]);
174
199
  await runManagers?.[0].handleLLMEnd({
175
200
  generations,
176
- llmOutput: {},
201
+ llmOutput,
177
202
  });
178
203
  }
179
204
  catch (e) {
@@ -1,5 +1,5 @@
1
1
  import { z } from "zod";
2
- import { type BaseMessage, BaseMessageChunk, type BaseMessageLike } from "../messages/index.js";
2
+ import { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from "../messages/index.js";
3
3
  import type { BasePromptValueInterface } from "../prompt_values.js";
4
4
  import { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from "../outputs.js";
5
5
  import { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from "./base.js";
@@ -66,7 +66,7 @@ export type BindToolsInput = StructuredToolInterface | Record<string, any> | Too
66
66
  * Base class for chat models. It extends the BaseLanguageModel class and
67
67
  * provides methods for generating chat based on input messages.
68
68
  */
69
- export declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, OutputMessageType extends BaseMessageChunk = BaseMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {
69
+ export declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {
70
70
  ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, "signal" | "timeout" | "maxConcurrency">>;
71
71
  lc_namespace: string[];
72
72
  constructor(fields: BaseChatModelParams);
@@ -1,5 +1,5 @@
1
1
  import { zodToJsonSchema } from "zod-to-json-schema";
2
- import { AIMessage, HumanMessage, coerceMessageLikeToMessage, } from "../messages/index.js";
2
+ import { AIMessage, HumanMessage, coerceMessageLikeToMessage, isAIMessageChunk, } from "../messages/index.js";
3
3
  import { RUN_KEY, } from "../outputs.js";
4
4
  import { BaseLanguageModel, } from "./base.js";
5
5
  import { CallbackManager, } from "../callbacks/manager.js";
@@ -84,6 +84,8 @@ export class BaseChatModel extends BaseLanguageModel {
84
84
  };
85
85
  const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [messages], runnableConfig.runId, undefined, extra, undefined, undefined, runnableConfig.runName);
86
86
  let generationChunk;
87
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
88
+ let llmOutput;
87
89
  try {
88
90
  for await (const chunk of this._streamResponseChunks(messages, callOptions, runManagers?.[0])) {
89
91
  if (chunk.message.id == null) {
@@ -102,6 +104,16 @@ export class BaseChatModel extends BaseLanguageModel {
102
104
  else {
103
105
  generationChunk = generationChunk.concat(chunk);
104
106
  }
107
+ if (isAIMessageChunk(chunk.message) &&
108
+ chunk.message.usage_metadata !== undefined) {
109
+ llmOutput = {
110
+ tokenUsage: {
111
+ promptTokens: chunk.message.usage_metadata.input_tokens,
112
+ completionTokens: chunk.message.usage_metadata.output_tokens,
113
+ totalTokens: chunk.message.usage_metadata.total_tokens,
114
+ },
115
+ };
116
+ }
105
117
  }
106
118
  }
107
119
  catch (err) {
@@ -111,6 +123,7 @@ export class BaseChatModel extends BaseLanguageModel {
111
123
  await Promise.all((runManagers ?? []).map((runManager) => runManager?.handleLLMEnd({
112
124
  // TODO: Remove cast after figuring out inheritance
113
125
  generations: [[generationChunk]],
126
+ llmOutput,
114
127
  })));
115
128
  }
116
129
  }
@@ -150,6 +163,8 @@ export class BaseChatModel extends BaseLanguageModel {
150
163
  try {
151
164
  const stream = await this._streamResponseChunks(baseMessages[0], parsedOptions, runManagers?.[0]);
152
165
  let aggregated;
166
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
167
+ let llmOutput;
153
168
  for await (const chunk of stream) {
154
169
  if (chunk.message.id == null) {
155
170
  const runId = runManagers?.at(0)?.runId;
@@ -162,6 +177,16 @@ export class BaseChatModel extends BaseLanguageModel {
162
177
  else {
163
178
  aggregated = concat(aggregated, chunk);
164
179
  }
180
+ if (isAIMessageChunk(chunk.message) &&
181
+ chunk.message.usage_metadata !== undefined) {
182
+ llmOutput = {
183
+ tokenUsage: {
184
+ promptTokens: chunk.message.usage_metadata.input_tokens,
185
+ completionTokens: chunk.message.usage_metadata.output_tokens,
186
+ totalTokens: chunk.message.usage_metadata.total_tokens,
187
+ },
188
+ };
189
+ }
165
190
  }
166
191
  if (aggregated === undefined) {
167
192
  throw new Error("Received empty response from chat model call.");
@@ -169,7 +194,7 @@ export class BaseChatModel extends BaseLanguageModel {
169
194
  generations.push([aggregated]);
170
195
  await runManagers?.[0].handleLLMEnd({
171
196
  generations,
172
- llmOutput: {},
197
+ llmOutput,
173
198
  });
174
199
  }
175
200
  catch (e) {
@@ -38,7 +38,7 @@ type NonAlphanumeric = " " | "\t" | "\n" | "\r" | '"' | "'" | "{" | "[" | "(" |
38
38
  */
39
39
  type ExtractTemplateParamsRecursive<T extends string, Result extends string[] = []> = T extends `${string}{${infer Param}}${infer Rest}` ? Param extends `${NonAlphanumeric}${string}` ? ExtractTemplateParamsRecursive<Rest, Result> : ExtractTemplateParamsRecursive<Rest, [...Result, Param]> : Result;
40
40
  export type ParamsFromFString<T extends string> = {
41
- [Key in ExtractTemplateParamsRecursive<T>[number] | (string & Record<never, never>)]: string;
41
+ [Key in ExtractTemplateParamsRecursive<T>[number] | (string & Record<never, never>)]: any;
42
42
  };
43
43
  export type ExtractedFStringParams<T extends string, RunInput extends InputValues = Symbol> = RunInput extends Symbol ? ParamsFromFString<T> : RunInput;
44
44
  /**
@@ -86,15 +86,20 @@ const parseMustache = (template) => {
86
86
  return mustacheTemplateToNodes(parsed);
87
87
  };
88
88
  exports.parseMustache = parseMustache;
89
- const interpolateFString = (template, values) => (0, exports.parseFString)(template).reduce((res, node) => {
90
- if (node.type === "variable") {
91
- if (node.name in values) {
92
- return res + values[node.name];
89
+ const interpolateFString = (template, values) => {
90
+ return (0, exports.parseFString)(template).reduce((res, node) => {
91
+ if (node.type === "variable") {
92
+ if (node.name in values) {
93
+ const stringValue = typeof values[node.name] === "string"
94
+ ? values[node.name]
95
+ : JSON.stringify(values[node.name]);
96
+ return res + stringValue;
97
+ }
98
+ throw new Error(`(f-string) Missing value for input ${node.name}`);
93
99
  }
94
- throw new Error(`(f-string) Missing value for input ${node.name}`);
95
- }
96
- return res + node.text;
97
- }, "");
100
+ return res + node.text;
101
+ }, "");
102
+ };
98
103
  exports.interpolateFString = interpolateFString;
99
104
  const interpolateMustache = (template, values) => {
100
105
  configureMustache();
@@ -78,15 +78,20 @@ export const parseMustache = (template) => {
78
78
  const parsed = mustache.parse(template);
79
79
  return mustacheTemplateToNodes(parsed);
80
80
  };
81
- export const interpolateFString = (template, values) => parseFString(template).reduce((res, node) => {
82
- if (node.type === "variable") {
83
- if (node.name in values) {
84
- return res + values[node.name];
81
+ export const interpolateFString = (template, values) => {
82
+ return parseFString(template).reduce((res, node) => {
83
+ if (node.type === "variable") {
84
+ if (node.name in values) {
85
+ const stringValue = typeof values[node.name] === "string"
86
+ ? values[node.name]
87
+ : JSON.stringify(values[node.name]);
88
+ return res + stringValue;
89
+ }
90
+ throw new Error(`(f-string) Missing value for input ${node.name}`);
85
91
  }
86
- throw new Error(`(f-string) Missing value for input ${node.name}`);
87
- }
88
- return res + node.text;
89
- }, "");
92
+ return res + node.text;
93
+ }, "");
94
+ };
90
95
  export const interpolateMustache = (template, values) => {
91
96
  configureMustache();
92
97
  return mustache.render(template, values);
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@langchain/core",
3
- "version": "0.3.3",
3
+ "version": "0.3.5",
4
4
  "description": "Core LangChain.js abstractions and schemas",
5
5
  "type": "module",
6
6
  "engines": {