@langchain/core 0.3.4 → 0.3.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -88,6 +88,8 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
88
88
|
};
|
|
89
89
|
const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [messages], runnableConfig.runId, undefined, extra, undefined, undefined, runnableConfig.runName);
|
|
90
90
|
let generationChunk;
|
|
91
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
92
|
+
let llmOutput;
|
|
91
93
|
try {
|
|
92
94
|
for await (const chunk of this._streamResponseChunks(messages, callOptions, runManagers?.[0])) {
|
|
93
95
|
if (chunk.message.id == null) {
|
|
@@ -106,6 +108,16 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
106
108
|
else {
|
|
107
109
|
generationChunk = generationChunk.concat(chunk);
|
|
108
110
|
}
|
|
111
|
+
if ((0, index_js_1.isAIMessageChunk)(chunk.message) &&
|
|
112
|
+
chunk.message.usage_metadata !== undefined) {
|
|
113
|
+
llmOutput = {
|
|
114
|
+
tokenUsage: {
|
|
115
|
+
promptTokens: chunk.message.usage_metadata.input_tokens,
|
|
116
|
+
completionTokens: chunk.message.usage_metadata.output_tokens,
|
|
117
|
+
totalTokens: chunk.message.usage_metadata.total_tokens,
|
|
118
|
+
},
|
|
119
|
+
};
|
|
120
|
+
}
|
|
109
121
|
}
|
|
110
122
|
}
|
|
111
123
|
catch (err) {
|
|
@@ -115,6 +127,7 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
115
127
|
await Promise.all((runManagers ?? []).map((runManager) => runManager?.handleLLMEnd({
|
|
116
128
|
// TODO: Remove cast after figuring out inheritance
|
|
117
129
|
generations: [[generationChunk]],
|
|
130
|
+
llmOutput,
|
|
118
131
|
})));
|
|
119
132
|
}
|
|
120
133
|
}
|
|
@@ -154,6 +167,8 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
154
167
|
try {
|
|
155
168
|
const stream = await this._streamResponseChunks(baseMessages[0], parsedOptions, runManagers?.[0]);
|
|
156
169
|
let aggregated;
|
|
170
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
171
|
+
let llmOutput;
|
|
157
172
|
for await (const chunk of stream) {
|
|
158
173
|
if (chunk.message.id == null) {
|
|
159
174
|
const runId = runManagers?.at(0)?.runId;
|
|
@@ -166,6 +181,16 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
166
181
|
else {
|
|
167
182
|
aggregated = (0, stream_js_1.concat)(aggregated, chunk);
|
|
168
183
|
}
|
|
184
|
+
if ((0, index_js_1.isAIMessageChunk)(chunk.message) &&
|
|
185
|
+
chunk.message.usage_metadata !== undefined) {
|
|
186
|
+
llmOutput = {
|
|
187
|
+
tokenUsage: {
|
|
188
|
+
promptTokens: chunk.message.usage_metadata.input_tokens,
|
|
189
|
+
completionTokens: chunk.message.usage_metadata.output_tokens,
|
|
190
|
+
totalTokens: chunk.message.usage_metadata.total_tokens,
|
|
191
|
+
},
|
|
192
|
+
};
|
|
193
|
+
}
|
|
169
194
|
}
|
|
170
195
|
if (aggregated === undefined) {
|
|
171
196
|
throw new Error("Received empty response from chat model call.");
|
|
@@ -173,7 +198,7 @@ class BaseChatModel extends base_js_1.BaseLanguageModel {
|
|
|
173
198
|
generations.push([aggregated]);
|
|
174
199
|
await runManagers?.[0].handleLLMEnd({
|
|
175
200
|
generations,
|
|
176
|
-
llmOutput
|
|
201
|
+
llmOutput,
|
|
177
202
|
});
|
|
178
203
|
}
|
|
179
204
|
catch (e) {
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { z } from "zod";
|
|
2
|
-
import { type BaseMessage, BaseMessageChunk, type BaseMessageLike } from "../messages/index.js";
|
|
2
|
+
import { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from "../messages/index.js";
|
|
3
3
|
import type { BasePromptValueInterface } from "../prompt_values.js";
|
|
4
4
|
import { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from "../outputs.js";
|
|
5
5
|
import { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from "./base.js";
|
|
@@ -66,7 +66,7 @@ export type BindToolsInput = StructuredToolInterface | Record<string, any> | Too
|
|
|
66
66
|
* Base class for chat models. It extends the BaseLanguageModel class and
|
|
67
67
|
* provides methods for generating chat based on input messages.
|
|
68
68
|
*/
|
|
69
|
-
export declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, OutputMessageType extends BaseMessageChunk =
|
|
69
|
+
export declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {
|
|
70
70
|
ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, "signal" | "timeout" | "maxConcurrency">>;
|
|
71
71
|
lc_namespace: string[];
|
|
72
72
|
constructor(fields: BaseChatModelParams);
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { zodToJsonSchema } from "zod-to-json-schema";
|
|
2
|
-
import { AIMessage, HumanMessage, coerceMessageLikeToMessage, } from "../messages/index.js";
|
|
2
|
+
import { AIMessage, HumanMessage, coerceMessageLikeToMessage, isAIMessageChunk, } from "../messages/index.js";
|
|
3
3
|
import { RUN_KEY, } from "../outputs.js";
|
|
4
4
|
import { BaseLanguageModel, } from "./base.js";
|
|
5
5
|
import { CallbackManager, } from "../callbacks/manager.js";
|
|
@@ -84,6 +84,8 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
84
84
|
};
|
|
85
85
|
const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [messages], runnableConfig.runId, undefined, extra, undefined, undefined, runnableConfig.runName);
|
|
86
86
|
let generationChunk;
|
|
87
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
88
|
+
let llmOutput;
|
|
87
89
|
try {
|
|
88
90
|
for await (const chunk of this._streamResponseChunks(messages, callOptions, runManagers?.[0])) {
|
|
89
91
|
if (chunk.message.id == null) {
|
|
@@ -102,6 +104,16 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
102
104
|
else {
|
|
103
105
|
generationChunk = generationChunk.concat(chunk);
|
|
104
106
|
}
|
|
107
|
+
if (isAIMessageChunk(chunk.message) &&
|
|
108
|
+
chunk.message.usage_metadata !== undefined) {
|
|
109
|
+
llmOutput = {
|
|
110
|
+
tokenUsage: {
|
|
111
|
+
promptTokens: chunk.message.usage_metadata.input_tokens,
|
|
112
|
+
completionTokens: chunk.message.usage_metadata.output_tokens,
|
|
113
|
+
totalTokens: chunk.message.usage_metadata.total_tokens,
|
|
114
|
+
},
|
|
115
|
+
};
|
|
116
|
+
}
|
|
105
117
|
}
|
|
106
118
|
}
|
|
107
119
|
catch (err) {
|
|
@@ -111,6 +123,7 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
111
123
|
await Promise.all((runManagers ?? []).map((runManager) => runManager?.handleLLMEnd({
|
|
112
124
|
// TODO: Remove cast after figuring out inheritance
|
|
113
125
|
generations: [[generationChunk]],
|
|
126
|
+
llmOutput,
|
|
114
127
|
})));
|
|
115
128
|
}
|
|
116
129
|
}
|
|
@@ -150,6 +163,8 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
150
163
|
try {
|
|
151
164
|
const stream = await this._streamResponseChunks(baseMessages[0], parsedOptions, runManagers?.[0]);
|
|
152
165
|
let aggregated;
|
|
166
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
167
|
+
let llmOutput;
|
|
153
168
|
for await (const chunk of stream) {
|
|
154
169
|
if (chunk.message.id == null) {
|
|
155
170
|
const runId = runManagers?.at(0)?.runId;
|
|
@@ -162,6 +177,16 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
162
177
|
else {
|
|
163
178
|
aggregated = concat(aggregated, chunk);
|
|
164
179
|
}
|
|
180
|
+
if (isAIMessageChunk(chunk.message) &&
|
|
181
|
+
chunk.message.usage_metadata !== undefined) {
|
|
182
|
+
llmOutput = {
|
|
183
|
+
tokenUsage: {
|
|
184
|
+
promptTokens: chunk.message.usage_metadata.input_tokens,
|
|
185
|
+
completionTokens: chunk.message.usage_metadata.output_tokens,
|
|
186
|
+
totalTokens: chunk.message.usage_metadata.total_tokens,
|
|
187
|
+
},
|
|
188
|
+
};
|
|
189
|
+
}
|
|
165
190
|
}
|
|
166
191
|
if (aggregated === undefined) {
|
|
167
192
|
throw new Error("Received empty response from chat model call.");
|
|
@@ -169,7 +194,7 @@ export class BaseChatModel extends BaseLanguageModel {
|
|
|
169
194
|
generations.push([aggregated]);
|
|
170
195
|
await runManagers?.[0].handleLLMEnd({
|
|
171
196
|
generations,
|
|
172
|
-
llmOutput
|
|
197
|
+
llmOutput,
|
|
173
198
|
});
|
|
174
199
|
}
|
|
175
200
|
catch (e) {
|