@langchain/core 1.0.0-alpha.6 → 1.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +131 -0
- package/LICENSE +6 -6
- package/README.md +2 -23
- package/dist/agents.d.ts.map +1 -1
- package/dist/caches/base.d.cts.map +1 -1
- package/dist/callbacks/base.d.cts.map +1 -1
- package/dist/callbacks/base.d.ts.map +1 -1
- package/dist/callbacks/manager.cjs +9 -64
- package/dist/callbacks/manager.cjs.map +1 -1
- package/dist/callbacks/manager.d.cts +1 -23
- package/dist/callbacks/manager.d.cts.map +1 -1
- package/dist/callbacks/manager.d.ts +1 -23
- package/dist/callbacks/manager.d.ts.map +1 -1
- package/dist/callbacks/manager.js +10 -63
- package/dist/callbacks/manager.js.map +1 -1
- package/dist/chat_history.cjs +0 -4
- package/dist/chat_history.cjs.map +1 -1
- package/dist/chat_history.d.cts +1 -3
- package/dist/chat_history.d.cts.map +1 -1
- package/dist/chat_history.d.ts +1 -3
- package/dist/chat_history.d.ts.map +1 -1
- package/dist/chat_history.js +0 -4
- package/dist/chat_history.js.map +1 -1
- package/dist/document_loaders/base.cjs +1 -13
- package/dist/document_loaders/base.cjs.map +1 -1
- package/dist/document_loaders/base.d.cts +0 -9
- package/dist/document_loaders/base.d.cts.map +1 -1
- package/dist/document_loaders/base.d.ts +0 -9
- package/dist/document_loaders/base.d.ts.map +1 -1
- package/dist/document_loaders/base.js +1 -13
- package/dist/document_loaders/base.js.map +1 -1
- package/dist/document_loaders/langsmith.d.cts +1 -1
- package/dist/language_models/base.cjs.map +1 -1
- package/dist/language_models/base.d.cts +0 -16
- package/dist/language_models/base.d.cts.map +1 -1
- package/dist/language_models/base.d.ts +0 -16
- package/dist/language_models/base.d.ts.map +1 -1
- package/dist/language_models/base.js.map +1 -1
- package/dist/language_models/chat_models.cjs +10 -83
- package/dist/language_models/chat_models.cjs.map +1 -1
- package/dist/language_models/chat_models.d.cts +21 -55
- package/dist/language_models/chat_models.d.cts.map +1 -1
- package/dist/language_models/chat_models.d.ts +21 -55
- package/dist/language_models/chat_models.d.ts.map +1 -1
- package/dist/language_models/chat_models.js +11 -83
- package/dist/language_models/chat_models.js.map +1 -1
- package/dist/language_models/llms.cjs +0 -56
- package/dist/language_models/llms.cjs.map +1 -1
- package/dist/language_models/llms.d.cts +1 -43
- package/dist/language_models/llms.d.cts.map +1 -1
- package/dist/language_models/llms.d.ts +1 -43
- package/dist/language_models/llms.d.ts.map +1 -1
- package/dist/language_models/llms.js +0 -56
- package/dist/language_models/llms.js.map +1 -1
- package/dist/load/import_map.cjs +1 -7
- package/dist/load/import_map.cjs.map +1 -1
- package/dist/load/import_map.js +1 -7
- package/dist/load/import_map.js.map +1 -1
- package/dist/messages/ai.cjs +4 -0
- package/dist/messages/ai.cjs.map +1 -1
- package/dist/messages/ai.js +4 -0
- package/dist/messages/ai.js.map +1 -1
- package/dist/messages/block_translators/anthropic.cjs +191 -144
- package/dist/messages/block_translators/anthropic.cjs.map +1 -1
- package/dist/messages/block_translators/anthropic.js +191 -144
- package/dist/messages/block_translators/anthropic.js.map +1 -1
- package/dist/messages/block_translators/index.cjs +2 -2
- package/dist/messages/block_translators/index.cjs.map +1 -1
- package/dist/messages/block_translators/index.js +4 -4
- package/dist/messages/block_translators/index.js.map +1 -1
- package/dist/messages/block_translators/openai.cjs +78 -22
- package/dist/messages/block_translators/openai.cjs.map +1 -1
- package/dist/messages/block_translators/openai.js +78 -22
- package/dist/messages/block_translators/openai.js.map +1 -1
- package/dist/messages/content/tools.cjs +1 -5
- package/dist/messages/content/tools.cjs.map +1 -1
- package/dist/messages/content/tools.d.cts +1 -85
- package/dist/messages/content/tools.d.cts.map +1 -1
- package/dist/messages/content/tools.d.ts +1 -85
- package/dist/messages/content/tools.d.ts.map +1 -1
- package/dist/messages/content/tools.js +1 -5
- package/dist/messages/content/tools.js.map +1 -1
- package/dist/messages/metadata.cjs.map +1 -1
- package/dist/messages/metadata.d.cts +3 -0
- package/dist/messages/metadata.d.cts.map +1 -1
- package/dist/messages/metadata.d.ts +3 -0
- package/dist/messages/metadata.d.ts.map +1 -1
- package/dist/messages/metadata.js.map +1 -1
- package/dist/messages/tool.cjs +2 -0
- package/dist/messages/tool.cjs.map +1 -1
- package/dist/messages/tool.d.cts +2 -0
- package/dist/messages/tool.d.cts.map +1 -1
- package/dist/messages/tool.d.ts +2 -0
- package/dist/messages/tool.d.ts.map +1 -1
- package/dist/messages/tool.js +2 -0
- package/dist/messages/tool.js.map +1 -1
- package/dist/output_parsers/json.cjs +5 -0
- package/dist/output_parsers/json.cjs.map +1 -1
- package/dist/output_parsers/json.d.cts +2 -0
- package/dist/output_parsers/json.d.cts.map +1 -1
- package/dist/output_parsers/json.d.ts +2 -0
- package/dist/output_parsers/json.d.ts.map +1 -1
- package/dist/output_parsers/json.js +5 -0
- package/dist/output_parsers/json.js.map +1 -1
- package/dist/prompts/base.cjs +0 -36
- package/dist/prompts/base.cjs.map +1 -1
- package/dist/prompts/base.d.cts +0 -16
- package/dist/prompts/base.d.cts.map +1 -1
- package/dist/prompts/base.d.ts +0 -16
- package/dist/prompts/base.d.ts.map +1 -1
- package/dist/prompts/base.js +0 -36
- package/dist/prompts/base.js.map +1 -1
- package/dist/prompts/chat.cjs +1 -5
- package/dist/prompts/chat.cjs.map +1 -1
- package/dist/prompts/chat.d.cts +1 -4
- package/dist/prompts/chat.d.cts.map +1 -1
- package/dist/prompts/chat.d.ts +1 -4
- package/dist/prompts/chat.d.ts.map +1 -1
- package/dist/prompts/chat.js +1 -5
- package/dist/prompts/chat.js.map +1 -1
- package/dist/prompts/dict.d.cts +1 -1
- package/dist/prompts/dict.d.ts +1 -1
- package/dist/prompts/few_shot.d.cts +2 -2
- package/dist/prompts/few_shot.d.ts +2 -2
- package/dist/prompts/image.d.cts +1 -1
- package/dist/prompts/image.d.ts +1 -1
- package/dist/prompts/index.d.cts +2 -2
- package/dist/prompts/index.d.ts +2 -2
- package/dist/prompts/pipeline.d.cts +1 -1
- package/dist/prompts/pipeline.d.ts +1 -1
- package/dist/prompts/prompt.d.cts +2 -2
- package/dist/prompts/prompt.d.ts +2 -2
- package/dist/retrievers/index.cjs +3 -18
- package/dist/retrievers/index.cjs.map +1 -1
- package/dist/retrievers/index.d.cts +2 -27
- package/dist/retrievers/index.d.cts.map +1 -1
- package/dist/retrievers/index.d.ts +2 -27
- package/dist/retrievers/index.d.ts.map +1 -1
- package/dist/retrievers/index.js +3 -18
- package/dist/retrievers/index.js.map +1 -1
- package/dist/runnables/base.cjs +24 -63
- package/dist/runnables/base.cjs.map +1 -1
- package/dist/runnables/base.d.cts +10 -39
- package/dist/runnables/base.d.cts.map +1 -1
- package/dist/runnables/base.d.ts +10 -39
- package/dist/runnables/base.d.ts.map +1 -1
- package/dist/runnables/base.js +24 -63
- package/dist/runnables/base.js.map +1 -1
- package/dist/runnables/graph.cjs +1 -1
- package/dist/runnables/graph.cjs.map +1 -1
- package/dist/runnables/graph.js +2 -2
- package/dist/runnables/graph.js.map +1 -1
- package/dist/runnables/graph_mermaid.cjs +1 -10
- package/dist/runnables/graph_mermaid.cjs.map +1 -1
- package/dist/runnables/graph_mermaid.js +1 -10
- package/dist/runnables/graph_mermaid.js.map +1 -1
- package/dist/runnables/history.cjs +1 -1
- package/dist/runnables/history.cjs.map +1 -1
- package/dist/runnables/history.d.cts +2 -2
- package/dist/runnables/history.d.cts.map +1 -1
- package/dist/runnables/history.d.ts +2 -2
- package/dist/runnables/history.d.ts.map +1 -1
- package/dist/runnables/history.js +1 -1
- package/dist/runnables/history.js.map +1 -1
- package/dist/stores.cjs.map +1 -1
- package/dist/stores.d.cts +3 -29
- package/dist/stores.d.cts.map +1 -1
- package/dist/stores.d.ts +3 -29
- package/dist/stores.d.ts.map +1 -1
- package/dist/stores.js.map +1 -1
- package/dist/tools/index.cjs +12 -4
- package/dist/tools/index.cjs.map +1 -1
- package/dist/tools/index.js +12 -4
- package/dist/tools/index.js.map +1 -1
- package/dist/tools/types.cjs.map +1 -1
- package/dist/tools/types.d.cts +4 -0
- package/dist/tools/types.d.cts.map +1 -1
- package/dist/tools/types.d.ts +4 -0
- package/dist/tools/types.d.ts.map +1 -1
- package/dist/tools/types.js.map +1 -1
- package/dist/tracers/base.cjs +1 -1
- package/dist/tracers/base.cjs.map +1 -1
- package/dist/tracers/base.d.cts +1 -1
- package/dist/tracers/base.js +2 -2
- package/dist/tracers/base.js.map +1 -1
- package/dist/tracers/log_stream.d.cts +1 -1
- package/dist/tracers/log_stream.d.ts +1 -1
- package/dist/tracers/tracer_langchain.cjs +1 -0
- package/dist/tracers/tracer_langchain.cjs.map +1 -1
- package/dist/tracers/tracer_langchain.d.cts +2 -2
- package/dist/tracers/tracer_langchain.js +1 -0
- package/dist/tracers/tracer_langchain.js.map +1 -1
- package/dist/utils/env.cjs +1 -9
- package/dist/utils/env.cjs.map +1 -1
- package/dist/utils/env.d.cts +2 -6
- package/dist/utils/env.d.cts.map +1 -1
- package/dist/utils/env.d.ts +2 -6
- package/dist/utils/env.d.ts.map +1 -1
- package/dist/utils/env.js +2 -9
- package/dist/utils/env.js.map +1 -1
- package/dist/utils/testing/message_history.cjs +1 -1
- package/dist/utils/testing/message_history.cjs.map +1 -1
- package/dist/utils/testing/message_history.d.cts +1 -1
- package/dist/utils/testing/message_history.d.cts.map +1 -1
- package/dist/utils/testing/message_history.d.ts +1 -1
- package/dist/utils/testing/message_history.d.ts.map +1 -1
- package/dist/utils/testing/message_history.js +1 -1
- package/dist/utils/testing/message_history.js.map +1 -1
- package/dist/utils/types/index.cjs +6 -0
- package/dist/utils/types/index.d.cts +2 -2
- package/dist/utils/types/index.d.ts +2 -2
- package/dist/utils/types/index.js +5 -2
- package/dist/utils/types/zod.cjs +23 -0
- package/dist/utils/types/zod.cjs.map +1 -1
- package/dist/utils/types/zod.d.cts +11 -1
- package/dist/utils/types/zod.d.cts.map +1 -1
- package/dist/utils/types/zod.d.ts +11 -1
- package/dist/utils/types/zod.d.ts.map +1 -1
- package/dist/utils/types/zod.js +21 -1
- package/dist/utils/types/zod.js.map +1 -1
- package/package.json +121 -154
- package/dist/runnables/remote.cjs +0 -399
- package/dist/runnables/remote.cjs.map +0 -1
- package/dist/runnables/remote.d.cts +0 -73
- package/dist/runnables/remote.d.cts.map +0 -1
- package/dist/runnables/remote.d.ts +0 -73
- package/dist/runnables/remote.d.ts.map +0 -1
- package/dist/runnables/remote.js +0 -393
- package/dist/runnables/remote.js.map +0 -1
- package/dist/tracers/initialize.cjs +0 -46
- package/dist/tracers/initialize.cjs.map +0 -1
- package/dist/tracers/initialize.d.cts +0 -26
- package/dist/tracers/initialize.d.cts.map +0 -1
- package/dist/tracers/initialize.d.ts +0 -26
- package/dist/tracers/initialize.d.ts.map +0 -1
- package/dist/tracers/initialize.js +0 -39
- package/dist/tracers/initialize.js.map +0 -1
- package/dist/tracers/tracer_langchain_v1.cjs +0 -168
- package/dist/tracers/tracer_langchain_v1.cjs.map +0 -1
- package/dist/tracers/tracer_langchain_v1.d.cts +0 -64
- package/dist/tracers/tracer_langchain_v1.d.cts.map +0 -1
- package/dist/tracers/tracer_langchain_v1.d.ts +0 -64
- package/dist/tracers/tracer_langchain_v1.d.ts.map +0 -1
- package/dist/tracers/tracer_langchain_v1.js +0 -162
- package/dist/tracers/tracer_langchain_v1.js.map +0 -1
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { BaseMessage, BaseMessageChunk, BaseMessageLike } from "../messages/base.cjs";
|
|
2
2
|
import { AIMessageChunk } from "../messages/ai.cjs";
|
|
3
|
-
import {
|
|
3
|
+
import { MessageOutputVersion } from "../messages/message.cjs";
|
|
4
4
|
import { ChatGenerationChunk, ChatResult, Generation, LLMResult } from "../outputs.cjs";
|
|
5
5
|
import { BaseCache } from "../caches/base.cjs";
|
|
6
6
|
import { CallbackManagerForLLMRun, Callbacks } from "../callbacks/manager.cjs";
|
|
@@ -59,7 +59,7 @@ type BaseChatModelParams = BaseLanguageModelParams & {
|
|
|
59
59
|
*
|
|
60
60
|
* @default "v0"
|
|
61
61
|
*/
|
|
62
|
-
outputVersion?:
|
|
62
|
+
outputVersion?: MessageOutputVersion;
|
|
63
63
|
};
|
|
64
64
|
/**
|
|
65
65
|
* Represents the call options for a base chat model.
|
|
@@ -80,13 +80,23 @@ type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {
|
|
|
80
80
|
* if used with an unsupported model.
|
|
81
81
|
*/
|
|
82
82
|
tool_choice?: ToolChoice;
|
|
83
|
+
/**
|
|
84
|
+
* Version of `AIMessage` output format to store in message content.
|
|
85
|
+
*
|
|
86
|
+
* `AIMessage.contentBlocks` will lazily parse the contents of `content` into a
|
|
87
|
+
* standard format. This flag can be used to additionally store the standard format
|
|
88
|
+
* as the message content, e.g., for serialization purposes.
|
|
89
|
+
*
|
|
90
|
+
* - "v0": provider-specific format in content (can lazily parse with `.contentBlocks`)
|
|
91
|
+
* - "v1": standardized format in content (consistent with `.contentBlocks`)
|
|
92
|
+
*
|
|
93
|
+
* You can also set `LC_OUTPUT_VERSION` as an environment variable to "v1" to
|
|
94
|
+
* enable this by default.
|
|
95
|
+
*
|
|
96
|
+
* @default "v0"
|
|
97
|
+
*/
|
|
98
|
+
outputVersion?: MessageOutputVersion;
|
|
83
99
|
};
|
|
84
|
-
/**
|
|
85
|
-
* Creates a transform stream for encoding chat message chunks.
|
|
86
|
-
* @deprecated Use {@link BytesOutputParser} instead
|
|
87
|
-
* @returns A TransformStream instance that encodes chat message chunks.
|
|
88
|
-
*/
|
|
89
|
-
declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<MessageStructure, MessageType>, any>;
|
|
90
100
|
type LangSmithParams = {
|
|
91
101
|
ls_provider?: string;
|
|
92
102
|
ls_model_name?: string;
|
|
@@ -110,7 +120,8 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
110
120
|
// Only ever instantiated in main LangChain
|
|
111
121
|
lc_namespace: string[];
|
|
112
122
|
disableStreaming: boolean;
|
|
113
|
-
outputVersion?:
|
|
123
|
+
outputVersion?: MessageOutputVersion;
|
|
124
|
+
get callKeys(): string[];
|
|
114
125
|
constructor(fields: BaseChatModelParams);
|
|
115
126
|
_combineLLMOutput?(...llmOutputs: LLMResult["llmOutput"][]): LLMResult["llmOutput"];
|
|
116
127
|
protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this["ParsedCallOptions"]];
|
|
@@ -168,11 +179,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
168
179
|
invocationParams(_options?: this["ParsedCallOptions"]): any;
|
|
169
180
|
_modelType(): string;
|
|
170
181
|
abstract _llmType(): string;
|
|
171
|
-
/**
|
|
172
|
-
* @deprecated
|
|
173
|
-
* Return a json-like object representing this LLM.
|
|
174
|
-
*/
|
|
175
|
-
serialize(): SerializedLLM;
|
|
176
182
|
/**
|
|
177
183
|
* Generates a prompt based on the input prompt values.
|
|
178
184
|
* @param promptValues An array of BasePromptValue instances.
|
|
@@ -182,46 +188,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
182
188
|
*/
|
|
183
189
|
generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;
|
|
184
190
|
abstract _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
|
|
185
|
-
/**
|
|
186
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
187
|
-
*
|
|
188
|
-
* Makes a single call to the chat model.
|
|
189
|
-
* @param messages An array of BaseMessage instances.
|
|
190
|
-
* @param options The call options or an array of stop sequences.
|
|
191
|
-
* @param callbacks The callbacks for the language model.
|
|
192
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
193
|
-
*/
|
|
194
|
-
call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
195
|
-
/**
|
|
196
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
197
|
-
*
|
|
198
|
-
* Makes a single call to the chat model with a prompt value.
|
|
199
|
-
* @param promptValue The value of the prompt.
|
|
200
|
-
* @param options The call options or an array of stop sequences.
|
|
201
|
-
* @param callbacks The callbacks for the language model.
|
|
202
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
203
|
-
*/
|
|
204
|
-
callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
205
|
-
/**
|
|
206
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
207
|
-
*
|
|
208
|
-
* Predicts the next message based on the input messages.
|
|
209
|
-
* @param messages An array of BaseMessage instances.
|
|
210
|
-
* @param options The call options or an array of stop sequences.
|
|
211
|
-
* @param callbacks The callbacks for the language model.
|
|
212
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
213
|
-
*/
|
|
214
|
-
predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
215
|
-
/**
|
|
216
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
217
|
-
*
|
|
218
|
-
* Predicts the next message based on a text input.
|
|
219
|
-
* @param text The text input.
|
|
220
|
-
* @param options The call options or an array of stop sequences.
|
|
221
|
-
* @param callbacks The callbacks for the language model.
|
|
222
|
-
* @returns A Promise that resolves to a string.
|
|
223
|
-
*/
|
|
224
|
-
predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;
|
|
225
191
|
withStructuredOutput<
|
|
226
192
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
227
193
|
RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: $ZodType<RunOutput>
|
|
@@ -258,5 +224,5 @@ declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOpti
|
|
|
258
224
|
_generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
|
|
259
225
|
}
|
|
260
226
|
//#endregion
|
|
261
|
-
export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice
|
|
227
|
+
export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice };
|
|
262
228
|
//# sourceMappingURL=chat_models.d.cts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat_models.d.cts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","createChatMessageChunkEncoderStream","___messages_message_js0","MessageStructure","MessageType","TransformStream","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: \"v0\" | \"v1\";\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n};\n/**\n * Creates a transform stream for encoding chat message chunks.\n * @deprecated Use {@link BytesOutputParser} instead\n * @returns A TransformStream instance that encodes chat message chunks.\n */\nexport declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<import(\"../messages/message.js\").MessageStructure, import(\"../messages/message.js\").MessageType>, any>;\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: \"v0\" | \"v1\";\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * @deprecated\n * Return a json-like object representing this LLM.\n */\n serialize(): SerializedLLM;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model with a prompt value.\n * @param promptValue The value of the prompt.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on the input messages.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on a text input.\n * @param text The text input.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a string.\n */\n predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY2B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;AA+BZ,CAAA,GA3CIH,MA2CQI,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;;;;AAegB;AAOJC,KA5DZH,aAAAA,GA4DYG;EAAmC,MAAA,EAAA,MAAA;EAAA,KAAA,EAAA,MAAA;CAAsF,GAzD7IL,MAyD6I,CAAA,MAAA,EAAA,GAAA,CAAA;;;AAAnE;AAClEU,KAtDAP,mBAAAA,GAAsBb,uBA4Df,GAAA;EAEPsB;;;;;;;AAEqE;AAKjF;EAA2C,gBAAA,CAAA,EAAA,OAAA;EAAA;;;;;;;;;;;;;;;EAWiC,aAAiBnB,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;CAAc;;;;AASFqB,KA1D7FV,wBAAAA,GAA2BhB,4BA0DkE0B,GAAAA;EAAiB;;;;;;;;;;;;;;EAUjC,WACpCJ,CAAAA,EAtDnCX,UAsDmCW;CAAe;;;;;;AAGpCY,iBAlDRjB,mCAAAA,CAAAA,CAkDQiB,EAlD+Bb,eAkD/Ba,CAlD+C7C,gBAkD/C6C,CAlD2B,gBAAA,EAAsF,WAAA,CAkDjHA,EAAAA,GAAAA,CAAAA;AAAOC,KAjD3Bb,eAAAA,GAiD2Ba;EAAY,WAAEC,CAAAA,EAAAA,MAAAA;EAAa,aAAEC,CAAAA,EAAAA,MAAAA;EAAc,aAChE/C,EAAAA,MAAAA;EAAe,cACRM,CAAAA,EAAAA,MAAAA;EAAU,aAApBU,CAAAA,EAAAA,MAAAA;EAAS,OAIAD,CAAAA,EAjDVkB,KAiDUlB,CAAAA,MAAAA,CAAAA;CAAc;AAGTF,KAlDjBqB,cAAAA,GAAiBjB;;EAE1BK,MAyDoBtB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GAzDES,cAyDFT,GAzDmBoB,gBAyDnBpB,GAzDsCkB,oBAyDtClB;;;;;AAYNwB,uBAhEaW,aAgEbX,CAAAA,oBAhE+CE,wBAgE/CF,GAhE0EE,wBAgE1EF;;0BA9DSzB,gBAsEwDsC,GAtErCpC,cAsEqCoC,CAAAA,SAtEb9B,iBAsEa8B,CAtEKD,iBAsELC,EAtEwBA,WAsExBA,CAAAA,CAAAA;EAAW;EAAuB,iBAAWlC,EApExGoC,IAoEwGpC,CApEnGkC,WAoEmGlC,EApEtFmC,OAoEsFnC,CAAAA,MApExEY,cAoEwEZ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAS;EAAV,YAC7FL,EAAAA,MAAAA,EAAAA;EAAW,gBAAqDe,EAAAA,OAAAA;EAAwB,aAAWR,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;EAAU,WAAlBoC,CAAAA,MAAAA,EAhEpGhB,mBAgEoGgB;EAAO,iBAUhHzC,CAAAA,CAAAA,GAAAA,UAAAA,EAzEmBG,SAyEnBH,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EAzE8CG,SAyE9CH,CAAAA,WAAAA,CAAAA;EAAe,UAAyBqC,4CAAAA,CAAAA,OAAAA,CAAAA,EAxEUG,OAwEVH,CAxEkBA,WAwElBA,CAAAA,CAAAA,EAAAA,CAxEkCtB,cAwElCsB,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAAW;;;;;;;;EAUiD,SAUzFvC,CAAAA,CAAAA,KAAAA,EAnFRoC,cAmFQpC,EAAAA,EAAAA,MAAAA,CAAAA,EAnFmB0C,OAmFnB1C,CAnF2BuC,WAmF3BvC,CAAAA,CAAAA,EAnF0CqB,QAmF1CrB,CAnFmDa,sBAmFnDb,EAnF2EsC,iBAmF3EtC,EAnF8FuC,WAmF9FvC,CAAAA;EAAW;;;;;;EAUwC,MAAG2C,CAAAA,KAAAA,EAtFlE9B,sBAsFkE8B,EAAAA,OAAAA,CAAAA,EAtFhCJ,WAsFgCI,CAAAA,EAtFlBA,OAsFkBA,CAtFVL,iBAsFUK,CAAAA;EAAO;EAG/D,qBAAgBnB,CAAAA,SAAAA,EAvFPxB,WAuFOwB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EAvF2DT,wBAuF3DS,CAAAA,EAvFsFoB,cAuFtFpB,CAvFqGlB,mBAuFrGkB,CAAAA;EAAM,eAAuC0B,CAAAA,KAAAA,EAtF9DrC,sBAsF8DqC,EAAAA,OAAAA,CAAAA,EAtF5BX,WAsF4BW,CAAAA,EAtFdN,cAsFcM,CAtFCZ,iBAsFDY,CAAAA;EAAS,WAAnBnD,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EArF1BmC,eAqF0BnC;EAAS;EAE3E,iBAAwBW,CAAAA,QAAAA,EArFLR,eAqFKQ,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EArF0EO,cAqF1EP,EAAAA,kBAAAA,CAAAA,EArF+GK,wBAqF/GL,EAAAA,CAAAA,EArF4IiC,OAqF5IjC,CArFoJL,SAqFpJK,CAAAA;EAA6B,eAAmBG,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAGG,CAHHA,EAAAA;IAAwBqC,QAAAA,EAnF3FhD,eAmF2FgD,EAAAA,EAAAA;IAAjC7B,KAAAA,EAlF7DH,SAkF6DG,CAlFnDb,UAkFmDa,EAAAA,CAAAA;IAGtDG,YAAAA,EAAAA,MAAAA;IAAsBA;IAA6C0B,aAAAA,EAAAA,GAAAA;IAAVnD,cAAAA,EAjFvDkB,cAiFuDlB;EAAS,CAAA,CAAA,EAhFhF4C,OAkFDnB,CAlFSnB,SAkFTmB,GAAAA;IAA8Bd,oBAAAA,EAAAA,MAAAA,EAAAA;IAA+CG,kBAAAA,CAAAA,EAhFvDE,wBAgFuDF,EAAAA;EAAsB,CAAA,CAAA;EAClF;;;;;;;EAOX,QAAwBH,CAAAA,QAAAA,EA/EdR,eA+EcQ,EAAAA,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GA/E4B6B,WA+E5B7B,EAAAA,SAAAA,CAAAA,EA/EqDM,SA+ErDN,CAAAA,EA/EiEiC,OA+EjEjC,CA/EyEL,SA+EzEK,CAAAA;EAA6B;;;EAAkB;EAGxD,gBAAgBc,CAAAA,QAAAA,CAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EAAAA,GAAAA;EAAM,UAAuC0B,CAAAA,CAAAA,EAAAA,MAAAA;EAAS,SAAnBrD,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA;EAAS;;;;EAGhE,SACRqD,CAAAA,CAAAA,EA1ECxB,aA0EDwB;EAAS;;AAxI6D;AA+ItF;;;;EAAqH,cAAwBX,CAAAA,YAAAA,EAzE5GnC,wBAyE4GmC,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GAzE3DA,WAyE2DA,EAAAA,SAAAA,CAAAA,EAzElCvB,SAyEkCuB,CAAAA,EAzEtBI,OAyEsBJ,CAzEdlC,SAyEckC,CAAAA;EAAW,SAC3HvC,SAAAA,CAAAA,QAAAA,EAzEIA,WAyEJA,EAAAA,EAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,UAAAA,CAAAA,EAzEoEe,wBAyEpEf,CAAAA,EAzE+F2C,OAyE/F3C,CAzEuGO,UAyEvGP,CAAAA;EAAW;;;;;;;AADoG;;iBA9DzHE,wCAAwCqC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;0BAU5EI,+CAA+CmC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;4BAU1FA,oCAAoCuC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;6CAUhEuC,yBAAyBvB,YAAY2B;;;oBAG9DnB,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;oBAIM1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;;;;uBAOcC,oCAAoCvB,2BAA2BA,kCAAkCS,cAAcE;2BAChHvC,gEAAgEe,2BAA2B4B;sBAChG3C,gEAAgEe,2BAA2B4B,QAAQpC"}
|
|
1
|
+
{"version":3,"file":"chat_models.d.cts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","MessageOutputVersion","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk, MessageOutputVersion } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: MessageOutputVersion;\n get callKeys(): string[];\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY4B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;CAAmB,GAZ3BH,MAY2B,CAAA,MAAA,EAAA,GAAA,CAAA;;;AA0BS;AAKxC;AAAoC,KAtCxBE,aAAAA,GAsCwB;EAAA,MAAGd,EAAAA,MAAAA;EAA4B,KAejDW,EAAAA,MAAAA;CAAU,GAlDxBC,MAkEgBrB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;AAAoB;AAExC;AAQA;AAA0B,KAxEdwB,mBAAAA,GAAsBb,uBAwER,GAAA;EAAA;;;;;AAEuD;AAKjF;;;EAAwF,gBAAGc,CAAAA,EAAAA,OAAAA;EAAwB;;;;;;;;;;;;;;;EAYR,aASrFG,CAAAA,EA1EF5B,oBA0EE4B;CAAc;;;;AAAwFG,KArEhHN,wBAAAA,GAA2BhB,4BAqEqFsB,GAAAA;EAAW;;;;;;;;;;;;;;EAWnE,WAEpCjC,CAAAA,EAnEdsB,UAmEctB;EAAe;;;;;;;;;;;;;;;EAUM,aAF7CqC,CAAAA,EA3DYnC,oBA2DZmC;CAAO;AAWkDJ,KApErDL,eAAAA,GAoEqDK;EAAW,WAAclB,CAAAA,EAAAA,MAAAA;EAAS,aAAWX,CAAAA,EAAAA,MAAAA;EAAS,aAAjBiC,EAAAA,MAAAA;EAAO,cAe5ElC,CAAAA,EAAAA,MAAAA;EAAwB,aAAyB8B,CAAAA,EAAAA,MAAAA;EAAW,OAAclB,CAAAA,EA7E7Fc,KA6E6Fd,CAAAA,MAAAA,CAAAA;CAAS;AAAGsB,KA3E3GP,cAAAA,GAAiBZ;;EAE1BK,MA0E8FT,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GA1ExEJ,cA0EwEI,GA1EvDO,gBA0EuDP,GA1EpCK,oBA0EoCL;;;;;AAGR8B,uBAxE3Db,aAwE2Da,CAAAA,oBAxEzBjB,wBAwEyBiB,GAxEEjB,wBAwEFiB;;0BAtE/D7C,gBAwEnBwB,GAxEsCtB,cAwEtCsB,CAAAA,SAxE8Df,iBAwE9De,CAxEgFS,iBAwEhFT,EAxEmGU,WAwEnGV,CAAAA,CAAAA;EAAM;EAAqD,iBAAmBX,EAtE9DuB,IAsE8DvB,CAtEzDqB,WAsEyDrB,EAtE5CsB,OAsE4CtB,CAAAA,MAtE9BI,cAsE8BJ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAsB;EAAW,YAA1CQ,EAAAA,MAAAA,EAAAA;EAAQ,gBAG9DG,EAAAA,OAAAA;EAAM,aAAgBA,CAAAA,EArExBrB,oBAqEwBqB;EAAM,IAAuCqB,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA,EAAAA;EAAS,WAAnB/C,CAAAA,MAAAA,EAnEvD6B,mBAmEuD7B;EAAS,iBAEjF0B,CAAAA,CAAAA,GAAAA,UAAAA,EApE+BnB,SAoE/BmB,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EApE0DnB,SAoE1DmB,CAAAA,WAAAA,CAAAA;EAAM,UAAwBd,4CAAAA,CAAAA,OAAAA,CAAAA,EAnEgC2B,OAmEhC3B,CAnEwCwB,WAmExCxB,CAAAA,CAAAA,EAAAA,CAnEwDO,cAmExDP,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAA6B;;;;;;;;EAMsB,SAEjFc,CAAAA,CAAAA,KAAAA,EAlEeO,cAkEfP,EAAAA,EAAAA,MAAAA,CAAAA,EAlE0Ca,OAkE1Cb,CAlEkDU,WAkElDV,CAAAA,CAAAA,EAlEiEH,QAkEjEG,CAlE0EX,sBAkE1EW,EAlEkGS,iBAkElGT,EAlEqHU,WAkErHV,CAAAA;EAAM;;;;;;EAGqC,MAAuCqB,CAAAA,KAAAA,EA9DvEhC,sBA8DuEgC,EAAAA,OAAAA,CAAAA,EA9DrCX,WA8DqCW,CAAAA,EA9DvBP,OA8DuBO,CA9DfZ,iBA8DeY,CAAAA;EAAS;EAAV,qBAEjFrB,CAAAA,SAAAA,EA9D8BzB,WA8D9ByB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EA9DgGT,wBA8DhGS,CAAAA,EA9D2He,cA8D3Hf,CA9D0IlB,mBA8D1IkB,CAAAA;EAAM,eAAwBd,CAAAA,KAAAA,EA7DVG,sBA6DUH,EAAAA,OAAAA,CAAAA,EA7DwBwB,WA6DxBxB,CAAAA,EA7DsC6B,cA6DtC7B,CA7DqDuB,iBA6DrDvB,CAAAA;EAA6B,WAAkBG,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EA5D/BgB,eA4D+BhB;EAAsB;EAClF,iBACRgC,CAAAA,QAAAA,EA5DgB5C,eA4DhB4C,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EA5D+F5B,cA4D/F4B,EAAAA,kBAAAA,CAAAA,EA5DoI9B,wBA4DpI8B,EAAAA,CAAAA,EA5DiKP,OA4DjKO,CA5DyKxC,SA4DzKwC,CAAAA;EAAS,eAFkDxB,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAS0C,CAT1CA,EAAAA;IA1FNZ,QAAAA,EAkCnDR,eAlCmDQ,EAAAA,EAAAA;IAAiB,KAAA,EAmCvES,SAnCuE,CAmC7DV,UAnC6D,EAAA,CAAA;IAmGxDsC,YAAAA,EAAe,MAAA;IAAA;IAAqBlB,aAAAA,EAAAA,GAAAA;IAA2BA,cAAAA,EA5DrEX,cA4DqEW;EAAwB,CAAA,CAAA,EA3D7GU,OA2DqIJ,CA3D7H7B,SA2D6H6B,GAAAA;IAChHnC,oBAAAA,EAAAA,MAAAA,EAAAA;IAAgEgB,kBAAAA,CAAAA,EA1DhEA,wBA0DgEA,EAAAA;EAAwB,CAAA,CAAA;EAAU;;;;;AADa;;qBAhDrHd,0CAA0CiC,yBAAyBlB,YAAYsB,QAAQjC;;;;;;;;;;;;;;;+BAe7ED,iDAAiD8B,yBAAyBlB,YAAYsB,QAAQjC;+BAC9FN,gEAAgEgB,2BAA2BuB,QAAQ/B;;;oBAG9GiB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;oBAIMrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;;;;uBAOcC,oCAAoClB,2BAA2BA,kCAAkCI,cAAcE;2BAChHnC,gEAAgEgB,2BAA2BuB;sBAChGvC,gEAAgEgB,2BAA2BuB,QAAQ/B"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { BaseMessage, BaseMessageChunk, BaseMessageLike } from "../messages/base.js";
|
|
2
2
|
import { AIMessageChunk } from "../messages/ai.js";
|
|
3
|
-
import {
|
|
3
|
+
import { MessageOutputVersion } from "../messages/message.js";
|
|
4
4
|
import { ChatGenerationChunk, ChatResult, Generation, LLMResult } from "../outputs.js";
|
|
5
5
|
import { BaseCache } from "../caches/base.js";
|
|
6
6
|
import { CallbackManagerForLLMRun, Callbacks } from "../callbacks/manager.js";
|
|
@@ -59,7 +59,7 @@ type BaseChatModelParams = BaseLanguageModelParams & {
|
|
|
59
59
|
*
|
|
60
60
|
* @default "v0"
|
|
61
61
|
*/
|
|
62
|
-
outputVersion?:
|
|
62
|
+
outputVersion?: MessageOutputVersion;
|
|
63
63
|
};
|
|
64
64
|
/**
|
|
65
65
|
* Represents the call options for a base chat model.
|
|
@@ -80,13 +80,23 @@ type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {
|
|
|
80
80
|
* if used with an unsupported model.
|
|
81
81
|
*/
|
|
82
82
|
tool_choice?: ToolChoice;
|
|
83
|
+
/**
|
|
84
|
+
* Version of `AIMessage` output format to store in message content.
|
|
85
|
+
*
|
|
86
|
+
* `AIMessage.contentBlocks` will lazily parse the contents of `content` into a
|
|
87
|
+
* standard format. This flag can be used to additionally store the standard format
|
|
88
|
+
* as the message content, e.g., for serialization purposes.
|
|
89
|
+
*
|
|
90
|
+
* - "v0": provider-specific format in content (can lazily parse with `.contentBlocks`)
|
|
91
|
+
* - "v1": standardized format in content (consistent with `.contentBlocks`)
|
|
92
|
+
*
|
|
93
|
+
* You can also set `LC_OUTPUT_VERSION` as an environment variable to "v1" to
|
|
94
|
+
* enable this by default.
|
|
95
|
+
*
|
|
96
|
+
* @default "v0"
|
|
97
|
+
*/
|
|
98
|
+
outputVersion?: MessageOutputVersion;
|
|
83
99
|
};
|
|
84
|
-
/**
|
|
85
|
-
* Creates a transform stream for encoding chat message chunks.
|
|
86
|
-
* @deprecated Use {@link BytesOutputParser} instead
|
|
87
|
-
* @returns A TransformStream instance that encodes chat message chunks.
|
|
88
|
-
*/
|
|
89
|
-
declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<MessageStructure, MessageType>, any>;
|
|
90
100
|
type LangSmithParams = {
|
|
91
101
|
ls_provider?: string;
|
|
92
102
|
ls_model_name?: string;
|
|
@@ -110,7 +120,8 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
110
120
|
// Only ever instantiated in main LangChain
|
|
111
121
|
lc_namespace: string[];
|
|
112
122
|
disableStreaming: boolean;
|
|
113
|
-
outputVersion?:
|
|
123
|
+
outputVersion?: MessageOutputVersion;
|
|
124
|
+
get callKeys(): string[];
|
|
114
125
|
constructor(fields: BaseChatModelParams);
|
|
115
126
|
_combineLLMOutput?(...llmOutputs: LLMResult["llmOutput"][]): LLMResult["llmOutput"];
|
|
116
127
|
protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this["ParsedCallOptions"]];
|
|
@@ -168,11 +179,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
168
179
|
invocationParams(_options?: this["ParsedCallOptions"]): any;
|
|
169
180
|
_modelType(): string;
|
|
170
181
|
abstract _llmType(): string;
|
|
171
|
-
/**
|
|
172
|
-
* @deprecated
|
|
173
|
-
* Return a json-like object representing this LLM.
|
|
174
|
-
*/
|
|
175
|
-
serialize(): SerializedLLM;
|
|
176
182
|
/**
|
|
177
183
|
* Generates a prompt based on the input prompt values.
|
|
178
184
|
* @param promptValues An array of BasePromptValue instances.
|
|
@@ -182,46 +188,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
|
|
|
182
188
|
*/
|
|
183
189
|
generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;
|
|
184
190
|
abstract _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
|
|
185
|
-
/**
|
|
186
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
187
|
-
*
|
|
188
|
-
* Makes a single call to the chat model.
|
|
189
|
-
* @param messages An array of BaseMessage instances.
|
|
190
|
-
* @param options The call options or an array of stop sequences.
|
|
191
|
-
* @param callbacks The callbacks for the language model.
|
|
192
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
193
|
-
*/
|
|
194
|
-
call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
195
|
-
/**
|
|
196
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
197
|
-
*
|
|
198
|
-
* Makes a single call to the chat model with a prompt value.
|
|
199
|
-
* @param promptValue The value of the prompt.
|
|
200
|
-
* @param options The call options or an array of stop sequences.
|
|
201
|
-
* @param callbacks The callbacks for the language model.
|
|
202
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
203
|
-
*/
|
|
204
|
-
callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
205
|
-
/**
|
|
206
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
207
|
-
*
|
|
208
|
-
* Predicts the next message based on the input messages.
|
|
209
|
-
* @param messages An array of BaseMessage instances.
|
|
210
|
-
* @param options The call options or an array of stop sequences.
|
|
211
|
-
* @param callbacks The callbacks for the language model.
|
|
212
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
213
|
-
*/
|
|
214
|
-
predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
|
|
215
|
-
/**
|
|
216
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
217
|
-
*
|
|
218
|
-
* Predicts the next message based on a text input.
|
|
219
|
-
* @param text The text input.
|
|
220
|
-
* @param options The call options or an array of stop sequences.
|
|
221
|
-
* @param callbacks The callbacks for the language model.
|
|
222
|
-
* @returns A Promise that resolves to a string.
|
|
223
|
-
*/
|
|
224
|
-
predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;
|
|
225
191
|
withStructuredOutput<
|
|
226
192
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
227
193
|
RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: $ZodType<RunOutput>
|
|
@@ -258,5 +224,5 @@ declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOpti
|
|
|
258
224
|
_generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
|
|
259
225
|
}
|
|
260
226
|
//#endregion
|
|
261
|
-
export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice
|
|
227
|
+
export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice };
|
|
262
228
|
//# sourceMappingURL=chat_models.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat_models.d.ts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","createChatMessageChunkEncoderStream","___messages_message_js0","MessageStructure","MessageType","TransformStream","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: \"v0\" | \"v1\";\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n};\n/**\n * Creates a transform stream for encoding chat message chunks.\n * @deprecated Use {@link BytesOutputParser} instead\n * @returns A TransformStream instance that encodes chat message chunks.\n */\nexport declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<import(\"../messages/message.js\").MessageStructure, import(\"../messages/message.js\").MessageType>, any>;\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: \"v0\" | \"v1\";\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * @deprecated\n * Return a json-like object representing this LLM.\n */\n serialize(): SerializedLLM;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model with a prompt value.\n * @param promptValue The value of the prompt.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on the input messages.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on a text input.\n * @param text The text input.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a string.\n */\n predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY2B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;AA+BZ,CAAA,GA3CIH,MA2CQI,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;;;;AAegB;AAOJC,KA5DZH,aAAAA,GA4DYG;EAAmC,MAAA,EAAA,MAAA;EAAA,KAAA,EAAA,MAAA;CAAsF,GAzD7IL,MAyD6I,CAAA,MAAA,EAAA,GAAA,CAAA;;;AAAnE;AAClEU,KAtDAP,mBAAAA,GAAsBb,uBA4Df,GAAA;EAEPsB;;;;;;;AAEqE;AAKjF;EAA2C,gBAAA,CAAA,EAAA,OAAA;EAAA;;;;;;;;;;;;;;;EAWiC,aAAiBnB,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;CAAc;;;;AASFqB,KA1D7FV,wBAAAA,GAA2BhB,4BA0DkE0B,GAAAA;EAAiB;;;;;;;;;;;;;;EAUjC,WACpCJ,CAAAA,EAtDnCX,UAsDmCW;CAAe;;;;;;AAGpCY,iBAlDRjB,mCAAAA,CAAAA,CAkDQiB,EAlD+Bb,eAkD/Ba,CAlD+C7C,gBAkD/C6C,CAlD2B,gBAAA,EAAsF,WAAA,CAkDjHA,EAAAA,GAAAA,CAAAA;AAAOC,KAjD3Bb,eAAAA,GAiD2Ba;EAAY,WAAEC,CAAAA,EAAAA,MAAAA;EAAa,aAAEC,CAAAA,EAAAA,MAAAA;EAAc,aAChE/C,EAAAA,MAAAA;EAAe,cACRM,CAAAA,EAAAA,MAAAA;EAAU,aAApBU,CAAAA,EAAAA,MAAAA;EAAS,OAIAD,CAAAA,EAjDVkB,KAiDUlB,CAAAA,MAAAA,CAAAA;CAAc;AAGTF,KAlDjBqB,cAAAA,GAAiBjB;;EAE1BK,MAyDoBtB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GAzDES,cAyDFT,GAzDmBoB,gBAyDnBpB,GAzDsCkB,oBAyDtClB;;;;;AAYNwB,uBAhEaW,aAgEbX,CAAAA,oBAhE+CE,wBAgE/CF,GAhE0EE,wBAgE1EF;;0BA9DSzB,gBAsEwDsC,GAtErCpC,cAsEqCoC,CAAAA,SAtEb9B,iBAsEa8B,CAtEKD,iBAsELC,EAtEwBA,WAsExBA,CAAAA,CAAAA;EAAW;EAAuB,iBAAWlC,EApExGoC,IAoEwGpC,CApEnGkC,WAoEmGlC,EApEtFmC,OAoEsFnC,CAAAA,MApExEY,cAoEwEZ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAS;EAAV,YAC7FL,EAAAA,MAAAA,EAAAA;EAAW,gBAAqDe,EAAAA,OAAAA;EAAwB,aAAWR,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;EAAU,WAAlBoC,CAAAA,MAAAA,EAhEpGhB,mBAgEoGgB;EAAO,iBAUhHzC,CAAAA,CAAAA,GAAAA,UAAAA,EAzEmBG,SAyEnBH,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EAzE8CG,SAyE9CH,CAAAA,WAAAA,CAAAA;EAAe,UAAyBqC,4CAAAA,CAAAA,OAAAA,CAAAA,EAxEUG,OAwEVH,CAxEkBA,WAwElBA,CAAAA,CAAAA,EAAAA,CAxEkCtB,cAwElCsB,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAAW;;;;;;;;EAUiD,SAUzFvC,CAAAA,CAAAA,KAAAA,EAnFRoC,cAmFQpC,EAAAA,EAAAA,MAAAA,CAAAA,EAnFmB0C,OAmFnB1C,CAnF2BuC,WAmF3BvC,CAAAA,CAAAA,EAnF0CqB,QAmF1CrB,CAnFmDa,sBAmFnDb,EAnF2EsC,iBAmF3EtC,EAnF8FuC,WAmF9FvC,CAAAA;EAAW;;;;;;EAUwC,MAAG2C,CAAAA,KAAAA,EAtFlE9B,sBAsFkE8B,EAAAA,OAAAA,CAAAA,EAtFhCJ,WAsFgCI,CAAAA,EAtFlBA,OAsFkBA,CAtFVL,iBAsFUK,CAAAA;EAAO;EAG/D,qBAAgBnB,CAAAA,SAAAA,EAvFPxB,WAuFOwB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EAvF2DT,wBAuF3DS,CAAAA,EAvFsFoB,cAuFtFpB,CAvFqGlB,mBAuFrGkB,CAAAA;EAAM,eAAuC0B,CAAAA,KAAAA,EAtF9DrC,sBAsF8DqC,EAAAA,OAAAA,CAAAA,EAtF5BX,WAsF4BW,CAAAA,EAtFdN,cAsFcM,CAtFCZ,iBAsFDY,CAAAA;EAAS,WAAnBnD,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EArF1BmC,eAqF0BnC;EAAS;EAE3E,iBAAwBW,CAAAA,QAAAA,EArFLR,eAqFKQ,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EArF0EO,cAqF1EP,EAAAA,kBAAAA,CAAAA,EArF+GK,wBAqF/GL,EAAAA,CAAAA,EArF4IiC,OAqF5IjC,CArFoJL,SAqFpJK,CAAAA;EAA6B,eAAmBG,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAGG,CAHHA,EAAAA;IAAwBqC,QAAAA,EAnF3FhD,eAmF2FgD,EAAAA,EAAAA;IAAjC7B,KAAAA,EAlF7DH,SAkF6DG,CAlFnDb,UAkFmDa,EAAAA,CAAAA;IAGtDG,YAAAA,EAAAA,MAAAA;IAAsBA;IAA6C0B,aAAAA,EAAAA,GAAAA;IAAVnD,cAAAA,EAjFvDkB,cAiFuDlB;EAAS,CAAA,CAAA,EAhFhF4C,OAkFDnB,CAlFSnB,SAkFTmB,GAAAA;IAA8Bd,oBAAAA,EAAAA,MAAAA,EAAAA;IAA+CG,kBAAAA,CAAAA,EAhFvDE,wBAgFuDF,EAAAA;EAAsB,CAAA,CAAA;EAClF;;;;;;;EAOX,QAAwBH,CAAAA,QAAAA,EA/EdR,eA+EcQ,EAAAA,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GA/E4B6B,WA+E5B7B,EAAAA,SAAAA,CAAAA,EA/EqDM,SA+ErDN,CAAAA,EA/EiEiC,OA+EjEjC,CA/EyEL,SA+EzEK,CAAAA;EAA6B;;;EAAkB;EAGxD,gBAAgBc,CAAAA,QAAAA,CAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EAAAA,GAAAA;EAAM,UAAuC0B,CAAAA,CAAAA,EAAAA,MAAAA;EAAS,SAAnBrD,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA;EAAS;;;;EAGhE,SACRqD,CAAAA,CAAAA,EA1ECxB,aA0EDwB;EAAS;;AAxI6D;AA+ItF;;;;EAAqH,cAAwBX,CAAAA,YAAAA,EAzE5GnC,wBAyE4GmC,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GAzE3DA,WAyE2DA,EAAAA,SAAAA,CAAAA,EAzElCvB,SAyEkCuB,CAAAA,EAzEtBI,OAyEsBJ,CAzEdlC,SAyEckC,CAAAA;EAAW,SAC3HvC,SAAAA,CAAAA,QAAAA,EAzEIA,WAyEJA,EAAAA,EAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,UAAAA,CAAAA,EAzEoEe,wBAyEpEf,CAAAA,EAzE+F2C,OAyE/F3C,CAzEuGO,UAyEvGP,CAAAA;EAAW;;;;;;;AADoG;;iBA9DzHE,wCAAwCqC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;0BAU5EI,+CAA+CmC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;4BAU1FA,oCAAoCuC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;6CAUhEuC,yBAAyBvB,YAAY2B;;;oBAG9DnB,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;oBAIM1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;;;;uBAOcC,oCAAoCvB,2BAA2BA,kCAAkCS,cAAcE;2BAChHvC,gEAAgEe,2BAA2B4B;sBAChG3C,gEAAgEe,2BAA2B4B,QAAQpC"}
|
|
1
|
+
{"version":3,"file":"chat_models.d.ts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","MessageOutputVersion","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk, MessageOutputVersion } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: MessageOutputVersion;\n get callKeys(): string[];\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY4B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;CAAmB,GAZ3BH,MAY2B,CAAA,MAAA,EAAA,GAAA,CAAA;;;AA0BS;AAKxC;AAAoC,KAtCxBE,aAAAA,GAsCwB;EAAA,MAAGd,EAAAA,MAAAA;EAA4B,KAejDW,EAAAA,MAAAA;CAAU,GAlDxBC,MAkEgBrB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;AAAoB;AAExC;AAQA;AAA0B,KAxEdwB,mBAAAA,GAAsBb,uBAwER,GAAA;EAAA;;;;;AAEuD;AAKjF;;;EAAwF,gBAAGc,CAAAA,EAAAA,OAAAA;EAAwB;;;;;;;;;;;;;;;EAYR,aASrFG,CAAAA,EA1EF5B,oBA0EE4B;CAAc;;;;AAAwFG,KArEhHN,wBAAAA,GAA2BhB,4BAqEqFsB,GAAAA;EAAW;;;;;;;;;;;;;;EAWnE,WAEpCjC,CAAAA,EAnEdsB,UAmEctB;EAAe;;;;;;;;;;;;;;;EAUM,aAF7CqC,CAAAA,EA3DYnC,oBA2DZmC;CAAO;AAWkDJ,KApErDL,eAAAA,GAoEqDK;EAAW,WAAclB,CAAAA,EAAAA,MAAAA;EAAS,aAAWX,CAAAA,EAAAA,MAAAA;EAAS,aAAjBiC,EAAAA,MAAAA;EAAO,cAe5ElC,CAAAA,EAAAA,MAAAA;EAAwB,aAAyB8B,CAAAA,EAAAA,MAAAA;EAAW,OAAclB,CAAAA,EA7E7Fc,KA6E6Fd,CAAAA,MAAAA,CAAAA;CAAS;AAAGsB,KA3E3GP,cAAAA,GAAiBZ;;EAE1BK,MA0E8FT,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GA1ExEJ,cA0EwEI,GA1EvDO,gBA0EuDP,GA1EpCK,oBA0EoCL;;;;;AAGR8B,uBAxE3Db,aAwE2Da,CAAAA,oBAxEzBjB,wBAwEyBiB,GAxEEjB,wBAwEFiB;;0BAtE/D7C,gBAwEnBwB,GAxEsCtB,cAwEtCsB,CAAAA,SAxE8Df,iBAwE9De,CAxEgFS,iBAwEhFT,EAxEmGU,WAwEnGV,CAAAA,CAAAA;EAAM;EAAqD,iBAAmBX,EAtE9DuB,IAsE8DvB,CAtEzDqB,WAsEyDrB,EAtE5CsB,OAsE4CtB,CAAAA,MAtE9BI,cAsE8BJ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAsB;EAAW,YAA1CQ,EAAAA,MAAAA,EAAAA;EAAQ,gBAG9DG,EAAAA,OAAAA;EAAM,aAAgBA,CAAAA,EArExBrB,oBAqEwBqB;EAAM,IAAuCqB,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA,EAAAA;EAAS,WAAnB/C,CAAAA,MAAAA,EAnEvD6B,mBAmEuD7B;EAAS,iBAEjF0B,CAAAA,CAAAA,GAAAA,UAAAA,EApE+BnB,SAoE/BmB,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EApE0DnB,SAoE1DmB,CAAAA,WAAAA,CAAAA;EAAM,UAAwBd,4CAAAA,CAAAA,OAAAA,CAAAA,EAnEgC2B,OAmEhC3B,CAnEwCwB,WAmExCxB,CAAAA,CAAAA,EAAAA,CAnEwDO,cAmExDP,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAA6B;;;;;;;;EAMsB,SAEjFc,CAAAA,CAAAA,KAAAA,EAlEeO,cAkEfP,EAAAA,EAAAA,MAAAA,CAAAA,EAlE0Ca,OAkE1Cb,CAlEkDU,WAkElDV,CAAAA,CAAAA,EAlEiEH,QAkEjEG,CAlE0EX,sBAkE1EW,EAlEkGS,iBAkElGT,EAlEqHU,WAkErHV,CAAAA;EAAM;;;;;;EAGqC,MAAuCqB,CAAAA,KAAAA,EA9DvEhC,sBA8DuEgC,EAAAA,OAAAA,CAAAA,EA9DrCX,WA8DqCW,CAAAA,EA9DvBP,OA8DuBO,CA9DfZ,iBA8DeY,CAAAA;EAAS;EAAV,qBAEjFrB,CAAAA,SAAAA,EA9D8BzB,WA8D9ByB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EA9DgGT,wBA8DhGS,CAAAA,EA9D2He,cA8D3Hf,CA9D0IlB,mBA8D1IkB,CAAAA;EAAM,eAAwBd,CAAAA,KAAAA,EA7DVG,sBA6DUH,EAAAA,OAAAA,CAAAA,EA7DwBwB,WA6DxBxB,CAAAA,EA7DsC6B,cA6DtC7B,CA7DqDuB,iBA6DrDvB,CAAAA;EAA6B,WAAkBG,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EA5D/BgB,eA4D+BhB;EAAsB;EAClF,iBACRgC,CAAAA,QAAAA,EA5DgB5C,eA4DhB4C,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EA5D+F5B,cA4D/F4B,EAAAA,kBAAAA,CAAAA,EA5DoI9B,wBA4DpI8B,EAAAA,CAAAA,EA5DiKP,OA4DjKO,CA5DyKxC,SA4DzKwC,CAAAA;EAAS,eAFkDxB,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAS0C,CAT1CA,EAAAA;IA1FNZ,QAAAA,EAkCnDR,eAlCmDQ,EAAAA,EAAAA;IAAiB,KAAA,EAmCvES,SAnCuE,CAmC7DV,UAnC6D,EAAA,CAAA;IAmGxDsC,YAAAA,EAAe,MAAA;IAAA;IAAqBlB,aAAAA,EAAAA,GAAAA;IAA2BA,cAAAA,EA5DrEX,cA4DqEW;EAAwB,CAAA,CAAA,EA3D7GU,OA2DqIJ,CA3D7H7B,SA2D6H6B,GAAAA;IAChHnC,oBAAAA,EAAAA,MAAAA,EAAAA;IAAgEgB,kBAAAA,CAAAA,EA1DhEA,wBA0DgEA,EAAAA;EAAwB,CAAA,CAAA;EAAU;;;;;AADa;;qBAhDrHd,0CAA0CiC,yBAAyBlB,YAAYsB,QAAQjC;;;;;;;;;;;;;;;+BAe7ED,iDAAiD8B,yBAAyBlB,YAAYsB,QAAQjC;+BAC9FN,gEAAgEgB,2BAA2BuB,QAAQ/B;;;oBAG9GiB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;oBAIMrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;;;;uBAOcC,oCAAoClB,2BAA2BA,kCAAkCI,cAAcE;2BAChHnC,gEAAgEgB,2BAA2BuB;sBAChGvC,gEAAgEgB,2BAA2BuB,QAAQ/B"}
|
|
@@ -2,7 +2,6 @@ import { __export } from "../_virtual/rolldown_runtime.js";
|
|
|
2
2
|
import { convertToOpenAIImageBlock, isBase64ContentBlock, isURLContentBlock } from "../messages/content/data.js";
|
|
3
3
|
import { isBaseMessage } from "../messages/base.js";
|
|
4
4
|
import { AIMessage, AIMessageChunk, isAIMessage, isAIMessageChunk } from "../messages/ai.js";
|
|
5
|
-
import { HumanMessage } from "../messages/human.js";
|
|
6
5
|
import { coerceMessageLikeToMessage } from "../messages/utils.js";
|
|
7
6
|
import { getEnvironmentVariable } from "../utils/env.js";
|
|
8
7
|
import { callbackHandlerPrefersStreaming } from "../callbacks/base.js";
|
|
@@ -21,20 +20,8 @@ import { castStandardMessageContent, iife } from "./utils.js";
|
|
|
21
20
|
var chat_models_exports = {};
|
|
22
21
|
__export(chat_models_exports, {
|
|
23
22
|
BaseChatModel: () => BaseChatModel,
|
|
24
|
-
SimpleChatModel: () => SimpleChatModel
|
|
25
|
-
createChatMessageChunkEncoderStream: () => createChatMessageChunkEncoderStream
|
|
23
|
+
SimpleChatModel: () => SimpleChatModel
|
|
26
24
|
});
|
|
27
|
-
/**
|
|
28
|
-
* Creates a transform stream for encoding chat message chunks.
|
|
29
|
-
* @deprecated Use {@link BytesOutputParser} instead
|
|
30
|
-
* @returns A TransformStream instance that encodes chat message chunks.
|
|
31
|
-
*/
|
|
32
|
-
function createChatMessageChunkEncoderStream() {
|
|
33
|
-
const textEncoder = new TextEncoder();
|
|
34
|
-
return new TransformStream({ transform(chunk, controller) {
|
|
35
|
-
controller.enqueue(textEncoder.encode(typeof chunk.content === "string" ? chunk.content : JSON.stringify(chunk.content)));
|
|
36
|
-
} });
|
|
37
|
-
}
|
|
38
25
|
function _formatForTracing(messages) {
|
|
39
26
|
const messagesToTrace = [];
|
|
40
27
|
for (const message of messages) {
|
|
@@ -68,6 +55,9 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
68
55
|
];
|
|
69
56
|
disableStreaming = false;
|
|
70
57
|
outputVersion;
|
|
58
|
+
get callKeys() {
|
|
59
|
+
return [...super.callKeys, "outputVersion"];
|
|
60
|
+
}
|
|
71
61
|
constructor(fields) {
|
|
72
62
|
super(fields);
|
|
73
63
|
this.outputVersion = iife(() => {
|
|
@@ -112,6 +102,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
112
102
|
invocation_params: this?.invocationParams(callOptions),
|
|
113
103
|
batch_size: 1
|
|
114
104
|
};
|
|
105
|
+
const outputVersion = callOptions.outputVersion ?? this.outputVersion;
|
|
115
106
|
const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [_formatForTracing(messages)], runnableConfig.runId, void 0, extra, void 0, void 0, runnableConfig.runName);
|
|
116
107
|
let generationChunk;
|
|
117
108
|
let llmOutput;
|
|
@@ -125,7 +116,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
125
116
|
...chunk.generationInfo,
|
|
126
117
|
...chunk.message.response_metadata
|
|
127
118
|
};
|
|
128
|
-
if (
|
|
119
|
+
if (outputVersion === "v1") yield castStandardMessageContent(chunk.message);
|
|
129
120
|
else yield chunk.message;
|
|
130
121
|
if (!generationChunk) generationChunk = chunk;
|
|
131
122
|
else generationChunk = generationChunk.concat(chunk);
|
|
@@ -171,6 +162,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
171
162
|
};
|
|
172
163
|
runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), baseMessages.map(_formatForTracing), handledOptions.runId, void 0, extra, void 0, void 0, handledOptions.runName);
|
|
173
164
|
}
|
|
165
|
+
const outputVersion = parsedOptions.outputVersion ?? this.outputVersion;
|
|
174
166
|
const generations = [];
|
|
175
167
|
const llmOutputs = [];
|
|
176
168
|
const hasStreamingHandler = !!runManagers?.[0].handlers.find(callbackHandlerPrefersStreaming);
|
|
@@ -207,7 +199,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
207
199
|
...parsedOptions,
|
|
208
200
|
promptIndex: i
|
|
209
201
|
}, runManagers?.[i]);
|
|
210
|
-
if (
|
|
202
|
+
if (outputVersion === "v1") for (const generation of generateResults.generations) generation.message = castStandardMessageContent(generation.message);
|
|
211
203
|
return generateResults;
|
|
212
204
|
}));
|
|
213
205
|
await Promise.all(results.map(async (pResult, i) => {
|
|
@@ -273,6 +265,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
273
265
|
result,
|
|
274
266
|
runManager: runManagers?.[index]
|
|
275
267
|
})).filter(({ result }) => result.status === "fulfilled" && result.value != null || result.status === "rejected");
|
|
268
|
+
const outputVersion = parsedOptions.outputVersion ?? this.outputVersion;
|
|
276
269
|
const generations = [];
|
|
277
270
|
await Promise.all(cachedResults.map(async ({ result: promiseResult, runManager }, i) => {
|
|
278
271
|
if (promiseResult.status === "fulfilled") {
|
|
@@ -284,7 +277,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
284
277
|
output_tokens: 0,
|
|
285
278
|
total_tokens: 0
|
|
286
279
|
};
|
|
287
|
-
if (
|
|
280
|
+
if (outputVersion === "v1") result$1.message = castStandardMessageContent(result$1.message);
|
|
288
281
|
}
|
|
289
282
|
result$1.generationInfo = {
|
|
290
283
|
...result$1.generationInfo,
|
|
@@ -360,17 +353,6 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
360
353
|
return "base_chat_model";
|
|
361
354
|
}
|
|
362
355
|
/**
|
|
363
|
-
* @deprecated
|
|
364
|
-
* Return a json-like object representing this LLM.
|
|
365
|
-
*/
|
|
366
|
-
serialize() {
|
|
367
|
-
return {
|
|
368
|
-
...this.invocationParams(),
|
|
369
|
-
_type: this._llmType(),
|
|
370
|
-
_model: this._modelType()
|
|
371
|
-
};
|
|
372
|
-
}
|
|
373
|
-
/**
|
|
374
356
|
* Generates a prompt based on the input prompt values.
|
|
375
357
|
* @param promptValues An array of BasePromptValue instances.
|
|
376
358
|
* @param options The call options or an array of stop sequences.
|
|
@@ -381,60 +363,6 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
|
|
|
381
363
|
const promptMessages = promptValues.map((promptValue) => promptValue.toChatMessages());
|
|
382
364
|
return this.generate(promptMessages, options, callbacks);
|
|
383
365
|
}
|
|
384
|
-
/**
|
|
385
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
386
|
-
*
|
|
387
|
-
* Makes a single call to the chat model.
|
|
388
|
-
* @param messages An array of BaseMessage instances.
|
|
389
|
-
* @param options The call options or an array of stop sequences.
|
|
390
|
-
* @param callbacks The callbacks for the language model.
|
|
391
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
392
|
-
*/
|
|
393
|
-
async call(messages, options, callbacks) {
|
|
394
|
-
const result = await this.generate([messages.map(coerceMessageLikeToMessage)], options, callbacks);
|
|
395
|
-
const generations = result.generations;
|
|
396
|
-
return generations[0][0].message;
|
|
397
|
-
}
|
|
398
|
-
/**
|
|
399
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
400
|
-
*
|
|
401
|
-
* Makes a single call to the chat model with a prompt value.
|
|
402
|
-
* @param promptValue The value of the prompt.
|
|
403
|
-
* @param options The call options or an array of stop sequences.
|
|
404
|
-
* @param callbacks The callbacks for the language model.
|
|
405
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
406
|
-
*/
|
|
407
|
-
async callPrompt(promptValue, options, callbacks) {
|
|
408
|
-
const promptMessages = promptValue.toChatMessages();
|
|
409
|
-
return this.call(promptMessages, options, callbacks);
|
|
410
|
-
}
|
|
411
|
-
/**
|
|
412
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
413
|
-
*
|
|
414
|
-
* Predicts the next message based on the input messages.
|
|
415
|
-
* @param messages An array of BaseMessage instances.
|
|
416
|
-
* @param options The call options or an array of stop sequences.
|
|
417
|
-
* @param callbacks The callbacks for the language model.
|
|
418
|
-
* @returns A Promise that resolves to a BaseMessage.
|
|
419
|
-
*/
|
|
420
|
-
async predictMessages(messages, options, callbacks) {
|
|
421
|
-
return this.call(messages, options, callbacks);
|
|
422
|
-
}
|
|
423
|
-
/**
|
|
424
|
-
* @deprecated Use .invoke() instead. Will be removed in 0.2.0.
|
|
425
|
-
*
|
|
426
|
-
* Predicts the next message based on a text input.
|
|
427
|
-
* @param text The text input.
|
|
428
|
-
* @param options The call options or an array of stop sequences.
|
|
429
|
-
* @param callbacks The callbacks for the language model.
|
|
430
|
-
* @returns A Promise that resolves to a string.
|
|
431
|
-
*/
|
|
432
|
-
async predict(text, options, callbacks) {
|
|
433
|
-
const message = new HumanMessage(text);
|
|
434
|
-
const result = await this.call([message], options, callbacks);
|
|
435
|
-
if (typeof result.content !== "string") throw new Error("Cannot use predict when output is not a string.");
|
|
436
|
-
return result.content;
|
|
437
|
-
}
|
|
438
366
|
withStructuredOutput(outputSchema, config) {
|
|
439
367
|
if (typeof this.bindTools !== "function") throw new Error(`Chat model must implement ".bindTools()" to use withStructuredOutput.`);
|
|
440
368
|
if (config?.strict) throw new Error(`"strict" mode is not supported for this model by default.`);
|
|
@@ -497,5 +425,5 @@ var SimpleChatModel = class extends BaseChatModel {
|
|
|
497
425
|
};
|
|
498
426
|
|
|
499
427
|
//#endregion
|
|
500
|
-
export { BaseChatModel, SimpleChatModel, chat_models_exports
|
|
428
|
+
export { BaseChatModel, SimpleChatModel, chat_models_exports };
|
|
501
429
|
//# sourceMappingURL=chat_models.js.map
|