@langchain/core 1.0.0-alpha.6 → 1.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (245) hide show
  1. package/CHANGELOG.md +131 -0
  2. package/LICENSE +6 -6
  3. package/README.md +2 -23
  4. package/dist/agents.d.ts.map +1 -1
  5. package/dist/caches/base.d.cts.map +1 -1
  6. package/dist/callbacks/base.d.cts.map +1 -1
  7. package/dist/callbacks/base.d.ts.map +1 -1
  8. package/dist/callbacks/manager.cjs +9 -64
  9. package/dist/callbacks/manager.cjs.map +1 -1
  10. package/dist/callbacks/manager.d.cts +1 -23
  11. package/dist/callbacks/manager.d.cts.map +1 -1
  12. package/dist/callbacks/manager.d.ts +1 -23
  13. package/dist/callbacks/manager.d.ts.map +1 -1
  14. package/dist/callbacks/manager.js +10 -63
  15. package/dist/callbacks/manager.js.map +1 -1
  16. package/dist/chat_history.cjs +0 -4
  17. package/dist/chat_history.cjs.map +1 -1
  18. package/dist/chat_history.d.cts +1 -3
  19. package/dist/chat_history.d.cts.map +1 -1
  20. package/dist/chat_history.d.ts +1 -3
  21. package/dist/chat_history.d.ts.map +1 -1
  22. package/dist/chat_history.js +0 -4
  23. package/dist/chat_history.js.map +1 -1
  24. package/dist/document_loaders/base.cjs +1 -13
  25. package/dist/document_loaders/base.cjs.map +1 -1
  26. package/dist/document_loaders/base.d.cts +0 -9
  27. package/dist/document_loaders/base.d.cts.map +1 -1
  28. package/dist/document_loaders/base.d.ts +0 -9
  29. package/dist/document_loaders/base.d.ts.map +1 -1
  30. package/dist/document_loaders/base.js +1 -13
  31. package/dist/document_loaders/base.js.map +1 -1
  32. package/dist/document_loaders/langsmith.d.cts +1 -1
  33. package/dist/language_models/base.cjs.map +1 -1
  34. package/dist/language_models/base.d.cts +0 -16
  35. package/dist/language_models/base.d.cts.map +1 -1
  36. package/dist/language_models/base.d.ts +0 -16
  37. package/dist/language_models/base.d.ts.map +1 -1
  38. package/dist/language_models/base.js.map +1 -1
  39. package/dist/language_models/chat_models.cjs +10 -83
  40. package/dist/language_models/chat_models.cjs.map +1 -1
  41. package/dist/language_models/chat_models.d.cts +21 -55
  42. package/dist/language_models/chat_models.d.cts.map +1 -1
  43. package/dist/language_models/chat_models.d.ts +21 -55
  44. package/dist/language_models/chat_models.d.ts.map +1 -1
  45. package/dist/language_models/chat_models.js +11 -83
  46. package/dist/language_models/chat_models.js.map +1 -1
  47. package/dist/language_models/llms.cjs +0 -56
  48. package/dist/language_models/llms.cjs.map +1 -1
  49. package/dist/language_models/llms.d.cts +1 -43
  50. package/dist/language_models/llms.d.cts.map +1 -1
  51. package/dist/language_models/llms.d.ts +1 -43
  52. package/dist/language_models/llms.d.ts.map +1 -1
  53. package/dist/language_models/llms.js +0 -56
  54. package/dist/language_models/llms.js.map +1 -1
  55. package/dist/load/import_map.cjs +1 -7
  56. package/dist/load/import_map.cjs.map +1 -1
  57. package/dist/load/import_map.js +1 -7
  58. package/dist/load/import_map.js.map +1 -1
  59. package/dist/messages/ai.cjs +4 -0
  60. package/dist/messages/ai.cjs.map +1 -1
  61. package/dist/messages/ai.js +4 -0
  62. package/dist/messages/ai.js.map +1 -1
  63. package/dist/messages/block_translators/anthropic.cjs +191 -144
  64. package/dist/messages/block_translators/anthropic.cjs.map +1 -1
  65. package/dist/messages/block_translators/anthropic.js +191 -144
  66. package/dist/messages/block_translators/anthropic.js.map +1 -1
  67. package/dist/messages/block_translators/index.cjs +2 -2
  68. package/dist/messages/block_translators/index.cjs.map +1 -1
  69. package/dist/messages/block_translators/index.js +4 -4
  70. package/dist/messages/block_translators/index.js.map +1 -1
  71. package/dist/messages/block_translators/openai.cjs +78 -22
  72. package/dist/messages/block_translators/openai.cjs.map +1 -1
  73. package/dist/messages/block_translators/openai.js +78 -22
  74. package/dist/messages/block_translators/openai.js.map +1 -1
  75. package/dist/messages/content/tools.cjs +1 -5
  76. package/dist/messages/content/tools.cjs.map +1 -1
  77. package/dist/messages/content/tools.d.cts +1 -85
  78. package/dist/messages/content/tools.d.cts.map +1 -1
  79. package/dist/messages/content/tools.d.ts +1 -85
  80. package/dist/messages/content/tools.d.ts.map +1 -1
  81. package/dist/messages/content/tools.js +1 -5
  82. package/dist/messages/content/tools.js.map +1 -1
  83. package/dist/messages/metadata.cjs.map +1 -1
  84. package/dist/messages/metadata.d.cts +3 -0
  85. package/dist/messages/metadata.d.cts.map +1 -1
  86. package/dist/messages/metadata.d.ts +3 -0
  87. package/dist/messages/metadata.d.ts.map +1 -1
  88. package/dist/messages/metadata.js.map +1 -1
  89. package/dist/messages/tool.cjs +2 -0
  90. package/dist/messages/tool.cjs.map +1 -1
  91. package/dist/messages/tool.d.cts +2 -0
  92. package/dist/messages/tool.d.cts.map +1 -1
  93. package/dist/messages/tool.d.ts +2 -0
  94. package/dist/messages/tool.d.ts.map +1 -1
  95. package/dist/messages/tool.js +2 -0
  96. package/dist/messages/tool.js.map +1 -1
  97. package/dist/output_parsers/json.cjs +5 -0
  98. package/dist/output_parsers/json.cjs.map +1 -1
  99. package/dist/output_parsers/json.d.cts +2 -0
  100. package/dist/output_parsers/json.d.cts.map +1 -1
  101. package/dist/output_parsers/json.d.ts +2 -0
  102. package/dist/output_parsers/json.d.ts.map +1 -1
  103. package/dist/output_parsers/json.js +5 -0
  104. package/dist/output_parsers/json.js.map +1 -1
  105. package/dist/prompts/base.cjs +0 -36
  106. package/dist/prompts/base.cjs.map +1 -1
  107. package/dist/prompts/base.d.cts +0 -16
  108. package/dist/prompts/base.d.cts.map +1 -1
  109. package/dist/prompts/base.d.ts +0 -16
  110. package/dist/prompts/base.d.ts.map +1 -1
  111. package/dist/prompts/base.js +0 -36
  112. package/dist/prompts/base.js.map +1 -1
  113. package/dist/prompts/chat.cjs +1 -5
  114. package/dist/prompts/chat.cjs.map +1 -1
  115. package/dist/prompts/chat.d.cts +1 -4
  116. package/dist/prompts/chat.d.cts.map +1 -1
  117. package/dist/prompts/chat.d.ts +1 -4
  118. package/dist/prompts/chat.d.ts.map +1 -1
  119. package/dist/prompts/chat.js +1 -5
  120. package/dist/prompts/chat.js.map +1 -1
  121. package/dist/prompts/dict.d.cts +1 -1
  122. package/dist/prompts/dict.d.ts +1 -1
  123. package/dist/prompts/few_shot.d.cts +2 -2
  124. package/dist/prompts/few_shot.d.ts +2 -2
  125. package/dist/prompts/image.d.cts +1 -1
  126. package/dist/prompts/image.d.ts +1 -1
  127. package/dist/prompts/index.d.cts +2 -2
  128. package/dist/prompts/index.d.ts +2 -2
  129. package/dist/prompts/pipeline.d.cts +1 -1
  130. package/dist/prompts/pipeline.d.ts +1 -1
  131. package/dist/prompts/prompt.d.cts +2 -2
  132. package/dist/prompts/prompt.d.ts +2 -2
  133. package/dist/retrievers/index.cjs +3 -18
  134. package/dist/retrievers/index.cjs.map +1 -1
  135. package/dist/retrievers/index.d.cts +2 -27
  136. package/dist/retrievers/index.d.cts.map +1 -1
  137. package/dist/retrievers/index.d.ts +2 -27
  138. package/dist/retrievers/index.d.ts.map +1 -1
  139. package/dist/retrievers/index.js +3 -18
  140. package/dist/retrievers/index.js.map +1 -1
  141. package/dist/runnables/base.cjs +24 -63
  142. package/dist/runnables/base.cjs.map +1 -1
  143. package/dist/runnables/base.d.cts +10 -39
  144. package/dist/runnables/base.d.cts.map +1 -1
  145. package/dist/runnables/base.d.ts +10 -39
  146. package/dist/runnables/base.d.ts.map +1 -1
  147. package/dist/runnables/base.js +24 -63
  148. package/dist/runnables/base.js.map +1 -1
  149. package/dist/runnables/graph.cjs +1 -1
  150. package/dist/runnables/graph.cjs.map +1 -1
  151. package/dist/runnables/graph.js +2 -2
  152. package/dist/runnables/graph.js.map +1 -1
  153. package/dist/runnables/graph_mermaid.cjs +1 -10
  154. package/dist/runnables/graph_mermaid.cjs.map +1 -1
  155. package/dist/runnables/graph_mermaid.js +1 -10
  156. package/dist/runnables/graph_mermaid.js.map +1 -1
  157. package/dist/runnables/history.cjs +1 -1
  158. package/dist/runnables/history.cjs.map +1 -1
  159. package/dist/runnables/history.d.cts +2 -2
  160. package/dist/runnables/history.d.cts.map +1 -1
  161. package/dist/runnables/history.d.ts +2 -2
  162. package/dist/runnables/history.d.ts.map +1 -1
  163. package/dist/runnables/history.js +1 -1
  164. package/dist/runnables/history.js.map +1 -1
  165. package/dist/stores.cjs.map +1 -1
  166. package/dist/stores.d.cts +3 -29
  167. package/dist/stores.d.cts.map +1 -1
  168. package/dist/stores.d.ts +3 -29
  169. package/dist/stores.d.ts.map +1 -1
  170. package/dist/stores.js.map +1 -1
  171. package/dist/tools/index.cjs +12 -4
  172. package/dist/tools/index.cjs.map +1 -1
  173. package/dist/tools/index.js +12 -4
  174. package/dist/tools/index.js.map +1 -1
  175. package/dist/tools/types.cjs.map +1 -1
  176. package/dist/tools/types.d.cts +4 -0
  177. package/dist/tools/types.d.cts.map +1 -1
  178. package/dist/tools/types.d.ts +4 -0
  179. package/dist/tools/types.d.ts.map +1 -1
  180. package/dist/tools/types.js.map +1 -1
  181. package/dist/tracers/base.cjs +1 -1
  182. package/dist/tracers/base.cjs.map +1 -1
  183. package/dist/tracers/base.d.cts +1 -1
  184. package/dist/tracers/base.js +2 -2
  185. package/dist/tracers/base.js.map +1 -1
  186. package/dist/tracers/log_stream.d.cts +1 -1
  187. package/dist/tracers/log_stream.d.ts +1 -1
  188. package/dist/tracers/tracer_langchain.cjs +1 -0
  189. package/dist/tracers/tracer_langchain.cjs.map +1 -1
  190. package/dist/tracers/tracer_langchain.d.cts +2 -2
  191. package/dist/tracers/tracer_langchain.js +1 -0
  192. package/dist/tracers/tracer_langchain.js.map +1 -1
  193. package/dist/utils/env.cjs +1 -9
  194. package/dist/utils/env.cjs.map +1 -1
  195. package/dist/utils/env.d.cts +2 -6
  196. package/dist/utils/env.d.cts.map +1 -1
  197. package/dist/utils/env.d.ts +2 -6
  198. package/dist/utils/env.d.ts.map +1 -1
  199. package/dist/utils/env.js +2 -9
  200. package/dist/utils/env.js.map +1 -1
  201. package/dist/utils/testing/message_history.cjs +1 -1
  202. package/dist/utils/testing/message_history.cjs.map +1 -1
  203. package/dist/utils/testing/message_history.d.cts +1 -1
  204. package/dist/utils/testing/message_history.d.cts.map +1 -1
  205. package/dist/utils/testing/message_history.d.ts +1 -1
  206. package/dist/utils/testing/message_history.d.ts.map +1 -1
  207. package/dist/utils/testing/message_history.js +1 -1
  208. package/dist/utils/testing/message_history.js.map +1 -1
  209. package/dist/utils/types/index.cjs +6 -0
  210. package/dist/utils/types/index.d.cts +2 -2
  211. package/dist/utils/types/index.d.ts +2 -2
  212. package/dist/utils/types/index.js +5 -2
  213. package/dist/utils/types/zod.cjs +23 -0
  214. package/dist/utils/types/zod.cjs.map +1 -1
  215. package/dist/utils/types/zod.d.cts +11 -1
  216. package/dist/utils/types/zod.d.cts.map +1 -1
  217. package/dist/utils/types/zod.d.ts +11 -1
  218. package/dist/utils/types/zod.d.ts.map +1 -1
  219. package/dist/utils/types/zod.js +21 -1
  220. package/dist/utils/types/zod.js.map +1 -1
  221. package/package.json +121 -154
  222. package/dist/runnables/remote.cjs +0 -399
  223. package/dist/runnables/remote.cjs.map +0 -1
  224. package/dist/runnables/remote.d.cts +0 -73
  225. package/dist/runnables/remote.d.cts.map +0 -1
  226. package/dist/runnables/remote.d.ts +0 -73
  227. package/dist/runnables/remote.d.ts.map +0 -1
  228. package/dist/runnables/remote.js +0 -393
  229. package/dist/runnables/remote.js.map +0 -1
  230. package/dist/tracers/initialize.cjs +0 -46
  231. package/dist/tracers/initialize.cjs.map +0 -1
  232. package/dist/tracers/initialize.d.cts +0 -26
  233. package/dist/tracers/initialize.d.cts.map +0 -1
  234. package/dist/tracers/initialize.d.ts +0 -26
  235. package/dist/tracers/initialize.d.ts.map +0 -1
  236. package/dist/tracers/initialize.js +0 -39
  237. package/dist/tracers/initialize.js.map +0 -1
  238. package/dist/tracers/tracer_langchain_v1.cjs +0 -168
  239. package/dist/tracers/tracer_langchain_v1.cjs.map +0 -1
  240. package/dist/tracers/tracer_langchain_v1.d.cts +0 -64
  241. package/dist/tracers/tracer_langchain_v1.d.cts.map +0 -1
  242. package/dist/tracers/tracer_langchain_v1.d.ts +0 -64
  243. package/dist/tracers/tracer_langchain_v1.d.ts.map +0 -1
  244. package/dist/tracers/tracer_langchain_v1.js +0 -162
  245. package/dist/tracers/tracer_langchain_v1.js.map +0 -1
@@ -1,6 +1,6 @@
1
1
  import { BaseMessage, BaseMessageChunk, BaseMessageLike } from "../messages/base.cjs";
2
2
  import { AIMessageChunk } from "../messages/ai.cjs";
3
- import { MessageStructure, MessageType } from "../messages/message.cjs";
3
+ import { MessageOutputVersion } from "../messages/message.cjs";
4
4
  import { ChatGenerationChunk, ChatResult, Generation, LLMResult } from "../outputs.cjs";
5
5
  import { BaseCache } from "../caches/base.cjs";
6
6
  import { CallbackManagerForLLMRun, Callbacks } from "../callbacks/manager.cjs";
@@ -59,7 +59,7 @@ type BaseChatModelParams = BaseLanguageModelParams & {
59
59
  *
60
60
  * @default "v0"
61
61
  */
62
- outputVersion?: "v0" | "v1";
62
+ outputVersion?: MessageOutputVersion;
63
63
  };
64
64
  /**
65
65
  * Represents the call options for a base chat model.
@@ -80,13 +80,23 @@ type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {
80
80
  * if used with an unsupported model.
81
81
  */
82
82
  tool_choice?: ToolChoice;
83
+ /**
84
+ * Version of `AIMessage` output format to store in message content.
85
+ *
86
+ * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a
87
+ * standard format. This flag can be used to additionally store the standard format
88
+ * as the message content, e.g., for serialization purposes.
89
+ *
90
+ * - "v0": provider-specific format in content (can lazily parse with `.contentBlocks`)
91
+ * - "v1": standardized format in content (consistent with `.contentBlocks`)
92
+ *
93
+ * You can also set `LC_OUTPUT_VERSION` as an environment variable to "v1" to
94
+ * enable this by default.
95
+ *
96
+ * @default "v0"
97
+ */
98
+ outputVersion?: MessageOutputVersion;
83
99
  };
84
- /**
85
- * Creates a transform stream for encoding chat message chunks.
86
- * @deprecated Use {@link BytesOutputParser} instead
87
- * @returns A TransformStream instance that encodes chat message chunks.
88
- */
89
- declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<MessageStructure, MessageType>, any>;
90
100
  type LangSmithParams = {
91
101
  ls_provider?: string;
92
102
  ls_model_name?: string;
@@ -110,7 +120,8 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
110
120
  // Only ever instantiated in main LangChain
111
121
  lc_namespace: string[];
112
122
  disableStreaming: boolean;
113
- outputVersion?: "v0" | "v1";
123
+ outputVersion?: MessageOutputVersion;
124
+ get callKeys(): string[];
114
125
  constructor(fields: BaseChatModelParams);
115
126
  _combineLLMOutput?(...llmOutputs: LLMResult["llmOutput"][]): LLMResult["llmOutput"];
116
127
  protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this["ParsedCallOptions"]];
@@ -168,11 +179,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
168
179
  invocationParams(_options?: this["ParsedCallOptions"]): any;
169
180
  _modelType(): string;
170
181
  abstract _llmType(): string;
171
- /**
172
- * @deprecated
173
- * Return a json-like object representing this LLM.
174
- */
175
- serialize(): SerializedLLM;
176
182
  /**
177
183
  * Generates a prompt based on the input prompt values.
178
184
  * @param promptValues An array of BasePromptValue instances.
@@ -182,46 +188,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
182
188
  */
183
189
  generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;
184
190
  abstract _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
185
- /**
186
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
187
- *
188
- * Makes a single call to the chat model.
189
- * @param messages An array of BaseMessage instances.
190
- * @param options The call options or an array of stop sequences.
191
- * @param callbacks The callbacks for the language model.
192
- * @returns A Promise that resolves to a BaseMessage.
193
- */
194
- call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
195
- /**
196
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
197
- *
198
- * Makes a single call to the chat model with a prompt value.
199
- * @param promptValue The value of the prompt.
200
- * @param options The call options or an array of stop sequences.
201
- * @param callbacks The callbacks for the language model.
202
- * @returns A Promise that resolves to a BaseMessage.
203
- */
204
- callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
205
- /**
206
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
207
- *
208
- * Predicts the next message based on the input messages.
209
- * @param messages An array of BaseMessage instances.
210
- * @param options The call options or an array of stop sequences.
211
- * @param callbacks The callbacks for the language model.
212
- * @returns A Promise that resolves to a BaseMessage.
213
- */
214
- predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
215
- /**
216
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
217
- *
218
- * Predicts the next message based on a text input.
219
- * @param text The text input.
220
- * @param options The call options or an array of stop sequences.
221
- * @param callbacks The callbacks for the language model.
222
- * @returns A Promise that resolves to a string.
223
- */
224
- predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;
225
191
  withStructuredOutput<
226
192
  // eslint-disable-next-line @typescript-eslint/no-explicit-any
227
193
  RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: $ZodType<RunOutput>
@@ -258,5 +224,5 @@ declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOpti
258
224
  _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
259
225
  }
260
226
  //#endregion
261
- export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice, createChatMessageChunkEncoderStream };
227
+ export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice };
262
228
  //# sourceMappingURL=chat_models.d.cts.map
@@ -1 +1 @@
1
- {"version":3,"file":"chat_models.d.cts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","createChatMessageChunkEncoderStream","___messages_message_js0","MessageStructure","MessageType","TransformStream","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: \"v0\" | \"v1\";\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n};\n/**\n * Creates a transform stream for encoding chat message chunks.\n * @deprecated Use {@link BytesOutputParser} instead\n * @returns A TransformStream instance that encodes chat message chunks.\n */\nexport declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<import(\"../messages/message.js\").MessageStructure, import(\"../messages/message.js\").MessageType>, any>;\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: \"v0\" | \"v1\";\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * @deprecated\n * Return a json-like object representing this LLM.\n */\n serialize(): SerializedLLM;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model with a prompt value.\n * @param promptValue The value of the prompt.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on the input messages.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on a text input.\n * @param text The text input.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a string.\n */\n predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY2B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;AA+BZ,CAAA,GA3CIH,MA2CQI,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;;;;AAegB;AAOJC,KA5DZH,aAAAA,GA4DYG;EAAmC,MAAA,EAAA,MAAA;EAAA,KAAA,EAAA,MAAA;CAAsF,GAzD7IL,MAyD6I,CAAA,MAAA,EAAA,GAAA,CAAA;;;AAAnE;AAClEU,KAtDAP,mBAAAA,GAAsBb,uBA4Df,GAAA;EAEPsB;;;;;;;AAEqE;AAKjF;EAA2C,gBAAA,CAAA,EAAA,OAAA;EAAA;;;;;;;;;;;;;;;EAWiC,aAAiBnB,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;CAAc;;;;AASFqB,KA1D7FV,wBAAAA,GAA2BhB,4BA0DkE0B,GAAAA;EAAiB;;;;;;;;;;;;;;EAUjC,WACpCJ,CAAAA,EAtDnCX,UAsDmCW;CAAe;;;;;;AAGpCY,iBAlDRjB,mCAAAA,CAAAA,CAkDQiB,EAlD+Bb,eAkD/Ba,CAlD+C7C,gBAkD/C6C,CAlD2B,gBAAA,EAAsF,WAAA,CAkDjHA,EAAAA,GAAAA,CAAAA;AAAOC,KAjD3Bb,eAAAA,GAiD2Ba;EAAY,WAAEC,CAAAA,EAAAA,MAAAA;EAAa,aAAEC,CAAAA,EAAAA,MAAAA;EAAc,aAChE/C,EAAAA,MAAAA;EAAe,cACRM,CAAAA,EAAAA,MAAAA;EAAU,aAApBU,CAAAA,EAAAA,MAAAA;EAAS,OAIAD,CAAAA,EAjDVkB,KAiDUlB,CAAAA,MAAAA,CAAAA;CAAc;AAGTF,KAlDjBqB,cAAAA,GAAiBjB;;EAE1BK,MAyDoBtB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GAzDES,cAyDFT,GAzDmBoB,gBAyDnBpB,GAzDsCkB,oBAyDtClB;;;;;AAYNwB,uBAhEaW,aAgEbX,CAAAA,oBAhE+CE,wBAgE/CF,GAhE0EE,wBAgE1EF;;0BA9DSzB,gBAsEwDsC,GAtErCpC,cAsEqCoC,CAAAA,SAtEb9B,iBAsEa8B,CAtEKD,iBAsELC,EAtEwBA,WAsExBA,CAAAA,CAAAA;EAAW;EAAuB,iBAAWlC,EApExGoC,IAoEwGpC,CApEnGkC,WAoEmGlC,EApEtFmC,OAoEsFnC,CAAAA,MApExEY,cAoEwEZ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAS;EAAV,YAC7FL,EAAAA,MAAAA,EAAAA;EAAW,gBAAqDe,EAAAA,OAAAA;EAAwB,aAAWR,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;EAAU,WAAlBoC,CAAAA,MAAAA,EAhEpGhB,mBAgEoGgB;EAAO,iBAUhHzC,CAAAA,CAAAA,GAAAA,UAAAA,EAzEmBG,SAyEnBH,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EAzE8CG,SAyE9CH,CAAAA,WAAAA,CAAAA;EAAe,UAAyBqC,4CAAAA,CAAAA,OAAAA,CAAAA,EAxEUG,OAwEVH,CAxEkBA,WAwElBA,CAAAA,CAAAA,EAAAA,CAxEkCtB,cAwElCsB,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAAW;;;;;;;;EAUiD,SAUzFvC,CAAAA,CAAAA,KAAAA,EAnFRoC,cAmFQpC,EAAAA,EAAAA,MAAAA,CAAAA,EAnFmB0C,OAmFnB1C,CAnF2BuC,WAmF3BvC,CAAAA,CAAAA,EAnF0CqB,QAmF1CrB,CAnFmDa,sBAmFnDb,EAnF2EsC,iBAmF3EtC,EAnF8FuC,WAmF9FvC,CAAAA;EAAW;;;;;;EAUwC,MAAG2C,CAAAA,KAAAA,EAtFlE9B,sBAsFkE8B,EAAAA,OAAAA,CAAAA,EAtFhCJ,WAsFgCI,CAAAA,EAtFlBA,OAsFkBA,CAtFVL,iBAsFUK,CAAAA;EAAO;EAG/D,qBAAgBnB,CAAAA,SAAAA,EAvFPxB,WAuFOwB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EAvF2DT,wBAuF3DS,CAAAA,EAvFsFoB,cAuFtFpB,CAvFqGlB,mBAuFrGkB,CAAAA;EAAM,eAAuC0B,CAAAA,KAAAA,EAtF9DrC,sBAsF8DqC,EAAAA,OAAAA,CAAAA,EAtF5BX,WAsF4BW,CAAAA,EAtFdN,cAsFcM,CAtFCZ,iBAsFDY,CAAAA;EAAS,WAAnBnD,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EArF1BmC,eAqF0BnC;EAAS;EAE3E,iBAAwBW,CAAAA,QAAAA,EArFLR,eAqFKQ,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EArF0EO,cAqF1EP,EAAAA,kBAAAA,CAAAA,EArF+GK,wBAqF/GL,EAAAA,CAAAA,EArF4IiC,OAqF5IjC,CArFoJL,SAqFpJK,CAAAA;EAA6B,eAAmBG,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAGG,CAHHA,EAAAA;IAAwBqC,QAAAA,EAnF3FhD,eAmF2FgD,EAAAA,EAAAA;IAAjC7B,KAAAA,EAlF7DH,SAkF6DG,CAlFnDb,UAkFmDa,EAAAA,CAAAA;IAGtDG,YAAAA,EAAAA,MAAAA;IAAsBA;IAA6C0B,aAAAA,EAAAA,GAAAA;IAAVnD,cAAAA,EAjFvDkB,cAiFuDlB;EAAS,CAAA,CAAA,EAhFhF4C,OAkFDnB,CAlFSnB,SAkFTmB,GAAAA;IAA8Bd,oBAAAA,EAAAA,MAAAA,EAAAA;IAA+CG,kBAAAA,CAAAA,EAhFvDE,wBAgFuDF,EAAAA;EAAsB,CAAA,CAAA;EAClF;;;;;;;EAOX,QAAwBH,CAAAA,QAAAA,EA/EdR,eA+EcQ,EAAAA,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GA/E4B6B,WA+E5B7B,EAAAA,SAAAA,CAAAA,EA/EqDM,SA+ErDN,CAAAA,EA/EiEiC,OA+EjEjC,CA/EyEL,SA+EzEK,CAAAA;EAA6B;;;EAAkB;EAGxD,gBAAgBc,CAAAA,QAAAA,CAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EAAAA,GAAAA;EAAM,UAAuC0B,CAAAA,CAAAA,EAAAA,MAAAA;EAAS,SAAnBrD,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA;EAAS;;;;EAGhE,SACRqD,CAAAA,CAAAA,EA1ECxB,aA0EDwB;EAAS;;AAxI6D;AA+ItF;;;;EAAqH,cAAwBX,CAAAA,YAAAA,EAzE5GnC,wBAyE4GmC,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GAzE3DA,WAyE2DA,EAAAA,SAAAA,CAAAA,EAzElCvB,SAyEkCuB,CAAAA,EAzEtBI,OAyEsBJ,CAzEdlC,SAyEckC,CAAAA;EAAW,SAC3HvC,SAAAA,CAAAA,QAAAA,EAzEIA,WAyEJA,EAAAA,EAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,UAAAA,CAAAA,EAzEoEe,wBAyEpEf,CAAAA,EAzE+F2C,OAyE/F3C,CAzEuGO,UAyEvGP,CAAAA;EAAW;;;;;;;AADoG;;iBA9DzHE,wCAAwCqC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;0BAU5EI,+CAA+CmC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;4BAU1FA,oCAAoCuC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;6CAUhEuC,yBAAyBvB,YAAY2B;;;oBAG9DnB,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;oBAIM1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;;;;uBAOcC,oCAAoCvB,2BAA2BA,kCAAkCS,cAAcE;2BAChHvC,gEAAgEe,2BAA2B4B;sBAChG3C,gEAAgEe,2BAA2B4B,QAAQpC"}
1
+ {"version":3,"file":"chat_models.d.cts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","MessageOutputVersion","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk, MessageOutputVersion } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: MessageOutputVersion;\n get callKeys(): string[];\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY4B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;CAAmB,GAZ3BH,MAY2B,CAAA,MAAA,EAAA,GAAA,CAAA;;;AA0BS;AAKxC;AAAoC,KAtCxBE,aAAAA,GAsCwB;EAAA,MAAGd,EAAAA,MAAAA;EAA4B,KAejDW,EAAAA,MAAAA;CAAU,GAlDxBC,MAkEgBrB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;AAAoB;AAExC;AAQA;AAA0B,KAxEdwB,mBAAAA,GAAsBb,uBAwER,GAAA;EAAA;;;;;AAEuD;AAKjF;;;EAAwF,gBAAGc,CAAAA,EAAAA,OAAAA;EAAwB;;;;;;;;;;;;;;;EAYR,aASrFG,CAAAA,EA1EF5B,oBA0EE4B;CAAc;;;;AAAwFG,KArEhHN,wBAAAA,GAA2BhB,4BAqEqFsB,GAAAA;EAAW;;;;;;;;;;;;;;EAWnE,WAEpCjC,CAAAA,EAnEdsB,UAmEctB;EAAe;;;;;;;;;;;;;;;EAUM,aAF7CqC,CAAAA,EA3DYnC,oBA2DZmC;CAAO;AAWkDJ,KApErDL,eAAAA,GAoEqDK;EAAW,WAAclB,CAAAA,EAAAA,MAAAA;EAAS,aAAWX,CAAAA,EAAAA,MAAAA;EAAS,aAAjBiC,EAAAA,MAAAA;EAAO,cAe5ElC,CAAAA,EAAAA,MAAAA;EAAwB,aAAyB8B,CAAAA,EAAAA,MAAAA;EAAW,OAAclB,CAAAA,EA7E7Fc,KA6E6Fd,CAAAA,MAAAA,CAAAA;CAAS;AAAGsB,KA3E3GP,cAAAA,GAAiBZ;;EAE1BK,MA0E8FT,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GA1ExEJ,cA0EwEI,GA1EvDO,gBA0EuDP,GA1EpCK,oBA0EoCL;;;;;AAGR8B,uBAxE3Db,aAwE2Da,CAAAA,oBAxEzBjB,wBAwEyBiB,GAxEEjB,wBAwEFiB;;0BAtE/D7C,gBAwEnBwB,GAxEsCtB,cAwEtCsB,CAAAA,SAxE8Df,iBAwE9De,CAxEgFS,iBAwEhFT,EAxEmGU,WAwEnGV,CAAAA,CAAAA;EAAM;EAAqD,iBAAmBX,EAtE9DuB,IAsE8DvB,CAtEzDqB,WAsEyDrB,EAtE5CsB,OAsE4CtB,CAAAA,MAtE9BI,cAsE8BJ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAsB;EAAW,YAA1CQ,EAAAA,MAAAA,EAAAA;EAAQ,gBAG9DG,EAAAA,OAAAA;EAAM,aAAgBA,CAAAA,EArExBrB,oBAqEwBqB;EAAM,IAAuCqB,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA,EAAAA;EAAS,WAAnB/C,CAAAA,MAAAA,EAnEvD6B,mBAmEuD7B;EAAS,iBAEjF0B,CAAAA,CAAAA,GAAAA,UAAAA,EApE+BnB,SAoE/BmB,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EApE0DnB,SAoE1DmB,CAAAA,WAAAA,CAAAA;EAAM,UAAwBd,4CAAAA,CAAAA,OAAAA,CAAAA,EAnEgC2B,OAmEhC3B,CAnEwCwB,WAmExCxB,CAAAA,CAAAA,EAAAA,CAnEwDO,cAmExDP,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAA6B;;;;;;;;EAMsB,SAEjFc,CAAAA,CAAAA,KAAAA,EAlEeO,cAkEfP,EAAAA,EAAAA,MAAAA,CAAAA,EAlE0Ca,OAkE1Cb,CAlEkDU,WAkElDV,CAAAA,CAAAA,EAlEiEH,QAkEjEG,CAlE0EX,sBAkE1EW,EAlEkGS,iBAkElGT,EAlEqHU,WAkErHV,CAAAA;EAAM;;;;;;EAGqC,MAAuCqB,CAAAA,KAAAA,EA9DvEhC,sBA8DuEgC,EAAAA,OAAAA,CAAAA,EA9DrCX,WA8DqCW,CAAAA,EA9DvBP,OA8DuBO,CA9DfZ,iBA8DeY,CAAAA;EAAS;EAAV,qBAEjFrB,CAAAA,SAAAA,EA9D8BzB,WA8D9ByB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EA9DgGT,wBA8DhGS,CAAAA,EA9D2He,cA8D3Hf,CA9D0IlB,mBA8D1IkB,CAAAA;EAAM,eAAwBd,CAAAA,KAAAA,EA7DVG,sBA6DUH,EAAAA,OAAAA,CAAAA,EA7DwBwB,WA6DxBxB,CAAAA,EA7DsC6B,cA6DtC7B,CA7DqDuB,iBA6DrDvB,CAAAA;EAA6B,WAAkBG,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EA5D/BgB,eA4D+BhB;EAAsB;EAClF,iBACRgC,CAAAA,QAAAA,EA5DgB5C,eA4DhB4C,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EA5D+F5B,cA4D/F4B,EAAAA,kBAAAA,CAAAA,EA5DoI9B,wBA4DpI8B,EAAAA,CAAAA,EA5DiKP,OA4DjKO,CA5DyKxC,SA4DzKwC,CAAAA;EAAS,eAFkDxB,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAS0C,CAT1CA,EAAAA;IA1FNZ,QAAAA,EAkCnDR,eAlCmDQ,EAAAA,EAAAA;IAAiB,KAAA,EAmCvES,SAnCuE,CAmC7DV,UAnC6D,EAAA,CAAA;IAmGxDsC,YAAAA,EAAe,MAAA;IAAA;IAAqBlB,aAAAA,EAAAA,GAAAA;IAA2BA,cAAAA,EA5DrEX,cA4DqEW;EAAwB,CAAA,CAAA,EA3D7GU,OA2DqIJ,CA3D7H7B,SA2D6H6B,GAAAA;IAChHnC,oBAAAA,EAAAA,MAAAA,EAAAA;IAAgEgB,kBAAAA,CAAAA,EA1DhEA,wBA0DgEA,EAAAA;EAAwB,CAAA,CAAA;EAAU;;;;;AADa;;qBAhDrHd,0CAA0CiC,yBAAyBlB,YAAYsB,QAAQjC;;;;;;;;;;;;;;;+BAe7ED,iDAAiD8B,yBAAyBlB,YAAYsB,QAAQjC;+BAC9FN,gEAAgEgB,2BAA2BuB,QAAQ/B;;;oBAG9GiB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;oBAIMrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;;;;uBAOcC,oCAAoClB,2BAA2BA,kCAAkCI,cAAcE;2BAChHnC,gEAAgEgB,2BAA2BuB;sBAChGvC,gEAAgEgB,2BAA2BuB,QAAQ/B"}
@@ -1,6 +1,6 @@
1
1
  import { BaseMessage, BaseMessageChunk, BaseMessageLike } from "../messages/base.js";
2
2
  import { AIMessageChunk } from "../messages/ai.js";
3
- import { MessageStructure, MessageType } from "../messages/message.js";
3
+ import { MessageOutputVersion } from "../messages/message.js";
4
4
  import { ChatGenerationChunk, ChatResult, Generation, LLMResult } from "../outputs.js";
5
5
  import { BaseCache } from "../caches/base.js";
6
6
  import { CallbackManagerForLLMRun, Callbacks } from "../callbacks/manager.js";
@@ -59,7 +59,7 @@ type BaseChatModelParams = BaseLanguageModelParams & {
59
59
  *
60
60
  * @default "v0"
61
61
  */
62
- outputVersion?: "v0" | "v1";
62
+ outputVersion?: MessageOutputVersion;
63
63
  };
64
64
  /**
65
65
  * Represents the call options for a base chat model.
@@ -80,13 +80,23 @@ type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {
80
80
  * if used with an unsupported model.
81
81
  */
82
82
  tool_choice?: ToolChoice;
83
+ /**
84
+ * Version of `AIMessage` output format to store in message content.
85
+ *
86
+ * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a
87
+ * standard format. This flag can be used to additionally store the standard format
88
+ * as the message content, e.g., for serialization purposes.
89
+ *
90
+ * - "v0": provider-specific format in content (can lazily parse with `.contentBlocks`)
91
+ * - "v1": standardized format in content (consistent with `.contentBlocks`)
92
+ *
93
+ * You can also set `LC_OUTPUT_VERSION` as an environment variable to "v1" to
94
+ * enable this by default.
95
+ *
96
+ * @default "v0"
97
+ */
98
+ outputVersion?: MessageOutputVersion;
83
99
  };
84
- /**
85
- * Creates a transform stream for encoding chat message chunks.
86
- * @deprecated Use {@link BytesOutputParser} instead
87
- * @returns A TransformStream instance that encodes chat message chunks.
88
- */
89
- declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<MessageStructure, MessageType>, any>;
90
100
  type LangSmithParams = {
91
101
  ls_provider?: string;
92
102
  ls_model_name?: string;
@@ -110,7 +120,8 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
110
120
  // Only ever instantiated in main LangChain
111
121
  lc_namespace: string[];
112
122
  disableStreaming: boolean;
113
- outputVersion?: "v0" | "v1";
123
+ outputVersion?: MessageOutputVersion;
124
+ get callKeys(): string[];
114
125
  constructor(fields: BaseChatModelParams);
115
126
  _combineLLMOutput?(...llmOutputs: LLMResult["llmOutput"][]): LLMResult["llmOutput"];
116
127
  protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this["ParsedCallOptions"]];
@@ -168,11 +179,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
168
179
  invocationParams(_options?: this["ParsedCallOptions"]): any;
169
180
  _modelType(): string;
170
181
  abstract _llmType(): string;
171
- /**
172
- * @deprecated
173
- * Return a json-like object representing this LLM.
174
- */
175
- serialize(): SerializedLLM;
176
182
  /**
177
183
  * Generates a prompt based on the input prompt values.
178
184
  * @param promptValues An array of BasePromptValue instances.
@@ -182,46 +188,6 @@ OutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguag
182
188
  */
183
189
  generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;
184
190
  abstract _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
185
- /**
186
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
187
- *
188
- * Makes a single call to the chat model.
189
- * @param messages An array of BaseMessage instances.
190
- * @param options The call options or an array of stop sequences.
191
- * @param callbacks The callbacks for the language model.
192
- * @returns A Promise that resolves to a BaseMessage.
193
- */
194
- call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
195
- /**
196
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
197
- *
198
- * Makes a single call to the chat model with a prompt value.
199
- * @param promptValue The value of the prompt.
200
- * @param options The call options or an array of stop sequences.
201
- * @param callbacks The callbacks for the language model.
202
- * @returns A Promise that resolves to a BaseMessage.
203
- */
204
- callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
205
- /**
206
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
207
- *
208
- * Predicts the next message based on the input messages.
209
- * @param messages An array of BaseMessage instances.
210
- * @param options The call options or an array of stop sequences.
211
- * @param callbacks The callbacks for the language model.
212
- * @returns A Promise that resolves to a BaseMessage.
213
- */
214
- predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;
215
- /**
216
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
217
- *
218
- * Predicts the next message based on a text input.
219
- * @param text The text input.
220
- * @param options The call options or an array of stop sequences.
221
- * @param callbacks The callbacks for the language model.
222
- * @returns A Promise that resolves to a string.
223
- */
224
- predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;
225
191
  withStructuredOutput<
226
192
  // eslint-disable-next-line @typescript-eslint/no-explicit-any
227
193
  RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: $ZodType<RunOutput>
@@ -258,5 +224,5 @@ declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOpti
258
224
  _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;
259
225
  }
260
226
  //#endregion
261
- export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice, createChatMessageChunkEncoderStream };
227
+ export { BaseChatModel, BaseChatModelCallOptions, BaseChatModelParams, BindToolsInput, LangSmithParams, SerializedChatModel, SerializedLLM, SimpleChatModel, ToolChoice };
262
228
  //# sourceMappingURL=chat_models.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"chat_models.d.ts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","createChatMessageChunkEncoderStream","___messages_message_js0","MessageStructure","MessageType","TransformStream","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: \"v0\" | \"v1\";\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n};\n/**\n * Creates a transform stream for encoding chat message chunks.\n * @deprecated Use {@link BytesOutputParser} instead\n * @returns A TransformStream instance that encodes chat message chunks.\n */\nexport declare function createChatMessageChunkEncoderStream(): TransformStream<BaseMessageChunk<import(\"../messages/message.js\").MessageStructure, import(\"../messages/message.js\").MessageType>, any>;\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: \"v0\" | \"v1\";\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * @deprecated\n * Return a json-like object representing this LLM.\n */\n serialize(): SerializedLLM;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n call(messages: BaseMessageLike[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Makes a single call to the chat model with a prompt value.\n * @param promptValue The value of the prompt.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n callPrompt(promptValue: BasePromptValueInterface, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on the input messages.\n * @param messages An array of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a BaseMessage.\n */\n predictMessages(messages: BaseMessage[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<BaseMessage>;\n /**\n * @deprecated Use .invoke() instead. Will be removed in 0.2.0.\n *\n * Predicts the next message based on a text input.\n * @param text The text input.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to a string.\n */\n predict(text: string, options?: string[] | CallOptions, callbacks?: Callbacks): Promise<string>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY2B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;AA+BZ,CAAA,GA3CIH,MA2CQI,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;;;;AAegB;AAOJC,KA5DZH,aAAAA,GA4DYG;EAAmC,MAAA,EAAA,MAAA;EAAA,KAAA,EAAA,MAAA;CAAsF,GAzD7IL,MAyD6I,CAAA,MAAA,EAAA,GAAA,CAAA;;;AAAnE;AAClEU,KAtDAP,mBAAAA,GAAsBb,uBA4Df,GAAA;EAEPsB;;;;;;;AAEqE;AAKjF;EAA2C,gBAAA,CAAA,EAAA,OAAA;EAAA;;;;;;;;;;;;;;;EAWiC,aAAiBnB,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;CAAc;;;;AASFqB,KA1D7FV,wBAAAA,GAA2BhB,4BA0DkE0B,GAAAA;EAAiB;;;;;;;;;;;;;;EAUjC,WACpCJ,CAAAA,EAtDnCX,UAsDmCW;CAAe;;;;;;AAGpCY,iBAlDRjB,mCAAAA,CAAAA,CAkDQiB,EAlD+Bb,eAkD/Ba,CAlD+C7C,gBAkD/C6C,CAlD2B,gBAAA,EAAsF,WAAA,CAkDjHA,EAAAA,GAAAA,CAAAA;AAAOC,KAjD3Bb,eAAAA,GAiD2Ba;EAAY,WAAEC,CAAAA,EAAAA,MAAAA;EAAa,aAAEC,CAAAA,EAAAA,MAAAA;EAAc,aAChE/C,EAAAA,MAAAA;EAAe,cACRM,CAAAA,EAAAA,MAAAA;EAAU,aAApBU,CAAAA,EAAAA,MAAAA;EAAS,OAIAD,CAAAA,EAjDVkB,KAiDUlB,CAAAA,MAAAA,CAAAA;CAAc;AAGTF,KAlDjBqB,cAAAA,GAAiBjB;;EAE1BK,MAyDoBtB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GAzDES,cAyDFT,GAzDmBoB,gBAyDnBpB,GAzDsCkB,oBAyDtClB;;;;;AAYNwB,uBAhEaW,aAgEbX,CAAAA,oBAhE+CE,wBAgE/CF,GAhE0EE,wBAgE1EF;;0BA9DSzB,gBAsEwDsC,GAtErCpC,cAsEqCoC,CAAAA,SAtEb9B,iBAsEa8B,CAtEKD,iBAsELC,EAtEwBA,WAsExBA,CAAAA,CAAAA;EAAW;EAAuB,iBAAWlC,EApExGoC,IAoEwGpC,CApEnGkC,WAoEmGlC,EApEtFmC,OAoEsFnC,CAAAA,MApExEY,cAoEwEZ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAS;EAAV,YAC7FL,EAAAA,MAAAA,EAAAA;EAAW,gBAAqDe,EAAAA,OAAAA;EAAwB,aAAWR,CAAAA,EAAAA,IAAAA,GAAAA,IAAAA;EAAU,WAAlBoC,CAAAA,MAAAA,EAhEpGhB,mBAgEoGgB;EAAO,iBAUhHzC,CAAAA,CAAAA,GAAAA,UAAAA,EAzEmBG,SAyEnBH,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EAzE8CG,SAyE9CH,CAAAA,WAAAA,CAAAA;EAAe,UAAyBqC,4CAAAA,CAAAA,OAAAA,CAAAA,EAxEUG,OAwEVH,CAxEkBA,WAwElBA,CAAAA,CAAAA,EAAAA,CAxEkCtB,cAwElCsB,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAAW;;;;;;;;EAUiD,SAUzFvC,CAAAA,CAAAA,KAAAA,EAnFRoC,cAmFQpC,EAAAA,EAAAA,MAAAA,CAAAA,EAnFmB0C,OAmFnB1C,CAnF2BuC,WAmF3BvC,CAAAA,CAAAA,EAnF0CqB,QAmF1CrB,CAnFmDa,sBAmFnDb,EAnF2EsC,iBAmF3EtC,EAnF8FuC,WAmF9FvC,CAAAA;EAAW;;;;;;EAUwC,MAAG2C,CAAAA,KAAAA,EAtFlE9B,sBAsFkE8B,EAAAA,OAAAA,CAAAA,EAtFhCJ,WAsFgCI,CAAAA,EAtFlBA,OAsFkBA,CAtFVL,iBAsFUK,CAAAA;EAAO;EAG/D,qBAAgBnB,CAAAA,SAAAA,EAvFPxB,WAuFOwB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EAvF2DT,wBAuF3DS,CAAAA,EAvFsFoB,cAuFtFpB,CAvFqGlB,mBAuFrGkB,CAAAA;EAAM,eAAuC0B,CAAAA,KAAAA,EAtF9DrC,sBAsF8DqC,EAAAA,OAAAA,CAAAA,EAtF5BX,WAsF4BW,CAAAA,EAtFdN,cAsFcM,CAtFCZ,iBAsFDY,CAAAA;EAAS,WAAnBnD,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EArF1BmC,eAqF0BnC;EAAS;EAE3E,iBAAwBW,CAAAA,QAAAA,EArFLR,eAqFKQ,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EArF0EO,cAqF1EP,EAAAA,kBAAAA,CAAAA,EArF+GK,wBAqF/GL,EAAAA,CAAAA,EArF4IiC,OAqF5IjC,CArFoJL,SAqFpJK,CAAAA;EAA6B,eAAmBG,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAGG,CAHHA,EAAAA;IAAwBqC,QAAAA,EAnF3FhD,eAmF2FgD,EAAAA,EAAAA;IAAjC7B,KAAAA,EAlF7DH,SAkF6DG,CAlFnDb,UAkFmDa,EAAAA,CAAAA;IAGtDG,YAAAA,EAAAA,MAAAA;IAAsBA;IAA6C0B,aAAAA,EAAAA,GAAAA;IAAVnD,cAAAA,EAjFvDkB,cAiFuDlB;EAAS,CAAA,CAAA,EAhFhF4C,OAkFDnB,CAlFSnB,SAkFTmB,GAAAA;IAA8Bd,oBAAAA,EAAAA,MAAAA,EAAAA;IAA+CG,kBAAAA,CAAAA,EAhFvDE,wBAgFuDF,EAAAA;EAAsB,CAAA,CAAA;EAClF;;;;;;;EAOX,QAAwBH,CAAAA,QAAAA,EA/EdR,eA+EcQ,EAAAA,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GA/E4B6B,WA+E5B7B,EAAAA,SAAAA,CAAAA,EA/EqDM,SA+ErDN,CAAAA,EA/EiEiC,OA+EjEjC,CA/EyEL,SA+EzEK,CAAAA;EAA6B;;;EAAkB;EAGxD,gBAAgBc,CAAAA,QAAAA,CAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EAAAA,GAAAA;EAAM,UAAuC0B,CAAAA,CAAAA,EAAAA,MAAAA;EAAS,SAAnBrD,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA;EAAS;;;;EAGhE,SACRqD,CAAAA,CAAAA,EA1ECxB,aA0EDwB;EAAS;;AAxI6D;AA+ItF;;;;EAAqH,cAAwBX,CAAAA,YAAAA,EAzE5GnC,wBAyE4GmC,EAAAA,EAAAA,OAAAA,CAAAA,EAAAA,MAAAA,EAAAA,GAzE3DA,WAyE2DA,EAAAA,SAAAA,CAAAA,EAzElCvB,SAyEkCuB,CAAAA,EAzEtBI,OAyEsBJ,CAzEdlC,SAyEckC,CAAAA;EAAW,SAC3HvC,SAAAA,CAAAA,QAAAA,EAzEIA,WAyEJA,EAAAA,EAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,UAAAA,CAAAA,EAzEoEe,wBAyEpEf,CAAAA,EAzE+F2C,OAyE/F3C,CAzEuGO,UAyEvGP,CAAAA;EAAW;;;;;;;AADoG;;iBA9DzHE,wCAAwCqC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;0BAU5EI,+CAA+CmC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;4BAU1FA,oCAAoCuC,yBAAyBvB,YAAY2B,QAAQ3C;;;;;;;;;;6CAUhEuC,yBAAyBvB,YAAY2B;;;oBAG9DnB,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmCzB,SAAUmD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;oBAIM1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,uCAAuCW,SAASR,wBAAwBqC;;;oBAGvF1B,sBAAsBA,mCAAmC3B,QAAUqD;;IAElF1B,8BAA8Bd,sCAAsCW,SAASR;SACvEb;YACGkD;;;;;;;uBAOcC,oCAAoCvB,2BAA2BA,kCAAkCS,cAAcE;2BAChHvC,gEAAgEe,2BAA2B4B;sBAChG3C,gEAAgEe,2BAA2B4B,QAAQpC"}
1
+ {"version":3,"file":"chat_models.d.ts","names":["ZodType","ZodTypeV3","$ZodType","ZodTypeV4","BaseMessage","BaseMessageChunk","BaseMessageLike","AIMessageChunk","MessageOutputVersion","BasePromptValueInterface","LLMResult","ChatGenerationChunk","ChatResult","Generation","BaseLanguageModel","StructuredOutputMethodOptions","ToolDefinition","BaseLanguageModelCallOptions","BaseLanguageModelInput","BaseLanguageModelParams","CallbackManagerForLLMRun","Callbacks","RunnableConfig","BaseCache","StructuredToolInterface","StructuredToolParams","Runnable","RunnableToolLike","ToolChoice","Record","SerializedChatModel","SerializedLLM","BaseChatModelParams","BaseChatModelCallOptions","LangSmithParams","Array","BindToolsInput","BaseChatModel","OutputMessageType","CallOptions","Exclude","Omit","Partial","Promise","AsyncGenerator","messages","cache","llmStringKey","parsedOptions","handledOptions","RunOutput","SimpleChatModel"],"sources":["../../src/language_models/chat_models.d.ts"],"sourcesContent":["import type { ZodType as ZodTypeV3 } from \"zod/v3\";\nimport type { $ZodType as ZodTypeV4 } from \"zod/v4/core\";\nimport { type BaseMessage, BaseMessageChunk, type BaseMessageLike, AIMessageChunk, MessageOutputVersion } from \"../messages/index.js\";\nimport type { BasePromptValueInterface } from \"../prompt_values.js\";\nimport { LLMResult, ChatGenerationChunk, type ChatResult, type Generation } from \"../outputs.js\";\nimport { BaseLanguageModel, type StructuredOutputMethodOptions, type ToolDefinition, type BaseLanguageModelCallOptions, type BaseLanguageModelInput, type BaseLanguageModelParams } from \"./base.js\";\nimport { type CallbackManagerForLLMRun, type Callbacks } from \"../callbacks/manager.js\";\nimport type { RunnableConfig } from \"../runnables/config.js\";\nimport type { BaseCache } from \"../caches/base.js\";\nimport { StructuredToolInterface, StructuredToolParams } from \"../tools/index.js\";\nimport { Runnable, RunnableToolLike } from \"../runnables/base.js\";\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type ToolChoice = string | Record<string, any> | \"auto\" | \"any\";\n/**\n * Represents a serialized chat model.\n */\nexport type SerializedChatModel = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n// todo?\n/**\n * Represents a serialized large language model.\n */\nexport type SerializedLLM = {\n _model: string;\n _type: string;\n} & Record<string, any>;\n/**\n * Represents the parameters for a base chat model.\n */\nexport type BaseChatModelParams = BaseLanguageModelParams & {\n /**\n * Whether to disable streaming.\n *\n * If streaming is bypassed, then `stream()` will defer to\n * `invoke()`.\n *\n * - If true, will always bypass streaming case.\n * - If false (default), will always use streaming case if available.\n */\n disableStreaming?: boolean;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\n/**\n * Represents the call options for a base chat model.\n */\nexport type BaseChatModelCallOptions = BaseLanguageModelCallOptions & {\n /**\n * Specifies how the chat model should use tools.\n * @default undefined\n *\n * Possible values:\n * - \"auto\": The model may choose to use any of the provided tools, or none.\n * - \"any\": The model must use one of the provided tools.\n * - \"none\": The model must not use any tools.\n * - A string (not \"auto\", \"any\", or \"none\"): The name of a specific tool the model must use.\n * - An object: A custom schema specifying tool choice parameters. Specific to the provider.\n *\n * Note: Not all providers support tool_choice. An error will be thrown\n * if used with an unsupported model.\n */\n tool_choice?: ToolChoice;\n /**\n * Version of `AIMessage` output format to store in message content.\n *\n * `AIMessage.contentBlocks` will lazily parse the contents of `content` into a\n * standard format. This flag can be used to additionally store the standard format\n * as the message content, e.g., for serialization purposes.\n *\n * - \"v0\": provider-specific format in content (can lazily parse with `.contentBlocks`)\n * - \"v1\": standardized format in content (consistent with `.contentBlocks`)\n *\n * You can also set `LC_OUTPUT_VERSION` as an environment variable to \"v1\" to\n * enable this by default.\n *\n * @default \"v0\"\n */\n outputVersion?: MessageOutputVersion;\n};\nexport type LangSmithParams = {\n ls_provider?: string;\n ls_model_name?: string;\n ls_model_type: \"chat\";\n ls_temperature?: number;\n ls_max_tokens?: number;\n ls_stop?: Array<string>;\n};\nexport type BindToolsInput = StructuredToolInterface\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any> | ToolDefinition | RunnableToolLike | StructuredToolParams;\n/**\n * Base class for chat models. It extends the BaseLanguageModel class and\n * provides methods for generating chat based on input messages.\n */\nexport declare abstract class BaseChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions, \n// TODO: Fix the parameter order on the next minor version.\nOutputMessageType extends BaseMessageChunk = AIMessageChunk> extends BaseLanguageModel<OutputMessageType, CallOptions> {\n // Backwards compatibility since fields have been moved to RunnableConfig\n ParsedCallOptions: Omit<CallOptions, Exclude<keyof RunnableConfig, \"signal\" | \"timeout\" | \"maxConcurrency\">>;\n // Only ever instantiated in main LangChain\n lc_namespace: string[];\n disableStreaming: boolean;\n outputVersion?: MessageOutputVersion;\n get callKeys(): string[];\n constructor(fields: BaseChatModelParams);\n _combineLLMOutput?(...llmOutputs: LLMResult[\"llmOutput\"][]): LLMResult[\"llmOutput\"];\n protected _separateRunnableConfigFromCallOptionsCompat(options?: Partial<CallOptions>): [RunnableConfig, this[\"ParsedCallOptions\"]];\n /**\n * Bind tool-like objects to this chat model.\n *\n * @param tools A list of tool definitions to bind to this chat model.\n * Can be a structured tool, an OpenAI formatted tool, or an object\n * matching the provider's specific tool schema.\n * @param kwargs Any additional parameters to bind.\n */\n bindTools?(tools: BindToolsInput[], kwargs?: Partial<CallOptions>): Runnable<BaseLanguageModelInput, OutputMessageType, CallOptions>;\n /**\n * Invokes the chat model with a single input.\n * @param input The input for the language model.\n * @param options The call options.\n * @returns A Promise that resolves to a BaseMessageChunk.\n */\n invoke(input: BaseLanguageModelInput, options?: CallOptions): Promise<OutputMessageType>;\n // eslint-disable-next-line require-yield\n _streamResponseChunks(_messages: BaseMessage[], _options: this[\"ParsedCallOptions\"], _runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>;\n _streamIterator(input: BaseLanguageModelInput, options?: CallOptions): AsyncGenerator<OutputMessageType>;\n getLsParams(options: this[\"ParsedCallOptions\"]): LangSmithParams;\n /** @ignore */\n _generateUncached(messages: BaseMessageLike[][], parsedOptions: this[\"ParsedCallOptions\"], handledOptions: RunnableConfig, startedRunManagers?: CallbackManagerForLLMRun[]): Promise<LLMResult>;\n _generateCached({ messages, cache, llmStringKey, parsedOptions, handledOptions }: {\n messages: BaseMessageLike[][];\n cache: BaseCache<Generation[]>;\n llmStringKey: string;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n parsedOptions: any;\n handledOptions: RunnableConfig;\n }): Promise<LLMResult & {\n missingPromptIndices: number[];\n startedRunManagers?: CallbackManagerForLLMRun[];\n }>;\n /**\n * Generates chat based on the input messages.\n * @param messages An array of arrays of BaseMessage instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generate(messages: BaseMessageLike[][], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n /**\n * Get the parameters used to invoke the model\n */\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n invocationParams(_options?: this[\"ParsedCallOptions\"]): any;\n _modelType(): string;\n abstract _llmType(): string;\n /**\n * Generates a prompt based on the input prompt values.\n * @param promptValues An array of BasePromptValue instances.\n * @param options The call options or an array of stop sequences.\n * @param callbacks The callbacks for the language model.\n * @returns A Promise that resolves to an LLMResult.\n */\n generatePrompt(promptValues: BasePromptValueInterface[], options?: string[] | CallOptions, callbacks?: Callbacks): Promise<LLMResult>;\n abstract _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV4<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<false>): Runnable<BaseLanguageModelInput, RunOutput>;\n withStructuredOutput<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>>(outputSchema: ZodTypeV3<RunOutput>\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | Record<string, any>, config?: StructuredOutputMethodOptions<true>): Runnable<BaseLanguageModelInput, {\n raw: BaseMessage;\n parsed: RunOutput;\n }>;\n}\n/**\n * An abstract class that extends BaseChatModel and provides a simple\n * implementation of _generate.\n */\nexport declare abstract class SimpleChatModel<CallOptions extends BaseChatModelCallOptions = BaseChatModelCallOptions> extends BaseChatModel<CallOptions> {\n abstract _call(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<string>;\n _generate(messages: BaseMessage[], options: this[\"ParsedCallOptions\"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>;\n}\n"],"mappings":";;;;;;;;;;;;;;;;KAYY4B,UAAAA,YAAsBC;;;AAAlC;AAIYC,KAAAA,mBAAAA,GAAmB;EAQnBC,MAAAA,EAAAA,MAAAA;EAOAC,KAAAA,EAAAA,MAAAA;CAAmB,GAZ3BH,MAY2B,CAAA,MAAA,EAAA,GAAA,CAAA;;;AA0BS;AAKxC;AAAoC,KAtCxBE,aAAAA,GAsCwB;EAAA,MAAGd,EAAAA,MAAAA;EAA4B,KAejDW,EAAAA,MAAAA;CAAU,GAlDxBC,MAkEgBrB,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA;AAAoB;AAExC;AAQA;AAA0B,KAxEdwB,mBAAAA,GAAsBb,uBAwER,GAAA;EAAA;;;;;AAEuD;AAKjF;;;EAAwF,gBAAGc,CAAAA,EAAAA,OAAAA;EAAwB;;;;;;;;;;;;;;;EAYR,aASrFG,CAAAA,EA1EF5B,oBA0EE4B;CAAc;;;;AAAwFG,KArEhHN,wBAAAA,GAA2BhB,4BAqEqFsB,GAAAA;EAAW;;;;;;;;;;;;;;EAWnE,WAEpCjC,CAAAA,EAnEdsB,UAmEctB;EAAe;;;;;;;;;;;;;;;EAUM,aAF7CqC,CAAAA,EA3DYnC,oBA2DZmC;CAAO;AAWkDJ,KApErDL,eAAAA,GAoEqDK;EAAW,WAAclB,CAAAA,EAAAA,MAAAA;EAAS,aAAWX,CAAAA,EAAAA,MAAAA;EAAS,aAAjBiC,EAAAA,MAAAA;EAAO,cAe5ElC,CAAAA,EAAAA,MAAAA;EAAwB,aAAyB8B,CAAAA,EAAAA,MAAAA;EAAW,OAAclB,CAAAA,EA7E7Fc,KA6E6Fd,CAAAA,MAAAA,CAAAA;CAAS;AAAGsB,KA3E3GP,cAAAA,GAAiBZ;;EAE1BK,MA0E8FT,CAAAA,MAAAA,EAAAA,GAAAA,CAAAA,GA1ExEJ,cA0EwEI,GA1EvDO,gBA0EuDP,GA1EpCK,oBA0EoCL;;;;;AAGR8B,uBAxE3Db,aAwE2Da,CAAAA,oBAxEzBjB,wBAwEyBiB,GAxEEjB,wBAwEFiB;;0BAtE/D7C,gBAwEnBwB,GAxEsCtB,cAwEtCsB,CAAAA,SAxE8Df,iBAwE9De,CAxEgFS,iBAwEhFT,EAxEmGU,WAwEnGV,CAAAA,CAAAA;EAAM;EAAqD,iBAAmBX,EAtE9DuB,IAsE8DvB,CAtEzDqB,WAsEyDrB,EAtE5CsB,OAsE4CtB,CAAAA,MAtE9BI,cAsE8BJ,EAAAA,QAAAA,GAAAA,SAAAA,GAAAA,gBAAAA,CAAAA,CAAAA;EAAsB;EAAW,YAA1CQ,EAAAA,MAAAA,EAAAA;EAAQ,gBAG9DG,EAAAA,OAAAA;EAAM,aAAgBA,CAAAA,EArExBrB,oBAqEwBqB;EAAM,IAAuCqB,QAAAA,CAAAA,CAAAA,EAAAA,MAAAA,EAAAA;EAAS,WAAnB/C,CAAAA,MAAAA,EAnEvD6B,mBAmEuD7B;EAAS,iBAEjF0B,CAAAA,CAAAA,GAAAA,UAAAA,EApE+BnB,SAoE/BmB,CAAAA,WAAAA,CAAAA,EAAAA,CAAAA,EApE0DnB,SAoE1DmB,CAAAA,WAAAA,CAAAA;EAAM,UAAwBd,4CAAAA,CAAAA,OAAAA,CAAAA,EAnEgC2B,OAmEhC3B,CAnEwCwB,WAmExCxB,CAAAA,CAAAA,EAAAA,CAnEwDO,cAmExDP,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA;EAA6B;;;;;;;;EAMsB,SAEjFc,CAAAA,CAAAA,KAAAA,EAlEeO,cAkEfP,EAAAA,EAAAA,MAAAA,CAAAA,EAlE0Ca,OAkE1Cb,CAlEkDU,WAkElDV,CAAAA,CAAAA,EAlEiEH,QAkEjEG,CAlE0EX,sBAkE1EW,EAlEkGS,iBAkElGT,EAlEqHU,WAkErHV,CAAAA;EAAM;;;;;;EAGqC,MAAuCqB,CAAAA,KAAAA,EA9DvEhC,sBA8DuEgC,EAAAA,OAAAA,CAAAA,EA9DrCX,WA8DqCW,CAAAA,EA9DvBP,OA8DuBO,CA9DfZ,iBA8DeY,CAAAA;EAAS;EAAV,qBAEjFrB,CAAAA,SAAAA,EA9D8BzB,WA8D9ByB,EAAAA,EAAAA,QAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,WAAAA,CAAAA,EA9DgGT,wBA8DhGS,CAAAA,EA9D2He,cA8D3Hf,CA9D0IlB,mBA8D1IkB,CAAAA;EAAM,eAAwBd,CAAAA,KAAAA,EA7DVG,sBA6DUH,EAAAA,OAAAA,CAAAA,EA7DwBwB,WA6DxBxB,CAAAA,EA7DsC6B,cA6DtC7B,CA7DqDuB,iBA6DrDvB,CAAAA;EAA6B,WAAkBG,CAAAA,OAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,CAAAA,EA5D/BgB,eA4D+BhB;EAAsB;EAClF,iBACRgC,CAAAA,QAAAA,EA5DgB5C,eA4DhB4C,EAAAA,EAAAA,EAAAA,aAAAA,EAAAA,IAAAA,CAAAA,mBAAAA,CAAAA,EAAAA,cAAAA,EA5D+F5B,cA4D/F4B,EAAAA,kBAAAA,CAAAA,EA5DoI9B,wBA4DpI8B,EAAAA,CAAAA,EA5DiKP,OA4DjKO,CA5DyKxC,SA4DzKwC,CAAAA;EAAS,eAFkDxB,CAAAA;IAAAA,QAAAA;IAAAA,KAAAA;IAAAA,YAAAA;IAAAA,aAAAA;IAAAA;EAS0C,CAT1CA,EAAAA;IA1FNZ,QAAAA,EAkCnDR,eAlCmDQ,EAAAA,EAAAA;IAAiB,KAAA,EAmCvES,SAnCuE,CAmC7DV,UAnC6D,EAAA,CAAA;IAmGxDsC,YAAAA,EAAe,MAAA;IAAA;IAAqBlB,aAAAA,EAAAA,GAAAA;IAA2BA,cAAAA,EA5DrEX,cA4DqEW;EAAwB,CAAA,CAAA,EA3D7GU,OA2DqIJ,CA3D7H7B,SA2D6H6B,GAAAA;IAChHnC,oBAAAA,EAAAA,MAAAA,EAAAA;IAAgEgB,kBAAAA,CAAAA,EA1DhEA,wBA0DgEA,EAAAA;EAAwB,CAAA,CAAA;EAAU;;;;;AADa;;qBAhDrHd,0CAA0CiC,yBAAyBlB,YAAYsB,QAAQjC;;;;;;;;;;;;;;;+BAe7ED,iDAAiD8B,yBAAyBlB,YAAYsB,QAAQjC;+BAC9FN,gEAAgEgB,2BAA2BuB,QAAQ/B;;;oBAG9GiB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC1B,SAAU+C;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;oBAIMrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,uCAAuCW,SAASR,wBAAwBgC;;;oBAGvFrB,sBAAsBA,mCAAmC5B,QAAUiD;;IAElFrB,8BAA8Bd,sCAAsCW,SAASR;SACvEd;YACG8C;;;;;;;uBAOcC,oCAAoClB,2BAA2BA,kCAAkCI,cAAcE;2BAChHnC,gEAAgEgB,2BAA2BuB;sBAChGvC,gEAAgEgB,2BAA2BuB,QAAQ/B"}
@@ -2,7 +2,6 @@ import { __export } from "../_virtual/rolldown_runtime.js";
2
2
  import { convertToOpenAIImageBlock, isBase64ContentBlock, isURLContentBlock } from "../messages/content/data.js";
3
3
  import { isBaseMessage } from "../messages/base.js";
4
4
  import { AIMessage, AIMessageChunk, isAIMessage, isAIMessageChunk } from "../messages/ai.js";
5
- import { HumanMessage } from "../messages/human.js";
6
5
  import { coerceMessageLikeToMessage } from "../messages/utils.js";
7
6
  import { getEnvironmentVariable } from "../utils/env.js";
8
7
  import { callbackHandlerPrefersStreaming } from "../callbacks/base.js";
@@ -21,20 +20,8 @@ import { castStandardMessageContent, iife } from "./utils.js";
21
20
  var chat_models_exports = {};
22
21
  __export(chat_models_exports, {
23
22
  BaseChatModel: () => BaseChatModel,
24
- SimpleChatModel: () => SimpleChatModel,
25
- createChatMessageChunkEncoderStream: () => createChatMessageChunkEncoderStream
23
+ SimpleChatModel: () => SimpleChatModel
26
24
  });
27
- /**
28
- * Creates a transform stream for encoding chat message chunks.
29
- * @deprecated Use {@link BytesOutputParser} instead
30
- * @returns A TransformStream instance that encodes chat message chunks.
31
- */
32
- function createChatMessageChunkEncoderStream() {
33
- const textEncoder = new TextEncoder();
34
- return new TransformStream({ transform(chunk, controller) {
35
- controller.enqueue(textEncoder.encode(typeof chunk.content === "string" ? chunk.content : JSON.stringify(chunk.content)));
36
- } });
37
- }
38
25
  function _formatForTracing(messages) {
39
26
  const messagesToTrace = [];
40
27
  for (const message of messages) {
@@ -68,6 +55,9 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
68
55
  ];
69
56
  disableStreaming = false;
70
57
  outputVersion;
58
+ get callKeys() {
59
+ return [...super.callKeys, "outputVersion"];
60
+ }
71
61
  constructor(fields) {
72
62
  super(fields);
73
63
  this.outputVersion = iife(() => {
@@ -112,6 +102,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
112
102
  invocation_params: this?.invocationParams(callOptions),
113
103
  batch_size: 1
114
104
  };
105
+ const outputVersion = callOptions.outputVersion ?? this.outputVersion;
115
106
  const runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), [_formatForTracing(messages)], runnableConfig.runId, void 0, extra, void 0, void 0, runnableConfig.runName);
116
107
  let generationChunk;
117
108
  let llmOutput;
@@ -125,7 +116,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
125
116
  ...chunk.generationInfo,
126
117
  ...chunk.message.response_metadata
127
118
  };
128
- if (this.outputVersion === "v1") yield castStandardMessageContent(chunk.message);
119
+ if (outputVersion === "v1") yield castStandardMessageContent(chunk.message);
129
120
  else yield chunk.message;
130
121
  if (!generationChunk) generationChunk = chunk;
131
122
  else generationChunk = generationChunk.concat(chunk);
@@ -171,6 +162,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
171
162
  };
172
163
  runManagers = await callbackManager_?.handleChatModelStart(this.toJSON(), baseMessages.map(_formatForTracing), handledOptions.runId, void 0, extra, void 0, void 0, handledOptions.runName);
173
164
  }
165
+ const outputVersion = parsedOptions.outputVersion ?? this.outputVersion;
174
166
  const generations = [];
175
167
  const llmOutputs = [];
176
168
  const hasStreamingHandler = !!runManagers?.[0].handlers.find(callbackHandlerPrefersStreaming);
@@ -207,7 +199,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
207
199
  ...parsedOptions,
208
200
  promptIndex: i
209
201
  }, runManagers?.[i]);
210
- if (this.outputVersion === "v1") for (const generation of generateResults.generations) generation.message = castStandardMessageContent(generation.message);
202
+ if (outputVersion === "v1") for (const generation of generateResults.generations) generation.message = castStandardMessageContent(generation.message);
211
203
  return generateResults;
212
204
  }));
213
205
  await Promise.all(results.map(async (pResult, i) => {
@@ -273,6 +265,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
273
265
  result,
274
266
  runManager: runManagers?.[index]
275
267
  })).filter(({ result }) => result.status === "fulfilled" && result.value != null || result.status === "rejected");
268
+ const outputVersion = parsedOptions.outputVersion ?? this.outputVersion;
276
269
  const generations = [];
277
270
  await Promise.all(cachedResults.map(async ({ result: promiseResult, runManager }, i) => {
278
271
  if (promiseResult.status === "fulfilled") {
@@ -284,7 +277,7 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
284
277
  output_tokens: 0,
285
278
  total_tokens: 0
286
279
  };
287
- if (this.outputVersion === "v1") result$1.message = castStandardMessageContent(result$1.message);
280
+ if (outputVersion === "v1") result$1.message = castStandardMessageContent(result$1.message);
288
281
  }
289
282
  result$1.generationInfo = {
290
283
  ...result$1.generationInfo,
@@ -360,17 +353,6 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
360
353
  return "base_chat_model";
361
354
  }
362
355
  /**
363
- * @deprecated
364
- * Return a json-like object representing this LLM.
365
- */
366
- serialize() {
367
- return {
368
- ...this.invocationParams(),
369
- _type: this._llmType(),
370
- _model: this._modelType()
371
- };
372
- }
373
- /**
374
356
  * Generates a prompt based on the input prompt values.
375
357
  * @param promptValues An array of BasePromptValue instances.
376
358
  * @param options The call options or an array of stop sequences.
@@ -381,60 +363,6 @@ var BaseChatModel = class BaseChatModel extends BaseLanguageModel {
381
363
  const promptMessages = promptValues.map((promptValue) => promptValue.toChatMessages());
382
364
  return this.generate(promptMessages, options, callbacks);
383
365
  }
384
- /**
385
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
386
- *
387
- * Makes a single call to the chat model.
388
- * @param messages An array of BaseMessage instances.
389
- * @param options The call options or an array of stop sequences.
390
- * @param callbacks The callbacks for the language model.
391
- * @returns A Promise that resolves to a BaseMessage.
392
- */
393
- async call(messages, options, callbacks) {
394
- const result = await this.generate([messages.map(coerceMessageLikeToMessage)], options, callbacks);
395
- const generations = result.generations;
396
- return generations[0][0].message;
397
- }
398
- /**
399
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
400
- *
401
- * Makes a single call to the chat model with a prompt value.
402
- * @param promptValue The value of the prompt.
403
- * @param options The call options or an array of stop sequences.
404
- * @param callbacks The callbacks for the language model.
405
- * @returns A Promise that resolves to a BaseMessage.
406
- */
407
- async callPrompt(promptValue, options, callbacks) {
408
- const promptMessages = promptValue.toChatMessages();
409
- return this.call(promptMessages, options, callbacks);
410
- }
411
- /**
412
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
413
- *
414
- * Predicts the next message based on the input messages.
415
- * @param messages An array of BaseMessage instances.
416
- * @param options The call options or an array of stop sequences.
417
- * @param callbacks The callbacks for the language model.
418
- * @returns A Promise that resolves to a BaseMessage.
419
- */
420
- async predictMessages(messages, options, callbacks) {
421
- return this.call(messages, options, callbacks);
422
- }
423
- /**
424
- * @deprecated Use .invoke() instead. Will be removed in 0.2.0.
425
- *
426
- * Predicts the next message based on a text input.
427
- * @param text The text input.
428
- * @param options The call options or an array of stop sequences.
429
- * @param callbacks The callbacks for the language model.
430
- * @returns A Promise that resolves to a string.
431
- */
432
- async predict(text, options, callbacks) {
433
- const message = new HumanMessage(text);
434
- const result = await this.call([message], options, callbacks);
435
- if (typeof result.content !== "string") throw new Error("Cannot use predict when output is not a string.");
436
- return result.content;
437
- }
438
366
  withStructuredOutput(outputSchema, config) {
439
367
  if (typeof this.bindTools !== "function") throw new Error(`Chat model must implement ".bindTools()" to use withStructuredOutput.`);
440
368
  if (config?.strict) throw new Error(`"strict" mode is not supported for this model by default.`);
@@ -497,5 +425,5 @@ var SimpleChatModel = class extends BaseChatModel {
497
425
  };
498
426
 
499
427
  //#endregion
500
- export { BaseChatModel, SimpleChatModel, chat_models_exports, createChatMessageChunkEncoderStream };
428
+ export { BaseChatModel, SimpleChatModel, chat_models_exports };
501
429
  //# sourceMappingURL=chat_models.js.map