@langchain/classic 1.0.27 → 1.0.28

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (298) hide show
  1. package/CHANGELOG.md +13 -0
  2. package/dist/agents/agent.cjs.map +1 -1
  3. package/dist/agents/agent.js.map +1 -1
  4. package/dist/agents/executor.cjs.map +1 -1
  5. package/dist/agents/executor.js.map +1 -1
  6. package/dist/agents/openai_functions/index.cjs.map +1 -1
  7. package/dist/agents/openai_functions/index.d.cts.map +1 -1
  8. package/dist/agents/openai_functions/index.d.ts.map +1 -1
  9. package/dist/agents/openai_functions/index.js.map +1 -1
  10. package/dist/agents/openai_tools/index.cjs.map +1 -1
  11. package/dist/agents/openai_tools/index.d.cts +2 -2
  12. package/dist/agents/openai_tools/index.d.cts.map +1 -1
  13. package/dist/agents/openai_tools/index.d.ts +2 -2
  14. package/dist/agents/openai_tools/index.d.ts.map +1 -1
  15. package/dist/agents/openai_tools/index.js.map +1 -1
  16. package/dist/agents/react/index.d.cts +2 -2
  17. package/dist/agents/react/index.d.cts.map +1 -1
  18. package/dist/agents/react/index.d.ts +2 -2
  19. package/dist/agents/react/index.d.ts.map +1 -1
  20. package/dist/agents/structured_chat/index.d.cts +2 -2
  21. package/dist/agents/structured_chat/index.d.cts.map +1 -1
  22. package/dist/agents/structured_chat/index.d.ts +2 -2
  23. package/dist/agents/structured_chat/index.d.ts.map +1 -1
  24. package/dist/agents/tool_calling/index.d.cts +2 -2
  25. package/dist/agents/tool_calling/index.d.cts.map +1 -1
  26. package/dist/agents/tool_calling/index.d.ts +2 -2
  27. package/dist/agents/tool_calling/index.d.ts.map +1 -1
  28. package/dist/agents/tool_calling/output_parser.cjs.map +1 -1
  29. package/dist/agents/tool_calling/output_parser.d.cts.map +1 -1
  30. package/dist/agents/tool_calling/output_parser.js.map +1 -1
  31. package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts +2 -2
  32. package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts.map +1 -1
  33. package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts +2 -2
  34. package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts.map +1 -1
  35. package/dist/agents/xml/index.d.cts.map +1 -1
  36. package/dist/agents/xml/index.d.ts.map +1 -1
  37. package/dist/cache/file_system.cjs.map +1 -1
  38. package/dist/cache/file_system.js.map +1 -1
  39. package/dist/chains/analyze_documents_chain.cjs.map +1 -1
  40. package/dist/chains/analyze_documents_chain.js.map +1 -1
  41. package/dist/chains/base.cjs.map +1 -1
  42. package/dist/chains/base.d.cts +2 -2
  43. package/dist/chains/base.d.cts.map +1 -1
  44. package/dist/chains/base.d.ts +2 -2
  45. package/dist/chains/base.d.ts.map +1 -1
  46. package/dist/chains/base.js.map +1 -1
  47. package/dist/chains/chat_vector_db_chain.cjs.map +1 -1
  48. package/dist/chains/chat_vector_db_chain.js.map +1 -1
  49. package/dist/chains/combine_docs_chain.d.cts +2 -2
  50. package/dist/chains/combine_docs_chain.d.cts.map +1 -1
  51. package/dist/chains/combine_docs_chain.d.ts +2 -2
  52. package/dist/chains/combine_docs_chain.d.ts.map +1 -1
  53. package/dist/chains/combine_documents/reduce.cjs.map +1 -1
  54. package/dist/chains/combine_documents/reduce.js.map +1 -1
  55. package/dist/chains/conversational_retrieval_chain.cjs.map +1 -1
  56. package/dist/chains/conversational_retrieval_chain.js.map +1 -1
  57. package/dist/chains/graph_qa/cypher.cjs.map +1 -1
  58. package/dist/chains/graph_qa/cypher.js.map +1 -1
  59. package/dist/chains/llm_chain.cjs.map +1 -1
  60. package/dist/chains/llm_chain.js.map +1 -1
  61. package/dist/chains/load.d.cts +1 -2
  62. package/dist/chains/load.d.cts.map +1 -1
  63. package/dist/chains/load.d.ts +1 -2
  64. package/dist/chains/load.d.ts.map +1 -1
  65. package/dist/chains/openai_functions/base.cjs.map +1 -1
  66. package/dist/chains/openai_functions/base.js.map +1 -1
  67. package/dist/chains/openai_functions/openapi.cjs.map +1 -1
  68. package/dist/chains/openai_functions/openapi.js.map +1 -1
  69. package/dist/chains/openai_moderation.cjs.map +1 -1
  70. package/dist/chains/openai_moderation.js.map +1 -1
  71. package/dist/chains/query_constructor/index.cjs.map +1 -1
  72. package/dist/chains/query_constructor/index.js.map +1 -1
  73. package/dist/chains/question_answering/load.d.cts +1 -2
  74. package/dist/chains/question_answering/load.d.cts.map +1 -1
  75. package/dist/chains/question_answering/load.d.ts +1 -2
  76. package/dist/chains/question_answering/load.d.ts.map +1 -1
  77. package/dist/chains/retrieval.cjs.map +1 -1
  78. package/dist/chains/retrieval.js.map +1 -1
  79. package/dist/chains/retrieval_qa.cjs.map +1 -1
  80. package/dist/chains/retrieval_qa.js.map +1 -1
  81. package/dist/chains/router/utils.cjs.map +1 -1
  82. package/dist/chains/router/utils.js.map +1 -1
  83. package/dist/chains/summarization/load.d.cts +1 -2
  84. package/dist/chains/summarization/load.d.cts.map +1 -1
  85. package/dist/chains/summarization/load.d.ts +1 -2
  86. package/dist/chains/summarization/load.d.ts.map +1 -1
  87. package/dist/chains/vector_db_qa.cjs.map +1 -1
  88. package/dist/chains/vector_db_qa.js.map +1 -1
  89. package/dist/chat_models/universal.cjs +8 -10
  90. package/dist/chat_models/universal.cjs.map +1 -1
  91. package/dist/chat_models/universal.d.cts +3 -5
  92. package/dist/chat_models/universal.d.cts.map +1 -1
  93. package/dist/chat_models/universal.d.ts +3 -5
  94. package/dist/chat_models/universal.d.ts.map +1 -1
  95. package/dist/chat_models/universal.js +8 -10
  96. package/dist/chat_models/universal.js.map +1 -1
  97. package/dist/document_loaders/fs/directory.d.cts.map +1 -1
  98. package/dist/document_loaders/fs/directory.d.ts.map +1 -1
  99. package/dist/document_loaders/fs/json.cjs.map +1 -1
  100. package/dist/document_loaders/fs/json.js.map +1 -1
  101. package/dist/document_transformers/openai_functions.d.cts.map +1 -1
  102. package/dist/evaluation/agents/index.d.cts +1 -0
  103. package/dist/evaluation/agents/trajectory.d.cts +2 -2
  104. package/dist/evaluation/agents/trajectory.d.cts.map +1 -1
  105. package/dist/evaluation/agents/trajectory.d.ts +2 -2
  106. package/dist/evaluation/agents/trajectory.d.ts.map +1 -1
  107. package/dist/evaluation/base.cjs.map +1 -1
  108. package/dist/evaluation/base.d.cts.map +1 -1
  109. package/dist/evaluation/base.js.map +1 -1
  110. package/dist/evaluation/comparison/index.d.cts +1 -0
  111. package/dist/evaluation/comparison/pairwise.cjs.map +1 -1
  112. package/dist/evaluation/comparison/pairwise.d.cts +3 -3
  113. package/dist/evaluation/comparison/pairwise.d.cts.map +1 -1
  114. package/dist/evaluation/comparison/pairwise.d.ts +3 -3
  115. package/dist/evaluation/comparison/pairwise.d.ts.map +1 -1
  116. package/dist/evaluation/comparison/pairwise.js.map +1 -1
  117. package/dist/evaluation/criteria/criteria.cjs.map +1 -1
  118. package/dist/evaluation/criteria/criteria.d.cts +3 -3
  119. package/dist/evaluation/criteria/criteria.d.cts.map +1 -1
  120. package/dist/evaluation/criteria/criteria.d.ts +3 -3
  121. package/dist/evaluation/criteria/criteria.d.ts.map +1 -1
  122. package/dist/evaluation/criteria/criteria.js.map +1 -1
  123. package/dist/evaluation/criteria/index.d.cts +1 -0
  124. package/dist/evaluation/embedding_distance/index.d.cts +1 -0
  125. package/dist/evaluation/loader.cjs.map +1 -1
  126. package/dist/evaluation/loader.d.cts.map +1 -1
  127. package/dist/evaluation/loader.js.map +1 -1
  128. package/dist/evaluation/qa/index.d.cts +1 -0
  129. package/dist/experimental/autogpt/prompt.d.cts +2 -2
  130. package/dist/experimental/autogpt/prompt.d.cts.map +1 -1
  131. package/dist/experimental/autogpt/prompt.d.ts +2 -2
  132. package/dist/experimental/autogpt/prompt.d.ts.map +1 -1
  133. package/dist/experimental/autogpt/prompt_generator.cjs.map +1 -1
  134. package/dist/experimental/autogpt/prompt_generator.js.map +1 -1
  135. package/dist/experimental/autogpt/schema.cjs.map +1 -1
  136. package/dist/experimental/autogpt/schema.js.map +1 -1
  137. package/dist/experimental/generative_agents/generative_agent_memory.cjs.map +1 -1
  138. package/dist/experimental/generative_agents/generative_agent_memory.js.map +1 -1
  139. package/dist/experimental/masking/parser.cjs.map +1 -1
  140. package/dist/experimental/masking/parser.js.map +1 -1
  141. package/dist/experimental/masking/regex_masking_transformer.cjs.map +1 -1
  142. package/dist/experimental/masking/regex_masking_transformer.js.map +1 -1
  143. package/dist/experimental/openai_assistant/index.cjs.map +1 -1
  144. package/dist/experimental/openai_assistant/index.d.cts +3 -3
  145. package/dist/experimental/openai_assistant/index.d.ts +3 -3
  146. package/dist/experimental/openai_assistant/index.js.map +1 -1
  147. package/dist/experimental/openai_files/index.d.cts +4 -4
  148. package/dist/experimental/openai_files/index.d.ts +4 -4
  149. package/dist/experimental/plan_and_execute/agent_executor.cjs.map +1 -1
  150. package/dist/experimental/plan_and_execute/agent_executor.js.map +1 -1
  151. package/dist/experimental/prompts/custom_format.cjs.map +1 -1
  152. package/dist/experimental/prompts/custom_format.js.map +1 -1
  153. package/dist/experimental/prompts/handlebars.cjs.map +1 -1
  154. package/dist/experimental/prompts/handlebars.d.cts +2 -2
  155. package/dist/experimental/prompts/handlebars.d.cts.map +1 -1
  156. package/dist/experimental/prompts/handlebars.d.ts +2 -2
  157. package/dist/experimental/prompts/handlebars.d.ts.map +1 -1
  158. package/dist/experimental/prompts/handlebars.js.map +1 -1
  159. package/dist/hub/base.cjs.map +1 -1
  160. package/dist/hub/base.d.cts +0 -1
  161. package/dist/hub/base.d.cts.map +1 -1
  162. package/dist/hub/base.d.ts +0 -1
  163. package/dist/hub/base.d.ts.map +1 -1
  164. package/dist/hub/base.js.map +1 -1
  165. package/dist/hub/index.cjs.map +1 -1
  166. package/dist/hub/index.js.map +1 -1
  167. package/dist/load/index.cjs.map +1 -1
  168. package/dist/load/index.js.map +1 -1
  169. package/dist/output_parsers/combining.cjs.map +1 -1
  170. package/dist/output_parsers/combining.js.map +1 -1
  171. package/dist/output_parsers/expression_type_handlers/base.cjs.map +1 -1
  172. package/dist/output_parsers/expression_type_handlers/base.d.cts.map +1 -1
  173. package/dist/output_parsers/expression_type_handlers/base.d.ts.map +1 -1
  174. package/dist/output_parsers/expression_type_handlers/base.js.map +1 -1
  175. package/dist/output_parsers/fix.cjs.map +1 -1
  176. package/dist/output_parsers/fix.js.map +1 -1
  177. package/dist/output_parsers/openai_functions.d.cts.map +1 -1
  178. package/dist/output_parsers/openai_functions.d.ts.map +1 -1
  179. package/dist/output_parsers/openai_tools.cjs.map +1 -1
  180. package/dist/output_parsers/openai_tools.js.map +1 -1
  181. package/dist/output_parsers/regex.cjs.map +1 -1
  182. package/dist/output_parsers/regex.js.map +1 -1
  183. package/dist/output_parsers/structured.d.cts +2 -2
  184. package/dist/output_parsers/structured.d.cts.map +1 -1
  185. package/dist/output_parsers/structured.d.ts +2 -2
  186. package/dist/output_parsers/structured.d.ts.map +1 -1
  187. package/dist/retrievers/document_compressors/index.cjs.map +1 -1
  188. package/dist/retrievers/document_compressors/index.js.map +1 -1
  189. package/dist/retrievers/matryoshka_retriever.cjs.map +1 -1
  190. package/dist/retrievers/matryoshka_retriever.d.cts +1 -1
  191. package/dist/retrievers/matryoshka_retriever.d.cts.map +1 -1
  192. package/dist/retrievers/matryoshka_retriever.d.ts +1 -1
  193. package/dist/retrievers/matryoshka_retriever.d.ts.map +1 -1
  194. package/dist/retrievers/matryoshka_retriever.js.map +1 -1
  195. package/dist/retrievers/multi_query.cjs.map +1 -1
  196. package/dist/retrievers/multi_query.js.map +1 -1
  197. package/dist/retrievers/parent_document.cjs.map +1 -1
  198. package/dist/retrievers/parent_document.js.map +1 -1
  199. package/dist/schema/prompt_template.d.cts +1 -2
  200. package/dist/schema/prompt_template.d.cts.map +1 -1
  201. package/dist/schema/prompt_template.d.ts +1 -2
  202. package/dist/schema/prompt_template.d.ts.map +1 -1
  203. package/dist/smith/config.cjs.map +1 -1
  204. package/dist/smith/config.d.cts +1 -1
  205. package/dist/smith/config.d.ts +1 -1
  206. package/dist/smith/config.d.ts.map +1 -1
  207. package/dist/smith/config.js.map +1 -1
  208. package/dist/smith/runner_utils.cjs.map +1 -1
  209. package/dist/smith/runner_utils.d.cts +1 -1
  210. package/dist/smith/runner_utils.d.cts.map +1 -1
  211. package/dist/smith/runner_utils.d.ts +1 -1
  212. package/dist/smith/runner_utils.d.ts.map +1 -1
  213. package/dist/smith/runner_utils.js.map +1 -1
  214. package/dist/sql_db.d.cts +2 -2
  215. package/dist/sql_db.d.cts.map +1 -1
  216. package/dist/sql_db.d.ts +2 -2
  217. package/dist/sql_db.d.ts.map +1 -1
  218. package/dist/storage/encoder_backed.cjs.map +1 -1
  219. package/dist/storage/encoder_backed.js.map +1 -1
  220. package/dist/storage/file_system.cjs.map +1 -1
  221. package/dist/storage/file_system.js.map +1 -1
  222. package/dist/stores/doc/in_memory.cjs.map +1 -1
  223. package/dist/stores/doc/in_memory.js.map +1 -1
  224. package/dist/tools/json.cjs.map +1 -1
  225. package/dist/tools/json.js.map +1 -1
  226. package/dist/tools/webbrowser.cjs.map +1 -1
  227. package/dist/tools/webbrowser.d.cts.map +1 -1
  228. package/dist/tools/webbrowser.d.ts.map +1 -1
  229. package/dist/tools/webbrowser.js.map +1 -1
  230. package/dist/util/document.d.cts +1 -1
  231. package/dist/util/document.d.cts.map +1 -1
  232. package/dist/util/document.d.ts +1 -1
  233. package/dist/util/document.d.ts.map +1 -1
  234. package/dist/util/is-network-error/index.cjs.map +1 -1
  235. package/dist/util/is-network-error/index.js.map +1 -1
  236. package/dist/util/load.cjs.map +1 -1
  237. package/dist/util/load.js.map +1 -1
  238. package/dist/util/openapi.cjs.map +1 -1
  239. package/dist/util/openapi.d.cts +32 -32
  240. package/dist/util/openapi.d.cts.map +1 -1
  241. package/dist/util/openapi.d.ts +32 -32
  242. package/dist/util/openapi.d.ts.map +1 -1
  243. package/dist/util/openapi.js.map +1 -1
  244. package/dist/util/p-retry/index.cjs.map +1 -1
  245. package/dist/util/p-retry/index.js.map +1 -1
  246. package/dist/vectorstores/memory.cjs.map +1 -1
  247. package/dist/vectorstores/memory.js.map +1 -1
  248. package/package.json +25 -48
  249. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.cts +0 -1494
  250. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.cts.map +0 -1
  251. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.ts +0 -1494
  252. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.ts.map +0 -1
  253. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/_runner.d.cts +0 -1
  254. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/_runner.d.ts +0 -1
  255. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluate_comparative.d.cts +0 -1
  256. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluate_comparative.d.ts +0 -1
  257. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.cts +0 -66
  258. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.cts.map +0 -1
  259. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.ts +0 -66
  260. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.ts.map +0 -1
  261. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/index.d.cts +0 -1
  262. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/index.d.ts +0 -1
  263. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/string_evaluator.d.cts +0 -1
  264. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/string_evaluator.d.ts +0 -1
  265. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/index.d.cts +0 -4
  266. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/index.d.ts +0 -4
  267. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.cts +0 -145
  268. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.cts.map +0 -1
  269. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.ts +0 -145
  270. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.ts.map +0 -1
  271. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.cts +0 -437
  272. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.cts.map +0 -1
  273. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.ts +0 -437
  274. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.ts.map +0 -1
  275. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.cts +0 -7
  276. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.cts.map +0 -1
  277. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.ts +0 -7
  278. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.ts.map +0 -1
  279. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.cts +0 -38
  280. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.cts.map +0 -1
  281. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.ts +0 -38
  282. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.ts.map +0 -1
  283. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.cts +0 -25
  284. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.cts.map +0 -1
  285. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.ts +0 -25
  286. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.ts.map +0 -1
  287. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/p-queue.d.cts +0 -1
  288. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/p-queue.d.ts +0 -1
  289. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.cts +0 -129
  290. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.cts.map +0 -1
  291. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.ts +0 -129
  292. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.ts.map +0 -1
  293. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/evaluation.d.cts +0 -1
  294. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/evaluation.d.ts +0 -1
  295. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/singletons/traceable.d.cts +0 -2
  296. package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/singletons/traceable.d.ts +0 -2
  297. package/dist/node_modules/.pnpm/p-queue@6.6.2/node_modules/p-queue/dist/index.d.cts +0 -1
  298. package/dist/node_modules/.pnpm/p-queue@6.6.2/node_modules/p-queue/dist/index.d.ts +0 -1
@@ -1 +1 @@
1
- {"version":3,"file":"config.cjs","names":[],"sources":["../../src/smith/config.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { RunnableConfig } from \"@langchain/core/runnables\";\nimport { Example, Run } from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport {\n Criteria as CriteriaType,\n type EmbeddingDistanceEvalChainInput,\n} from \"../evaluation/index.js\";\nimport { LoadEvaluatorOptions } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\n\nexport type EvaluatorInputs = {\n input?: string | unknown;\n prediction: string | unknown;\n reference?: string | unknown;\n};\n\nexport type EvaluatorInputFormatter = ({\n rawInput,\n rawPrediction,\n rawReferenceOutput,\n run,\n}: {\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawInput: any;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawPrediction: any;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawReferenceOutput?: any;\n run: Run;\n}) => EvaluatorInputs;\n\nexport type DynamicRunEvaluatorParams<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Input extends Record<string, any> = Record<string, unknown>,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Prediction extends Record<string, any> = Record<string, unknown>,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Reference extends Record<string, any> = Record<string, unknown>,\n> = {\n input: Input;\n prediction?: Prediction;\n reference?: Reference;\n run: Run;\n example?: Example;\n};\n\n/**\n * Type of a function that can be coerced into a RunEvaluator function.\n * While we have the class-based RunEvaluator, it's often more convenient to directly\n * pass a function to the runner. This type allows us to do that.\n */\nexport type RunEvaluatorLike =\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => Promise<EvaluationResult>)\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => EvaluationResult);\n\nexport function isOffTheShelfEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is T | EvalConfig {\n return typeof evaluator === \"string\" || \"evaluatorType\" in evaluator;\n}\n\nexport function isCustomEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is U {\n return !isOffTheShelfEvaluator(evaluator);\n}\n\nexport type RunEvalType<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = T | EvalConfig | U;\n\n/**\n * Configuration class for running evaluations on datasets.\n *\n * @remarks\n * RunEvalConfig in LangSmith is a configuration class for running evaluations on datasets. Its primary purpose is to define the parameters and evaluators that will be applied during the evaluation of a dataset. This configuration can include various evaluators, custom evaluators, and different keys for inputs, predictions, and references.\n *\n * @typeparam T - The type of evaluators.\n * @typeparam U - The type of custom evaluators.\n */\nexport type RunEvalConfig<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = {\n /**\n * Evaluators to apply to a dataset run.\n * You can optionally specify these by name, or by\n * configuring them with an EvalConfig object.\n */\n evaluators?: RunEvalType<T, U>[];\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs?: EvaluatorInputFormatter;\n\n /**\n * Custom evaluators to apply to a dataset run.\n * Each evaluator is provided with a run trace containing the model\n * outputs, as well as an \"example\" object representing a record\n * in the dataset.\n *\n * @deprecated Use `evaluators` instead.\n */\n customEvaluators?: U[];\n};\n\nexport interface EvalConfig extends LoadEvaluatorOptions {\n /**\n * The name of the evaluator to use.\n * Example: labeled_criteria, criteria, etc.\n */\n evaluatorType: keyof EvaluatorType;\n\n /**\n * The feedback (or metric) name to use for the logged\n * evaluation results. If none provided, we default to\n * the evaluationName.\n */\n feedbackKey?: string;\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs: EvaluatorInputFormatter;\n}\n\nconst isStringifiableValue = (\n value: unknown\n): value is string | number | boolean | bigint =>\n typeof value === \"string\" ||\n typeof value === \"number\" ||\n typeof value === \"boolean\" ||\n typeof value === \"bigint\";\n\nconst getSingleStringifiedValue = (value: unknown) => {\n if (isStringifiableValue(value)) {\n return `${value}`;\n }\n\n if (typeof value === \"object\" && value != null && !Array.isArray(value)) {\n const entries = Object.entries(value);\n\n if (entries.length === 1 && isStringifiableValue(entries[0][1])) {\n return `${entries[0][1]}`;\n }\n }\n\n console.warn(\"Non-stringifiable value found when coercing\", value);\n return `${value}`;\n};\n\n/**\n * Configuration to load a \"CriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [Criteria(\"helpfulness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * Criteria({\n * \"isCompliant\": \"Does the submission comply with the requirements of XYZ\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: \"helpfulness\"\n * formatEvaluatorInputs: ...\n * }]\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: { \"isCompliant\": \"Does the submission comply with the requirements of XYZ\" },\n * formatEvaluatorInputs: ...\n * }]\n * };\n */\nexport type Criteria = EvalConfig & {\n evaluatorType: \"criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/criteria-evaluator\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\n// for compatibility reasons\nexport type CriteriaEvalChainConfig = Criteria;\n\nexport function Criteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): EvalConfig {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"LabeledCriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria and also\n * provides a \"ground truth\" label for the evaluator to incorporate\n * in its evaluation.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [LabeledCriteria(\"correctness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * LabeledCriteria({\n * \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: \"correctness\",\n * formatEvaluatorInputs: ...\n * }],\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: { \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\" },\n * formatEvaluatorInputs: ...\n * }],\n * };\n */\nexport type LabeledCriteria = EvalConfig & {\n evaluatorType: \"labeled_criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/labeled-criteria\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\nexport function LabeledCriteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): LabeledCriteria {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one labeled criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"labeled_criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"EmbeddingDistanceEvalChain\" evaluator,\n * which embeds distances to score semantic difference between\n * a prediction and reference.\n */\nexport type EmbeddingDistance = EvalConfig &\n EmbeddingDistanceEvalChainInput & { evaluatorType: \"embedding_distance\" };\n\nexport function EmbeddingDistance(\n distanceMetric: EmbeddingDistanceEvalChainInput[\"distanceMetric\"],\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"embedding\" | \"feedbackKey\"\n >\n): EmbeddingDistance {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n return {\n evaluatorType: \"embedding_distance\",\n embedding: config?.embedding,\n distanceMetric,\n feedbackKey: config?.feedbackKey ?? \"embedding_distance\",\n formatEvaluatorInputs,\n };\n}\n"],"mappings":";AA8DA,SAAgB,uBAGd,WAA4D;AAC5D,QAAO,OAAO,cAAc,YAAY,mBAAmB;;AAG7D,SAAgB,kBAGd,WAA+C;AAC/C,QAAO,CAAC,uBAAuB,UAAU;;AA8G3C,MAAM,wBACJ,UAEA,OAAO,UAAU,YACjB,OAAO,UAAU,YACjB,OAAO,UAAU,aACjB,OAAO,UAAU;AAEnB,MAAM,6BAA6B,UAAmB;AACpD,KAAI,qBAAqB,MAAM,CAC7B,QAAO,GAAG;AAGZ,KAAI,OAAO,UAAU,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,MAAM,EAAE;EACvE,MAAM,UAAU,OAAO,QAAQ,MAAM;AAErC,MAAI,QAAQ,WAAW,KAAK,qBAAqB,QAAQ,GAAG,GAAG,CAC7D,QAAO,GAAG,QAAQ,GAAG;;AAIzB,SAAQ,KAAK,+CAA+C,MAAM;AAClE,QAAO,GAAG;;AAgEZ,SAAgB,SACd,UACA,QAIY;CACZ,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EACnD;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,oEACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AA+DH,SAAgB,gBACd,UACA,QAIiB;CACjB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EAClD,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,4EACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AAWH,SAAgB,kBACd,gBACA,QAImB;CACnB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,QAAO;EACL,eAAe;EACf,WAAW,QAAQ;EACnB;EACA,aAAa,QAAQ,eAAe;EACpC;EACD"}
1
+ {"version":3,"file":"config.cjs","names":[],"sources":["../../src/smith/config.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { RunnableConfig } from \"@langchain/core/runnables\";\nimport { Example, Run } from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport {\n Criteria as CriteriaType,\n type EmbeddingDistanceEvalChainInput,\n} from \"../evaluation/index.js\";\nimport { LoadEvaluatorOptions } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\n\nexport type EvaluatorInputs = {\n input?: string | unknown;\n prediction: string | unknown;\n reference?: string | unknown;\n};\n\nexport type EvaluatorInputFormatter = ({\n rawInput,\n rawPrediction,\n rawReferenceOutput,\n run,\n}: {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawInput: any;\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawPrediction: any;\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawReferenceOutput?: any;\n run: Run;\n}) => EvaluatorInputs;\n\nexport type DynamicRunEvaluatorParams<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Input extends Record<string, any> = Record<string, unknown>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Prediction extends Record<string, any> = Record<string, unknown>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Reference extends Record<string, any> = Record<string, unknown>,\n> = {\n input: Input;\n prediction?: Prediction;\n reference?: Reference;\n run: Run;\n example?: Example;\n};\n\n/**\n * Type of a function that can be coerced into a RunEvaluator function.\n * While we have the class-based RunEvaluator, it's often more convenient to directly\n * pass a function to the runner. This type allows us to do that.\n */\nexport type RunEvaluatorLike =\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => Promise<EvaluationResult>)\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => EvaluationResult);\n\nexport function isOffTheShelfEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is T | EvalConfig {\n return typeof evaluator === \"string\" || \"evaluatorType\" in evaluator;\n}\n\nexport function isCustomEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is U {\n return !isOffTheShelfEvaluator(evaluator);\n}\n\nexport type RunEvalType<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = T | EvalConfig | U;\n\n/**\n * Configuration class for running evaluations on datasets.\n *\n * @remarks\n * RunEvalConfig in LangSmith is a configuration class for running evaluations on datasets. Its primary purpose is to define the parameters and evaluators that will be applied during the evaluation of a dataset. This configuration can include various evaluators, custom evaluators, and different keys for inputs, predictions, and references.\n *\n * @typeparam T - The type of evaluators.\n * @typeparam U - The type of custom evaluators.\n */\nexport type RunEvalConfig<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = {\n /**\n * Evaluators to apply to a dataset run.\n * You can optionally specify these by name, or by\n * configuring them with an EvalConfig object.\n */\n evaluators?: RunEvalType<T, U>[];\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs?: EvaluatorInputFormatter;\n\n /**\n * Custom evaluators to apply to a dataset run.\n * Each evaluator is provided with a run trace containing the model\n * outputs, as well as an \"example\" object representing a record\n * in the dataset.\n *\n * @deprecated Use `evaluators` instead.\n */\n customEvaluators?: U[];\n};\n\nexport interface EvalConfig extends LoadEvaluatorOptions {\n /**\n * The name of the evaluator to use.\n * Example: labeled_criteria, criteria, etc.\n */\n evaluatorType: keyof EvaluatorType;\n\n /**\n * The feedback (or metric) name to use for the logged\n * evaluation results. If none provided, we default to\n * the evaluationName.\n */\n feedbackKey?: string;\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs: EvaluatorInputFormatter;\n}\n\nconst isStringifiableValue = (\n value: unknown\n): value is string | number | boolean | bigint =>\n typeof value === \"string\" ||\n typeof value === \"number\" ||\n typeof value === \"boolean\" ||\n typeof value === \"bigint\";\n\nconst getSingleStringifiedValue = (value: unknown) => {\n if (isStringifiableValue(value)) {\n return `${value}`;\n }\n\n if (typeof value === \"object\" && value != null && !Array.isArray(value)) {\n const entries = Object.entries(value);\n\n if (entries.length === 1 && isStringifiableValue(entries[0][1])) {\n return `${entries[0][1]}`;\n }\n }\n\n console.warn(\"Non-stringifiable value found when coercing\", value);\n return `${value}`;\n};\n\n/**\n * Configuration to load a \"CriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [Criteria(\"helpfulness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * Criteria({\n * \"isCompliant\": \"Does the submission comply with the requirements of XYZ\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: \"helpfulness\"\n * formatEvaluatorInputs: ...\n * }]\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: { \"isCompliant\": \"Does the submission comply with the requirements of XYZ\" },\n * formatEvaluatorInputs: ...\n * }]\n * };\n */\nexport type Criteria = EvalConfig & {\n evaluatorType: \"criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/criteria-evaluator\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\n// for compatibility reasons\nexport type CriteriaEvalChainConfig = Criteria;\n\nexport function Criteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): EvalConfig {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"LabeledCriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria and also\n * provides a \"ground truth\" label for the evaluator to incorporate\n * in its evaluation.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [LabeledCriteria(\"correctness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * LabeledCriteria({\n * \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: \"correctness\",\n * formatEvaluatorInputs: ...\n * }],\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: { \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\" },\n * formatEvaluatorInputs: ...\n * }],\n * };\n */\nexport type LabeledCriteria = EvalConfig & {\n evaluatorType: \"labeled_criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/labeled-criteria\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\nexport function LabeledCriteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): LabeledCriteria {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one labeled criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"labeled_criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"EmbeddingDistanceEvalChain\" evaluator,\n * which embeds distances to score semantic difference between\n * a prediction and reference.\n */\nexport type EmbeddingDistance = EvalConfig &\n EmbeddingDistanceEvalChainInput & { evaluatorType: \"embedding_distance\" };\n\nexport function EmbeddingDistance(\n distanceMetric: EmbeddingDistanceEvalChainInput[\"distanceMetric\"],\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"embedding\" | \"feedbackKey\"\n >\n): EmbeddingDistance {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n return {\n evaluatorType: \"embedding_distance\",\n embedding: config?.embedding,\n distanceMetric,\n feedbackKey: config?.feedbackKey ?? \"embedding_distance\",\n formatEvaluatorInputs,\n };\n}\n"],"mappings":";AA8DA,SAAgB,uBAGd,WAA4D;AAC5D,QAAO,OAAO,cAAc,YAAY,mBAAmB;;AAG7D,SAAgB,kBAGd,WAA+C;AAC/C,QAAO,CAAC,uBAAuB,UAAU;;AA8G3C,MAAM,wBACJ,UAEA,OAAO,UAAU,YACjB,OAAO,UAAU,YACjB,OAAO,UAAU,aACjB,OAAO,UAAU;AAEnB,MAAM,6BAA6B,UAAmB;AACpD,KAAI,qBAAqB,MAAM,CAC7B,QAAO,GAAG;AAGZ,KAAI,OAAO,UAAU,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,MAAM,EAAE;EACvE,MAAM,UAAU,OAAO,QAAQ,MAAM;AAErC,MAAI,QAAQ,WAAW,KAAK,qBAAqB,QAAQ,GAAG,GAAG,CAC7D,QAAO,GAAG,QAAQ,GAAG;;AAIzB,SAAQ,KAAK,+CAA+C,MAAM;AAClE,QAAO,GAAG;;AAgEZ,SAAgB,SACd,UACA,QAIY;CACZ,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EACnD;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,oEACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AA+DH,SAAgB,gBACd,UACA,QAIiB;CACjB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EAClD,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,4EACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AAWH,SAAgB,kBACd,gBACA,QAImB;CACnB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,QAAO;EACL,eAAe;EACf,WAAW,QAAQ;EACnB;EACA,aAAa,QAAQ,eAAe;EACpC;EACD"}
@@ -2,10 +2,10 @@ import { Criteria as Criteria$1 } from "../evaluation/criteria/criteria.cjs";
2
2
  import { EmbeddingDistanceEvalChainInput } from "../evaluation/embedding_distance/base.cjs";
3
3
  import { EvaluatorType } from "../evaluation/types.cjs";
4
4
  import { LoadEvaluatorOptions } from "../evaluation/loader.cjs";
5
- import { EvaluationResult, RunEvaluator } from "../node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.cjs";
6
5
  import { RunnableConfig } from "@langchain/core/runnables";
7
6
  import { BaseLanguageModel } from "@langchain/core/language_models/base";
8
7
  import { Example, Run } from "langsmith";
8
+ import { EvaluationResult, RunEvaluator } from "langsmith/evaluation";
9
9
 
10
10
  //#region src/smith/config.d.ts
11
11
  type EvaluatorInputs = {
@@ -2,10 +2,10 @@ import { Criteria as Criteria$1 } from "../evaluation/criteria/criteria.js";
2
2
  import { EmbeddingDistanceEvalChainInput } from "../evaluation/embedding_distance/base.js";
3
3
  import { EvaluatorType } from "../evaluation/types.js";
4
4
  import { LoadEvaluatorOptions } from "../evaluation/loader.js";
5
- import { EvaluationResult, RunEvaluator } from "../node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.js";
6
5
  import { RunnableConfig } from "@langchain/core/runnables";
7
6
  import { BaseLanguageModel } from "@langchain/core/language_models/base";
8
7
  import { Example, Run } from "langsmith";
8
+ import { EvaluationResult, RunEvaluator } from "langsmith/evaluation";
9
9
 
10
10
  //#region src/smith/config.d.ts
11
11
  type EvaluatorInputs = {
@@ -1 +1 @@
1
- {"version":3,"file":"config.d.ts","names":[],"sources":["../../src/smith/config.ts"],"mappings":";;;;;;;;;;KAWY,eAAA;EACV,KAAA;EACA,UAAA;EACA,SAAA;AAAA;AAAA,KAGU,uBAAA;EACV,QAAA;EACA,aAAA;EACA,kBAAA;EACA;AAAA;EAGA,QAAA;EAEA,aAAA;EAEA,kBAAA;EACA,GAAA,EAAK,GAAA;AAAA,MACD,eAAA;AAAA,KAEM,yBAAA,eAEI,MAAA,gBAAsB,MAAA,sCAEjB,MAAA,gBAAsB,MAAA,qCAEvB,MAAA,gBAAsB,MAAA;EAExC,KAAA,EAAO,KAAA;EACP,UAAA,GAAa,UAAA;EACb,SAAA,GAAY,SAAA;EACZ,GAAA,EAAK,GAAA;EACL,OAAA,GAAU,OAAA;AAAA;;;;;;KAQA,gBAAA,KAEN,KAAA,EAAO,yBAAA,EACP,OAAA,EAAS,cAAA,KACN,OAAA,CAAQ,gBAAA,OAEX,KAAA,EAAO,yBAAA,EACP,OAAA,EAAS,cAAA,KACN,gBAAA;AAAA,iBAEO,sBAAA,iBACE,aAAA,YACN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,CAAA,CAC3D,SAAA,EAAW,CAAA,GAAI,UAAA,GAAa,CAAA,GAAI,SAAA,IAAa,CAAA,GAAI,UAAA;AAAA,iBAInC,iBAAA,iBACE,aAAA,YACN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,CAAA,CAC3D,SAAA,EAAW,CAAA,GAAI,UAAA,GAAa,CAAA,GAAI,SAAA,IAAa,CAAA;AAAA,KAInC,WAAA,iBACM,aAAA,qEAIN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,IACzD,CAAA,GAAI,UAAA,GAAa,CAAA;;;;;;;;;;KAWT,aAAA,iBACM,aAAA,qEAIN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA;EArEtD;;;;AAGP;EAyEE,UAAA,GAAa,WAAA,CAAY,CAAA,EAAG,CAAA;EAzEO;;;;;;;;;;;;;;;;;;;;;;;EAkGnC,qBAAA,GAAwB,uBAAA;EA1FjB;;;;;;;;EAoGP,gBAAA,GAAmB,CAAA;AAAA;AAAA,UAGJ,UAAA,SAAmB,oBAAA;EA3FxB;;;;EAgGV,aAAA,QAAqB,aAAA;EA5FN;;;;;EAmGf,WAAA;EA/FuB;;;;;;;;;;;;;;AAEzB;;;;;;;;;EAsHE,qBAAA,EAAuB,uBAAA;AAAA;;;;;;;;;;;;;;;;;;;;;AA/GzB;;;;;;;;;;;;;;;;;;;;;KAoLY,QAAA,GAAW,UAAA;EACrB,aAAA;EAlL4B;;;;;;EA0L5B,QAAA,GAAW,UAAA,GAAe,MAAA;EAtLL;;;EA2LrB,GAAA,GAAM,iBAAA;AAAA;AAAA,KAII,uBAAA,GAA0B,QAAA;AAAA,iBAEtB,QAAA,CACd,QAAA,EAAU,UAAA,GAAe,MAAA,kBACzB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,sDAGT,UAAA;;;;;;;;;;;;;;;;;;AAtLH;;;;;;;;;;;;;;;;;;;;;;;;;;KA2PY,eAAA,GAAkB,UAAA;EAC5B,aAAA;EA7MA;;;;AAGF;;EAkNE,QAAA,GAAW,UAAA,GAAe,MAAA;EA7ML;;;EAkNrB,GAAA,GAAM,iBAAA;AAAA;AAAA,iBAGQ,eAAA,CACd,QAAA,EAAU,UAAA,GAAe,MAAA,kBACzB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,sDAGT,eAAA;;;;;;KAgCS,iBAAA,GAAoB,UAAA,GAC9B,+BAAA;EAAoC,aAAA;AAAA;AAAA,iBAEtB,iBAAA,CACd,cAAA,EAAgB,+BAAA,oBAChB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,4DAGT,iBAAA"}
1
+ {"version":3,"file":"config.d.ts","names":[],"sources":["../../src/smith/config.ts"],"mappings":";;;;;;;;;;KAWY,eAAA;EACV,KAAA;EACA,UAAA;EACA,SAAA;AAAA;AAAA,KAGU,uBAAA;EACV,QAAA;EACA,aAAA;EACA,kBAAA;EACA;AAAA;EAGA,QAAA;EAEA,aAAA;EAEA,kBAAA;EACA,GAAA,EAAK,GAAA;AAAA,MACD,eAAA;AAAA,KAEM,yBAAA,eAEI,MAAA,gBAAsB,MAAA,sCAEjB,MAAA,gBAAsB,MAAA,qCAEvB,MAAA,gBAAsB,MAAA;EAExC,KAAA,EAAO,KAAA;EACP,UAAA,GAAa,UAAA;EACb,SAAA,GAAY,SAAA;EACZ,GAAA,EAAK,GAAA;EACL,OAAA,GAAU,OAAA;AAAA;;;;;;KAQA,gBAAA,KAEN,KAAA,EAAO,yBAAA,EACP,OAAA,EAAS,cAAA,KACN,OAAA,CAAQ,gBAAA,OAEX,KAAA,EAAO,yBAAA,EACP,OAAA,EAAS,cAAA,KACN,gBAAA;AAAA,iBAEO,sBAAA,iBACE,aAAA,YACN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,CAAA,CAC3D,SAAA,EAAW,CAAA,GAAI,UAAA,GAAa,CAAA,GAAI,SAAA,IAAa,CAAA,GAAI,UAAA;AAAA,iBAInC,iBAAA,iBACE,aAAA,YACN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,CAAA,CAC3D,SAAA,EAAW,CAAA,GAAI,UAAA,GAAa,CAAA,GAAI,SAAA,IAAa,CAAA;AAAA,KAInC,WAAA,iBACM,aAAA,qEAIN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA,IACzD,CAAA,GAAI,UAAA,GAAa,CAAA;;;;;;;;;;KAWT,aAAA,iBACM,aAAA,qEAIN,YAAA,GAAe,gBAAA,GAAmB,YAAA,GAAe,gBAAA;EApEvD;;;AAEN;;EAyEE,UAAA,GAAa,WAAA,CAAY,CAAA,EAAG,CAAA;EAvEd;;;;;;;;;;;;;;;;;;;;;;;EAgGd,qBAAA,GAAwB,uBAAA;EAzFxB;;;;;;;;EAmGA,gBAAA,GAAmB,CAAA;AAAA;AAAA,UAGJ,UAAA,SAAmB,oBAAA;EA3FR;;;;EAgG1B,aAAA,QAAqB,aAAA;EA5Fd;;;;;EAmGP,WAAA;EArGW;;;;;;;;;;;;;AAQb;;;;;;;;;;EAsHE,qBAAA,EAAuB,uBAAA;AAAA;;;;;;;;;;;;;;;;;;;;AA/GzB;;;;;;;;;;;;;;;;;;;;;;KAoLY,QAAA,GAAW,UAAA;EACrB,aAAA;EAlLA;;;;;AAIF;EAsLE,QAAA,GAAW,UAAA,GAAe,MAAA;EAtLL;;;EA2LrB,GAAA,GAAM,iBAAA;AAAA;AAAA,KAII,uBAAA,GAA0B,QAAA;AAAA,iBAEtB,QAAA,CACd,QAAA,EAAU,UAAA,GAAe,MAAA,kBACzB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,sDAGT,UAAA;;;;;;;;;;;;;;;;;AAtLH;;;;;;;;;;;;;;;;;;;;;;;;;;;KA2PY,eAAA,GAAkB,UAAA;EAC5B,aAAA;EA7MmB;;;AAGrB;;;EAkNE,QAAA,GAAW,UAAA,GAAe,MAAA;EA7KH;;;EAkLvB,GAAA,GAAM,iBAAA;AAAA;AAAA,iBAGQ,eAAA,CACd,QAAA,EAAU,UAAA,GAAe,MAAA,kBACzB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,sDAGT,eAAA;;;;;;KAgCS,iBAAA,GAAoB,UAAA,GAC9B,+BAAA;EAAoC,aAAA;AAAA;AAAA,iBAEtB,iBAAA,CACd,cAAA,EAAgB,+BAAA,oBAChB,MAAA,GAAS,IAAA,CACP,OAAA,CAAQ,eAAA,4DAGT,iBAAA"}
@@ -1 +1 @@
1
- {"version":3,"file":"config.js","names":[],"sources":["../../src/smith/config.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { RunnableConfig } from \"@langchain/core/runnables\";\nimport { Example, Run } from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport {\n Criteria as CriteriaType,\n type EmbeddingDistanceEvalChainInput,\n} from \"../evaluation/index.js\";\nimport { LoadEvaluatorOptions } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\n\nexport type EvaluatorInputs = {\n input?: string | unknown;\n prediction: string | unknown;\n reference?: string | unknown;\n};\n\nexport type EvaluatorInputFormatter = ({\n rawInput,\n rawPrediction,\n rawReferenceOutput,\n run,\n}: {\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawInput: any;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawPrediction: any;\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n rawReferenceOutput?: any;\n run: Run;\n}) => EvaluatorInputs;\n\nexport type DynamicRunEvaluatorParams<\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Input extends Record<string, any> = Record<string, unknown>,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Prediction extends Record<string, any> = Record<string, unknown>,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n Reference extends Record<string, any> = Record<string, unknown>,\n> = {\n input: Input;\n prediction?: Prediction;\n reference?: Reference;\n run: Run;\n example?: Example;\n};\n\n/**\n * Type of a function that can be coerced into a RunEvaluator function.\n * While we have the class-based RunEvaluator, it's often more convenient to directly\n * pass a function to the runner. This type allows us to do that.\n */\nexport type RunEvaluatorLike =\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => Promise<EvaluationResult>)\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => EvaluationResult);\n\nexport function isOffTheShelfEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is T | EvalConfig {\n return typeof evaluator === \"string\" || \"evaluatorType\" in evaluator;\n}\n\nexport function isCustomEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is U {\n return !isOffTheShelfEvaluator(evaluator);\n}\n\nexport type RunEvalType<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = T | EvalConfig | U;\n\n/**\n * Configuration class for running evaluations on datasets.\n *\n * @remarks\n * RunEvalConfig in LangSmith is a configuration class for running evaluations on datasets. Its primary purpose is to define the parameters and evaluators that will be applied during the evaluation of a dataset. This configuration can include various evaluators, custom evaluators, and different keys for inputs, predictions, and references.\n *\n * @typeparam T - The type of evaluators.\n * @typeparam U - The type of custom evaluators.\n */\nexport type RunEvalConfig<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = {\n /**\n * Evaluators to apply to a dataset run.\n * You can optionally specify these by name, or by\n * configuring them with an EvalConfig object.\n */\n evaluators?: RunEvalType<T, U>[];\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs?: EvaluatorInputFormatter;\n\n /**\n * Custom evaluators to apply to a dataset run.\n * Each evaluator is provided with a run trace containing the model\n * outputs, as well as an \"example\" object representing a record\n * in the dataset.\n *\n * @deprecated Use `evaluators` instead.\n */\n customEvaluators?: U[];\n};\n\nexport interface EvalConfig extends LoadEvaluatorOptions {\n /**\n * The name of the evaluator to use.\n * Example: labeled_criteria, criteria, etc.\n */\n evaluatorType: keyof EvaluatorType;\n\n /**\n * The feedback (or metric) name to use for the logged\n * evaluation results. If none provided, we default to\n * the evaluationName.\n */\n feedbackKey?: string;\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs: EvaluatorInputFormatter;\n}\n\nconst isStringifiableValue = (\n value: unknown\n): value is string | number | boolean | bigint =>\n typeof value === \"string\" ||\n typeof value === \"number\" ||\n typeof value === \"boolean\" ||\n typeof value === \"bigint\";\n\nconst getSingleStringifiedValue = (value: unknown) => {\n if (isStringifiableValue(value)) {\n return `${value}`;\n }\n\n if (typeof value === \"object\" && value != null && !Array.isArray(value)) {\n const entries = Object.entries(value);\n\n if (entries.length === 1 && isStringifiableValue(entries[0][1])) {\n return `${entries[0][1]}`;\n }\n }\n\n console.warn(\"Non-stringifiable value found when coercing\", value);\n return `${value}`;\n};\n\n/**\n * Configuration to load a \"CriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [Criteria(\"helpfulness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * Criteria({\n * \"isCompliant\": \"Does the submission comply with the requirements of XYZ\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: \"helpfulness\"\n * formatEvaluatorInputs: ...\n * }]\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: { \"isCompliant\": \"Does the submission comply with the requirements of XYZ\" },\n * formatEvaluatorInputs: ...\n * }]\n * };\n */\nexport type Criteria = EvalConfig & {\n evaluatorType: \"criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/criteria-evaluator\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\n// for compatibility reasons\nexport type CriteriaEvalChainConfig = Criteria;\n\nexport function Criteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): EvalConfig {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"LabeledCriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria and also\n * provides a \"ground truth\" label for the evaluator to incorporate\n * in its evaluation.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [LabeledCriteria(\"correctness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * LabeledCriteria({\n * \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: \"correctness\",\n * formatEvaluatorInputs: ...\n * }],\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: { \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\" },\n * formatEvaluatorInputs: ...\n * }],\n * };\n */\nexport type LabeledCriteria = EvalConfig & {\n evaluatorType: \"labeled_criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/labeled-criteria\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\nexport function LabeledCriteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): LabeledCriteria {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one labeled criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"labeled_criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"EmbeddingDistanceEvalChain\" evaluator,\n * which embeds distances to score semantic difference between\n * a prediction and reference.\n */\nexport type EmbeddingDistance = EvalConfig &\n EmbeddingDistanceEvalChainInput & { evaluatorType: \"embedding_distance\" };\n\nexport function EmbeddingDistance(\n distanceMetric: EmbeddingDistanceEvalChainInput[\"distanceMetric\"],\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"embedding\" | \"feedbackKey\"\n >\n): EmbeddingDistance {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n return {\n evaluatorType: \"embedding_distance\",\n embedding: config?.embedding,\n distanceMetric,\n feedbackKey: config?.feedbackKey ?? \"embedding_distance\",\n formatEvaluatorInputs,\n };\n}\n"],"mappings":";AA8DA,SAAgB,uBAGd,WAA4D;AAC5D,QAAO,OAAO,cAAc,YAAY,mBAAmB;;AAG7D,SAAgB,kBAGd,WAA+C;AAC/C,QAAO,CAAC,uBAAuB,UAAU;;AA8G3C,MAAM,wBACJ,UAEA,OAAO,UAAU,YACjB,OAAO,UAAU,YACjB,OAAO,UAAU,aACjB,OAAO,UAAU;AAEnB,MAAM,6BAA6B,UAAmB;AACpD,KAAI,qBAAqB,MAAM,CAC7B,QAAO,GAAG;AAGZ,KAAI,OAAO,UAAU,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,MAAM,EAAE;EACvE,MAAM,UAAU,OAAO,QAAQ,MAAM;AAErC,MAAI,QAAQ,WAAW,KAAK,qBAAqB,QAAQ,GAAG,GAAG,CAC7D,QAAO,GAAG,QAAQ,GAAG;;AAIzB,SAAQ,KAAK,+CAA+C,MAAM;AAClE,QAAO,GAAG;;AAgEZ,SAAgB,SACd,UACA,QAIY;CACZ,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EACnD;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,oEACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AA+DH,SAAgB,gBACd,UACA,QAIiB;CACjB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EAClD,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,4EACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AAWH,SAAgB,kBACd,gBACA,QAImB;CACnB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,QAAO;EACL,eAAe;EACf,WAAW,QAAQ;EACnB;EACA,aAAa,QAAQ,eAAe;EACpC;EACD"}
1
+ {"version":3,"file":"config.js","names":[],"sources":["../../src/smith/config.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { RunnableConfig } from \"@langchain/core/runnables\";\nimport { Example, Run } from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport {\n Criteria as CriteriaType,\n type EmbeddingDistanceEvalChainInput,\n} from \"../evaluation/index.js\";\nimport { LoadEvaluatorOptions } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\n\nexport type EvaluatorInputs = {\n input?: string | unknown;\n prediction: string | unknown;\n reference?: string | unknown;\n};\n\nexport type EvaluatorInputFormatter = ({\n rawInput,\n rawPrediction,\n rawReferenceOutput,\n run,\n}: {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawInput: any;\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawPrediction: any;\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n rawReferenceOutput?: any;\n run: Run;\n}) => EvaluatorInputs;\n\nexport type DynamicRunEvaluatorParams<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Input extends Record<string, any> = Record<string, unknown>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Prediction extends Record<string, any> = Record<string, unknown>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n Reference extends Record<string, any> = Record<string, unknown>,\n> = {\n input: Input;\n prediction?: Prediction;\n reference?: Reference;\n run: Run;\n example?: Example;\n};\n\n/**\n * Type of a function that can be coerced into a RunEvaluator function.\n * While we have the class-based RunEvaluator, it's often more convenient to directly\n * pass a function to the runner. This type allows us to do that.\n */\nexport type RunEvaluatorLike =\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => Promise<EvaluationResult>)\n | ((\n props: DynamicRunEvaluatorParams,\n options: RunnableConfig\n ) => EvaluationResult);\n\nexport function isOffTheShelfEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is T | EvalConfig {\n return typeof evaluator === \"string\" || \"evaluatorType\" in evaluator;\n}\n\nexport function isCustomEvaluator<\n T extends keyof EvaluatorType,\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n>(evaluator: T | EvalConfig | U): evaluator is U {\n return !isOffTheShelfEvaluator(evaluator);\n}\n\nexport type RunEvalType<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = T | EvalConfig | U;\n\n/**\n * Configuration class for running evaluations on datasets.\n *\n * @remarks\n * RunEvalConfig in LangSmith is a configuration class for running evaluations on datasets. Its primary purpose is to define the parameters and evaluators that will be applied during the evaluation of a dataset. This configuration can include various evaluators, custom evaluators, and different keys for inputs, predictions, and references.\n *\n * @typeparam T - The type of evaluators.\n * @typeparam U - The type of custom evaluators.\n */\nexport type RunEvalConfig<\n T extends keyof EvaluatorType =\n | \"criteria\"\n | \"labeled_criteria\"\n | \"embedding_distance\",\n U extends RunEvaluator | RunEvaluatorLike = RunEvaluator | RunEvaluatorLike,\n> = {\n /**\n * Evaluators to apply to a dataset run.\n * You can optionally specify these by name, or by\n * configuring them with an EvalConfig object.\n */\n evaluators?: RunEvalType<T, U>[];\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs?: EvaluatorInputFormatter;\n\n /**\n * Custom evaluators to apply to a dataset run.\n * Each evaluator is provided with a run trace containing the model\n * outputs, as well as an \"example\" object representing a record\n * in the dataset.\n *\n * @deprecated Use `evaluators` instead.\n */\n customEvaluators?: U[];\n};\n\nexport interface EvalConfig extends LoadEvaluatorOptions {\n /**\n * The name of the evaluator to use.\n * Example: labeled_criteria, criteria, etc.\n */\n evaluatorType: keyof EvaluatorType;\n\n /**\n * The feedback (or metric) name to use for the logged\n * evaluation results. If none provided, we default to\n * the evaluationName.\n */\n feedbackKey?: string;\n\n /**\n * Convert the evaluation data into formats that can be used by the evaluator.\n * This should most commonly be a string.\n * Parameters are the raw input from the run, the raw output, raw reference output, and the raw run.\n * @example\n * ```ts\n * // Chain input: { input: \"some string\" }\n * // Chain output: { output: \"some output\" }\n * // Reference example output format: { output: \"some reference output\" }\n * const formatEvaluatorInputs = ({\n * rawInput,\n * rawPrediction,\n * rawReferenceOutput,\n * }) => {\n * return {\n * input: rawInput.input,\n * prediction: rawPrediction.output,\n * reference: rawReferenceOutput.output,\n * };\n * };\n * ```\n * @returns The prepared data.\n */\n formatEvaluatorInputs: EvaluatorInputFormatter;\n}\n\nconst isStringifiableValue = (\n value: unknown\n): value is string | number | boolean | bigint =>\n typeof value === \"string\" ||\n typeof value === \"number\" ||\n typeof value === \"boolean\" ||\n typeof value === \"bigint\";\n\nconst getSingleStringifiedValue = (value: unknown) => {\n if (isStringifiableValue(value)) {\n return `${value}`;\n }\n\n if (typeof value === \"object\" && value != null && !Array.isArray(value)) {\n const entries = Object.entries(value);\n\n if (entries.length === 1 && isStringifiableValue(entries[0][1])) {\n return `${entries[0][1]}`;\n }\n }\n\n console.warn(\"Non-stringifiable value found when coercing\", value);\n return `${value}`;\n};\n\n/**\n * Configuration to load a \"CriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [Criteria(\"helpfulness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * Criteria({\n * \"isCompliant\": \"Does the submission comply with the requirements of XYZ\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: \"helpfulness\"\n * formatEvaluatorInputs: ...\n * }]\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"criteria\",\n * criteria: { \"isCompliant\": \"Does the submission comply with the requirements of XYZ\" },\n * formatEvaluatorInputs: ...\n * }]\n * };\n */\nexport type Criteria = EvalConfig & {\n evaluatorType: \"criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/criteria-evaluator\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\n// for compatibility reasons\nexport type CriteriaEvalChainConfig = Criteria;\n\nexport function Criteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): EvalConfig {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"LabeledCriteriaEvalChain\" evaluator,\n * which prompts an LLM to determine whether the model's\n * prediction complies with the provided criteria and also\n * provides a \"ground truth\" label for the evaluator to incorporate\n * in its evaluation.\n * @param criteria - The criteria to use for the evaluator.\n * @param llm - The language model to use for the evaluator.\n * @returns The configuration for the evaluator.\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [LabeledCriteria(\"correctness\")],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [\n * LabeledCriteria({\n * \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\"\n * })\n * ],\n * };\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: \"correctness\",\n * formatEvaluatorInputs: ...\n * }],\n * };\n * ```\n * @example\n * ```ts\n * const evalConfig = {\n * evaluators: [{\n * evaluatorType: \"labeled_criteria\",\n * criteria: { \"mentionsAllFacts\": \"Does the include all facts provided in the reference?\" },\n * formatEvaluatorInputs: ...\n * }],\n * };\n */\nexport type LabeledCriteria = EvalConfig & {\n evaluatorType: \"labeled_criteria\";\n\n /**\n * The \"criteria\" to insert into the prompt template\n * used for evaluation. See the prompt at\n * https://smith.langchain.com/hub/langchain-ai/labeled-criteria\n * for more information.\n */\n criteria?: CriteriaType | Record<string, string>;\n\n /**\n * The language model to use as the evaluator, defaults to GPT-4\n */\n llm?: BaseLanguageModel;\n};\n\nexport function LabeledCriteria(\n criteria: CriteriaType | Record<string, string>,\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"llm\" | \"feedbackKey\"\n >\n): LabeledCriteria {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n input: getSingleStringifiedValue(payload.rawInput),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n if (typeof criteria !== \"string\" && Object.keys(criteria).length !== 1) {\n throw new Error(\n \"Only one labeled criteria key is allowed when specifying custom criteria.\"\n );\n }\n\n const criteriaKey =\n typeof criteria === \"string\" ? criteria : Object.keys(criteria)[0];\n\n return {\n evaluatorType: \"labeled_criteria\",\n criteria,\n feedbackKey: config?.feedbackKey ?? criteriaKey,\n llm: config?.llm,\n formatEvaluatorInputs,\n };\n}\n\n/**\n * Configuration to load a \"EmbeddingDistanceEvalChain\" evaluator,\n * which embeds distances to score semantic difference between\n * a prediction and reference.\n */\nexport type EmbeddingDistance = EvalConfig &\n EmbeddingDistanceEvalChainInput & { evaluatorType: \"embedding_distance\" };\n\nexport function EmbeddingDistance(\n distanceMetric: EmbeddingDistanceEvalChainInput[\"distanceMetric\"],\n config?: Pick<\n Partial<LabeledCriteria>,\n \"formatEvaluatorInputs\" | \"embedding\" | \"feedbackKey\"\n >\n): EmbeddingDistance {\n const formatEvaluatorInputs =\n config?.formatEvaluatorInputs ??\n ((payload) => ({\n prediction: getSingleStringifiedValue(payload.rawPrediction),\n reference: getSingleStringifiedValue(payload.rawReferenceOutput),\n }));\n\n return {\n evaluatorType: \"embedding_distance\",\n embedding: config?.embedding,\n distanceMetric,\n feedbackKey: config?.feedbackKey ?? \"embedding_distance\",\n formatEvaluatorInputs,\n };\n}\n"],"mappings":";AA8DA,SAAgB,uBAGd,WAA4D;AAC5D,QAAO,OAAO,cAAc,YAAY,mBAAmB;;AAG7D,SAAgB,kBAGd,WAA+C;AAC/C,QAAO,CAAC,uBAAuB,UAAU;;AA8G3C,MAAM,wBACJ,UAEA,OAAO,UAAU,YACjB,OAAO,UAAU,YACjB,OAAO,UAAU,aACjB,OAAO,UAAU;AAEnB,MAAM,6BAA6B,UAAmB;AACpD,KAAI,qBAAqB,MAAM,CAC7B,QAAO,GAAG;AAGZ,KAAI,OAAO,UAAU,YAAY,SAAS,QAAQ,CAAC,MAAM,QAAQ,MAAM,EAAE;EACvE,MAAM,UAAU,OAAO,QAAQ,MAAM;AAErC,MAAI,QAAQ,WAAW,KAAK,qBAAqB,QAAQ,GAAG,GAAG,CAC7D,QAAO,GAAG,QAAQ,GAAG;;AAIzB,SAAQ,KAAK,+CAA+C,MAAM;AAClE,QAAO,GAAG;;AAgEZ,SAAgB,SACd,UACA,QAIY;CACZ,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EACnD;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,oEACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AA+DH,SAAgB,gBACd,UACA,QAIiB;CACjB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,OAAO,0BAA0B,QAAQ,SAAS;EAClD,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,KAAI,OAAO,aAAa,YAAY,OAAO,KAAK,SAAS,CAAC,WAAW,EACnE,OAAM,IAAI,MACR,4EACD;CAGH,MAAM,cACJ,OAAO,aAAa,WAAW,WAAW,OAAO,KAAK,SAAS,CAAC;AAElE,QAAO;EACL,eAAe;EACf;EACA,aAAa,QAAQ,eAAe;EACpC,KAAK,QAAQ;EACb;EACD;;AAWH,SAAgB,kBACd,gBACA,QAImB;CACnB,MAAM,wBACJ,QAAQ,2BACN,aAAa;EACb,YAAY,0BAA0B,QAAQ,cAAc;EAC5D,WAAW,0BAA0B,QAAQ,mBAAmB;EACjE;AAEH,QAAO;EACL,eAAe;EACf,WAAW,QAAQ;EACnB;EACA,aAAa,QAAQ,eAAe;EACpC;EACD"}
@@ -1 +1 @@
1
- {"version":3,"file":"runner_utils.cjs","names":["BaseTracer","RunnableLambda","LangChainTracer","RunTree","Runnable","loadEvaluator","isCustomEvaluator","isOffTheShelfEvaluator","ProgressBar","AsyncCaller","Client","randomName"],"sources":["../../src/smith/runner_utils.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { Serialized } from \"@langchain/core/load/serializable\";\nimport { mapStoredMessagesToChatMessages } from \"@langchain/core/messages\";\nimport {\n Runnable,\n RunnableConfig,\n RunnableLambda,\n getCallbackManagerForConfig,\n} from \"@langchain/core/runnables\";\nimport { LangChainTracer } from \"@langchain/core/tracers/tracer_langchain\";\nimport { BaseTracer } from \"@langchain/core/tracers/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { AsyncCaller } from \"@langchain/core/utils/async_caller\";\nimport type {\n CallbackManager,\n CallbackManagerForChainRun,\n} from \"@langchain/core/callbacks/manager\";\nimport {\n Client,\n Example,\n Feedback,\n Run,\n RunTree,\n RunTreeConfig,\n} from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport { DataType } from \"langsmith/schemas\";\nimport type { TraceableFunction } from \"langsmith/singletons/traceable\";\nimport { LLMStringEvaluator } from \"../evaluation/base.js\";\nimport { loadEvaluator } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\nimport {\n isOffTheShelfEvaluator,\n type DynamicRunEvaluatorParams,\n type EvalConfig,\n type EvaluatorInputFormatter,\n type RunEvalConfig,\n type RunEvaluatorLike,\n isCustomEvaluator,\n} from \"./config.js\";\nimport { randomName } from \"./name_generation.js\";\nimport { ProgressBar } from \"./progress.js\";\n\nexport type ChainOrFactory =\n | Runnable\n | (() => Runnable)\n | AnyTraceableFunction\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => any)\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => Promise<any>)\n | (() => (obj: unknown) => unknown)\n | (() => (obj: unknown) => Promise<unknown>);\n\nclass SingleRunIdExtractor {\n runIdPromiseResolver: (runId: string) => void;\n\n runIdPromise: Promise<string>;\n\n constructor() {\n this.runIdPromise = new Promise<string>((extract) => {\n this.runIdPromiseResolver = extract;\n });\n }\n\n handleChainStart = (\n _chain: Serialized,\n _inputs: ChainValues,\n runId: string\n ) => {\n this.runIdPromiseResolver(runId);\n };\n\n async extract(): Promise<string> {\n return this.runIdPromise;\n }\n}\n\nclass SingleRunExtractor extends BaseTracer {\n runPromiseResolver: (run: Run) => void;\n\n runPromise: Promise<Run>;\n\n /** The name of the callback handler. */\n name = \"single_run_extractor\";\n\n constructor() {\n super();\n this.runPromise = new Promise<Run>((extract) => {\n this.runPromiseResolver = extract;\n });\n }\n\n async persistRun(run: Run) {\n this.runPromiseResolver(run);\n }\n\n async extract(): Promise<Run> {\n return this.runPromise;\n }\n}\n\n/**\n * Wraps an evaluator function + implements the RunEvaluator interface.\n */\nclass DynamicRunEvaluator implements RunEvaluator {\n evaluator: RunnableLambda<DynamicRunEvaluatorParams, EvaluationResult>;\n\n constructor(evaluator: RunEvaluatorLike) {\n this.evaluator = new RunnableLambda({ func: evaluator });\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n const result = await this.evaluator.invoke(\n {\n run,\n example,\n input: run.inputs,\n prediction: run.outputs,\n reference: example?.outputs,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n sourceRunId: runId,\n ...result,\n };\n }\n}\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nfunction isLLMStringEvaluator(evaluator: any): evaluator is LLMStringEvaluator {\n return evaluator && typeof evaluator.evaluateStrings === \"function\";\n}\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AnyTraceableFunction = TraceableFunction<(...any: any[]) => any>;\n\n/**\n * Internal implementation of RunTree, which uses the\n * provided callback manager instead of the internal LangSmith client.\n *\n * The goal of this class is to ensure seamless interop when intergrated\n * with other Runnables.\n */\nclass CallbackManagerRunTree extends RunTree {\n callbackManager: CallbackManager;\n\n activeCallbackManager: CallbackManagerForChainRun | undefined = undefined;\n\n constructor(config: RunTreeConfig, callbackManager: CallbackManager) {\n super(config);\n\n this.callbackManager = callbackManager;\n }\n\n createChild(config: RunTreeConfig): CallbackManagerRunTree {\n const child = new CallbackManagerRunTree(\n {\n ...config,\n parent_run: this,\n project_name: this.project_name,\n client: this.client,\n },\n this.activeCallbackManager?.getChild() ?? this.callbackManager\n );\n this.child_runs.push(child);\n return child;\n }\n\n async postRun(): Promise<void> {\n // how it is translated in comparison to basic RunTree?\n this.activeCallbackManager = await this.callbackManager.handleChainStart(\n typeof this.serialized !== \"object\" &&\n this.serialized != null &&\n \"lc\" in this.serialized\n ? this.serialized\n : {\n id: [\"langchain\", \"smith\", \"CallbackManagerRunTree\"],\n lc: 1,\n type: \"not_implemented\",\n },\n this.inputs,\n this.id,\n this.run_type,\n undefined,\n undefined,\n this.name\n );\n }\n\n async patchRun(): Promise<void> {\n if (this.error) {\n await this.activeCallbackManager?.handleChainError(\n this.error,\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n } else {\n await this.activeCallbackManager?.handleChainEnd(\n this.outputs ?? {},\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n }\n }\n}\n\nclass RunnableTraceable<RunInput, RunOutput> extends Runnable<\n RunInput,\n RunOutput\n> {\n lc_serializable = false;\n\n lc_namespace = [\"langchain_core\", \"runnables\"];\n\n protected func: AnyTraceableFunction;\n\n constructor(fields: { func: AnyTraceableFunction }) {\n super(fields);\n\n if (!isLangsmithTraceableFunction(fields.func)) {\n throw new Error(\n \"RunnableTraceable requires a function that is wrapped in traceable higher-order function\"\n );\n }\n\n this.func = fields.func;\n }\n\n async invoke(input: RunInput, options?: Partial<RunnableConfig>) {\n const [config] = this._getOptionsList(options ?? {}, 1);\n const callbackManager = await getCallbackManagerForConfig(config);\n\n const partialConfig =\n \"langsmith:traceable\" in this.func\n ? (this.func[\"langsmith:traceable\"] as RunTreeConfig)\n : { name: \"<lambda>\" };\n\n if (!callbackManager) throw new Error(\"CallbackManager not found\");\n const runTree = new CallbackManagerRunTree(\n {\n ...partialConfig,\n parent_run: callbackManager?._parentRunId\n ? new RunTree({ name: \"<parent>\", id: callbackManager?._parentRunId })\n : undefined,\n },\n callbackManager\n );\n\n if (\n typeof input === \"object\" &&\n input != null &&\n Object.keys(input).length === 1\n ) {\n if (\"args\" in input && Array.isArray(input)) {\n return (await this.func(runTree, ...input)) as RunOutput;\n }\n\n if (\n \"input\" in input &&\n !(\n typeof input === \"object\" &&\n input != null &&\n !Array.isArray(input) &&\n // eslint-disable-next-line no-instanceof/no-instanceof\n !(input instanceof Date)\n )\n ) {\n try {\n return (await this.func(runTree, input.input)) as RunOutput;\n } catch {\n return (await this.func(runTree, input)) as RunOutput;\n }\n }\n }\n\n return (await this.func(runTree, input)) as RunOutput;\n }\n}\n\n/**\n * Wraps an off-the-shelf evaluator (loaded using loadEvaluator; of EvaluatorType[T])\n * and composes with a prepareData function so the user can prepare the trace and\n * dataset data for the evaluator.\n */\nclass PreparedRunEvaluator implements RunEvaluator {\n evaluator: LLMStringEvaluator;\n\n formatEvaluatorInputs: EvaluatorInputFormatter;\n\n isStringEvaluator: boolean;\n\n evaluationName: string;\n\n constructor(\n evaluator: LLMStringEvaluator,\n evaluationName: string,\n formatEvaluatorInputs: EvaluatorInputFormatter\n ) {\n this.evaluator = evaluator;\n this.isStringEvaluator = typeof evaluator?.evaluateStrings === \"function\";\n this.evaluationName = evaluationName;\n this.formatEvaluatorInputs = formatEvaluatorInputs;\n }\n\n static async fromEvalConfig(\n config: EvalConfig | keyof EvaluatorType\n ): Promise<PreparedRunEvaluator> {\n const evaluatorType =\n typeof config === \"string\" ? config : config.evaluatorType;\n const evalConfig = typeof config === \"string\" ? ({} as EvalConfig) : config;\n const evaluator = await loadEvaluator(evaluatorType, evalConfig);\n const feedbackKey = evalConfig?.feedbackKey ?? evaluator?.evaluationName;\n if (!isLLMStringEvaluator(evaluator)) {\n throw new Error(\n `Evaluator of type ${evaluatorType} not yet supported. ` +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n if (!feedbackKey) {\n throw new Error(\n `Evaluator of type ${evaluatorType} must have an evaluationName` +\n ` or feedbackKey. Please manually provide a feedbackKey in the EvalConfig.`\n );\n }\n return new PreparedRunEvaluator(\n evaluator as LLMStringEvaluator,\n feedbackKey,\n evalConfig?.formatEvaluatorInputs\n );\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const { prediction, input, reference } = this.formatEvaluatorInputs({\n rawInput: run.inputs,\n rawPrediction: run.outputs,\n rawReferenceOutput: example?.outputs,\n run,\n });\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n if (this.isStringEvaluator) {\n const evalResult = await this.evaluator.evaluateStrings(\n {\n prediction: prediction as string,\n reference: reference as string,\n input: input as string,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n key: this.evaluationName,\n comment: evalResult?.reasoning,\n sourceRunId: runId,\n ...evalResult,\n };\n }\n throw new Error(\n \"Evaluator not yet supported. \" +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n}\n\nclass LoadedEvalConfig {\n constructor(public evaluators: (RunEvaluator | DynamicRunEvaluator)[]) {}\n\n static async fromRunEvalConfig(\n config: RunEvalConfig<keyof EvaluatorType>\n ): Promise<LoadedEvalConfig> {\n // Custom evaluators are applied \"as-is\"\n const customEvaluators = (\n config?.customEvaluators ?? config.evaluators?.filter(isCustomEvaluator)\n )?.map((evaluator) => {\n if (typeof evaluator === \"function\") {\n return new DynamicRunEvaluator(evaluator);\n } else {\n return evaluator;\n }\n });\n\n const offTheShelfEvaluators = await Promise.all(\n config?.evaluators\n ?.filter(isOffTheShelfEvaluator)\n ?.map(\n async (evaluator) =>\n await PreparedRunEvaluator.fromEvalConfig(evaluator)\n ) ?? []\n );\n return new LoadedEvalConfig(\n (customEvaluators ?? []).concat(offTheShelfEvaluators ?? [])\n );\n }\n}\n\nexport interface RunOnDatasetParams extends Omit<\n RunEvalConfig,\n \"customEvaluators\"\n> {\n /**\n * Name of the project for logging and tracking.\n */\n projectName?: string;\n\n /**\n * Additional metadata for the project.\n */\n projectMetadata?: Record<string, unknown>;\n\n /**\n * Client instance for LangSmith service interaction.\n */\n client?: Client;\n\n /**\n * Maximum concurrency level for dataset processing.\n */\n maxConcurrency?: number;\n\n /**\n * @deprecated Pass keys directly to the RunOnDatasetParams instead\n */\n evaluationConfig?: RunEvalConfig;\n}\n\n/**\n * Internals expect a constructor () -> Runnable. This function wraps/coerces\n * the provided LangChain object, custom function, or factory function into\n * a constructor of a runnable.\n * @param modelOrFactory The model or factory to create a wrapped model from.\n * @returns A function that returns the wrapped model.\n * @throws Error if the modelOrFactory is invalid.\n */\nconst createWrappedModel = async (modelOrFactory: ChainOrFactory) => {\n if (Runnable.isRunnable(modelOrFactory)) {\n return () => modelOrFactory;\n }\n if (typeof modelOrFactory === \"function\") {\n if (isLangsmithTraceableFunction(modelOrFactory)) {\n const wrappedModel = new RunnableTraceable({ func: modelOrFactory });\n return () => wrappedModel;\n }\n\n try {\n // If it works with no arguments, assume it's a factory\n let res = (modelOrFactory as () => Runnable)();\n if (\n res &&\n typeof (res as unknown as Promise<Runnable>).then === \"function\"\n ) {\n res = await res;\n }\n return modelOrFactory as () => Runnable;\n } catch {\n // Otherwise, it's a custom UDF, and we'll wrap\n // the function in a lambda\n const wrappedModel = new RunnableLambda({ func: modelOrFactory });\n return () => wrappedModel;\n }\n }\n throw new Error(\"Invalid modelOrFactory\");\n};\n\nconst loadExamples = async ({\n datasetName,\n client,\n projectName,\n}: {\n datasetName: string;\n client: Client;\n projectName: string;\n maxConcurrency: number;\n}) => {\n const exampleIterator = client.listExamples({ datasetName });\n const configs: RunnableConfig[] = [];\n const runExtractors = [];\n const examples = [];\n for await (const example of exampleIterator) {\n const runExtractor = new SingleRunExtractor();\n configs.push({\n callbacks: [\n new LangChainTracer({ exampleId: example.id, projectName }),\n runExtractor,\n ],\n });\n examples.push(example);\n runExtractors.push(runExtractor);\n }\n return {\n configs,\n examples,\n runExtractors,\n };\n};\n\nconst applyEvaluators = async ({\n evaluation,\n runs,\n examples,\n client,\n maxConcurrency,\n}: {\n evaluation: LoadedEvalConfig;\n runs: Run[];\n examples: Example[];\n client: Client;\n maxConcurrency: number;\n}): Promise<{\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n}> => {\n // TODO: Parallelize and/or put in callbacks to speed up evals.\n const { evaluators } = evaluation;\n const progress = new ProgressBar({\n total: examples.length,\n format: \"Running Evaluators: {bar} {percentage}% | {value}/{total}\\n\",\n });\n const caller = new AsyncCaller({\n maxConcurrency,\n });\n const requests = runs.map(\n async (\n run,\n i\n ): Promise<{\n run_id: string;\n execution_time?: number;\n feedback: Feedback[];\n }> =>\n caller.call(async () => {\n const evaluatorResults = await Promise.allSettled(\n evaluators.map((evaluator) =>\n client.evaluateRun(run, evaluator, {\n referenceExample: examples[i],\n loadChildRuns: false,\n })\n )\n );\n progress.increment();\n return {\n execution_time:\n run?.end_time && run.start_time\n ? new Date(run.end_time).getTime() -\n new Date(run.start_time).getTime()\n : undefined,\n feedback: evaluatorResults.map((evalResult) =>\n evalResult.status === \"fulfilled\"\n ? evalResult.value\n : evalResult.reason\n ),\n run_id: run.id,\n };\n })\n );\n const results = await Promise.all(requests);\n\n return results.reduce(\n (acc, result, i) => ({\n ...acc,\n [examples[i].id]: result,\n }),\n {}\n );\n};\n\nexport type EvalResults = {\n projectName: string;\n results: {\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n };\n};\n\nconst getExamplesInputs = (\n examples: Example[],\n chainOrFactory: ChainOrFactory,\n dataType?: DataType\n) => {\n if (dataType === \"chat\") {\n // For some batty reason, we store the chat dataset differently.\n // { type: \"system\", data: { content: inputs.input } },\n // But we need to create AIMesage, SystemMessage, etc.\n return examples.map(({ inputs }) =>\n mapStoredMessagesToChatMessages(inputs.input)\n );\n }\n // If it's a language model and ALL example inputs have a single value,\n // then we can be friendly and flatten the inputs to a list of strings.\n const isLanguageModel =\n typeof chainOrFactory === \"object\" &&\n typeof (chainOrFactory as BaseLanguageModel)._llmType === \"function\";\n if (\n isLanguageModel &&\n examples.every(({ inputs }) => Object.keys(inputs).length === 1)\n ) {\n return examples.map(({ inputs }) => Object.values(inputs)[0]);\n }\n return examples.map(({ inputs }) => inputs);\n};\n\n/**\n * Evaluates a given model or chain against a specified LangSmith dataset.\n *\n * This function fetches example records from the specified dataset,\n * runs the model or chain against each example, and returns the evaluation\n * results.\n *\n * @param chainOrFactory - A model or factory/constructor function to be evaluated. It can be a\n * Runnable instance, a factory function that returns a Runnable, or a user-defined\n * function or factory.\n *\n * @param datasetName - The name of the dataset against which the evaluation will be\n * performed. This dataset should already be defined and contain the relevant data\n * for evaluation.\n *\n * @param options - (Optional) Additional parameters for the evaluation process:\n * - `evaluators` (RunEvalType[]): Evaluators to apply to a dataset run.\n * - `formatEvaluatorInputs` (EvaluatorInputFormatter): Convert the evaluation data into formats that can be used by the evaluator.\n * - `projectName` (string): Name of the project for logging and tracking.\n * - `projectMetadata` (Record<string, unknown>): Additional metadata for the project.\n * - `client` (Client): Client instance for LangSmith service interaction.\n * - `maxConcurrency` (number): Maximum concurrency level for dataset processing.\n *\n * @returns A promise that resolves to an `EvalResults` object. This object includes\n * detailed results of the evaluation, such as execution time, run IDs, and feedback\n * for each entry in the dataset.\n *\n * @example\n * ```typescript\n * // Example usage for evaluating a model on a dataset\n * async function evaluateModel() {\n * const chain = /* ...create your model or chain...*\\//\n * const datasetName = 'example-dataset';\n * const client = new Client(/* ...config... *\\//);\n *\n * const results = await runOnDataset(chain, datasetName, {\n * evaluators: [/* ...evaluators... *\\//],\n * client,\n * });\n *\n * console.log('Evaluation Results:', results);\n * }\n *\n * evaluateModel();\n * ```\n * In this example, `runOnDataset` is used to evaluate a language model (or a chain of models) against\n * a dataset named 'example-dataset'. The evaluation process is configured using `RunOnDatasetParams[\"evaluators\"]`, which can\n * include both standard and custom evaluators. The `Client` instance is used to interact with LangChain services.\n * The function returns the evaluation results, which can be logged or further processed as needed.\n */\n\nexport async function runOnDataset(\n chainOrFactory: ChainOrFactory,\n datasetName: string,\n options?: RunOnDatasetParams\n) {\n const {\n projectName,\n projectMetadata,\n client,\n maxConcurrency,\n }: RunOnDatasetParams = options ?? {};\n\n const evaluationConfig: RunEvalConfig | undefined =\n options?.evaluationConfig ??\n (options?.evaluators != null\n ? {\n evaluators: options.evaluators,\n formatEvaluatorInputs: options.formatEvaluatorInputs,\n }\n : undefined);\n\n const wrappedModel = await createWrappedModel(chainOrFactory);\n const testClient = client ?? new Client();\n const testProjectName = projectName ?? randomName();\n const dataset = await testClient.readDataset({ datasetName });\n const datasetId = dataset.id;\n const testConcurrency = maxConcurrency ?? 5;\n const { configs, examples, runExtractors } = await loadExamples({\n datasetName,\n client: testClient,\n projectName: testProjectName,\n maxConcurrency: testConcurrency,\n });\n\n await testClient.createProject({\n projectName: testProjectName,\n referenceDatasetId: datasetId,\n projectExtra: { metadata: { ...projectMetadata } },\n });\n const wrappedRunnable: Runnable = new RunnableLambda({\n func: wrappedModel,\n }).withConfig({ runName: \"evaluationRun\" });\n const runInputs = getExamplesInputs(\n examples,\n chainOrFactory,\n dataset.data_type\n );\n const progress = new ProgressBar({\n total: runInputs.length,\n format: \"Predicting: {bar} {percentage}% | {value}/{total}\",\n });\n // TODO: Collect the runs as well.\n await wrappedRunnable\n .withListeners({\n onEnd: () => progress.increment(),\n })\n // TODO: Insert evaluation inline for immediate feedback.\n .batch(runInputs, configs, {\n maxConcurrency,\n returnExceptions: true,\n });\n\n progress.complete();\n const runs: Run[] = [];\n for (let i = 0; i < examples.length; i += 1) {\n runs.push(await runExtractors[i].extract());\n }\n let evalResults: Record<\n string,\n { run_id: string; execution_time?: number; feedback: Feedback[] }\n > = {};\n if (evaluationConfig) {\n const loadedEvalConfig =\n await LoadedEvalConfig.fromRunEvalConfig(evaluationConfig);\n evalResults = await applyEvaluators({\n evaluation: loadedEvalConfig,\n runs,\n examples,\n client: testClient,\n maxConcurrency: testConcurrency,\n });\n }\n const results: EvalResults = {\n projectName: testProjectName,\n results: evalResults ?? {},\n };\n return results;\n}\n\nfunction isLangsmithTraceableFunction(x: unknown): x is AnyTraceableFunction {\n return typeof x === \"function\" && \"langsmith:traceable\" in x;\n}\n"],"mappings":";;;;;;;;;;;;AAsDA,IAAM,uBAAN,MAA2B;CACzB;CAEA;CAEA,cAAc;AACZ,OAAK,eAAe,IAAI,SAAiB,YAAY;AACnD,QAAK,uBAAuB;IAC5B;;CAGJ,oBACE,QACA,SACA,UACG;AACH,OAAK,qBAAqB,MAAM;;CAGlC,MAAM,UAA2B;AAC/B,SAAO,KAAK;;;AAIhB,IAAM,qBAAN,cAAiCA,6BAAAA,WAAW;CAC1C;CAEA;;CAGA,OAAO;CAEP,cAAc;AACZ,SAAO;AACP,OAAK,aAAa,IAAI,SAAc,YAAY;AAC9C,QAAK,qBAAqB;IAC1B;;CAGJ,MAAM,WAAW,KAAU;AACzB,OAAK,mBAAmB,IAAI;;CAG9B,MAAM,UAAwB;AAC5B,SAAO,KAAK;;;;;;AAOhB,IAAM,sBAAN,MAAkD;CAChD;CAEA,YAAY,WAA6B;AACvC,OAAK,YAAY,IAAIC,0BAAAA,eAAe,EAAE,MAAM,WAAW,CAAC;;;;;;;;CAS1D,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAIC,yCAAAA,gBAAgB,EAAE,aAAa,cAAc,CAAC;EACjE,MAAM,SAAS,MAAM,KAAK,UAAU,OAClC;GACE;GACA;GACA,OAAO,IAAI;GACX,YAAY,IAAI;GAChB,WAAW,SAAS;GACrB,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;AAED,SAAO;GACL,aAFY,MAAM,UAAU,SAAS;GAGrC,GAAG;GACJ;;;AAKL,SAAS,qBAAqB,WAAiD;AAC7E,QAAO,aAAa,OAAO,UAAU,oBAAoB;;;;;;;;;AAa3D,IAAM,yBAAN,MAAM,+BAA+BC,UAAAA,QAAQ;CAC3C;CAEA,wBAAgE,KAAA;CAEhE,YAAY,QAAuB,iBAAkC;AACnE,QAAM,OAAO;AAEb,OAAK,kBAAkB;;CAGzB,YAAY,QAA+C;EACzD,MAAM,QAAQ,IAAI,uBAChB;GACE,GAAG;GACH,YAAY;GACZ,cAAc,KAAK;GACnB,QAAQ,KAAK;GACd,EACD,KAAK,uBAAuB,UAAU,IAAI,KAAK,gBAChD;AACD,OAAK,WAAW,KAAK,MAAM;AAC3B,SAAO;;CAGT,MAAM,UAAyB;AAE7B,OAAK,wBAAwB,MAAM,KAAK,gBAAgB,iBACtD,OAAO,KAAK,eAAe,YACzB,KAAK,cAAc,QACnB,QAAQ,KAAK,aACX,KAAK,aACL;GACE,IAAI;IAAC;IAAa;IAAS;IAAyB;GACpD,IAAI;GACJ,MAAM;GACP,EACL,KAAK,QACL,KAAK,IACL,KAAK,UACL,KAAA,GACA,KAAA,GACA,KAAK,KACN;;CAGH,MAAM,WAA0B;AAC9B,MAAI,KAAK,MACP,OAAM,KAAK,uBAAuB,iBAChC,KAAK,OACL,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;MAED,OAAM,KAAK,uBAAuB,eAChC,KAAK,WAAW,EAAE,EAClB,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;;;AAKP,IAAM,oBAAN,cAAqDC,0BAAAA,SAGnD;CACA,kBAAkB;CAElB,eAAe,CAAC,kBAAkB,YAAY;CAE9C;CAEA,YAAY,QAAwC;AAClD,QAAM,OAAO;AAEb,MAAI,CAAC,6BAA6B,OAAO,KAAK,CAC5C,OAAM,IAAI,MACR,2FACD;AAGH,OAAK,OAAO,OAAO;;CAGrB,MAAM,OAAO,OAAiB,SAAmC;EAC/D,MAAM,CAAC,UAAU,KAAK,gBAAgB,WAAW,EAAE,EAAE,EAAE;EACvD,MAAM,kBAAkB,OAAA,GAAA,0BAAA,6BAAkC,OAAO;EAEjE,MAAM,gBACJ,yBAAyB,KAAK,OACzB,KAAK,KAAK,yBACX,EAAE,MAAM,YAAY;AAE1B,MAAI,CAAC,gBAAiB,OAAM,IAAI,MAAM,4BAA4B;EAClE,MAAM,UAAU,IAAI,uBAClB;GACE,GAAG;GACH,YAAY,iBAAiB,eACzB,IAAID,UAAAA,QAAQ;IAAE,MAAM;IAAY,IAAI,iBAAiB;IAAc,CAAC,GACpE,KAAA;GACL,EACD,gBACD;AAED,MACE,OAAO,UAAU,YACjB,SAAS,QACT,OAAO,KAAK,MAAM,CAAC,WAAW,GAC9B;AACA,OAAI,UAAU,SAAS,MAAM,QAAQ,MAAM,CACzC,QAAQ,MAAM,KAAK,KAAK,SAAS,GAAG,MAAM;AAG5C,OACE,WAAW,SACX,EACE,OAAO,UAAU,YACjB,SAAS,QACT,CAAC,MAAM,QAAQ,MAAM,IAErB,EAAE,iBAAiB,OAGrB,KAAI;AACF,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM,MAAM;WACvC;AACN,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;AAK7C,SAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;;;;;;AAS3C,IAAM,uBAAN,MAAM,qBAA6C;CACjD;CAEA;CAEA;CAEA;CAEA,YACE,WACA,gBACA,uBACA;AACA,OAAK,YAAY;AACjB,OAAK,oBAAoB,OAAO,WAAW,oBAAoB;AAC/D,OAAK,iBAAiB;AACtB,OAAK,wBAAwB;;CAG/B,aAAa,eACX,QAC+B;EAC/B,MAAM,gBACJ,OAAO,WAAW,WAAW,SAAS,OAAO;EAC/C,MAAM,aAAa,OAAO,WAAW,WAAY,EAAE,GAAkB;EACrE,MAAM,YAAY,MAAME,eAAAA,cAAc,eAAe,WAAW;EAChE,MAAM,cAAc,YAAY,eAAe,WAAW;AAC1D,MAAI,CAAC,qBAAqB,UAAU,CAClC,OAAM,IAAI,MACR,qBAAqB,cAAc,8GAGpC;AAEH,MAAI,CAAC,YACH,OAAM,IAAI,MACR,qBAAqB,cAAc,uGAEpC;AAEH,SAAO,IAAI,qBACT,WACA,aACA,YAAY,sBACb;;;;;;;;CASH,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,EAAE,YAAY,OAAO,cAAc,KAAK,sBAAsB;GAClE,UAAU,IAAI;GACd,eAAe,IAAI;GACnB,oBAAoB,SAAS;GAC7B;GACD,CAAC;EACF,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAIH,yCAAAA,gBAAgB,EAAE,aAAa,cAAc,CAAC;AACjE,MAAI,KAAK,mBAAmB;GAC1B,MAAM,aAAa,MAAM,KAAK,UAAU,gBACtC;IACc;IACD;IACJ;IACR,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;GACD,MAAM,QAAQ,MAAM,UAAU,SAAS;AACvC,UAAO;IACL,KAAK,KAAK;IACV,SAAS,YAAY;IACrB,aAAa;IACb,GAAG;IACJ;;AAEH,QAAM,IAAI,MACR,wHAGD;;;AAIL,IAAM,mBAAN,MAAM,iBAAiB;CACrB,YAAY,YAA2D;AAApD,OAAA,aAAA;;CAEnB,aAAa,kBACX,QAC2B;EAE3B,MAAM,oBACJ,QAAQ,oBAAoB,OAAO,YAAY,OAAOI,eAAAA,kBAAkB,GACvE,KAAK,cAAc;AACpB,OAAI,OAAO,cAAc,WACvB,QAAO,IAAI,oBAAoB,UAAU;OAEzC,QAAO;IAET;EAEF,MAAM,wBAAwB,MAAM,QAAQ,IAC1C,QAAQ,YACJ,OAAOC,eAAAA,uBAAuB,EAC9B,IACA,OAAO,cACL,MAAM,qBAAqB,eAAe,UAAU,CACvD,IAAI,EAAE,CACV;AACD,SAAO,IAAI,kBACR,oBAAoB,EAAE,EAAE,OAAO,yBAAyB,EAAE,CAAC,CAC7D;;;;;;;;;;;AA0CL,MAAM,qBAAqB,OAAO,mBAAmC;AACnE,KAAIH,0BAAAA,SAAS,WAAW,eAAe,CACrC,cAAa;AAEf,KAAI,OAAO,mBAAmB,YAAY;AACxC,MAAI,6BAA6B,eAAe,EAAE;GAChD,MAAM,eAAe,IAAI,kBAAkB,EAAE,MAAM,gBAAgB,CAAC;AACpE,gBAAa;;AAGf,MAAI;GAEF,IAAI,MAAO,gBAAmC;AAC9C,OACE,OACA,OAAQ,IAAqC,SAAS,WAEtD,OAAM,MAAM;AAEd,UAAO;UACD;GAGN,MAAM,eAAe,IAAIH,0BAAAA,eAAe,EAAE,MAAM,gBAAgB,CAAC;AACjE,gBAAa;;;AAGjB,OAAM,IAAI,MAAM,yBAAyB;;AAG3C,MAAM,eAAe,OAAO,EAC1B,aACA,QACA,kBAMI;CACJ,MAAM,kBAAkB,OAAO,aAAa,EAAE,aAAa,CAAC;CAC5D,MAAM,UAA4B,EAAE;CACpC,MAAM,gBAAgB,EAAE;CACxB,MAAM,WAAW,EAAE;AACnB,YAAW,MAAM,WAAW,iBAAiB;EAC3C,MAAM,eAAe,IAAI,oBAAoB;AAC7C,UAAQ,KAAK,EACX,WAAW,CACT,IAAIC,yCAAAA,gBAAgB;GAAE,WAAW,QAAQ;GAAI;GAAa,CAAC,EAC3D,aACD,EACF,CAAC;AACF,WAAS,KAAK,QAAQ;AACtB,gBAAc,KAAK,aAAa;;AAElC,QAAO;EACL;EACA;EACA;EACD;;AAGH,MAAM,kBAAkB,OAAO,EAC7B,YACA,MACA,UACA,QACA,qBAaI;CAEJ,MAAM,EAAE,eAAe;CACvB,MAAM,WAAW,IAAIM,iBAAAA,YAAY;EAC/B,OAAO,SAAS;EAChB,QAAQ;EACT,CAAC;CACF,MAAM,SAAS,IAAIC,mCAAAA,YAAY,EAC7B,gBACD,CAAC;CACF,MAAM,WAAW,KAAK,IACpB,OACE,KACA,MAMA,OAAO,KAAK,YAAY;EACtB,MAAM,mBAAmB,MAAM,QAAQ,WACrC,WAAW,KAAK,cACd,OAAO,YAAY,KAAK,WAAW;GACjC,kBAAkB,SAAS;GAC3B,eAAe;GAChB,CAAC,CACH,CACF;AACD,WAAS,WAAW;AACpB,SAAO;GACL,gBACE,KAAK,YAAY,IAAI,aACjB,IAAI,KAAK,IAAI,SAAS,CAAC,SAAS,GAChC,IAAI,KAAK,IAAI,WAAW,CAAC,SAAS,GAClC,KAAA;GACN,UAAU,iBAAiB,KAAK,eAC9B,WAAW,WAAW,cAClB,WAAW,QACX,WAAW,OAChB;GACD,QAAQ,IAAI;GACb;GACD,CACL;AAGD,SAFgB,MAAM,QAAQ,IAAI,SAAS,EAE5B,QACZ,KAAK,QAAQ,OAAO;EACnB,GAAG;GACF,SAAS,GAAG,KAAK;EACnB,GACD,EAAE,CACH;;AAcH,MAAM,qBACJ,UACA,gBACA,aACG;AACH,KAAI,aAAa,OAIf,QAAO,SAAS,KAAK,EAAE,cAAA,GAAA,yBAAA,iCACW,OAAO,MAAM,CAC9C;AAOH,KAFE,OAAO,mBAAmB,YAC1B,OAAQ,eAAqC,aAAa,cAG1D,SAAS,OAAO,EAAE,aAAa,OAAO,KAAK,OAAO,CAAC,WAAW,EAAE,CAEhE,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO,OAAO,OAAO,CAAC,GAAG;AAE/D,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsD7C,eAAsB,aACpB,gBACA,aACA,SACA;CACA,MAAM,EACJ,aACA,iBACA,QACA,mBACsB,WAAW,EAAE;CAErC,MAAM,mBACJ,SAAS,qBACR,SAAS,cAAc,OACpB;EACE,YAAY,QAAQ;EACpB,uBAAuB,QAAQ;EAChC,GACD,KAAA;CAEN,MAAM,eAAe,MAAM,mBAAmB,eAAe;CAC7D,MAAM,aAAa,UAAU,IAAIC,UAAAA,QAAQ;CACzC,MAAM,kBAAkB,eAAeC,wBAAAA,YAAY;CACnD,MAAM,UAAU,MAAM,WAAW,YAAY,EAAE,aAAa,CAAC;CAC7D,MAAM,YAAY,QAAQ;CAC1B,MAAM,kBAAkB,kBAAkB;CAC1C,MAAM,EAAE,SAAS,UAAU,kBAAkB,MAAM,aAAa;EAC9D;EACA,QAAQ;EACR,aAAa;EACb,gBAAgB;EACjB,CAAC;AAEF,OAAM,WAAW,cAAc;EAC7B,aAAa;EACb,oBAAoB;EACpB,cAAc,EAAE,UAAU,EAAE,GAAG,iBAAiB,EAAE;EACnD,CAAC;CACF,MAAM,kBAA4B,IAAIV,0BAAAA,eAAe,EACnD,MAAM,cACP,CAAC,CAAC,WAAW,EAAE,SAAS,iBAAiB,CAAC;CAC3C,MAAM,YAAY,kBAChB,UACA,gBACA,QAAQ,UACT;CACD,MAAM,WAAW,IAAIO,iBAAAA,YAAY;EAC/B,OAAO,UAAU;EACjB,QAAQ;EACT,CAAC;AAEF,OAAM,gBACH,cAAc,EACb,aAAa,SAAS,WAAW,EAClC,CAAC,CAED,MAAM,WAAW,SAAS;EACzB;EACA,kBAAkB;EACnB,CAAC;AAEJ,UAAS,UAAU;CACnB,MAAM,OAAc,EAAE;AACtB,MAAK,IAAI,IAAI,GAAG,IAAI,SAAS,QAAQ,KAAK,EACxC,MAAK,KAAK,MAAM,cAAc,GAAG,SAAS,CAAC;CAE7C,IAAI,cAGA,EAAE;AACN,KAAI,iBAGF,eAAc,MAAM,gBAAgB;EAClC,YAFA,MAAM,iBAAiB,kBAAkB,iBAAiB;EAG1D;EACA;EACA,QAAQ;EACR,gBAAgB;EACjB,CAAC;AAMJ,QAJ6B;EAC3B,aAAa;EACb,SAAS,eAAe,EAAE;EAC3B;;AAIH,SAAS,6BAA6B,GAAuC;AAC3E,QAAO,OAAO,MAAM,cAAc,yBAAyB"}
1
+ {"version":3,"file":"runner_utils.cjs","names":["BaseTracer","RunnableLambda","LangChainTracer","RunTree","Runnable","loadEvaluator","isCustomEvaluator","isOffTheShelfEvaluator","ProgressBar","AsyncCaller","Client","randomName"],"sources":["../../src/smith/runner_utils.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { Serialized } from \"@langchain/core/load/serializable\";\nimport { mapStoredMessagesToChatMessages } from \"@langchain/core/messages\";\nimport {\n Runnable,\n RunnableConfig,\n RunnableLambda,\n getCallbackManagerForConfig,\n} from \"@langchain/core/runnables\";\nimport { LangChainTracer } from \"@langchain/core/tracers/tracer_langchain\";\nimport { BaseTracer } from \"@langchain/core/tracers/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { AsyncCaller } from \"@langchain/core/utils/async_caller\";\nimport type {\n CallbackManager,\n CallbackManagerForChainRun,\n} from \"@langchain/core/callbacks/manager\";\nimport {\n Client,\n Example,\n Feedback,\n Run,\n RunTree,\n RunTreeConfig,\n} from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport { DataType } from \"langsmith/schemas\";\nimport type { TraceableFunction } from \"langsmith/singletons/traceable\";\nimport { LLMStringEvaluator } from \"../evaluation/base.js\";\nimport { loadEvaluator } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\nimport {\n isOffTheShelfEvaluator,\n type DynamicRunEvaluatorParams,\n type EvalConfig,\n type EvaluatorInputFormatter,\n type RunEvalConfig,\n type RunEvaluatorLike,\n isCustomEvaluator,\n} from \"./config.js\";\nimport { randomName } from \"./name_generation.js\";\nimport { ProgressBar } from \"./progress.js\";\n\nexport type ChainOrFactory =\n | Runnable\n | (() => Runnable)\n | AnyTraceableFunction\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => any)\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => Promise<any>)\n | (() => (obj: unknown) => unknown)\n | (() => (obj: unknown) => Promise<unknown>);\n\nclass SingleRunIdExtractor {\n runIdPromiseResolver: (runId: string) => void;\n\n runIdPromise: Promise<string>;\n\n constructor() {\n this.runIdPromise = new Promise<string>((extract) => {\n this.runIdPromiseResolver = extract;\n });\n }\n\n handleChainStart = (\n _chain: Serialized,\n _inputs: ChainValues,\n runId: string\n ) => {\n this.runIdPromiseResolver(runId);\n };\n\n async extract(): Promise<string> {\n return this.runIdPromise;\n }\n}\n\nclass SingleRunExtractor extends BaseTracer {\n runPromiseResolver: (run: Run) => void;\n\n runPromise: Promise<Run>;\n\n /** The name of the callback handler. */\n name = \"single_run_extractor\";\n\n constructor() {\n super();\n this.runPromise = new Promise<Run>((extract) => {\n this.runPromiseResolver = extract;\n });\n }\n\n async persistRun(run: Run) {\n this.runPromiseResolver(run);\n }\n\n async extract(): Promise<Run> {\n return this.runPromise;\n }\n}\n\n/**\n * Wraps an evaluator function + implements the RunEvaluator interface.\n */\nclass DynamicRunEvaluator implements RunEvaluator {\n evaluator: RunnableLambda<DynamicRunEvaluatorParams, EvaluationResult>;\n\n constructor(evaluator: RunEvaluatorLike) {\n this.evaluator = new RunnableLambda({ func: evaluator });\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n const result = await this.evaluator.invoke(\n {\n run,\n example,\n input: run.inputs,\n prediction: run.outputs,\n reference: example?.outputs,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n sourceRunId: runId,\n ...result,\n };\n }\n}\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\nfunction isLLMStringEvaluator(evaluator: any): evaluator is LLMStringEvaluator {\n return evaluator && typeof evaluator.evaluateStrings === \"function\";\n}\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\ntype AnyTraceableFunction = TraceableFunction<(...any: any[]) => any>;\n\n/**\n * Internal implementation of RunTree, which uses the\n * provided callback manager instead of the internal LangSmith client.\n *\n * The goal of this class is to ensure seamless interop when intergrated\n * with other Runnables.\n */\nclass CallbackManagerRunTree extends RunTree {\n callbackManager: CallbackManager;\n\n activeCallbackManager: CallbackManagerForChainRun | undefined = undefined;\n\n constructor(config: RunTreeConfig, callbackManager: CallbackManager) {\n super(config);\n\n this.callbackManager = callbackManager;\n }\n\n createChild(config: RunTreeConfig): CallbackManagerRunTree {\n const child = new CallbackManagerRunTree(\n {\n ...config,\n parent_run: this,\n project_name: this.project_name,\n client: this.client,\n },\n this.activeCallbackManager?.getChild() ?? this.callbackManager\n );\n this.child_runs.push(child);\n return child;\n }\n\n async postRun(): Promise<void> {\n // how it is translated in comparison to basic RunTree?\n this.activeCallbackManager = await this.callbackManager.handleChainStart(\n typeof this.serialized !== \"object\" &&\n this.serialized != null &&\n \"lc\" in this.serialized\n ? this.serialized\n : {\n id: [\"langchain\", \"smith\", \"CallbackManagerRunTree\"],\n lc: 1,\n type: \"not_implemented\",\n },\n this.inputs,\n this.id,\n this.run_type,\n undefined,\n undefined,\n this.name\n );\n }\n\n async patchRun(): Promise<void> {\n if (this.error) {\n await this.activeCallbackManager?.handleChainError(\n this.error,\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n } else {\n await this.activeCallbackManager?.handleChainEnd(\n this.outputs ?? {},\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n }\n }\n}\n\nclass RunnableTraceable<RunInput, RunOutput> extends Runnable<\n RunInput,\n RunOutput\n> {\n lc_serializable = false;\n\n lc_namespace = [\"langchain_core\", \"runnables\"];\n\n protected func: AnyTraceableFunction;\n\n constructor(fields: { func: AnyTraceableFunction }) {\n super(fields);\n\n if (!isLangsmithTraceableFunction(fields.func)) {\n throw new Error(\n \"RunnableTraceable requires a function that is wrapped in traceable higher-order function\"\n );\n }\n\n this.func = fields.func;\n }\n\n async invoke(input: RunInput, options?: Partial<RunnableConfig>) {\n const [config] = this._getOptionsList(options ?? {}, 1);\n const callbackManager = await getCallbackManagerForConfig(config);\n\n const partialConfig =\n \"langsmith:traceable\" in this.func\n ? (this.func[\"langsmith:traceable\"] as RunTreeConfig)\n : { name: \"<lambda>\" };\n\n if (!callbackManager) throw new Error(\"CallbackManager not found\");\n const runTree = new CallbackManagerRunTree(\n {\n ...partialConfig,\n parent_run: callbackManager?._parentRunId\n ? new RunTree({ name: \"<parent>\", id: callbackManager?._parentRunId })\n : undefined,\n },\n callbackManager\n );\n\n if (\n typeof input === \"object\" &&\n input != null &&\n Object.keys(input).length === 1\n ) {\n if (\"args\" in input && Array.isArray(input)) {\n return (await this.func(runTree, ...input)) as RunOutput;\n }\n\n if (\n \"input\" in input &&\n !(\n typeof input === \"object\" &&\n input != null &&\n !Array.isArray(input) &&\n // oxlint-disable-next-line no-instanceof/no-instanceof\n !(input instanceof Date)\n )\n ) {\n try {\n return (await this.func(runTree, input.input)) as RunOutput;\n } catch {\n return (await this.func(runTree, input)) as RunOutput;\n }\n }\n }\n\n return (await this.func(runTree, input)) as RunOutput;\n }\n}\n\n/**\n * Wraps an off-the-shelf evaluator (loaded using loadEvaluator; of EvaluatorType[T])\n * and composes with a prepareData function so the user can prepare the trace and\n * dataset data for the evaluator.\n */\nclass PreparedRunEvaluator implements RunEvaluator {\n evaluator: LLMStringEvaluator;\n\n formatEvaluatorInputs: EvaluatorInputFormatter;\n\n isStringEvaluator: boolean;\n\n evaluationName: string;\n\n constructor(\n evaluator: LLMStringEvaluator,\n evaluationName: string,\n formatEvaluatorInputs: EvaluatorInputFormatter\n ) {\n this.evaluator = evaluator;\n this.isStringEvaluator = typeof evaluator?.evaluateStrings === \"function\";\n this.evaluationName = evaluationName;\n this.formatEvaluatorInputs = formatEvaluatorInputs;\n }\n\n static async fromEvalConfig(\n config: EvalConfig | keyof EvaluatorType\n ): Promise<PreparedRunEvaluator> {\n const evaluatorType =\n typeof config === \"string\" ? config : config.evaluatorType;\n const evalConfig = typeof config === \"string\" ? ({} as EvalConfig) : config;\n const evaluator = await loadEvaluator(evaluatorType, evalConfig);\n const feedbackKey = evalConfig?.feedbackKey ?? evaluator?.evaluationName;\n if (!isLLMStringEvaluator(evaluator)) {\n throw new Error(\n `Evaluator of type ${evaluatorType} not yet supported. ` +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n if (!feedbackKey) {\n throw new Error(\n `Evaluator of type ${evaluatorType} must have an evaluationName` +\n ` or feedbackKey. Please manually provide a feedbackKey in the EvalConfig.`\n );\n }\n return new PreparedRunEvaluator(\n evaluator as LLMStringEvaluator,\n feedbackKey,\n evalConfig?.formatEvaluatorInputs\n );\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const { prediction, input, reference } = this.formatEvaluatorInputs({\n rawInput: run.inputs,\n rawPrediction: run.outputs,\n rawReferenceOutput: example?.outputs,\n run,\n });\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n if (this.isStringEvaluator) {\n const evalResult = await this.evaluator.evaluateStrings(\n {\n prediction: prediction as string,\n reference: reference as string,\n input: input as string,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n key: this.evaluationName,\n comment: evalResult?.reasoning,\n sourceRunId: runId,\n ...evalResult,\n };\n }\n throw new Error(\n \"Evaluator not yet supported. \" +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n}\n\nclass LoadedEvalConfig {\n constructor(public evaluators: (RunEvaluator | DynamicRunEvaluator)[]) {}\n\n static async fromRunEvalConfig(\n config: RunEvalConfig<keyof EvaluatorType>\n ): Promise<LoadedEvalConfig> {\n // Custom evaluators are applied \"as-is\"\n const customEvaluators = (\n config?.customEvaluators ?? config.evaluators?.filter(isCustomEvaluator)\n )?.map((evaluator) => {\n if (typeof evaluator === \"function\") {\n return new DynamicRunEvaluator(evaluator);\n } else {\n return evaluator;\n }\n });\n\n const offTheShelfEvaluators = await Promise.all(\n config?.evaluators\n ?.filter(isOffTheShelfEvaluator)\n ?.map(\n async (evaluator) =>\n await PreparedRunEvaluator.fromEvalConfig(evaluator)\n ) ?? []\n );\n return new LoadedEvalConfig(\n (customEvaluators ?? []).concat(offTheShelfEvaluators ?? [])\n );\n }\n}\n\nexport interface RunOnDatasetParams extends Omit<\n RunEvalConfig,\n \"customEvaluators\"\n> {\n /**\n * Name of the project for logging and tracking.\n */\n projectName?: string;\n\n /**\n * Additional metadata for the project.\n */\n projectMetadata?: Record<string, unknown>;\n\n /**\n * Client instance for LangSmith service interaction.\n */\n client?: Client;\n\n /**\n * Maximum concurrency level for dataset processing.\n */\n maxConcurrency?: number;\n\n /**\n * @deprecated Pass keys directly to the RunOnDatasetParams instead\n */\n evaluationConfig?: RunEvalConfig;\n}\n\n/**\n * Internals expect a constructor () -> Runnable. This function wraps/coerces\n * the provided LangChain object, custom function, or factory function into\n * a constructor of a runnable.\n * @param modelOrFactory The model or factory to create a wrapped model from.\n * @returns A function that returns the wrapped model.\n * @throws Error if the modelOrFactory is invalid.\n */\nconst createWrappedModel = async (modelOrFactory: ChainOrFactory) => {\n if (Runnable.isRunnable(modelOrFactory)) {\n return () => modelOrFactory;\n }\n if (typeof modelOrFactory === \"function\") {\n if (isLangsmithTraceableFunction(modelOrFactory)) {\n const wrappedModel = new RunnableTraceable({ func: modelOrFactory });\n return () => wrappedModel;\n }\n\n try {\n // If it works with no arguments, assume it's a factory\n let res = (modelOrFactory as () => Runnable)();\n if (\n res &&\n typeof (res as unknown as Promise<Runnable>).then === \"function\"\n ) {\n res = await res;\n }\n return modelOrFactory as () => Runnable;\n } catch {\n // Otherwise, it's a custom UDF, and we'll wrap\n // the function in a lambda\n const wrappedModel = new RunnableLambda({ func: modelOrFactory });\n return () => wrappedModel;\n }\n }\n throw new Error(\"Invalid modelOrFactory\");\n};\n\nconst loadExamples = async ({\n datasetName,\n client,\n projectName,\n}: {\n datasetName: string;\n client: Client;\n projectName: string;\n maxConcurrency: number;\n}) => {\n const exampleIterator = client.listExamples({ datasetName });\n const configs: RunnableConfig[] = [];\n const runExtractors = [];\n const examples = [];\n for await (const example of exampleIterator) {\n const runExtractor = new SingleRunExtractor();\n configs.push({\n callbacks: [\n new LangChainTracer({ exampleId: example.id, projectName }),\n runExtractor,\n ],\n });\n examples.push(example);\n runExtractors.push(runExtractor);\n }\n return {\n configs,\n examples,\n runExtractors,\n };\n};\n\nconst applyEvaluators = async ({\n evaluation,\n runs,\n examples,\n client,\n maxConcurrency,\n}: {\n evaluation: LoadedEvalConfig;\n runs: Run[];\n examples: Example[];\n client: Client;\n maxConcurrency: number;\n}): Promise<{\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n}> => {\n // TODO: Parallelize and/or put in callbacks to speed up evals.\n const { evaluators } = evaluation;\n const progress = new ProgressBar({\n total: examples.length,\n format: \"Running Evaluators: {bar} {percentage}% | {value}/{total}\\n\",\n });\n const caller = new AsyncCaller({\n maxConcurrency,\n });\n const requests = runs.map(\n async (\n run,\n i\n ): Promise<{\n run_id: string;\n execution_time?: number;\n feedback: Feedback[];\n }> =>\n caller.call(async () => {\n const evaluatorResults = await Promise.allSettled(\n evaluators.map((evaluator) =>\n client.evaluateRun(run, evaluator, {\n referenceExample: examples[i],\n loadChildRuns: false,\n })\n )\n );\n progress.increment();\n return {\n execution_time:\n run?.end_time && run.start_time\n ? new Date(run.end_time).getTime() -\n new Date(run.start_time).getTime()\n : undefined,\n feedback: evaluatorResults.map((evalResult) =>\n evalResult.status === \"fulfilled\"\n ? evalResult.value\n : evalResult.reason\n ),\n run_id: run.id,\n };\n })\n );\n const results = await Promise.all(requests);\n\n return results.reduce(\n (acc, result, i) => ({\n ...acc,\n [examples[i].id]: result,\n }),\n {}\n );\n};\n\nexport type EvalResults = {\n projectName: string;\n results: {\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n };\n};\n\nconst getExamplesInputs = (\n examples: Example[],\n chainOrFactory: ChainOrFactory,\n dataType?: DataType\n) => {\n if (dataType === \"chat\") {\n // For some batty reason, we store the chat dataset differently.\n // { type: \"system\", data: { content: inputs.input } },\n // But we need to create AIMesage, SystemMessage, etc.\n return examples.map(({ inputs }) =>\n mapStoredMessagesToChatMessages(inputs.input)\n );\n }\n // If it's a language model and ALL example inputs have a single value,\n // then we can be friendly and flatten the inputs to a list of strings.\n const isLanguageModel =\n typeof chainOrFactory === \"object\" &&\n typeof (chainOrFactory as BaseLanguageModel)._llmType === \"function\";\n if (\n isLanguageModel &&\n examples.every(({ inputs }) => Object.keys(inputs).length === 1)\n ) {\n return examples.map(({ inputs }) => Object.values(inputs)[0]);\n }\n return examples.map(({ inputs }) => inputs);\n};\n\n/**\n * Evaluates a given model or chain against a specified LangSmith dataset.\n *\n * This function fetches example records from the specified dataset,\n * runs the model or chain against each example, and returns the evaluation\n * results.\n *\n * @param chainOrFactory - A model or factory/constructor function to be evaluated. It can be a\n * Runnable instance, a factory function that returns a Runnable, or a user-defined\n * function or factory.\n *\n * @param datasetName - The name of the dataset against which the evaluation will be\n * performed. This dataset should already be defined and contain the relevant data\n * for evaluation.\n *\n * @param options - (Optional) Additional parameters for the evaluation process:\n * - `evaluators` (RunEvalType[]): Evaluators to apply to a dataset run.\n * - `formatEvaluatorInputs` (EvaluatorInputFormatter): Convert the evaluation data into formats that can be used by the evaluator.\n * - `projectName` (string): Name of the project for logging and tracking.\n * - `projectMetadata` (Record<string, unknown>): Additional metadata for the project.\n * - `client` (Client): Client instance for LangSmith service interaction.\n * - `maxConcurrency` (number): Maximum concurrency level for dataset processing.\n *\n * @returns A promise that resolves to an `EvalResults` object. This object includes\n * detailed results of the evaluation, such as execution time, run IDs, and feedback\n * for each entry in the dataset.\n *\n * @example\n * ```typescript\n * // Example usage for evaluating a model on a dataset\n * async function evaluateModel() {\n * const chain = /* ...create your model or chain...*\\//\n * const datasetName = 'example-dataset';\n * const client = new Client(/* ...config... *\\//);\n *\n * const results = await runOnDataset(chain, datasetName, {\n * evaluators: [/* ...evaluators... *\\//],\n * client,\n * });\n *\n * console.log('Evaluation Results:', results);\n * }\n *\n * evaluateModel();\n * ```\n * In this example, `runOnDataset` is used to evaluate a language model (or a chain of models) against\n * a dataset named 'example-dataset'. The evaluation process is configured using `RunOnDatasetParams[\"evaluators\"]`, which can\n * include both standard and custom evaluators. The `Client` instance is used to interact with LangChain services.\n * The function returns the evaluation results, which can be logged or further processed as needed.\n */\n\nexport async function runOnDataset(\n chainOrFactory: ChainOrFactory,\n datasetName: string,\n options?: RunOnDatasetParams\n) {\n const {\n projectName,\n projectMetadata,\n client,\n maxConcurrency,\n }: RunOnDatasetParams = options ?? {};\n\n const evaluationConfig: RunEvalConfig | undefined =\n options?.evaluationConfig ??\n (options?.evaluators != null\n ? {\n evaluators: options.evaluators,\n formatEvaluatorInputs: options.formatEvaluatorInputs,\n }\n : undefined);\n\n const wrappedModel = await createWrappedModel(chainOrFactory);\n const testClient = client ?? new Client();\n const testProjectName = projectName ?? randomName();\n const dataset = await testClient.readDataset({ datasetName });\n const datasetId = dataset.id;\n const testConcurrency = maxConcurrency ?? 5;\n const { configs, examples, runExtractors } = await loadExamples({\n datasetName,\n client: testClient,\n projectName: testProjectName,\n maxConcurrency: testConcurrency,\n });\n\n await testClient.createProject({\n projectName: testProjectName,\n referenceDatasetId: datasetId,\n projectExtra: { metadata: { ...projectMetadata } },\n });\n const wrappedRunnable: Runnable = new RunnableLambda({\n func: wrappedModel,\n }).withConfig({ runName: \"evaluationRun\" });\n const runInputs = getExamplesInputs(\n examples,\n chainOrFactory,\n dataset.data_type\n );\n const progress = new ProgressBar({\n total: runInputs.length,\n format: \"Predicting: {bar} {percentage}% | {value}/{total}\",\n });\n // TODO: Collect the runs as well.\n await wrappedRunnable\n .withListeners({\n onEnd: () => progress.increment(),\n })\n // TODO: Insert evaluation inline for immediate feedback.\n .batch(runInputs, configs, {\n maxConcurrency,\n returnExceptions: true,\n });\n\n progress.complete();\n const runs: Run[] = [];\n for (let i = 0; i < examples.length; i += 1) {\n runs.push(await runExtractors[i].extract());\n }\n let evalResults: Record<\n string,\n { run_id: string; execution_time?: number; feedback: Feedback[] }\n > = {};\n if (evaluationConfig) {\n const loadedEvalConfig =\n await LoadedEvalConfig.fromRunEvalConfig(evaluationConfig);\n evalResults = await applyEvaluators({\n evaluation: loadedEvalConfig,\n runs,\n examples,\n client: testClient,\n maxConcurrency: testConcurrency,\n });\n }\n const results: EvalResults = {\n projectName: testProjectName,\n results: evalResults ?? {},\n };\n return results;\n}\n\nfunction isLangsmithTraceableFunction(x: unknown): x is AnyTraceableFunction {\n return typeof x === \"function\" && \"langsmith:traceable\" in x;\n}\n"],"mappings":";;;;;;;;;;;;AAsDA,IAAM,uBAAN,MAA2B;CACzB;CAEA;CAEA,cAAc;AACZ,OAAK,eAAe,IAAI,SAAiB,YAAY;AACnD,QAAK,uBAAuB;IAC5B;;CAGJ,oBACE,QACA,SACA,UACG;AACH,OAAK,qBAAqB,MAAM;;CAGlC,MAAM,UAA2B;AAC/B,SAAO,KAAK;;;AAIhB,IAAM,qBAAN,cAAiCA,6BAAAA,WAAW;CAC1C;CAEA;;CAGA,OAAO;CAEP,cAAc;AACZ,SAAO;AACP,OAAK,aAAa,IAAI,SAAc,YAAY;AAC9C,QAAK,qBAAqB;IAC1B;;CAGJ,MAAM,WAAW,KAAU;AACzB,OAAK,mBAAmB,IAAI;;CAG9B,MAAM,UAAwB;AAC5B,SAAO,KAAK;;;;;;AAOhB,IAAM,sBAAN,MAAkD;CAChD;CAEA,YAAY,WAA6B;AACvC,OAAK,YAAY,IAAIC,0BAAAA,eAAe,EAAE,MAAM,WAAW,CAAC;;;;;;;;CAS1D,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAIC,yCAAAA,gBAAgB,EAAE,aAAa,cAAc,CAAC;EACjE,MAAM,SAAS,MAAM,KAAK,UAAU,OAClC;GACE;GACA;GACA,OAAO,IAAI;GACX,YAAY,IAAI;GAChB,WAAW,SAAS;GACrB,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;AAED,SAAO;GACL,aAFY,MAAM,UAAU,SAAS;GAGrC,GAAG;GACJ;;;AAKL,SAAS,qBAAqB,WAAiD;AAC7E,QAAO,aAAa,OAAO,UAAU,oBAAoB;;;;;;;;;AAa3D,IAAM,yBAAN,MAAM,+BAA+BC,UAAAA,QAAQ;CAC3C;CAEA,wBAAgE,KAAA;CAEhE,YAAY,QAAuB,iBAAkC;AACnE,QAAM,OAAO;AAEb,OAAK,kBAAkB;;CAGzB,YAAY,QAA+C;EACzD,MAAM,QAAQ,IAAI,uBAChB;GACE,GAAG;GACH,YAAY;GACZ,cAAc,KAAK;GACnB,QAAQ,KAAK;GACd,EACD,KAAK,uBAAuB,UAAU,IAAI,KAAK,gBAChD;AACD,OAAK,WAAW,KAAK,MAAM;AAC3B,SAAO;;CAGT,MAAM,UAAyB;AAE7B,OAAK,wBAAwB,MAAM,KAAK,gBAAgB,iBACtD,OAAO,KAAK,eAAe,YACzB,KAAK,cAAc,QACnB,QAAQ,KAAK,aACX,KAAK,aACL;GACE,IAAI;IAAC;IAAa;IAAS;IAAyB;GACpD,IAAI;GACJ,MAAM;GACP,EACL,KAAK,QACL,KAAK,IACL,KAAK,UACL,KAAA,GACA,KAAA,GACA,KAAK,KACN;;CAGH,MAAM,WAA0B;AAC9B,MAAI,KAAK,MACP,OAAM,KAAK,uBAAuB,iBAChC,KAAK,OACL,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;MAED,OAAM,KAAK,uBAAuB,eAChC,KAAK,WAAW,EAAE,EAClB,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;;;AAKP,IAAM,oBAAN,cAAqDC,0BAAAA,SAGnD;CACA,kBAAkB;CAElB,eAAe,CAAC,kBAAkB,YAAY;CAE9C;CAEA,YAAY,QAAwC;AAClD,QAAM,OAAO;AAEb,MAAI,CAAC,6BAA6B,OAAO,KAAK,CAC5C,OAAM,IAAI,MACR,2FACD;AAGH,OAAK,OAAO,OAAO;;CAGrB,MAAM,OAAO,OAAiB,SAAmC;EAC/D,MAAM,CAAC,UAAU,KAAK,gBAAgB,WAAW,EAAE,EAAE,EAAE;EACvD,MAAM,kBAAkB,OAAA,GAAA,0BAAA,6BAAkC,OAAO;EAEjE,MAAM,gBACJ,yBAAyB,KAAK,OACzB,KAAK,KAAK,yBACX,EAAE,MAAM,YAAY;AAE1B,MAAI,CAAC,gBAAiB,OAAM,IAAI,MAAM,4BAA4B;EAClE,MAAM,UAAU,IAAI,uBAClB;GACE,GAAG;GACH,YAAY,iBAAiB,eACzB,IAAID,UAAAA,QAAQ;IAAE,MAAM;IAAY,IAAI,iBAAiB;IAAc,CAAC,GACpE,KAAA;GACL,EACD,gBACD;AAED,MACE,OAAO,UAAU,YACjB,SAAS,QACT,OAAO,KAAK,MAAM,CAAC,WAAW,GAC9B;AACA,OAAI,UAAU,SAAS,MAAM,QAAQ,MAAM,CACzC,QAAQ,MAAM,KAAK,KAAK,SAAS,GAAG,MAAM;AAG5C,OACE,WAAW,SACX,EACE,OAAO,UAAU,YACjB,SAAS,QACT,CAAC,MAAM,QAAQ,MAAM,IAErB,EAAE,iBAAiB,OAGrB,KAAI;AACF,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM,MAAM;WACvC;AACN,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;AAK7C,SAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;;;;;;AAS3C,IAAM,uBAAN,MAAM,qBAA6C;CACjD;CAEA;CAEA;CAEA;CAEA,YACE,WACA,gBACA,uBACA;AACA,OAAK,YAAY;AACjB,OAAK,oBAAoB,OAAO,WAAW,oBAAoB;AAC/D,OAAK,iBAAiB;AACtB,OAAK,wBAAwB;;CAG/B,aAAa,eACX,QAC+B;EAC/B,MAAM,gBACJ,OAAO,WAAW,WAAW,SAAS,OAAO;EAC/C,MAAM,aAAa,OAAO,WAAW,WAAY,EAAE,GAAkB;EACrE,MAAM,YAAY,MAAME,eAAAA,cAAc,eAAe,WAAW;EAChE,MAAM,cAAc,YAAY,eAAe,WAAW;AAC1D,MAAI,CAAC,qBAAqB,UAAU,CAClC,OAAM,IAAI,MACR,qBAAqB,cAAc,8GAGpC;AAEH,MAAI,CAAC,YACH,OAAM,IAAI,MACR,qBAAqB,cAAc,uGAEpC;AAEH,SAAO,IAAI,qBACT,WACA,aACA,YAAY,sBACb;;;;;;;;CASH,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,EAAE,YAAY,OAAO,cAAc,KAAK,sBAAsB;GAClE,UAAU,IAAI;GACd,eAAe,IAAI;GACnB,oBAAoB,SAAS;GAC7B;GACD,CAAC;EACF,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAIH,yCAAAA,gBAAgB,EAAE,aAAa,cAAc,CAAC;AACjE,MAAI,KAAK,mBAAmB;GAC1B,MAAM,aAAa,MAAM,KAAK,UAAU,gBACtC;IACc;IACD;IACJ;IACR,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;GACD,MAAM,QAAQ,MAAM,UAAU,SAAS;AACvC,UAAO;IACL,KAAK,KAAK;IACV,SAAS,YAAY;IACrB,aAAa;IACb,GAAG;IACJ;;AAEH,QAAM,IAAI,MACR,wHAGD;;;AAIL,IAAM,mBAAN,MAAM,iBAAiB;CACrB,YAAY,YAA2D;AAApD,OAAA,aAAA;;CAEnB,aAAa,kBACX,QAC2B;EAE3B,MAAM,oBACJ,QAAQ,oBAAoB,OAAO,YAAY,OAAOI,eAAAA,kBAAkB,GACvE,KAAK,cAAc;AACpB,OAAI,OAAO,cAAc,WACvB,QAAO,IAAI,oBAAoB,UAAU;OAEzC,QAAO;IAET;EAEF,MAAM,wBAAwB,MAAM,QAAQ,IAC1C,QAAQ,YACJ,OAAOC,eAAAA,uBAAuB,EAC9B,IACA,OAAO,cACL,MAAM,qBAAqB,eAAe,UAAU,CACvD,IAAI,EAAE,CACV;AACD,SAAO,IAAI,kBACR,oBAAoB,EAAE,EAAE,OAAO,yBAAyB,EAAE,CAAC,CAC7D;;;;;;;;;;;AA0CL,MAAM,qBAAqB,OAAO,mBAAmC;AACnE,KAAIH,0BAAAA,SAAS,WAAW,eAAe,CACrC,cAAa;AAEf,KAAI,OAAO,mBAAmB,YAAY;AACxC,MAAI,6BAA6B,eAAe,EAAE;GAChD,MAAM,eAAe,IAAI,kBAAkB,EAAE,MAAM,gBAAgB,CAAC;AACpE,gBAAa;;AAGf,MAAI;GAEF,IAAI,MAAO,gBAAmC;AAC9C,OACE,OACA,OAAQ,IAAqC,SAAS,WAEtD,OAAM,MAAM;AAEd,UAAO;UACD;GAGN,MAAM,eAAe,IAAIH,0BAAAA,eAAe,EAAE,MAAM,gBAAgB,CAAC;AACjE,gBAAa;;;AAGjB,OAAM,IAAI,MAAM,yBAAyB;;AAG3C,MAAM,eAAe,OAAO,EAC1B,aACA,QACA,kBAMI;CACJ,MAAM,kBAAkB,OAAO,aAAa,EAAE,aAAa,CAAC;CAC5D,MAAM,UAA4B,EAAE;CACpC,MAAM,gBAAgB,EAAE;CACxB,MAAM,WAAW,EAAE;AACnB,YAAW,MAAM,WAAW,iBAAiB;EAC3C,MAAM,eAAe,IAAI,oBAAoB;AAC7C,UAAQ,KAAK,EACX,WAAW,CACT,IAAIC,yCAAAA,gBAAgB;GAAE,WAAW,QAAQ;GAAI;GAAa,CAAC,EAC3D,aACD,EACF,CAAC;AACF,WAAS,KAAK,QAAQ;AACtB,gBAAc,KAAK,aAAa;;AAElC,QAAO;EACL;EACA;EACA;EACD;;AAGH,MAAM,kBAAkB,OAAO,EAC7B,YACA,MACA,UACA,QACA,qBAaI;CAEJ,MAAM,EAAE,eAAe;CACvB,MAAM,WAAW,IAAIM,iBAAAA,YAAY;EAC/B,OAAO,SAAS;EAChB,QAAQ;EACT,CAAC;CACF,MAAM,SAAS,IAAIC,mCAAAA,YAAY,EAC7B,gBACD,CAAC;CACF,MAAM,WAAW,KAAK,IACpB,OACE,KACA,MAMA,OAAO,KAAK,YAAY;EACtB,MAAM,mBAAmB,MAAM,QAAQ,WACrC,WAAW,KAAK,cACd,OAAO,YAAY,KAAK,WAAW;GACjC,kBAAkB,SAAS;GAC3B,eAAe;GAChB,CAAC,CACH,CACF;AACD,WAAS,WAAW;AACpB,SAAO;GACL,gBACE,KAAK,YAAY,IAAI,aACjB,IAAI,KAAK,IAAI,SAAS,CAAC,SAAS,GAChC,IAAI,KAAK,IAAI,WAAW,CAAC,SAAS,GAClC,KAAA;GACN,UAAU,iBAAiB,KAAK,eAC9B,WAAW,WAAW,cAClB,WAAW,QACX,WAAW,OAChB;GACD,QAAQ,IAAI;GACb;GACD,CACL;AAGD,SAFgB,MAAM,QAAQ,IAAI,SAAS,EAE5B,QACZ,KAAK,QAAQ,OAAO;EACnB,GAAG;GACF,SAAS,GAAG,KAAK;EACnB,GACD,EAAE,CACH;;AAcH,MAAM,qBACJ,UACA,gBACA,aACG;AACH,KAAI,aAAa,OAIf,QAAO,SAAS,KAAK,EAAE,cAAA,GAAA,yBAAA,iCACW,OAAO,MAAM,CAC9C;AAOH,KAFE,OAAO,mBAAmB,YAC1B,OAAQ,eAAqC,aAAa,cAG1D,SAAS,OAAO,EAAE,aAAa,OAAO,KAAK,OAAO,CAAC,WAAW,EAAE,CAEhE,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO,OAAO,OAAO,CAAC,GAAG;AAE/D,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsD7C,eAAsB,aACpB,gBACA,aACA,SACA;CACA,MAAM,EACJ,aACA,iBACA,QACA,mBACsB,WAAW,EAAE;CAErC,MAAM,mBACJ,SAAS,qBACR,SAAS,cAAc,OACpB;EACE,YAAY,QAAQ;EACpB,uBAAuB,QAAQ;EAChC,GACD,KAAA;CAEN,MAAM,eAAe,MAAM,mBAAmB,eAAe;CAC7D,MAAM,aAAa,UAAU,IAAIC,UAAAA,QAAQ;CACzC,MAAM,kBAAkB,eAAeC,wBAAAA,YAAY;CACnD,MAAM,UAAU,MAAM,WAAW,YAAY,EAAE,aAAa,CAAC;CAC7D,MAAM,YAAY,QAAQ;CAC1B,MAAM,kBAAkB,kBAAkB;CAC1C,MAAM,EAAE,SAAS,UAAU,kBAAkB,MAAM,aAAa;EAC9D;EACA,QAAQ;EACR,aAAa;EACb,gBAAgB;EACjB,CAAC;AAEF,OAAM,WAAW,cAAc;EAC7B,aAAa;EACb,oBAAoB;EACpB,cAAc,EAAE,UAAU,EAAE,GAAG,iBAAiB,EAAE;EACnD,CAAC;CACF,MAAM,kBAA4B,IAAIV,0BAAAA,eAAe,EACnD,MAAM,cACP,CAAC,CAAC,WAAW,EAAE,SAAS,iBAAiB,CAAC;CAC3C,MAAM,YAAY,kBAChB,UACA,gBACA,QAAQ,UACT;CACD,MAAM,WAAW,IAAIO,iBAAAA,YAAY;EAC/B,OAAO,UAAU;EACjB,QAAQ;EACT,CAAC;AAEF,OAAM,gBACH,cAAc,EACb,aAAa,SAAS,WAAW,EAClC,CAAC,CAED,MAAM,WAAW,SAAS;EACzB;EACA,kBAAkB;EACnB,CAAC;AAEJ,UAAS,UAAU;CACnB,MAAM,OAAc,EAAE;AACtB,MAAK,IAAI,IAAI,GAAG,IAAI,SAAS,QAAQ,KAAK,EACxC,MAAK,KAAK,MAAM,cAAc,GAAG,SAAS,CAAC;CAE7C,IAAI,cAGA,EAAE;AACN,KAAI,iBAGF,eAAc,MAAM,gBAAgB;EAClC,YAFA,MAAM,iBAAiB,kBAAkB,iBAAiB;EAG1D;EACA;EACA,QAAQ;EACR,gBAAgB;EACjB,CAAC;AAMJ,QAJ6B;EAC3B,aAAa;EACb,SAAS,eAAe,EAAE;EAC3B;;AAIH,SAAS,6BAA6B,GAAuC;AAC3E,QAAO,OAAO,MAAM,cAAc,yBAAyB"}
@@ -1,7 +1,7 @@
1
- import { TraceableFunction } from "../node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.cjs";
2
1
  import { RunEvalConfig } from "./config.cjs";
3
2
  import { Runnable } from "@langchain/core/runnables";
4
3
  import { Client, Feedback } from "langsmith";
4
+ import { TraceableFunction } from "langsmith/singletons/traceable";
5
5
 
6
6
  //#region src/smith/runner_utils.d.ts
7
7
  type ChainOrFactory = Runnable | (() => Runnable) | AnyTraceableFunction | ((obj: any) => any) | ((obj: any) => Promise<any>) | (() => (obj: unknown) => unknown) | (() => (obj: unknown) => Promise<unknown>);
@@ -1 +1 @@
1
- {"version":3,"file":"runner_utils.d.cts","names":[],"sources":["../../src/smith/runner_utils.ts"],"mappings":";;;;;;KA2CY,cAAA,GACR,QAAA,UACO,QAAA,IACP,oBAAA,KAEE,GAAA,mBAEA,GAAA,UAAa,OAAA,iBACP,GAAA,iCACA,GAAA,cAAiB,OAAA;AAAA,KA+FxB,oBAAA,GAAuB,iBAAA,KAAsB,GAAA;AAAA,UAmRjC,kBAAA,SAA2B,IAAA,CAC1C,aAAA;EA5XU;;;EAkYV,WAAA;EAhYS;;;EAqYT,eAAA,GAAkB,MAAA;EA9XgB;;;EAmYlC,MAAA,GAAS,MAAA;EAzYP;;;EA8YF,cAAA;EAzYU;;;EA8YV,gBAAA,GAAmB,aAAA;AAAA;AAAA,KAkJT,WAAA;EACV,WAAA;EACA,OAAA;IAAA,CACG,GAAA;MACC,cAAA;MACA,MAAA;MACA,QAAA,EAAU,QAAA;IAAA;EAAA;AAAA;;;;;;;;;;;;;;;;;;AANhB;;;;;;;;;;;;;AAyFA;;;;;;;;;;;;;;;;;;;;iBAAsB,YAAA,CACpB,cAAA,EAAgB,cAAA,EAChB,WAAA,UACA,OAAA,GAAU,kBAAA,GAAkB,OAAA,CAAA,WAAA"}
1
+ {"version":3,"file":"runner_utils.d.cts","names":[],"sources":["../../src/smith/runner_utils.ts"],"mappings":";;;;;;KA2CY,cAAA,GACR,QAAA,UACO,QAAA,IACP,oBAAA,KAEE,GAAA,mBAEA,GAAA,UAAa,OAAA,iBACP,GAAA,iCACA,GAAA,cAAiB,OAAA;AAAA,KA+FxB,oBAAA,GAAuB,iBAAA,KAAsB,GAAA;AAAA,UAmRjC,kBAAA,SAA2B,IAAA,CAC1C,aAAA;EA5XwB;;;EAkYxB,WAAA;EA/XE;;;EAoYF,eAAA,GAAkB,MAAA;EA9XgB;;;EAmYlC,MAAA,GAAS,MAAA;EAvYL;;;EA4YJ,cAAA;EAxYU;;;EA6YV,gBAAA,GAAmB,aAAA;AAAA;AAAA,KAkJT,WAAA;EACV,WAAA;EACA,OAAA;IAAA,CACG,GAAA;MACC,cAAA;MACA,MAAA;MACA,QAAA,EAAU,QAAA;IAAA;EAAA;AAAA;;;;;;;;;;;;;;;;;AANhB;;;;;;;;;;;;;AAyFA;;;;;;;;;;;;;;;;;;;;;iBAAsB,YAAA,CACpB,cAAA,EAAgB,cAAA,EAChB,WAAA,UACA,OAAA,GAAU,kBAAA,GAAkB,OAAA,CAAA,WAAA"}
@@ -1,7 +1,7 @@
1
- import { TraceableFunction } from "../node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.js";
2
1
  import { RunEvalConfig } from "./config.js";
3
2
  import { Runnable } from "@langchain/core/runnables";
4
3
  import { Client, Feedback } from "langsmith";
4
+ import { TraceableFunction } from "langsmith/singletons/traceable";
5
5
 
6
6
  //#region src/smith/runner_utils.d.ts
7
7
  type ChainOrFactory = Runnable | (() => Runnable) | AnyTraceableFunction | ((obj: any) => any) | ((obj: any) => Promise<any>) | (() => (obj: unknown) => unknown) | (() => (obj: unknown) => Promise<unknown>);
@@ -1 +1 @@
1
- {"version":3,"file":"runner_utils.d.ts","names":[],"sources":["../../src/smith/runner_utils.ts"],"mappings":";;;;;;KA2CY,cAAA,GACR,QAAA,UACO,QAAA,IACP,oBAAA,KAEE,GAAA,mBAEA,GAAA,UAAa,OAAA,iBACP,GAAA,iCACA,GAAA,cAAiB,OAAA;AAAA,KA+FxB,oBAAA,GAAuB,iBAAA,KAAsB,GAAA;AAAA,UAmRjC,kBAAA,SAA2B,IAAA,CAC1C,aAAA;EA5XU;;;EAkYV,WAAA;EAhYS;;;EAqYT,eAAA,GAAkB,MAAA;EA9XgB;;;EAmYlC,MAAA,GAAS,MAAA;EAzYP;;;EA8YF,cAAA;EAzYU;;;EA8YV,gBAAA,GAAmB,aAAA;AAAA;AAAA,KAkJT,WAAA;EACV,WAAA;EACA,OAAA;IAAA,CACG,GAAA;MACC,cAAA;MACA,MAAA;MACA,QAAA,EAAU,QAAA;IAAA;EAAA;AAAA;;;;;;;;;;;;;;;;;;AANhB;;;;;;;;;;;;;AAyFA;;;;;;;;;;;;;;;;;;;;iBAAsB,YAAA,CACpB,cAAA,EAAgB,cAAA,EAChB,WAAA,UACA,OAAA,GAAU,kBAAA,GAAkB,OAAA,CAAA,WAAA"}
1
+ {"version":3,"file":"runner_utils.d.ts","names":[],"sources":["../../src/smith/runner_utils.ts"],"mappings":";;;;;;KA2CY,cAAA,GACR,QAAA,UACO,QAAA,IACP,oBAAA,KAEE,GAAA,mBAEA,GAAA,UAAa,OAAA,iBACP,GAAA,iCACA,GAAA,cAAiB,OAAA;AAAA,KA+FxB,oBAAA,GAAuB,iBAAA,KAAsB,GAAA;AAAA,UAmRjC,kBAAA,SAA2B,IAAA,CAC1C,aAAA;EA5XwB;;;EAkYxB,WAAA;EA/XE;;;EAoYF,eAAA,GAAkB,MAAA;EA9XgB;;;EAmYlC,MAAA,GAAS,MAAA;EAvYL;;;EA4YJ,cAAA;EAxYU;;;EA6YV,gBAAA,GAAmB,aAAA;AAAA;AAAA,KAkJT,WAAA;EACV,WAAA;EACA,OAAA;IAAA,CACG,GAAA;MACC,cAAA;MACA,MAAA;MACA,QAAA,EAAU,QAAA;IAAA;EAAA;AAAA;;;;;;;;;;;;;;;;;AANhB;;;;;;;;;;;;;AAyFA;;;;;;;;;;;;;;;;;;;;;iBAAsB,YAAA,CACpB,cAAA,EAAgB,cAAA,EAChB,WAAA,UACA,OAAA,GAAU,kBAAA,GAAkB,OAAA,CAAA,WAAA"}
@@ -1 +1 @@
1
- {"version":3,"file":"runner_utils.js","names":[],"sources":["../../src/smith/runner_utils.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { Serialized } from \"@langchain/core/load/serializable\";\nimport { mapStoredMessagesToChatMessages } from \"@langchain/core/messages\";\nimport {\n Runnable,\n RunnableConfig,\n RunnableLambda,\n getCallbackManagerForConfig,\n} from \"@langchain/core/runnables\";\nimport { LangChainTracer } from \"@langchain/core/tracers/tracer_langchain\";\nimport { BaseTracer } from \"@langchain/core/tracers/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { AsyncCaller } from \"@langchain/core/utils/async_caller\";\nimport type {\n CallbackManager,\n CallbackManagerForChainRun,\n} from \"@langchain/core/callbacks/manager\";\nimport {\n Client,\n Example,\n Feedback,\n Run,\n RunTree,\n RunTreeConfig,\n} from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport { DataType } from \"langsmith/schemas\";\nimport type { TraceableFunction } from \"langsmith/singletons/traceable\";\nimport { LLMStringEvaluator } from \"../evaluation/base.js\";\nimport { loadEvaluator } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\nimport {\n isOffTheShelfEvaluator,\n type DynamicRunEvaluatorParams,\n type EvalConfig,\n type EvaluatorInputFormatter,\n type RunEvalConfig,\n type RunEvaluatorLike,\n isCustomEvaluator,\n} from \"./config.js\";\nimport { randomName } from \"./name_generation.js\";\nimport { ProgressBar } from \"./progress.js\";\n\nexport type ChainOrFactory =\n | Runnable\n | (() => Runnable)\n | AnyTraceableFunction\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => any)\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => Promise<any>)\n | (() => (obj: unknown) => unknown)\n | (() => (obj: unknown) => Promise<unknown>);\n\nclass SingleRunIdExtractor {\n runIdPromiseResolver: (runId: string) => void;\n\n runIdPromise: Promise<string>;\n\n constructor() {\n this.runIdPromise = new Promise<string>((extract) => {\n this.runIdPromiseResolver = extract;\n });\n }\n\n handleChainStart = (\n _chain: Serialized,\n _inputs: ChainValues,\n runId: string\n ) => {\n this.runIdPromiseResolver(runId);\n };\n\n async extract(): Promise<string> {\n return this.runIdPromise;\n }\n}\n\nclass SingleRunExtractor extends BaseTracer {\n runPromiseResolver: (run: Run) => void;\n\n runPromise: Promise<Run>;\n\n /** The name of the callback handler. */\n name = \"single_run_extractor\";\n\n constructor() {\n super();\n this.runPromise = new Promise<Run>((extract) => {\n this.runPromiseResolver = extract;\n });\n }\n\n async persistRun(run: Run) {\n this.runPromiseResolver(run);\n }\n\n async extract(): Promise<Run> {\n return this.runPromise;\n }\n}\n\n/**\n * Wraps an evaluator function + implements the RunEvaluator interface.\n */\nclass DynamicRunEvaluator implements RunEvaluator {\n evaluator: RunnableLambda<DynamicRunEvaluatorParams, EvaluationResult>;\n\n constructor(evaluator: RunEvaluatorLike) {\n this.evaluator = new RunnableLambda({ func: evaluator });\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n const result = await this.evaluator.invoke(\n {\n run,\n example,\n input: run.inputs,\n prediction: run.outputs,\n reference: example?.outputs,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n sourceRunId: runId,\n ...result,\n };\n }\n}\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nfunction isLLMStringEvaluator(evaluator: any): evaluator is LLMStringEvaluator {\n return evaluator && typeof evaluator.evaluateStrings === \"function\";\n}\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AnyTraceableFunction = TraceableFunction<(...any: any[]) => any>;\n\n/**\n * Internal implementation of RunTree, which uses the\n * provided callback manager instead of the internal LangSmith client.\n *\n * The goal of this class is to ensure seamless interop when intergrated\n * with other Runnables.\n */\nclass CallbackManagerRunTree extends RunTree {\n callbackManager: CallbackManager;\n\n activeCallbackManager: CallbackManagerForChainRun | undefined = undefined;\n\n constructor(config: RunTreeConfig, callbackManager: CallbackManager) {\n super(config);\n\n this.callbackManager = callbackManager;\n }\n\n createChild(config: RunTreeConfig): CallbackManagerRunTree {\n const child = new CallbackManagerRunTree(\n {\n ...config,\n parent_run: this,\n project_name: this.project_name,\n client: this.client,\n },\n this.activeCallbackManager?.getChild() ?? this.callbackManager\n );\n this.child_runs.push(child);\n return child;\n }\n\n async postRun(): Promise<void> {\n // how it is translated in comparison to basic RunTree?\n this.activeCallbackManager = await this.callbackManager.handleChainStart(\n typeof this.serialized !== \"object\" &&\n this.serialized != null &&\n \"lc\" in this.serialized\n ? this.serialized\n : {\n id: [\"langchain\", \"smith\", \"CallbackManagerRunTree\"],\n lc: 1,\n type: \"not_implemented\",\n },\n this.inputs,\n this.id,\n this.run_type,\n undefined,\n undefined,\n this.name\n );\n }\n\n async patchRun(): Promise<void> {\n if (this.error) {\n await this.activeCallbackManager?.handleChainError(\n this.error,\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n } else {\n await this.activeCallbackManager?.handleChainEnd(\n this.outputs ?? {},\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n }\n }\n}\n\nclass RunnableTraceable<RunInput, RunOutput> extends Runnable<\n RunInput,\n RunOutput\n> {\n lc_serializable = false;\n\n lc_namespace = [\"langchain_core\", \"runnables\"];\n\n protected func: AnyTraceableFunction;\n\n constructor(fields: { func: AnyTraceableFunction }) {\n super(fields);\n\n if (!isLangsmithTraceableFunction(fields.func)) {\n throw new Error(\n \"RunnableTraceable requires a function that is wrapped in traceable higher-order function\"\n );\n }\n\n this.func = fields.func;\n }\n\n async invoke(input: RunInput, options?: Partial<RunnableConfig>) {\n const [config] = this._getOptionsList(options ?? {}, 1);\n const callbackManager = await getCallbackManagerForConfig(config);\n\n const partialConfig =\n \"langsmith:traceable\" in this.func\n ? (this.func[\"langsmith:traceable\"] as RunTreeConfig)\n : { name: \"<lambda>\" };\n\n if (!callbackManager) throw new Error(\"CallbackManager not found\");\n const runTree = new CallbackManagerRunTree(\n {\n ...partialConfig,\n parent_run: callbackManager?._parentRunId\n ? new RunTree({ name: \"<parent>\", id: callbackManager?._parentRunId })\n : undefined,\n },\n callbackManager\n );\n\n if (\n typeof input === \"object\" &&\n input != null &&\n Object.keys(input).length === 1\n ) {\n if (\"args\" in input && Array.isArray(input)) {\n return (await this.func(runTree, ...input)) as RunOutput;\n }\n\n if (\n \"input\" in input &&\n !(\n typeof input === \"object\" &&\n input != null &&\n !Array.isArray(input) &&\n // eslint-disable-next-line no-instanceof/no-instanceof\n !(input instanceof Date)\n )\n ) {\n try {\n return (await this.func(runTree, input.input)) as RunOutput;\n } catch {\n return (await this.func(runTree, input)) as RunOutput;\n }\n }\n }\n\n return (await this.func(runTree, input)) as RunOutput;\n }\n}\n\n/**\n * Wraps an off-the-shelf evaluator (loaded using loadEvaluator; of EvaluatorType[T])\n * and composes with a prepareData function so the user can prepare the trace and\n * dataset data for the evaluator.\n */\nclass PreparedRunEvaluator implements RunEvaluator {\n evaluator: LLMStringEvaluator;\n\n formatEvaluatorInputs: EvaluatorInputFormatter;\n\n isStringEvaluator: boolean;\n\n evaluationName: string;\n\n constructor(\n evaluator: LLMStringEvaluator,\n evaluationName: string,\n formatEvaluatorInputs: EvaluatorInputFormatter\n ) {\n this.evaluator = evaluator;\n this.isStringEvaluator = typeof evaluator?.evaluateStrings === \"function\";\n this.evaluationName = evaluationName;\n this.formatEvaluatorInputs = formatEvaluatorInputs;\n }\n\n static async fromEvalConfig(\n config: EvalConfig | keyof EvaluatorType\n ): Promise<PreparedRunEvaluator> {\n const evaluatorType =\n typeof config === \"string\" ? config : config.evaluatorType;\n const evalConfig = typeof config === \"string\" ? ({} as EvalConfig) : config;\n const evaluator = await loadEvaluator(evaluatorType, evalConfig);\n const feedbackKey = evalConfig?.feedbackKey ?? evaluator?.evaluationName;\n if (!isLLMStringEvaluator(evaluator)) {\n throw new Error(\n `Evaluator of type ${evaluatorType} not yet supported. ` +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n if (!feedbackKey) {\n throw new Error(\n `Evaluator of type ${evaluatorType} must have an evaluationName` +\n ` or feedbackKey. Please manually provide a feedbackKey in the EvalConfig.`\n );\n }\n return new PreparedRunEvaluator(\n evaluator as LLMStringEvaluator,\n feedbackKey,\n evalConfig?.formatEvaluatorInputs\n );\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const { prediction, input, reference } = this.formatEvaluatorInputs({\n rawInput: run.inputs,\n rawPrediction: run.outputs,\n rawReferenceOutput: example?.outputs,\n run,\n });\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n if (this.isStringEvaluator) {\n const evalResult = await this.evaluator.evaluateStrings(\n {\n prediction: prediction as string,\n reference: reference as string,\n input: input as string,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n key: this.evaluationName,\n comment: evalResult?.reasoning,\n sourceRunId: runId,\n ...evalResult,\n };\n }\n throw new Error(\n \"Evaluator not yet supported. \" +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n}\n\nclass LoadedEvalConfig {\n constructor(public evaluators: (RunEvaluator | DynamicRunEvaluator)[]) {}\n\n static async fromRunEvalConfig(\n config: RunEvalConfig<keyof EvaluatorType>\n ): Promise<LoadedEvalConfig> {\n // Custom evaluators are applied \"as-is\"\n const customEvaluators = (\n config?.customEvaluators ?? config.evaluators?.filter(isCustomEvaluator)\n )?.map((evaluator) => {\n if (typeof evaluator === \"function\") {\n return new DynamicRunEvaluator(evaluator);\n } else {\n return evaluator;\n }\n });\n\n const offTheShelfEvaluators = await Promise.all(\n config?.evaluators\n ?.filter(isOffTheShelfEvaluator)\n ?.map(\n async (evaluator) =>\n await PreparedRunEvaluator.fromEvalConfig(evaluator)\n ) ?? []\n );\n return new LoadedEvalConfig(\n (customEvaluators ?? []).concat(offTheShelfEvaluators ?? [])\n );\n }\n}\n\nexport interface RunOnDatasetParams extends Omit<\n RunEvalConfig,\n \"customEvaluators\"\n> {\n /**\n * Name of the project for logging and tracking.\n */\n projectName?: string;\n\n /**\n * Additional metadata for the project.\n */\n projectMetadata?: Record<string, unknown>;\n\n /**\n * Client instance for LangSmith service interaction.\n */\n client?: Client;\n\n /**\n * Maximum concurrency level for dataset processing.\n */\n maxConcurrency?: number;\n\n /**\n * @deprecated Pass keys directly to the RunOnDatasetParams instead\n */\n evaluationConfig?: RunEvalConfig;\n}\n\n/**\n * Internals expect a constructor () -> Runnable. This function wraps/coerces\n * the provided LangChain object, custom function, or factory function into\n * a constructor of a runnable.\n * @param modelOrFactory The model or factory to create a wrapped model from.\n * @returns A function that returns the wrapped model.\n * @throws Error if the modelOrFactory is invalid.\n */\nconst createWrappedModel = async (modelOrFactory: ChainOrFactory) => {\n if (Runnable.isRunnable(modelOrFactory)) {\n return () => modelOrFactory;\n }\n if (typeof modelOrFactory === \"function\") {\n if (isLangsmithTraceableFunction(modelOrFactory)) {\n const wrappedModel = new RunnableTraceable({ func: modelOrFactory });\n return () => wrappedModel;\n }\n\n try {\n // If it works with no arguments, assume it's a factory\n let res = (modelOrFactory as () => Runnable)();\n if (\n res &&\n typeof (res as unknown as Promise<Runnable>).then === \"function\"\n ) {\n res = await res;\n }\n return modelOrFactory as () => Runnable;\n } catch {\n // Otherwise, it's a custom UDF, and we'll wrap\n // the function in a lambda\n const wrappedModel = new RunnableLambda({ func: modelOrFactory });\n return () => wrappedModel;\n }\n }\n throw new Error(\"Invalid modelOrFactory\");\n};\n\nconst loadExamples = async ({\n datasetName,\n client,\n projectName,\n}: {\n datasetName: string;\n client: Client;\n projectName: string;\n maxConcurrency: number;\n}) => {\n const exampleIterator = client.listExamples({ datasetName });\n const configs: RunnableConfig[] = [];\n const runExtractors = [];\n const examples = [];\n for await (const example of exampleIterator) {\n const runExtractor = new SingleRunExtractor();\n configs.push({\n callbacks: [\n new LangChainTracer({ exampleId: example.id, projectName }),\n runExtractor,\n ],\n });\n examples.push(example);\n runExtractors.push(runExtractor);\n }\n return {\n configs,\n examples,\n runExtractors,\n };\n};\n\nconst applyEvaluators = async ({\n evaluation,\n runs,\n examples,\n client,\n maxConcurrency,\n}: {\n evaluation: LoadedEvalConfig;\n runs: Run[];\n examples: Example[];\n client: Client;\n maxConcurrency: number;\n}): Promise<{\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n}> => {\n // TODO: Parallelize and/or put in callbacks to speed up evals.\n const { evaluators } = evaluation;\n const progress = new ProgressBar({\n total: examples.length,\n format: \"Running Evaluators: {bar} {percentage}% | {value}/{total}\\n\",\n });\n const caller = new AsyncCaller({\n maxConcurrency,\n });\n const requests = runs.map(\n async (\n run,\n i\n ): Promise<{\n run_id: string;\n execution_time?: number;\n feedback: Feedback[];\n }> =>\n caller.call(async () => {\n const evaluatorResults = await Promise.allSettled(\n evaluators.map((evaluator) =>\n client.evaluateRun(run, evaluator, {\n referenceExample: examples[i],\n loadChildRuns: false,\n })\n )\n );\n progress.increment();\n return {\n execution_time:\n run?.end_time && run.start_time\n ? new Date(run.end_time).getTime() -\n new Date(run.start_time).getTime()\n : undefined,\n feedback: evaluatorResults.map((evalResult) =>\n evalResult.status === \"fulfilled\"\n ? evalResult.value\n : evalResult.reason\n ),\n run_id: run.id,\n };\n })\n );\n const results = await Promise.all(requests);\n\n return results.reduce(\n (acc, result, i) => ({\n ...acc,\n [examples[i].id]: result,\n }),\n {}\n );\n};\n\nexport type EvalResults = {\n projectName: string;\n results: {\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n };\n};\n\nconst getExamplesInputs = (\n examples: Example[],\n chainOrFactory: ChainOrFactory,\n dataType?: DataType\n) => {\n if (dataType === \"chat\") {\n // For some batty reason, we store the chat dataset differently.\n // { type: \"system\", data: { content: inputs.input } },\n // But we need to create AIMesage, SystemMessage, etc.\n return examples.map(({ inputs }) =>\n mapStoredMessagesToChatMessages(inputs.input)\n );\n }\n // If it's a language model and ALL example inputs have a single value,\n // then we can be friendly and flatten the inputs to a list of strings.\n const isLanguageModel =\n typeof chainOrFactory === \"object\" &&\n typeof (chainOrFactory as BaseLanguageModel)._llmType === \"function\";\n if (\n isLanguageModel &&\n examples.every(({ inputs }) => Object.keys(inputs).length === 1)\n ) {\n return examples.map(({ inputs }) => Object.values(inputs)[0]);\n }\n return examples.map(({ inputs }) => inputs);\n};\n\n/**\n * Evaluates a given model or chain against a specified LangSmith dataset.\n *\n * This function fetches example records from the specified dataset,\n * runs the model or chain against each example, and returns the evaluation\n * results.\n *\n * @param chainOrFactory - A model or factory/constructor function to be evaluated. It can be a\n * Runnable instance, a factory function that returns a Runnable, or a user-defined\n * function or factory.\n *\n * @param datasetName - The name of the dataset against which the evaluation will be\n * performed. This dataset should already be defined and contain the relevant data\n * for evaluation.\n *\n * @param options - (Optional) Additional parameters for the evaluation process:\n * - `evaluators` (RunEvalType[]): Evaluators to apply to a dataset run.\n * - `formatEvaluatorInputs` (EvaluatorInputFormatter): Convert the evaluation data into formats that can be used by the evaluator.\n * - `projectName` (string): Name of the project for logging and tracking.\n * - `projectMetadata` (Record<string, unknown>): Additional metadata for the project.\n * - `client` (Client): Client instance for LangSmith service interaction.\n * - `maxConcurrency` (number): Maximum concurrency level for dataset processing.\n *\n * @returns A promise that resolves to an `EvalResults` object. This object includes\n * detailed results of the evaluation, such as execution time, run IDs, and feedback\n * for each entry in the dataset.\n *\n * @example\n * ```typescript\n * // Example usage for evaluating a model on a dataset\n * async function evaluateModel() {\n * const chain = /* ...create your model or chain...*\\//\n * const datasetName = 'example-dataset';\n * const client = new Client(/* ...config... *\\//);\n *\n * const results = await runOnDataset(chain, datasetName, {\n * evaluators: [/* ...evaluators... *\\//],\n * client,\n * });\n *\n * console.log('Evaluation Results:', results);\n * }\n *\n * evaluateModel();\n * ```\n * In this example, `runOnDataset` is used to evaluate a language model (or a chain of models) against\n * a dataset named 'example-dataset'. The evaluation process is configured using `RunOnDatasetParams[\"evaluators\"]`, which can\n * include both standard and custom evaluators. The `Client` instance is used to interact with LangChain services.\n * The function returns the evaluation results, which can be logged or further processed as needed.\n */\n\nexport async function runOnDataset(\n chainOrFactory: ChainOrFactory,\n datasetName: string,\n options?: RunOnDatasetParams\n) {\n const {\n projectName,\n projectMetadata,\n client,\n maxConcurrency,\n }: RunOnDatasetParams = options ?? {};\n\n const evaluationConfig: RunEvalConfig | undefined =\n options?.evaluationConfig ??\n (options?.evaluators != null\n ? {\n evaluators: options.evaluators,\n formatEvaluatorInputs: options.formatEvaluatorInputs,\n }\n : undefined);\n\n const wrappedModel = await createWrappedModel(chainOrFactory);\n const testClient = client ?? new Client();\n const testProjectName = projectName ?? randomName();\n const dataset = await testClient.readDataset({ datasetName });\n const datasetId = dataset.id;\n const testConcurrency = maxConcurrency ?? 5;\n const { configs, examples, runExtractors } = await loadExamples({\n datasetName,\n client: testClient,\n projectName: testProjectName,\n maxConcurrency: testConcurrency,\n });\n\n await testClient.createProject({\n projectName: testProjectName,\n referenceDatasetId: datasetId,\n projectExtra: { metadata: { ...projectMetadata } },\n });\n const wrappedRunnable: Runnable = new RunnableLambda({\n func: wrappedModel,\n }).withConfig({ runName: \"evaluationRun\" });\n const runInputs = getExamplesInputs(\n examples,\n chainOrFactory,\n dataset.data_type\n );\n const progress = new ProgressBar({\n total: runInputs.length,\n format: \"Predicting: {bar} {percentage}% | {value}/{total}\",\n });\n // TODO: Collect the runs as well.\n await wrappedRunnable\n .withListeners({\n onEnd: () => progress.increment(),\n })\n // TODO: Insert evaluation inline for immediate feedback.\n .batch(runInputs, configs, {\n maxConcurrency,\n returnExceptions: true,\n });\n\n progress.complete();\n const runs: Run[] = [];\n for (let i = 0; i < examples.length; i += 1) {\n runs.push(await runExtractors[i].extract());\n }\n let evalResults: Record<\n string,\n { run_id: string; execution_time?: number; feedback: Feedback[] }\n > = {};\n if (evaluationConfig) {\n const loadedEvalConfig =\n await LoadedEvalConfig.fromRunEvalConfig(evaluationConfig);\n evalResults = await applyEvaluators({\n evaluation: loadedEvalConfig,\n runs,\n examples,\n client: testClient,\n maxConcurrency: testConcurrency,\n });\n }\n const results: EvalResults = {\n projectName: testProjectName,\n results: evalResults ?? {},\n };\n return results;\n}\n\nfunction isLangsmithTraceableFunction(x: unknown): x is AnyTraceableFunction {\n return typeof x === \"function\" && \"langsmith:traceable\" in x;\n}\n"],"mappings":";;;;;;;;;;;AAsDA,IAAM,uBAAN,MAA2B;CACzB;CAEA;CAEA,cAAc;AACZ,OAAK,eAAe,IAAI,SAAiB,YAAY;AACnD,QAAK,uBAAuB;IAC5B;;CAGJ,oBACE,QACA,SACA,UACG;AACH,OAAK,qBAAqB,MAAM;;CAGlC,MAAM,UAA2B;AAC/B,SAAO,KAAK;;;AAIhB,IAAM,qBAAN,cAAiC,WAAW;CAC1C;CAEA;;CAGA,OAAO;CAEP,cAAc;AACZ,SAAO;AACP,OAAK,aAAa,IAAI,SAAc,YAAY;AAC9C,QAAK,qBAAqB;IAC1B;;CAGJ,MAAM,WAAW,KAAU;AACzB,OAAK,mBAAmB,IAAI;;CAG9B,MAAM,UAAwB;AAC5B,SAAO,KAAK;;;;;;AAOhB,IAAM,sBAAN,MAAkD;CAChD;CAEA,YAAY,WAA6B;AACvC,OAAK,YAAY,IAAI,eAAe,EAAE,MAAM,WAAW,CAAC;;;;;;;;CAS1D,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAI,gBAAgB,EAAE,aAAa,cAAc,CAAC;EACjE,MAAM,SAAS,MAAM,KAAK,UAAU,OAClC;GACE;GACA;GACA,OAAO,IAAI;GACX,YAAY,IAAI;GAChB,WAAW,SAAS;GACrB,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;AAED,SAAO;GACL,aAFY,MAAM,UAAU,SAAS;GAGrC,GAAG;GACJ;;;AAKL,SAAS,qBAAqB,WAAiD;AAC7E,QAAO,aAAa,OAAO,UAAU,oBAAoB;;;;;;;;;AAa3D,IAAM,yBAAN,MAAM,+BAA+B,QAAQ;CAC3C;CAEA,wBAAgE,KAAA;CAEhE,YAAY,QAAuB,iBAAkC;AACnE,QAAM,OAAO;AAEb,OAAK,kBAAkB;;CAGzB,YAAY,QAA+C;EACzD,MAAM,QAAQ,IAAI,uBAChB;GACE,GAAG;GACH,YAAY;GACZ,cAAc,KAAK;GACnB,QAAQ,KAAK;GACd,EACD,KAAK,uBAAuB,UAAU,IAAI,KAAK,gBAChD;AACD,OAAK,WAAW,KAAK,MAAM;AAC3B,SAAO;;CAGT,MAAM,UAAyB;AAE7B,OAAK,wBAAwB,MAAM,KAAK,gBAAgB,iBACtD,OAAO,KAAK,eAAe,YACzB,KAAK,cAAc,QACnB,QAAQ,KAAK,aACX,KAAK,aACL;GACE,IAAI;IAAC;IAAa;IAAS;IAAyB;GACpD,IAAI;GACJ,MAAM;GACP,EACL,KAAK,QACL,KAAK,IACL,KAAK,UACL,KAAA,GACA,KAAA,GACA,KAAK,KACN;;CAGH,MAAM,WAA0B;AAC9B,MAAI,KAAK,MACP,OAAM,KAAK,uBAAuB,iBAChC,KAAK,OACL,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;MAED,OAAM,KAAK,uBAAuB,eAChC,KAAK,WAAW,EAAE,EAClB,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;;;AAKP,IAAM,oBAAN,cAAqD,SAGnD;CACA,kBAAkB;CAElB,eAAe,CAAC,kBAAkB,YAAY;CAE9C;CAEA,YAAY,QAAwC;AAClD,QAAM,OAAO;AAEb,MAAI,CAAC,6BAA6B,OAAO,KAAK,CAC5C,OAAM,IAAI,MACR,2FACD;AAGH,OAAK,OAAO,OAAO;;CAGrB,MAAM,OAAO,OAAiB,SAAmC;EAC/D,MAAM,CAAC,UAAU,KAAK,gBAAgB,WAAW,EAAE,EAAE,EAAE;EACvD,MAAM,kBAAkB,MAAM,4BAA4B,OAAO;EAEjE,MAAM,gBACJ,yBAAyB,KAAK,OACzB,KAAK,KAAK,yBACX,EAAE,MAAM,YAAY;AAE1B,MAAI,CAAC,gBAAiB,OAAM,IAAI,MAAM,4BAA4B;EAClE,MAAM,UAAU,IAAI,uBAClB;GACE,GAAG;GACH,YAAY,iBAAiB,eACzB,IAAI,QAAQ;IAAE,MAAM;IAAY,IAAI,iBAAiB;IAAc,CAAC,GACpE,KAAA;GACL,EACD,gBACD;AAED,MACE,OAAO,UAAU,YACjB,SAAS,QACT,OAAO,KAAK,MAAM,CAAC,WAAW,GAC9B;AACA,OAAI,UAAU,SAAS,MAAM,QAAQ,MAAM,CACzC,QAAQ,MAAM,KAAK,KAAK,SAAS,GAAG,MAAM;AAG5C,OACE,WAAW,SACX,EACE,OAAO,UAAU,YACjB,SAAS,QACT,CAAC,MAAM,QAAQ,MAAM,IAErB,EAAE,iBAAiB,OAGrB,KAAI;AACF,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM,MAAM;WACvC;AACN,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;AAK7C,SAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;;;;;;AAS3C,IAAM,uBAAN,MAAM,qBAA6C;CACjD;CAEA;CAEA;CAEA;CAEA,YACE,WACA,gBACA,uBACA;AACA,OAAK,YAAY;AACjB,OAAK,oBAAoB,OAAO,WAAW,oBAAoB;AAC/D,OAAK,iBAAiB;AACtB,OAAK,wBAAwB;;CAG/B,aAAa,eACX,QAC+B;EAC/B,MAAM,gBACJ,OAAO,WAAW,WAAW,SAAS,OAAO;EAC/C,MAAM,aAAa,OAAO,WAAW,WAAY,EAAE,GAAkB;EACrE,MAAM,YAAY,MAAM,cAAc,eAAe,WAAW;EAChE,MAAM,cAAc,YAAY,eAAe,WAAW;AAC1D,MAAI,CAAC,qBAAqB,UAAU,CAClC,OAAM,IAAI,MACR,qBAAqB,cAAc,8GAGpC;AAEH,MAAI,CAAC,YACH,OAAM,IAAI,MACR,qBAAqB,cAAc,uGAEpC;AAEH,SAAO,IAAI,qBACT,WACA,aACA,YAAY,sBACb;;;;;;;;CASH,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,EAAE,YAAY,OAAO,cAAc,KAAK,sBAAsB;GAClE,UAAU,IAAI;GACd,eAAe,IAAI;GACnB,oBAAoB,SAAS;GAC7B;GACD,CAAC;EACF,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAI,gBAAgB,EAAE,aAAa,cAAc,CAAC;AACjE,MAAI,KAAK,mBAAmB;GAC1B,MAAM,aAAa,MAAM,KAAK,UAAU,gBACtC;IACc;IACD;IACJ;IACR,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;GACD,MAAM,QAAQ,MAAM,UAAU,SAAS;AACvC,UAAO;IACL,KAAK,KAAK;IACV,SAAS,YAAY;IACrB,aAAa;IACb,GAAG;IACJ;;AAEH,QAAM,IAAI,MACR,wHAGD;;;AAIL,IAAM,mBAAN,MAAM,iBAAiB;CACrB,YAAY,YAA2D;AAApD,OAAA,aAAA;;CAEnB,aAAa,kBACX,QAC2B;EAE3B,MAAM,oBACJ,QAAQ,oBAAoB,OAAO,YAAY,OAAO,kBAAkB,GACvE,KAAK,cAAc;AACpB,OAAI,OAAO,cAAc,WACvB,QAAO,IAAI,oBAAoB,UAAU;OAEzC,QAAO;IAET;EAEF,MAAM,wBAAwB,MAAM,QAAQ,IAC1C,QAAQ,YACJ,OAAO,uBAAuB,EAC9B,IACA,OAAO,cACL,MAAM,qBAAqB,eAAe,UAAU,CACvD,IAAI,EAAE,CACV;AACD,SAAO,IAAI,kBACR,oBAAoB,EAAE,EAAE,OAAO,yBAAyB,EAAE,CAAC,CAC7D;;;;;;;;;;;AA0CL,MAAM,qBAAqB,OAAO,mBAAmC;AACnE,KAAI,SAAS,WAAW,eAAe,CACrC,cAAa;AAEf,KAAI,OAAO,mBAAmB,YAAY;AACxC,MAAI,6BAA6B,eAAe,EAAE;GAChD,MAAM,eAAe,IAAI,kBAAkB,EAAE,MAAM,gBAAgB,CAAC;AACpE,gBAAa;;AAGf,MAAI;GAEF,IAAI,MAAO,gBAAmC;AAC9C,OACE,OACA,OAAQ,IAAqC,SAAS,WAEtD,OAAM,MAAM;AAEd,UAAO;UACD;GAGN,MAAM,eAAe,IAAI,eAAe,EAAE,MAAM,gBAAgB,CAAC;AACjE,gBAAa;;;AAGjB,OAAM,IAAI,MAAM,yBAAyB;;AAG3C,MAAM,eAAe,OAAO,EAC1B,aACA,QACA,kBAMI;CACJ,MAAM,kBAAkB,OAAO,aAAa,EAAE,aAAa,CAAC;CAC5D,MAAM,UAA4B,EAAE;CACpC,MAAM,gBAAgB,EAAE;CACxB,MAAM,WAAW,EAAE;AACnB,YAAW,MAAM,WAAW,iBAAiB;EAC3C,MAAM,eAAe,IAAI,oBAAoB;AAC7C,UAAQ,KAAK,EACX,WAAW,CACT,IAAI,gBAAgB;GAAE,WAAW,QAAQ;GAAI;GAAa,CAAC,EAC3D,aACD,EACF,CAAC;AACF,WAAS,KAAK,QAAQ;AACtB,gBAAc,KAAK,aAAa;;AAElC,QAAO;EACL;EACA;EACA;EACD;;AAGH,MAAM,kBAAkB,OAAO,EAC7B,YACA,MACA,UACA,QACA,qBAaI;CAEJ,MAAM,EAAE,eAAe;CACvB,MAAM,WAAW,IAAI,YAAY;EAC/B,OAAO,SAAS;EAChB,QAAQ;EACT,CAAC;CACF,MAAM,SAAS,IAAI,YAAY,EAC7B,gBACD,CAAC;CACF,MAAM,WAAW,KAAK,IACpB,OACE,KACA,MAMA,OAAO,KAAK,YAAY;EACtB,MAAM,mBAAmB,MAAM,QAAQ,WACrC,WAAW,KAAK,cACd,OAAO,YAAY,KAAK,WAAW;GACjC,kBAAkB,SAAS;GAC3B,eAAe;GAChB,CAAC,CACH,CACF;AACD,WAAS,WAAW;AACpB,SAAO;GACL,gBACE,KAAK,YAAY,IAAI,aACjB,IAAI,KAAK,IAAI,SAAS,CAAC,SAAS,GAChC,IAAI,KAAK,IAAI,WAAW,CAAC,SAAS,GAClC,KAAA;GACN,UAAU,iBAAiB,KAAK,eAC9B,WAAW,WAAW,cAClB,WAAW,QACX,WAAW,OAChB;GACD,QAAQ,IAAI;GACb;GACD,CACL;AAGD,SAFgB,MAAM,QAAQ,IAAI,SAAS,EAE5B,QACZ,KAAK,QAAQ,OAAO;EACnB,GAAG;GACF,SAAS,GAAG,KAAK;EACnB,GACD,EAAE,CACH;;AAcH,MAAM,qBACJ,UACA,gBACA,aACG;AACH,KAAI,aAAa,OAIf,QAAO,SAAS,KAAK,EAAE,aACrB,gCAAgC,OAAO,MAAM,CAC9C;AAOH,KAFE,OAAO,mBAAmB,YAC1B,OAAQ,eAAqC,aAAa,cAG1D,SAAS,OAAO,EAAE,aAAa,OAAO,KAAK,OAAO,CAAC,WAAW,EAAE,CAEhE,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO,OAAO,OAAO,CAAC,GAAG;AAE/D,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsD7C,eAAsB,aACpB,gBACA,aACA,SACA;CACA,MAAM,EACJ,aACA,iBACA,QACA,mBACsB,WAAW,EAAE;CAErC,MAAM,mBACJ,SAAS,qBACR,SAAS,cAAc,OACpB;EACE,YAAY,QAAQ;EACpB,uBAAuB,QAAQ;EAChC,GACD,KAAA;CAEN,MAAM,eAAe,MAAM,mBAAmB,eAAe;CAC7D,MAAM,aAAa,UAAU,IAAI,QAAQ;CACzC,MAAM,kBAAkB,eAAe,YAAY;CACnD,MAAM,UAAU,MAAM,WAAW,YAAY,EAAE,aAAa,CAAC;CAC7D,MAAM,YAAY,QAAQ;CAC1B,MAAM,kBAAkB,kBAAkB;CAC1C,MAAM,EAAE,SAAS,UAAU,kBAAkB,MAAM,aAAa;EAC9D;EACA,QAAQ;EACR,aAAa;EACb,gBAAgB;EACjB,CAAC;AAEF,OAAM,WAAW,cAAc;EAC7B,aAAa;EACb,oBAAoB;EACpB,cAAc,EAAE,UAAU,EAAE,GAAG,iBAAiB,EAAE;EACnD,CAAC;CACF,MAAM,kBAA4B,IAAI,eAAe,EACnD,MAAM,cACP,CAAC,CAAC,WAAW,EAAE,SAAS,iBAAiB,CAAC;CAC3C,MAAM,YAAY,kBAChB,UACA,gBACA,QAAQ,UACT;CACD,MAAM,WAAW,IAAI,YAAY;EAC/B,OAAO,UAAU;EACjB,QAAQ;EACT,CAAC;AAEF,OAAM,gBACH,cAAc,EACb,aAAa,SAAS,WAAW,EAClC,CAAC,CAED,MAAM,WAAW,SAAS;EACzB;EACA,kBAAkB;EACnB,CAAC;AAEJ,UAAS,UAAU;CACnB,MAAM,OAAc,EAAE;AACtB,MAAK,IAAI,IAAI,GAAG,IAAI,SAAS,QAAQ,KAAK,EACxC,MAAK,KAAK,MAAM,cAAc,GAAG,SAAS,CAAC;CAE7C,IAAI,cAGA,EAAE;AACN,KAAI,iBAGF,eAAc,MAAM,gBAAgB;EAClC,YAFA,MAAM,iBAAiB,kBAAkB,iBAAiB;EAG1D;EACA;EACA,QAAQ;EACR,gBAAgB;EACjB,CAAC;AAMJ,QAJ6B;EAC3B,aAAa;EACb,SAAS,eAAe,EAAE;EAC3B;;AAIH,SAAS,6BAA6B,GAAuC;AAC3E,QAAO,OAAO,MAAM,cAAc,yBAAyB"}
1
+ {"version":3,"file":"runner_utils.js","names":[],"sources":["../../src/smith/runner_utils.ts"],"sourcesContent":["import { BaseLanguageModel } from \"@langchain/core/language_models/base\";\nimport { Serialized } from \"@langchain/core/load/serializable\";\nimport { mapStoredMessagesToChatMessages } from \"@langchain/core/messages\";\nimport {\n Runnable,\n RunnableConfig,\n RunnableLambda,\n getCallbackManagerForConfig,\n} from \"@langchain/core/runnables\";\nimport { LangChainTracer } from \"@langchain/core/tracers/tracer_langchain\";\nimport { BaseTracer } from \"@langchain/core/tracers/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { AsyncCaller } from \"@langchain/core/utils/async_caller\";\nimport type {\n CallbackManager,\n CallbackManagerForChainRun,\n} from \"@langchain/core/callbacks/manager\";\nimport {\n Client,\n Example,\n Feedback,\n Run,\n RunTree,\n RunTreeConfig,\n} from \"langsmith\";\nimport { EvaluationResult, RunEvaluator } from \"langsmith/evaluation\";\nimport { DataType } from \"langsmith/schemas\";\nimport type { TraceableFunction } from \"langsmith/singletons/traceable\";\nimport { LLMStringEvaluator } from \"../evaluation/base.js\";\nimport { loadEvaluator } from \"../evaluation/loader.js\";\nimport { EvaluatorType } from \"../evaluation/types.js\";\nimport {\n isOffTheShelfEvaluator,\n type DynamicRunEvaluatorParams,\n type EvalConfig,\n type EvaluatorInputFormatter,\n type RunEvalConfig,\n type RunEvaluatorLike,\n isCustomEvaluator,\n} from \"./config.js\";\nimport { randomName } from \"./name_generation.js\";\nimport { ProgressBar } from \"./progress.js\";\n\nexport type ChainOrFactory =\n | Runnable\n | (() => Runnable)\n | AnyTraceableFunction\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => any)\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n | ((obj: any) => Promise<any>)\n | (() => (obj: unknown) => unknown)\n | (() => (obj: unknown) => Promise<unknown>);\n\nclass SingleRunIdExtractor {\n runIdPromiseResolver: (runId: string) => void;\n\n runIdPromise: Promise<string>;\n\n constructor() {\n this.runIdPromise = new Promise<string>((extract) => {\n this.runIdPromiseResolver = extract;\n });\n }\n\n handleChainStart = (\n _chain: Serialized,\n _inputs: ChainValues,\n runId: string\n ) => {\n this.runIdPromiseResolver(runId);\n };\n\n async extract(): Promise<string> {\n return this.runIdPromise;\n }\n}\n\nclass SingleRunExtractor extends BaseTracer {\n runPromiseResolver: (run: Run) => void;\n\n runPromise: Promise<Run>;\n\n /** The name of the callback handler. */\n name = \"single_run_extractor\";\n\n constructor() {\n super();\n this.runPromise = new Promise<Run>((extract) => {\n this.runPromiseResolver = extract;\n });\n }\n\n async persistRun(run: Run) {\n this.runPromiseResolver(run);\n }\n\n async extract(): Promise<Run> {\n return this.runPromise;\n }\n}\n\n/**\n * Wraps an evaluator function + implements the RunEvaluator interface.\n */\nclass DynamicRunEvaluator implements RunEvaluator {\n evaluator: RunnableLambda<DynamicRunEvaluatorParams, EvaluationResult>;\n\n constructor(evaluator: RunEvaluatorLike) {\n this.evaluator = new RunnableLambda({ func: evaluator });\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n const result = await this.evaluator.invoke(\n {\n run,\n example,\n input: run.inputs,\n prediction: run.outputs,\n reference: example?.outputs,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n sourceRunId: runId,\n ...result,\n };\n }\n}\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\nfunction isLLMStringEvaluator(evaluator: any): evaluator is LLMStringEvaluator {\n return evaluator && typeof evaluator.evaluateStrings === \"function\";\n}\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\ntype AnyTraceableFunction = TraceableFunction<(...any: any[]) => any>;\n\n/**\n * Internal implementation of RunTree, which uses the\n * provided callback manager instead of the internal LangSmith client.\n *\n * The goal of this class is to ensure seamless interop when intergrated\n * with other Runnables.\n */\nclass CallbackManagerRunTree extends RunTree {\n callbackManager: CallbackManager;\n\n activeCallbackManager: CallbackManagerForChainRun | undefined = undefined;\n\n constructor(config: RunTreeConfig, callbackManager: CallbackManager) {\n super(config);\n\n this.callbackManager = callbackManager;\n }\n\n createChild(config: RunTreeConfig): CallbackManagerRunTree {\n const child = new CallbackManagerRunTree(\n {\n ...config,\n parent_run: this,\n project_name: this.project_name,\n client: this.client,\n },\n this.activeCallbackManager?.getChild() ?? this.callbackManager\n );\n this.child_runs.push(child);\n return child;\n }\n\n async postRun(): Promise<void> {\n // how it is translated in comparison to basic RunTree?\n this.activeCallbackManager = await this.callbackManager.handleChainStart(\n typeof this.serialized !== \"object\" &&\n this.serialized != null &&\n \"lc\" in this.serialized\n ? this.serialized\n : {\n id: [\"langchain\", \"smith\", \"CallbackManagerRunTree\"],\n lc: 1,\n type: \"not_implemented\",\n },\n this.inputs,\n this.id,\n this.run_type,\n undefined,\n undefined,\n this.name\n );\n }\n\n async patchRun(): Promise<void> {\n if (this.error) {\n await this.activeCallbackManager?.handleChainError(\n this.error,\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n } else {\n await this.activeCallbackManager?.handleChainEnd(\n this.outputs ?? {},\n this.id,\n this.parent_run?.id,\n undefined,\n undefined\n );\n }\n }\n}\n\nclass RunnableTraceable<RunInput, RunOutput> extends Runnable<\n RunInput,\n RunOutput\n> {\n lc_serializable = false;\n\n lc_namespace = [\"langchain_core\", \"runnables\"];\n\n protected func: AnyTraceableFunction;\n\n constructor(fields: { func: AnyTraceableFunction }) {\n super(fields);\n\n if (!isLangsmithTraceableFunction(fields.func)) {\n throw new Error(\n \"RunnableTraceable requires a function that is wrapped in traceable higher-order function\"\n );\n }\n\n this.func = fields.func;\n }\n\n async invoke(input: RunInput, options?: Partial<RunnableConfig>) {\n const [config] = this._getOptionsList(options ?? {}, 1);\n const callbackManager = await getCallbackManagerForConfig(config);\n\n const partialConfig =\n \"langsmith:traceable\" in this.func\n ? (this.func[\"langsmith:traceable\"] as RunTreeConfig)\n : { name: \"<lambda>\" };\n\n if (!callbackManager) throw new Error(\"CallbackManager not found\");\n const runTree = new CallbackManagerRunTree(\n {\n ...partialConfig,\n parent_run: callbackManager?._parentRunId\n ? new RunTree({ name: \"<parent>\", id: callbackManager?._parentRunId })\n : undefined,\n },\n callbackManager\n );\n\n if (\n typeof input === \"object\" &&\n input != null &&\n Object.keys(input).length === 1\n ) {\n if (\"args\" in input && Array.isArray(input)) {\n return (await this.func(runTree, ...input)) as RunOutput;\n }\n\n if (\n \"input\" in input &&\n !(\n typeof input === \"object\" &&\n input != null &&\n !Array.isArray(input) &&\n // oxlint-disable-next-line no-instanceof/no-instanceof\n !(input instanceof Date)\n )\n ) {\n try {\n return (await this.func(runTree, input.input)) as RunOutput;\n } catch {\n return (await this.func(runTree, input)) as RunOutput;\n }\n }\n }\n\n return (await this.func(runTree, input)) as RunOutput;\n }\n}\n\n/**\n * Wraps an off-the-shelf evaluator (loaded using loadEvaluator; of EvaluatorType[T])\n * and composes with a prepareData function so the user can prepare the trace and\n * dataset data for the evaluator.\n */\nclass PreparedRunEvaluator implements RunEvaluator {\n evaluator: LLMStringEvaluator;\n\n formatEvaluatorInputs: EvaluatorInputFormatter;\n\n isStringEvaluator: boolean;\n\n evaluationName: string;\n\n constructor(\n evaluator: LLMStringEvaluator,\n evaluationName: string,\n formatEvaluatorInputs: EvaluatorInputFormatter\n ) {\n this.evaluator = evaluator;\n this.isStringEvaluator = typeof evaluator?.evaluateStrings === \"function\";\n this.evaluationName = evaluationName;\n this.formatEvaluatorInputs = formatEvaluatorInputs;\n }\n\n static async fromEvalConfig(\n config: EvalConfig | keyof EvaluatorType\n ): Promise<PreparedRunEvaluator> {\n const evaluatorType =\n typeof config === \"string\" ? config : config.evaluatorType;\n const evalConfig = typeof config === \"string\" ? ({} as EvalConfig) : config;\n const evaluator = await loadEvaluator(evaluatorType, evalConfig);\n const feedbackKey = evalConfig?.feedbackKey ?? evaluator?.evaluationName;\n if (!isLLMStringEvaluator(evaluator)) {\n throw new Error(\n `Evaluator of type ${evaluatorType} not yet supported. ` +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n if (!feedbackKey) {\n throw new Error(\n `Evaluator of type ${evaluatorType} must have an evaluationName` +\n ` or feedbackKey. Please manually provide a feedbackKey in the EvalConfig.`\n );\n }\n return new PreparedRunEvaluator(\n evaluator as LLMStringEvaluator,\n feedbackKey,\n evalConfig?.formatEvaluatorInputs\n );\n }\n\n /**\n * Evaluates a run with an optional example and returns the evaluation result.\n * @param run The run to evaluate.\n * @param example The optional example to use for evaluation.\n * @returns A promise that extracts to the evaluation result.\n */\n async evaluateRun(run: Run, example?: Example): Promise<EvaluationResult> {\n const { prediction, input, reference } = this.formatEvaluatorInputs({\n rawInput: run.inputs,\n rawPrediction: run.outputs,\n rawReferenceOutput: example?.outputs,\n run,\n });\n const extractor = new SingleRunIdExtractor();\n const tracer = new LangChainTracer({ projectName: \"evaluators\" });\n if (this.isStringEvaluator) {\n const evalResult = await this.evaluator.evaluateStrings(\n {\n prediction: prediction as string,\n reference: reference as string,\n input: input as string,\n },\n {\n callbacks: [extractor, tracer],\n }\n );\n const runId = await extractor.extract();\n return {\n key: this.evaluationName,\n comment: evalResult?.reasoning,\n sourceRunId: runId,\n ...evalResult,\n };\n }\n throw new Error(\n \"Evaluator not yet supported. \" +\n \"Please use a string evaluator, or implement your \" +\n \"evaluation logic as a custom evaluator.\"\n );\n }\n}\n\nclass LoadedEvalConfig {\n constructor(public evaluators: (RunEvaluator | DynamicRunEvaluator)[]) {}\n\n static async fromRunEvalConfig(\n config: RunEvalConfig<keyof EvaluatorType>\n ): Promise<LoadedEvalConfig> {\n // Custom evaluators are applied \"as-is\"\n const customEvaluators = (\n config?.customEvaluators ?? config.evaluators?.filter(isCustomEvaluator)\n )?.map((evaluator) => {\n if (typeof evaluator === \"function\") {\n return new DynamicRunEvaluator(evaluator);\n } else {\n return evaluator;\n }\n });\n\n const offTheShelfEvaluators = await Promise.all(\n config?.evaluators\n ?.filter(isOffTheShelfEvaluator)\n ?.map(\n async (evaluator) =>\n await PreparedRunEvaluator.fromEvalConfig(evaluator)\n ) ?? []\n );\n return new LoadedEvalConfig(\n (customEvaluators ?? []).concat(offTheShelfEvaluators ?? [])\n );\n }\n}\n\nexport interface RunOnDatasetParams extends Omit<\n RunEvalConfig,\n \"customEvaluators\"\n> {\n /**\n * Name of the project for logging and tracking.\n */\n projectName?: string;\n\n /**\n * Additional metadata for the project.\n */\n projectMetadata?: Record<string, unknown>;\n\n /**\n * Client instance for LangSmith service interaction.\n */\n client?: Client;\n\n /**\n * Maximum concurrency level for dataset processing.\n */\n maxConcurrency?: number;\n\n /**\n * @deprecated Pass keys directly to the RunOnDatasetParams instead\n */\n evaluationConfig?: RunEvalConfig;\n}\n\n/**\n * Internals expect a constructor () -> Runnable. This function wraps/coerces\n * the provided LangChain object, custom function, or factory function into\n * a constructor of a runnable.\n * @param modelOrFactory The model or factory to create a wrapped model from.\n * @returns A function that returns the wrapped model.\n * @throws Error if the modelOrFactory is invalid.\n */\nconst createWrappedModel = async (modelOrFactory: ChainOrFactory) => {\n if (Runnable.isRunnable(modelOrFactory)) {\n return () => modelOrFactory;\n }\n if (typeof modelOrFactory === \"function\") {\n if (isLangsmithTraceableFunction(modelOrFactory)) {\n const wrappedModel = new RunnableTraceable({ func: modelOrFactory });\n return () => wrappedModel;\n }\n\n try {\n // If it works with no arguments, assume it's a factory\n let res = (modelOrFactory as () => Runnable)();\n if (\n res &&\n typeof (res as unknown as Promise<Runnable>).then === \"function\"\n ) {\n res = await res;\n }\n return modelOrFactory as () => Runnable;\n } catch {\n // Otherwise, it's a custom UDF, and we'll wrap\n // the function in a lambda\n const wrappedModel = new RunnableLambda({ func: modelOrFactory });\n return () => wrappedModel;\n }\n }\n throw new Error(\"Invalid modelOrFactory\");\n};\n\nconst loadExamples = async ({\n datasetName,\n client,\n projectName,\n}: {\n datasetName: string;\n client: Client;\n projectName: string;\n maxConcurrency: number;\n}) => {\n const exampleIterator = client.listExamples({ datasetName });\n const configs: RunnableConfig[] = [];\n const runExtractors = [];\n const examples = [];\n for await (const example of exampleIterator) {\n const runExtractor = new SingleRunExtractor();\n configs.push({\n callbacks: [\n new LangChainTracer({ exampleId: example.id, projectName }),\n runExtractor,\n ],\n });\n examples.push(example);\n runExtractors.push(runExtractor);\n }\n return {\n configs,\n examples,\n runExtractors,\n };\n};\n\nconst applyEvaluators = async ({\n evaluation,\n runs,\n examples,\n client,\n maxConcurrency,\n}: {\n evaluation: LoadedEvalConfig;\n runs: Run[];\n examples: Example[];\n client: Client;\n maxConcurrency: number;\n}): Promise<{\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n}> => {\n // TODO: Parallelize and/or put in callbacks to speed up evals.\n const { evaluators } = evaluation;\n const progress = new ProgressBar({\n total: examples.length,\n format: \"Running Evaluators: {bar} {percentage}% | {value}/{total}\\n\",\n });\n const caller = new AsyncCaller({\n maxConcurrency,\n });\n const requests = runs.map(\n async (\n run,\n i\n ): Promise<{\n run_id: string;\n execution_time?: number;\n feedback: Feedback[];\n }> =>\n caller.call(async () => {\n const evaluatorResults = await Promise.allSettled(\n evaluators.map((evaluator) =>\n client.evaluateRun(run, evaluator, {\n referenceExample: examples[i],\n loadChildRuns: false,\n })\n )\n );\n progress.increment();\n return {\n execution_time:\n run?.end_time && run.start_time\n ? new Date(run.end_time).getTime() -\n new Date(run.start_time).getTime()\n : undefined,\n feedback: evaluatorResults.map((evalResult) =>\n evalResult.status === \"fulfilled\"\n ? evalResult.value\n : evalResult.reason\n ),\n run_id: run.id,\n };\n })\n );\n const results = await Promise.all(requests);\n\n return results.reduce(\n (acc, result, i) => ({\n ...acc,\n [examples[i].id]: result,\n }),\n {}\n );\n};\n\nexport type EvalResults = {\n projectName: string;\n results: {\n [key: string]: {\n execution_time?: number;\n run_id: string;\n feedback: Feedback[];\n };\n };\n};\n\nconst getExamplesInputs = (\n examples: Example[],\n chainOrFactory: ChainOrFactory,\n dataType?: DataType\n) => {\n if (dataType === \"chat\") {\n // For some batty reason, we store the chat dataset differently.\n // { type: \"system\", data: { content: inputs.input } },\n // But we need to create AIMesage, SystemMessage, etc.\n return examples.map(({ inputs }) =>\n mapStoredMessagesToChatMessages(inputs.input)\n );\n }\n // If it's a language model and ALL example inputs have a single value,\n // then we can be friendly and flatten the inputs to a list of strings.\n const isLanguageModel =\n typeof chainOrFactory === \"object\" &&\n typeof (chainOrFactory as BaseLanguageModel)._llmType === \"function\";\n if (\n isLanguageModel &&\n examples.every(({ inputs }) => Object.keys(inputs).length === 1)\n ) {\n return examples.map(({ inputs }) => Object.values(inputs)[0]);\n }\n return examples.map(({ inputs }) => inputs);\n};\n\n/**\n * Evaluates a given model or chain against a specified LangSmith dataset.\n *\n * This function fetches example records from the specified dataset,\n * runs the model or chain against each example, and returns the evaluation\n * results.\n *\n * @param chainOrFactory - A model or factory/constructor function to be evaluated. It can be a\n * Runnable instance, a factory function that returns a Runnable, or a user-defined\n * function or factory.\n *\n * @param datasetName - The name of the dataset against which the evaluation will be\n * performed. This dataset should already be defined and contain the relevant data\n * for evaluation.\n *\n * @param options - (Optional) Additional parameters for the evaluation process:\n * - `evaluators` (RunEvalType[]): Evaluators to apply to a dataset run.\n * - `formatEvaluatorInputs` (EvaluatorInputFormatter): Convert the evaluation data into formats that can be used by the evaluator.\n * - `projectName` (string): Name of the project for logging and tracking.\n * - `projectMetadata` (Record<string, unknown>): Additional metadata for the project.\n * - `client` (Client): Client instance for LangSmith service interaction.\n * - `maxConcurrency` (number): Maximum concurrency level for dataset processing.\n *\n * @returns A promise that resolves to an `EvalResults` object. This object includes\n * detailed results of the evaluation, such as execution time, run IDs, and feedback\n * for each entry in the dataset.\n *\n * @example\n * ```typescript\n * // Example usage for evaluating a model on a dataset\n * async function evaluateModel() {\n * const chain = /* ...create your model or chain...*\\//\n * const datasetName = 'example-dataset';\n * const client = new Client(/* ...config... *\\//);\n *\n * const results = await runOnDataset(chain, datasetName, {\n * evaluators: [/* ...evaluators... *\\//],\n * client,\n * });\n *\n * console.log('Evaluation Results:', results);\n * }\n *\n * evaluateModel();\n * ```\n * In this example, `runOnDataset` is used to evaluate a language model (or a chain of models) against\n * a dataset named 'example-dataset'. The evaluation process is configured using `RunOnDatasetParams[\"evaluators\"]`, which can\n * include both standard and custom evaluators. The `Client` instance is used to interact with LangChain services.\n * The function returns the evaluation results, which can be logged or further processed as needed.\n */\n\nexport async function runOnDataset(\n chainOrFactory: ChainOrFactory,\n datasetName: string,\n options?: RunOnDatasetParams\n) {\n const {\n projectName,\n projectMetadata,\n client,\n maxConcurrency,\n }: RunOnDatasetParams = options ?? {};\n\n const evaluationConfig: RunEvalConfig | undefined =\n options?.evaluationConfig ??\n (options?.evaluators != null\n ? {\n evaluators: options.evaluators,\n formatEvaluatorInputs: options.formatEvaluatorInputs,\n }\n : undefined);\n\n const wrappedModel = await createWrappedModel(chainOrFactory);\n const testClient = client ?? new Client();\n const testProjectName = projectName ?? randomName();\n const dataset = await testClient.readDataset({ datasetName });\n const datasetId = dataset.id;\n const testConcurrency = maxConcurrency ?? 5;\n const { configs, examples, runExtractors } = await loadExamples({\n datasetName,\n client: testClient,\n projectName: testProjectName,\n maxConcurrency: testConcurrency,\n });\n\n await testClient.createProject({\n projectName: testProjectName,\n referenceDatasetId: datasetId,\n projectExtra: { metadata: { ...projectMetadata } },\n });\n const wrappedRunnable: Runnable = new RunnableLambda({\n func: wrappedModel,\n }).withConfig({ runName: \"evaluationRun\" });\n const runInputs = getExamplesInputs(\n examples,\n chainOrFactory,\n dataset.data_type\n );\n const progress = new ProgressBar({\n total: runInputs.length,\n format: \"Predicting: {bar} {percentage}% | {value}/{total}\",\n });\n // TODO: Collect the runs as well.\n await wrappedRunnable\n .withListeners({\n onEnd: () => progress.increment(),\n })\n // TODO: Insert evaluation inline for immediate feedback.\n .batch(runInputs, configs, {\n maxConcurrency,\n returnExceptions: true,\n });\n\n progress.complete();\n const runs: Run[] = [];\n for (let i = 0; i < examples.length; i += 1) {\n runs.push(await runExtractors[i].extract());\n }\n let evalResults: Record<\n string,\n { run_id: string; execution_time?: number; feedback: Feedback[] }\n > = {};\n if (evaluationConfig) {\n const loadedEvalConfig =\n await LoadedEvalConfig.fromRunEvalConfig(evaluationConfig);\n evalResults = await applyEvaluators({\n evaluation: loadedEvalConfig,\n runs,\n examples,\n client: testClient,\n maxConcurrency: testConcurrency,\n });\n }\n const results: EvalResults = {\n projectName: testProjectName,\n results: evalResults ?? {},\n };\n return results;\n}\n\nfunction isLangsmithTraceableFunction(x: unknown): x is AnyTraceableFunction {\n return typeof x === \"function\" && \"langsmith:traceable\" in x;\n}\n"],"mappings":";;;;;;;;;;;AAsDA,IAAM,uBAAN,MAA2B;CACzB;CAEA;CAEA,cAAc;AACZ,OAAK,eAAe,IAAI,SAAiB,YAAY;AACnD,QAAK,uBAAuB;IAC5B;;CAGJ,oBACE,QACA,SACA,UACG;AACH,OAAK,qBAAqB,MAAM;;CAGlC,MAAM,UAA2B;AAC/B,SAAO,KAAK;;;AAIhB,IAAM,qBAAN,cAAiC,WAAW;CAC1C;CAEA;;CAGA,OAAO;CAEP,cAAc;AACZ,SAAO;AACP,OAAK,aAAa,IAAI,SAAc,YAAY;AAC9C,QAAK,qBAAqB;IAC1B;;CAGJ,MAAM,WAAW,KAAU;AACzB,OAAK,mBAAmB,IAAI;;CAG9B,MAAM,UAAwB;AAC5B,SAAO,KAAK;;;;;;AAOhB,IAAM,sBAAN,MAAkD;CAChD;CAEA,YAAY,WAA6B;AACvC,OAAK,YAAY,IAAI,eAAe,EAAE,MAAM,WAAW,CAAC;;;;;;;;CAS1D,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAI,gBAAgB,EAAE,aAAa,cAAc,CAAC;EACjE,MAAM,SAAS,MAAM,KAAK,UAAU,OAClC;GACE;GACA;GACA,OAAO,IAAI;GACX,YAAY,IAAI;GAChB,WAAW,SAAS;GACrB,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;AAED,SAAO;GACL,aAFY,MAAM,UAAU,SAAS;GAGrC,GAAG;GACJ;;;AAKL,SAAS,qBAAqB,WAAiD;AAC7E,QAAO,aAAa,OAAO,UAAU,oBAAoB;;;;;;;;;AAa3D,IAAM,yBAAN,MAAM,+BAA+B,QAAQ;CAC3C;CAEA,wBAAgE,KAAA;CAEhE,YAAY,QAAuB,iBAAkC;AACnE,QAAM,OAAO;AAEb,OAAK,kBAAkB;;CAGzB,YAAY,QAA+C;EACzD,MAAM,QAAQ,IAAI,uBAChB;GACE,GAAG;GACH,YAAY;GACZ,cAAc,KAAK;GACnB,QAAQ,KAAK;GACd,EACD,KAAK,uBAAuB,UAAU,IAAI,KAAK,gBAChD;AACD,OAAK,WAAW,KAAK,MAAM;AAC3B,SAAO;;CAGT,MAAM,UAAyB;AAE7B,OAAK,wBAAwB,MAAM,KAAK,gBAAgB,iBACtD,OAAO,KAAK,eAAe,YACzB,KAAK,cAAc,QACnB,QAAQ,KAAK,aACX,KAAK,aACL;GACE,IAAI;IAAC;IAAa;IAAS;IAAyB;GACpD,IAAI;GACJ,MAAM;GACP,EACL,KAAK,QACL,KAAK,IACL,KAAK,UACL,KAAA,GACA,KAAA,GACA,KAAK,KACN;;CAGH,MAAM,WAA0B;AAC9B,MAAI,KAAK,MACP,OAAM,KAAK,uBAAuB,iBAChC,KAAK,OACL,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;MAED,OAAM,KAAK,uBAAuB,eAChC,KAAK,WAAW,EAAE,EAClB,KAAK,IACL,KAAK,YAAY,IACjB,KAAA,GACA,KAAA,EACD;;;AAKP,IAAM,oBAAN,cAAqD,SAGnD;CACA,kBAAkB;CAElB,eAAe,CAAC,kBAAkB,YAAY;CAE9C;CAEA,YAAY,QAAwC;AAClD,QAAM,OAAO;AAEb,MAAI,CAAC,6BAA6B,OAAO,KAAK,CAC5C,OAAM,IAAI,MACR,2FACD;AAGH,OAAK,OAAO,OAAO;;CAGrB,MAAM,OAAO,OAAiB,SAAmC;EAC/D,MAAM,CAAC,UAAU,KAAK,gBAAgB,WAAW,EAAE,EAAE,EAAE;EACvD,MAAM,kBAAkB,MAAM,4BAA4B,OAAO;EAEjE,MAAM,gBACJ,yBAAyB,KAAK,OACzB,KAAK,KAAK,yBACX,EAAE,MAAM,YAAY;AAE1B,MAAI,CAAC,gBAAiB,OAAM,IAAI,MAAM,4BAA4B;EAClE,MAAM,UAAU,IAAI,uBAClB;GACE,GAAG;GACH,YAAY,iBAAiB,eACzB,IAAI,QAAQ;IAAE,MAAM;IAAY,IAAI,iBAAiB;IAAc,CAAC,GACpE,KAAA;GACL,EACD,gBACD;AAED,MACE,OAAO,UAAU,YACjB,SAAS,QACT,OAAO,KAAK,MAAM,CAAC,WAAW,GAC9B;AACA,OAAI,UAAU,SAAS,MAAM,QAAQ,MAAM,CACzC,QAAQ,MAAM,KAAK,KAAK,SAAS,GAAG,MAAM;AAG5C,OACE,WAAW,SACX,EACE,OAAO,UAAU,YACjB,SAAS,QACT,CAAC,MAAM,QAAQ,MAAM,IAErB,EAAE,iBAAiB,OAGrB,KAAI;AACF,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM,MAAM;WACvC;AACN,WAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;AAK7C,SAAQ,MAAM,KAAK,KAAK,SAAS,MAAM;;;;;;;;AAS3C,IAAM,uBAAN,MAAM,qBAA6C;CACjD;CAEA;CAEA;CAEA;CAEA,YACE,WACA,gBACA,uBACA;AACA,OAAK,YAAY;AACjB,OAAK,oBAAoB,OAAO,WAAW,oBAAoB;AAC/D,OAAK,iBAAiB;AACtB,OAAK,wBAAwB;;CAG/B,aAAa,eACX,QAC+B;EAC/B,MAAM,gBACJ,OAAO,WAAW,WAAW,SAAS,OAAO;EAC/C,MAAM,aAAa,OAAO,WAAW,WAAY,EAAE,GAAkB;EACrE,MAAM,YAAY,MAAM,cAAc,eAAe,WAAW;EAChE,MAAM,cAAc,YAAY,eAAe,WAAW;AAC1D,MAAI,CAAC,qBAAqB,UAAU,CAClC,OAAM,IAAI,MACR,qBAAqB,cAAc,8GAGpC;AAEH,MAAI,CAAC,YACH,OAAM,IAAI,MACR,qBAAqB,cAAc,uGAEpC;AAEH,SAAO,IAAI,qBACT,WACA,aACA,YAAY,sBACb;;;;;;;;CASH,MAAM,YAAY,KAAU,SAA8C;EACxE,MAAM,EAAE,YAAY,OAAO,cAAc,KAAK,sBAAsB;GAClE,UAAU,IAAI;GACd,eAAe,IAAI;GACnB,oBAAoB,SAAS;GAC7B;GACD,CAAC;EACF,MAAM,YAAY,IAAI,sBAAsB;EAC5C,MAAM,SAAS,IAAI,gBAAgB,EAAE,aAAa,cAAc,CAAC;AACjE,MAAI,KAAK,mBAAmB;GAC1B,MAAM,aAAa,MAAM,KAAK,UAAU,gBACtC;IACc;IACD;IACJ;IACR,EACD,EACE,WAAW,CAAC,WAAW,OAAO,EAC/B,CACF;GACD,MAAM,QAAQ,MAAM,UAAU,SAAS;AACvC,UAAO;IACL,KAAK,KAAK;IACV,SAAS,YAAY;IACrB,aAAa;IACb,GAAG;IACJ;;AAEH,QAAM,IAAI,MACR,wHAGD;;;AAIL,IAAM,mBAAN,MAAM,iBAAiB;CACrB,YAAY,YAA2D;AAApD,OAAA,aAAA;;CAEnB,aAAa,kBACX,QAC2B;EAE3B,MAAM,oBACJ,QAAQ,oBAAoB,OAAO,YAAY,OAAO,kBAAkB,GACvE,KAAK,cAAc;AACpB,OAAI,OAAO,cAAc,WACvB,QAAO,IAAI,oBAAoB,UAAU;OAEzC,QAAO;IAET;EAEF,MAAM,wBAAwB,MAAM,QAAQ,IAC1C,QAAQ,YACJ,OAAO,uBAAuB,EAC9B,IACA,OAAO,cACL,MAAM,qBAAqB,eAAe,UAAU,CACvD,IAAI,EAAE,CACV;AACD,SAAO,IAAI,kBACR,oBAAoB,EAAE,EAAE,OAAO,yBAAyB,EAAE,CAAC,CAC7D;;;;;;;;;;;AA0CL,MAAM,qBAAqB,OAAO,mBAAmC;AACnE,KAAI,SAAS,WAAW,eAAe,CACrC,cAAa;AAEf,KAAI,OAAO,mBAAmB,YAAY;AACxC,MAAI,6BAA6B,eAAe,EAAE;GAChD,MAAM,eAAe,IAAI,kBAAkB,EAAE,MAAM,gBAAgB,CAAC;AACpE,gBAAa;;AAGf,MAAI;GAEF,IAAI,MAAO,gBAAmC;AAC9C,OACE,OACA,OAAQ,IAAqC,SAAS,WAEtD,OAAM,MAAM;AAEd,UAAO;UACD;GAGN,MAAM,eAAe,IAAI,eAAe,EAAE,MAAM,gBAAgB,CAAC;AACjE,gBAAa;;;AAGjB,OAAM,IAAI,MAAM,yBAAyB;;AAG3C,MAAM,eAAe,OAAO,EAC1B,aACA,QACA,kBAMI;CACJ,MAAM,kBAAkB,OAAO,aAAa,EAAE,aAAa,CAAC;CAC5D,MAAM,UAA4B,EAAE;CACpC,MAAM,gBAAgB,EAAE;CACxB,MAAM,WAAW,EAAE;AACnB,YAAW,MAAM,WAAW,iBAAiB;EAC3C,MAAM,eAAe,IAAI,oBAAoB;AAC7C,UAAQ,KAAK,EACX,WAAW,CACT,IAAI,gBAAgB;GAAE,WAAW,QAAQ;GAAI;GAAa,CAAC,EAC3D,aACD,EACF,CAAC;AACF,WAAS,KAAK,QAAQ;AACtB,gBAAc,KAAK,aAAa;;AAElC,QAAO;EACL;EACA;EACA;EACD;;AAGH,MAAM,kBAAkB,OAAO,EAC7B,YACA,MACA,UACA,QACA,qBAaI;CAEJ,MAAM,EAAE,eAAe;CACvB,MAAM,WAAW,IAAI,YAAY;EAC/B,OAAO,SAAS;EAChB,QAAQ;EACT,CAAC;CACF,MAAM,SAAS,IAAI,YAAY,EAC7B,gBACD,CAAC;CACF,MAAM,WAAW,KAAK,IACpB,OACE,KACA,MAMA,OAAO,KAAK,YAAY;EACtB,MAAM,mBAAmB,MAAM,QAAQ,WACrC,WAAW,KAAK,cACd,OAAO,YAAY,KAAK,WAAW;GACjC,kBAAkB,SAAS;GAC3B,eAAe;GAChB,CAAC,CACH,CACF;AACD,WAAS,WAAW;AACpB,SAAO;GACL,gBACE,KAAK,YAAY,IAAI,aACjB,IAAI,KAAK,IAAI,SAAS,CAAC,SAAS,GAChC,IAAI,KAAK,IAAI,WAAW,CAAC,SAAS,GAClC,KAAA;GACN,UAAU,iBAAiB,KAAK,eAC9B,WAAW,WAAW,cAClB,WAAW,QACX,WAAW,OAChB;GACD,QAAQ,IAAI;GACb;GACD,CACL;AAGD,SAFgB,MAAM,QAAQ,IAAI,SAAS,EAE5B,QACZ,KAAK,QAAQ,OAAO;EACnB,GAAG;GACF,SAAS,GAAG,KAAK;EACnB,GACD,EAAE,CACH;;AAcH,MAAM,qBACJ,UACA,gBACA,aACG;AACH,KAAI,aAAa,OAIf,QAAO,SAAS,KAAK,EAAE,aACrB,gCAAgC,OAAO,MAAM,CAC9C;AAOH,KAFE,OAAO,mBAAmB,YAC1B,OAAQ,eAAqC,aAAa,cAG1D,SAAS,OAAO,EAAE,aAAa,OAAO,KAAK,OAAO,CAAC,WAAW,EAAE,CAEhE,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO,OAAO,OAAO,CAAC,GAAG;AAE/D,QAAO,SAAS,KAAK,EAAE,aAAa,OAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsD7C,eAAsB,aACpB,gBACA,aACA,SACA;CACA,MAAM,EACJ,aACA,iBACA,QACA,mBACsB,WAAW,EAAE;CAErC,MAAM,mBACJ,SAAS,qBACR,SAAS,cAAc,OACpB;EACE,YAAY,QAAQ;EACpB,uBAAuB,QAAQ;EAChC,GACD,KAAA;CAEN,MAAM,eAAe,MAAM,mBAAmB,eAAe;CAC7D,MAAM,aAAa,UAAU,IAAI,QAAQ;CACzC,MAAM,kBAAkB,eAAe,YAAY;CACnD,MAAM,UAAU,MAAM,WAAW,YAAY,EAAE,aAAa,CAAC;CAC7D,MAAM,YAAY,QAAQ;CAC1B,MAAM,kBAAkB,kBAAkB;CAC1C,MAAM,EAAE,SAAS,UAAU,kBAAkB,MAAM,aAAa;EAC9D;EACA,QAAQ;EACR,aAAa;EACb,gBAAgB;EACjB,CAAC;AAEF,OAAM,WAAW,cAAc;EAC7B,aAAa;EACb,oBAAoB;EACpB,cAAc,EAAE,UAAU,EAAE,GAAG,iBAAiB,EAAE;EACnD,CAAC;CACF,MAAM,kBAA4B,IAAI,eAAe,EACnD,MAAM,cACP,CAAC,CAAC,WAAW,EAAE,SAAS,iBAAiB,CAAC;CAC3C,MAAM,YAAY,kBAChB,UACA,gBACA,QAAQ,UACT;CACD,MAAM,WAAW,IAAI,YAAY;EAC/B,OAAO,UAAU;EACjB,QAAQ;EACT,CAAC;AAEF,OAAM,gBACH,cAAc,EACb,aAAa,SAAS,WAAW,EAClC,CAAC,CAED,MAAM,WAAW,SAAS;EACzB;EACA,kBAAkB;EACnB,CAAC;AAEJ,UAAS,UAAU;CACnB,MAAM,OAAc,EAAE;AACtB,MAAK,IAAI,IAAI,GAAG,IAAI,SAAS,QAAQ,KAAK,EACxC,MAAK,KAAK,MAAM,cAAc,GAAG,SAAS,CAAC;CAE7C,IAAI,cAGA,EAAE;AACN,KAAI,iBAGF,eAAc,MAAM,gBAAgB;EAClC,YAFA,MAAM,iBAAiB,kBAAkB,iBAAiB;EAG1D;EACA;EACA,QAAQ;EACR,gBAAgB;EACjB,CAAC;AAMJ,QAJ6B;EAC3B,aAAa;EACb,SAAS,eAAe,EAAE;EAC3B;;AAIH,SAAS,6BAA6B,GAAuC;AAC3E,QAAO,OAAO,MAAM,cAAc,yBAAyB"}
package/dist/sql_db.d.cts CHANGED
@@ -1,5 +1,5 @@
1
1
  import { SerializedSqlDatabase, SqlDatabaseDataSourceParams, SqlDatabaseOptionsParams, SqlTable } from "./util/sql_utils.cjs";
2
- import * as _langchain_core_load_serializable0 from "@langchain/core/load/serializable";
2
+ import * as _$_langchain_core_load_serializable0 from "@langchain/core/load/serializable";
3
3
  import { Serializable } from "@langchain/core/load/serializable";
4
4
  import { DataSource, DataSourceOptions } from "typeorm";
5
5
 
@@ -20,7 +20,7 @@ import { DataSource, DataSourceOptions } from "typeorm";
20
20
  */
21
21
  declare class SqlDatabase extends Serializable implements SqlDatabaseOptionsParams, SqlDatabaseDataSourceParams {
22
22
  lc_namespace: string[];
23
- toJSON(): _langchain_core_load_serializable0.SerializedNotImplemented;
23
+ toJSON(): _$_langchain_core_load_serializable0.SerializedNotImplemented;
24
24
  appDataSourceOptions: DataSourceOptions;
25
25
  appDataSource: DataSource;
26
26
  allTables: Array<SqlTable>;
@@ -1 +1 @@
1
- {"version":3,"file":"sql_db.d.cts","names":[],"sources":["../src/sql_db.ts"],"mappings":";;;;;;;;AA8BA;;;;;;;;;;;;cAAa,WAAA,SACH,YAAA,YACG,wBAAA,EAA0B,2BAAA;EAErC,YAAA;EAEA,MAAA,CAAA,GALA,kCAAA,CAKM,wBAAA;EAIN,oBAAA,EAAsB,iBAAA;EAEtB,aAAA,EAAe,UAAA;EAEf,SAAA,EAAW,KAAA,CAAM,QAAA;EAEjB,cAAA,EAAgB,KAAA;EAEhB,YAAA,EAAc,KAAA;EAEd,qBAAA;EAEA,iBAAA,GAAoB,MAAA;EAAA,UAEX,WAAA,CAAa,MAAA,EAAQ,2BAAA;EAAA,OAajB,oBAAA,CACX,MAAA,EAAQ,2BAAA,GACP,OAAA,CAAQ,WAAA;EAAA,OA0BE,iBAAA,CACX,MAAA,EAAQ,wBAAA,GACP,OAAA,CAAQ,WAAA;EAjEA;;;;;;;;;;EAoFL,YAAA,CAAa,YAAA,GAAe,KAAA,WAAgB,OAAA;EA5E5B;;;;;EAkHhB,GAAA,CAAI,OAAA,UAAiB,KAAA,mBAA+B,OAAA;EAe1D,SAAA,CAAA,GAAa,qBAAA;EA3HG;EAAA,OAsIH,OAAA,CAAA,GAAO,OAAA"}
1
+ {"version":3,"file":"sql_db.d.cts","names":[],"sources":["../src/sql_db.ts"],"mappings":";;;;;;;;AA8BA;;;;;;;;;;;;cAAa,WAAA,SACH,YAAA,YACG,wBAAA,EAA0B,2BAAA;EAErC,YAAA;EAEA,MAAA,CAAA,GALA,oCAAA,CAKM,wBAAA;EAIN,oBAAA,EAAsB,iBAAA;EAEtB,aAAA,EAAe,UAAA;EAEf,SAAA,EAAW,KAAA,CAAM,QAAA;EAEjB,cAAA,EAAgB,KAAA;EAEhB,YAAA,EAAc,KAAA;EAEd,qBAAA;EAEA,iBAAA,GAAoB,MAAA;EAAA,UAEX,WAAA,CAAa,MAAA,EAAQ,2BAAA;EAAA,OAajB,oBAAA,CACX,MAAA,EAAQ,2BAAA,GACP,OAAA,CAAQ,WAAA;EAAA,OA0BE,iBAAA,CACX,MAAA,EAAQ,wBAAA,GACP,OAAA,CAAQ,WAAA;EAjEA;;;;;;;;;;EAoFL,YAAA,CAAa,YAAA,GAAe,KAAA,WAAgB,OAAA;EA5E5B;;;;;EAkHhB,GAAA,CAAI,OAAA,UAAiB,KAAA,mBAA+B,OAAA;EAe1D,SAAA,CAAA,GAAa,qBAAA;EA3HG;EAAA,OAsIH,OAAA,CAAA,GAAO,OAAA"}
package/dist/sql_db.d.ts CHANGED
@@ -1,5 +1,5 @@
1
1
  import { SerializedSqlDatabase, SqlDatabaseDataSourceParams, SqlDatabaseOptionsParams, SqlTable } from "./util/sql_utils.js";
2
- import * as _langchain_core_load_serializable0 from "@langchain/core/load/serializable";
2
+ import * as _$_langchain_core_load_serializable0 from "@langchain/core/load/serializable";
3
3
  import { Serializable } from "@langchain/core/load/serializable";
4
4
  import { DataSource, DataSourceOptions } from "typeorm";
5
5
 
@@ -20,7 +20,7 @@ import { DataSource, DataSourceOptions } from "typeorm";
20
20
  */
21
21
  declare class SqlDatabase extends Serializable implements SqlDatabaseOptionsParams, SqlDatabaseDataSourceParams {
22
22
  lc_namespace: string[];
23
- toJSON(): _langchain_core_load_serializable0.SerializedNotImplemented;
23
+ toJSON(): _$_langchain_core_load_serializable0.SerializedNotImplemented;
24
24
  appDataSourceOptions: DataSourceOptions;
25
25
  appDataSource: DataSource;
26
26
  allTables: Array<SqlTable>;
@@ -1 +1 @@
1
- {"version":3,"file":"sql_db.d.ts","names":[],"sources":["../src/sql_db.ts"],"mappings":";;;;;;;;AA8BA;;;;;;;;;;;;cAAa,WAAA,SACH,YAAA,YACG,wBAAA,EAA0B,2BAAA;EAErC,YAAA;EAEA,MAAA,CAAA,GALA,kCAAA,CAKM,wBAAA;EAIN,oBAAA,EAAsB,iBAAA;EAEtB,aAAA,EAAe,UAAA;EAEf,SAAA,EAAW,KAAA,CAAM,QAAA;EAEjB,cAAA,EAAgB,KAAA;EAEhB,YAAA,EAAc,KAAA;EAEd,qBAAA;EAEA,iBAAA,GAAoB,MAAA;EAAA,UAEX,WAAA,CAAa,MAAA,EAAQ,2BAAA;EAAA,OAajB,oBAAA,CACX,MAAA,EAAQ,2BAAA,GACP,OAAA,CAAQ,WAAA;EAAA,OA0BE,iBAAA,CACX,MAAA,EAAQ,wBAAA,GACP,OAAA,CAAQ,WAAA;EAjEA;;;;;;;;;;EAoFL,YAAA,CAAa,YAAA,GAAe,KAAA,WAAgB,OAAA;EA5E5B;;;;;EAkHhB,GAAA,CAAI,OAAA,UAAiB,KAAA,mBAA+B,OAAA;EAe1D,SAAA,CAAA,GAAa,qBAAA;EA3HG;EAAA,OAsIH,OAAA,CAAA,GAAO,OAAA"}
1
+ {"version":3,"file":"sql_db.d.ts","names":[],"sources":["../src/sql_db.ts"],"mappings":";;;;;;;;AA8BA;;;;;;;;;;;;cAAa,WAAA,SACH,YAAA,YACG,wBAAA,EAA0B,2BAAA;EAErC,YAAA;EAEA,MAAA,CAAA,GALA,oCAAA,CAKM,wBAAA;EAIN,oBAAA,EAAsB,iBAAA;EAEtB,aAAA,EAAe,UAAA;EAEf,SAAA,EAAW,KAAA,CAAM,QAAA;EAEjB,cAAA,EAAgB,KAAA;EAEhB,YAAA,EAAc,KAAA;EAEd,qBAAA;EAEA,iBAAA,GAAoB,MAAA;EAAA,UAEX,WAAA,CAAa,MAAA,EAAQ,2BAAA;EAAA,OAajB,oBAAA,CACX,MAAA,EAAQ,2BAAA,GACP,OAAA,CAAQ,WAAA;EAAA,OA0BE,iBAAA,CACX,MAAA,EAAQ,wBAAA,GACP,OAAA,CAAQ,WAAA;EAjEA;;;;;;;;;;EAoFL,YAAA,CAAa,YAAA,GAAe,KAAA,WAAgB,OAAA;EA5E5B;;;;;EAkHhB,GAAA,CAAI,OAAA,UAAiB,KAAA,mBAA+B,OAAA;EAe1D,SAAA,CAAA,GAAa,qBAAA;EA3HG;EAAA,OAsIH,OAAA,CAAA,GAAO,OAAA"}