@langchain/classic 1.0.27 → 1.0.28
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +13 -0
- package/dist/agents/agent.cjs.map +1 -1
- package/dist/agents/agent.js.map +1 -1
- package/dist/agents/executor.cjs.map +1 -1
- package/dist/agents/executor.js.map +1 -1
- package/dist/agents/openai_functions/index.cjs.map +1 -1
- package/dist/agents/openai_functions/index.d.cts.map +1 -1
- package/dist/agents/openai_functions/index.d.ts.map +1 -1
- package/dist/agents/openai_functions/index.js.map +1 -1
- package/dist/agents/openai_tools/index.cjs.map +1 -1
- package/dist/agents/openai_tools/index.d.cts +2 -2
- package/dist/agents/openai_tools/index.d.cts.map +1 -1
- package/dist/agents/openai_tools/index.d.ts +2 -2
- package/dist/agents/openai_tools/index.d.ts.map +1 -1
- package/dist/agents/openai_tools/index.js.map +1 -1
- package/dist/agents/react/index.d.cts +2 -2
- package/dist/agents/react/index.d.cts.map +1 -1
- package/dist/agents/react/index.d.ts +2 -2
- package/dist/agents/react/index.d.ts.map +1 -1
- package/dist/agents/structured_chat/index.d.cts +2 -2
- package/dist/agents/structured_chat/index.d.cts.map +1 -1
- package/dist/agents/structured_chat/index.d.ts +2 -2
- package/dist/agents/structured_chat/index.d.ts.map +1 -1
- package/dist/agents/tool_calling/index.d.cts +2 -2
- package/dist/agents/tool_calling/index.d.cts.map +1 -1
- package/dist/agents/tool_calling/index.d.ts +2 -2
- package/dist/agents/tool_calling/index.d.ts.map +1 -1
- package/dist/agents/tool_calling/output_parser.cjs.map +1 -1
- package/dist/agents/tool_calling/output_parser.d.cts.map +1 -1
- package/dist/agents/tool_calling/output_parser.js.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts +2 -2
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts +2 -2
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts.map +1 -1
- package/dist/agents/xml/index.d.cts.map +1 -1
- package/dist/agents/xml/index.d.ts.map +1 -1
- package/dist/cache/file_system.cjs.map +1 -1
- package/dist/cache/file_system.js.map +1 -1
- package/dist/chains/analyze_documents_chain.cjs.map +1 -1
- package/dist/chains/analyze_documents_chain.js.map +1 -1
- package/dist/chains/base.cjs.map +1 -1
- package/dist/chains/base.d.cts +2 -2
- package/dist/chains/base.d.cts.map +1 -1
- package/dist/chains/base.d.ts +2 -2
- package/dist/chains/base.d.ts.map +1 -1
- package/dist/chains/base.js.map +1 -1
- package/dist/chains/chat_vector_db_chain.cjs.map +1 -1
- package/dist/chains/chat_vector_db_chain.js.map +1 -1
- package/dist/chains/combine_docs_chain.d.cts +2 -2
- package/dist/chains/combine_docs_chain.d.cts.map +1 -1
- package/dist/chains/combine_docs_chain.d.ts +2 -2
- package/dist/chains/combine_docs_chain.d.ts.map +1 -1
- package/dist/chains/combine_documents/reduce.cjs.map +1 -1
- package/dist/chains/combine_documents/reduce.js.map +1 -1
- package/dist/chains/conversational_retrieval_chain.cjs.map +1 -1
- package/dist/chains/conversational_retrieval_chain.js.map +1 -1
- package/dist/chains/graph_qa/cypher.cjs.map +1 -1
- package/dist/chains/graph_qa/cypher.js.map +1 -1
- package/dist/chains/llm_chain.cjs.map +1 -1
- package/dist/chains/llm_chain.js.map +1 -1
- package/dist/chains/load.d.cts +1 -2
- package/dist/chains/load.d.cts.map +1 -1
- package/dist/chains/load.d.ts +1 -2
- package/dist/chains/load.d.ts.map +1 -1
- package/dist/chains/openai_functions/base.cjs.map +1 -1
- package/dist/chains/openai_functions/base.js.map +1 -1
- package/dist/chains/openai_functions/openapi.cjs.map +1 -1
- package/dist/chains/openai_functions/openapi.js.map +1 -1
- package/dist/chains/openai_moderation.cjs.map +1 -1
- package/dist/chains/openai_moderation.js.map +1 -1
- package/dist/chains/query_constructor/index.cjs.map +1 -1
- package/dist/chains/query_constructor/index.js.map +1 -1
- package/dist/chains/question_answering/load.d.cts +1 -2
- package/dist/chains/question_answering/load.d.cts.map +1 -1
- package/dist/chains/question_answering/load.d.ts +1 -2
- package/dist/chains/question_answering/load.d.ts.map +1 -1
- package/dist/chains/retrieval.cjs.map +1 -1
- package/dist/chains/retrieval.js.map +1 -1
- package/dist/chains/retrieval_qa.cjs.map +1 -1
- package/dist/chains/retrieval_qa.js.map +1 -1
- package/dist/chains/router/utils.cjs.map +1 -1
- package/dist/chains/router/utils.js.map +1 -1
- package/dist/chains/summarization/load.d.cts +1 -2
- package/dist/chains/summarization/load.d.cts.map +1 -1
- package/dist/chains/summarization/load.d.ts +1 -2
- package/dist/chains/summarization/load.d.ts.map +1 -1
- package/dist/chains/vector_db_qa.cjs.map +1 -1
- package/dist/chains/vector_db_qa.js.map +1 -1
- package/dist/chat_models/universal.cjs +8 -10
- package/dist/chat_models/universal.cjs.map +1 -1
- package/dist/chat_models/universal.d.cts +3 -5
- package/dist/chat_models/universal.d.cts.map +1 -1
- package/dist/chat_models/universal.d.ts +3 -5
- package/dist/chat_models/universal.d.ts.map +1 -1
- package/dist/chat_models/universal.js +8 -10
- package/dist/chat_models/universal.js.map +1 -1
- package/dist/document_loaders/fs/directory.d.cts.map +1 -1
- package/dist/document_loaders/fs/directory.d.ts.map +1 -1
- package/dist/document_loaders/fs/json.cjs.map +1 -1
- package/dist/document_loaders/fs/json.js.map +1 -1
- package/dist/document_transformers/openai_functions.d.cts.map +1 -1
- package/dist/evaluation/agents/index.d.cts +1 -0
- package/dist/evaluation/agents/trajectory.d.cts +2 -2
- package/dist/evaluation/agents/trajectory.d.cts.map +1 -1
- package/dist/evaluation/agents/trajectory.d.ts +2 -2
- package/dist/evaluation/agents/trajectory.d.ts.map +1 -1
- package/dist/evaluation/base.cjs.map +1 -1
- package/dist/evaluation/base.d.cts.map +1 -1
- package/dist/evaluation/base.js.map +1 -1
- package/dist/evaluation/comparison/index.d.cts +1 -0
- package/dist/evaluation/comparison/pairwise.cjs.map +1 -1
- package/dist/evaluation/comparison/pairwise.d.cts +3 -3
- package/dist/evaluation/comparison/pairwise.d.cts.map +1 -1
- package/dist/evaluation/comparison/pairwise.d.ts +3 -3
- package/dist/evaluation/comparison/pairwise.d.ts.map +1 -1
- package/dist/evaluation/comparison/pairwise.js.map +1 -1
- package/dist/evaluation/criteria/criteria.cjs.map +1 -1
- package/dist/evaluation/criteria/criteria.d.cts +3 -3
- package/dist/evaluation/criteria/criteria.d.cts.map +1 -1
- package/dist/evaluation/criteria/criteria.d.ts +3 -3
- package/dist/evaluation/criteria/criteria.d.ts.map +1 -1
- package/dist/evaluation/criteria/criteria.js.map +1 -1
- package/dist/evaluation/criteria/index.d.cts +1 -0
- package/dist/evaluation/embedding_distance/index.d.cts +1 -0
- package/dist/evaluation/loader.cjs.map +1 -1
- package/dist/evaluation/loader.d.cts.map +1 -1
- package/dist/evaluation/loader.js.map +1 -1
- package/dist/evaluation/qa/index.d.cts +1 -0
- package/dist/experimental/autogpt/prompt.d.cts +2 -2
- package/dist/experimental/autogpt/prompt.d.cts.map +1 -1
- package/dist/experimental/autogpt/prompt.d.ts +2 -2
- package/dist/experimental/autogpt/prompt.d.ts.map +1 -1
- package/dist/experimental/autogpt/prompt_generator.cjs.map +1 -1
- package/dist/experimental/autogpt/prompt_generator.js.map +1 -1
- package/dist/experimental/autogpt/schema.cjs.map +1 -1
- package/dist/experimental/autogpt/schema.js.map +1 -1
- package/dist/experimental/generative_agents/generative_agent_memory.cjs.map +1 -1
- package/dist/experimental/generative_agents/generative_agent_memory.js.map +1 -1
- package/dist/experimental/masking/parser.cjs.map +1 -1
- package/dist/experimental/masking/parser.js.map +1 -1
- package/dist/experimental/masking/regex_masking_transformer.cjs.map +1 -1
- package/dist/experimental/masking/regex_masking_transformer.js.map +1 -1
- package/dist/experimental/openai_assistant/index.cjs.map +1 -1
- package/dist/experimental/openai_assistant/index.d.cts +3 -3
- package/dist/experimental/openai_assistant/index.d.ts +3 -3
- package/dist/experimental/openai_assistant/index.js.map +1 -1
- package/dist/experimental/openai_files/index.d.cts +4 -4
- package/dist/experimental/openai_files/index.d.ts +4 -4
- package/dist/experimental/plan_and_execute/agent_executor.cjs.map +1 -1
- package/dist/experimental/plan_and_execute/agent_executor.js.map +1 -1
- package/dist/experimental/prompts/custom_format.cjs.map +1 -1
- package/dist/experimental/prompts/custom_format.js.map +1 -1
- package/dist/experimental/prompts/handlebars.cjs.map +1 -1
- package/dist/experimental/prompts/handlebars.d.cts +2 -2
- package/dist/experimental/prompts/handlebars.d.cts.map +1 -1
- package/dist/experimental/prompts/handlebars.d.ts +2 -2
- package/dist/experimental/prompts/handlebars.d.ts.map +1 -1
- package/dist/experimental/prompts/handlebars.js.map +1 -1
- package/dist/hub/base.cjs.map +1 -1
- package/dist/hub/base.d.cts +0 -1
- package/dist/hub/base.d.cts.map +1 -1
- package/dist/hub/base.d.ts +0 -1
- package/dist/hub/base.d.ts.map +1 -1
- package/dist/hub/base.js.map +1 -1
- package/dist/hub/index.cjs.map +1 -1
- package/dist/hub/index.js.map +1 -1
- package/dist/load/index.cjs.map +1 -1
- package/dist/load/index.js.map +1 -1
- package/dist/output_parsers/combining.cjs.map +1 -1
- package/dist/output_parsers/combining.js.map +1 -1
- package/dist/output_parsers/expression_type_handlers/base.cjs.map +1 -1
- package/dist/output_parsers/expression_type_handlers/base.d.cts.map +1 -1
- package/dist/output_parsers/expression_type_handlers/base.d.ts.map +1 -1
- package/dist/output_parsers/expression_type_handlers/base.js.map +1 -1
- package/dist/output_parsers/fix.cjs.map +1 -1
- package/dist/output_parsers/fix.js.map +1 -1
- package/dist/output_parsers/openai_functions.d.cts.map +1 -1
- package/dist/output_parsers/openai_functions.d.ts.map +1 -1
- package/dist/output_parsers/openai_tools.cjs.map +1 -1
- package/dist/output_parsers/openai_tools.js.map +1 -1
- package/dist/output_parsers/regex.cjs.map +1 -1
- package/dist/output_parsers/regex.js.map +1 -1
- package/dist/output_parsers/structured.d.cts +2 -2
- package/dist/output_parsers/structured.d.cts.map +1 -1
- package/dist/output_parsers/structured.d.ts +2 -2
- package/dist/output_parsers/structured.d.ts.map +1 -1
- package/dist/retrievers/document_compressors/index.cjs.map +1 -1
- package/dist/retrievers/document_compressors/index.js.map +1 -1
- package/dist/retrievers/matryoshka_retriever.cjs.map +1 -1
- package/dist/retrievers/matryoshka_retriever.d.cts +1 -1
- package/dist/retrievers/matryoshka_retriever.d.cts.map +1 -1
- package/dist/retrievers/matryoshka_retriever.d.ts +1 -1
- package/dist/retrievers/matryoshka_retriever.d.ts.map +1 -1
- package/dist/retrievers/matryoshka_retriever.js.map +1 -1
- package/dist/retrievers/multi_query.cjs.map +1 -1
- package/dist/retrievers/multi_query.js.map +1 -1
- package/dist/retrievers/parent_document.cjs.map +1 -1
- package/dist/retrievers/parent_document.js.map +1 -1
- package/dist/schema/prompt_template.d.cts +1 -2
- package/dist/schema/prompt_template.d.cts.map +1 -1
- package/dist/schema/prompt_template.d.ts +1 -2
- package/dist/schema/prompt_template.d.ts.map +1 -1
- package/dist/smith/config.cjs.map +1 -1
- package/dist/smith/config.d.cts +1 -1
- package/dist/smith/config.d.ts +1 -1
- package/dist/smith/config.d.ts.map +1 -1
- package/dist/smith/config.js.map +1 -1
- package/dist/smith/runner_utils.cjs.map +1 -1
- package/dist/smith/runner_utils.d.cts +1 -1
- package/dist/smith/runner_utils.d.cts.map +1 -1
- package/dist/smith/runner_utils.d.ts +1 -1
- package/dist/smith/runner_utils.d.ts.map +1 -1
- package/dist/smith/runner_utils.js.map +1 -1
- package/dist/sql_db.d.cts +2 -2
- package/dist/sql_db.d.cts.map +1 -1
- package/dist/sql_db.d.ts +2 -2
- package/dist/sql_db.d.ts.map +1 -1
- package/dist/storage/encoder_backed.cjs.map +1 -1
- package/dist/storage/encoder_backed.js.map +1 -1
- package/dist/storage/file_system.cjs.map +1 -1
- package/dist/storage/file_system.js.map +1 -1
- package/dist/stores/doc/in_memory.cjs.map +1 -1
- package/dist/stores/doc/in_memory.js.map +1 -1
- package/dist/tools/json.cjs.map +1 -1
- package/dist/tools/json.js.map +1 -1
- package/dist/tools/webbrowser.cjs.map +1 -1
- package/dist/tools/webbrowser.d.cts.map +1 -1
- package/dist/tools/webbrowser.d.ts.map +1 -1
- package/dist/tools/webbrowser.js.map +1 -1
- package/dist/util/document.d.cts +1 -1
- package/dist/util/document.d.cts.map +1 -1
- package/dist/util/document.d.ts +1 -1
- package/dist/util/document.d.ts.map +1 -1
- package/dist/util/is-network-error/index.cjs.map +1 -1
- package/dist/util/is-network-error/index.js.map +1 -1
- package/dist/util/load.cjs.map +1 -1
- package/dist/util/load.js.map +1 -1
- package/dist/util/openapi.cjs.map +1 -1
- package/dist/util/openapi.d.cts +32 -32
- package/dist/util/openapi.d.cts.map +1 -1
- package/dist/util/openapi.d.ts +32 -32
- package/dist/util/openapi.d.ts.map +1 -1
- package/dist/util/openapi.js.map +1 -1
- package/dist/util/p-retry/index.cjs.map +1 -1
- package/dist/util/p-retry/index.js.map +1 -1
- package/dist/vectorstores/memory.cjs.map +1 -1
- package/dist/vectorstores/memory.js.map +1 -1
- package/package.json +25 -48
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.cts +0 -1494
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.ts +0 -1494
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/client.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/_runner.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/_runner.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluate_comparative.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluate_comparative.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.cts +0 -66
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.ts +0 -66
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/evaluator.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/index.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/index.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/string_evaluator.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/evaluation/string_evaluator.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/index.d.cts +0 -4
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/index.d.ts +0 -4
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.cts +0 -145
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.ts +0 -145
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/run_trees.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.cts +0 -437
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.ts +0 -437
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/schemas.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.cts +0 -7
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.ts +0 -7
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/traceable.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.cts +0 -38
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.ts +0 -38
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/singletons/types.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.cts +0 -25
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.ts +0 -25
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/async_caller.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/p-queue.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/p-queue.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.cts +0 -129
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.cts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.ts +0 -129
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/dist/utils/prompt_cache/index.d.ts.map +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/evaluation.d.cts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/evaluation.d.ts +0 -1
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/singletons/traceable.d.cts +0 -2
- package/dist/node_modules/.pnpm/langsmith@0.5.9_@opentelemetry_api@1.9.0_openai@6.22.0_ws@8.20.0_bufferutil@4.1.0__zod@4.3.6_/node_modules/langsmith/singletons/traceable.d.ts +0 -2
- package/dist/node_modules/.pnpm/p-queue@6.6.2/node_modules/p-queue/dist/index.d.cts +0 -1
- package/dist/node_modules/.pnpm/p-queue@6.6.2/node_modules/p-queue/dist/index.d.ts +0 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"conversational_retrieval_chain.js","names":[],"sources":["../../src/chains/conversational_retrieval_chain.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport type { BaseRetrieverInterface } from \"@langchain/core/retrievers\";\nimport { PromptTemplate } from \"@langchain/core/prompts\";\nimport { BaseMessage, HumanMessage, AIMessage } from \"@langchain/core/messages\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { SerializedChatVectorDBQAChain } from \"./serde.js\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { LLMChain } from \"./llm_chain.js\";\nimport { QAChainParams, loadQAChain } from \"./question_answering/load.js\";\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nexport type LoadValues = Record<string, any>;\n\nconst question_generator_template = `Given the following conversation and a follow up question, rephrase the follow up question to be a standalone question.\n\nChat History:\n{chat_history}\nFollow Up Input: {question}\nStandalone question:`;\n\n/**\n * Interface for the input parameters of the\n * ConversationalRetrievalQAChain class.\n */\nexport interface ConversationalRetrievalQAChainInput extends ChainInputs {\n retriever: BaseRetrieverInterface;\n combineDocumentsChain: BaseChain;\n questionGeneratorChain: LLMChain;\n returnSourceDocuments?: boolean;\n returnGeneratedQuestion?: boolean;\n inputKey?: string;\n}\n\n/**\n * Class for conducting conversational question-answering tasks with a\n * retrieval component. Extends the BaseChain class and implements the\n * ConversationalRetrievalQAChainInput interface.\n * @example\n * ```typescript\n * import { ChatAnthropic } from \"@langchain/anthropic\";\n * import {\n * ChatPromptTemplate,\n * MessagesPlaceholder,\n * } from \"@langchain/core/prompts\";\n * import { BaseMessage } from \"@langchain/core/messages\";\n * import { createStuffDocumentsChain } from \"@langchain/classic/chains/combine_documents\";\n * import { createHistoryAwareRetriever } from \"@langchain/classic/chains/history_aware_retriever\";\n * import { createRetrievalChain } from \"@langchain/classic/chains/retrieval\";\n *\n * const retriever = ...your retriever;\n * const llm = new ChatAnthropic();\n *\n * // Contextualize question\n * const contextualizeQSystemPrompt = `\n * Given a chat history and the latest user question\n * which might reference context in the chat history,\n * formulate a standalone question which can be understood\n * without the chat history. Do NOT answer the question, just\n * reformulate it if needed and otherwise return it as is.`;\n * const contextualizeQPrompt = ChatPromptTemplate.fromMessages([\n * [\"system\", contextualizeQSystemPrompt],\n * new MessagesPlaceholder(\"chat_history\"),\n * [\"human\", \"{input}\"],\n * ]);\n * const historyAwareRetriever = await createHistoryAwareRetriever({\n * llm,\n * retriever,\n * rephrasePrompt: contextualizeQPrompt,\n * });\n *\n * // Answer question\n * const qaSystemPrompt = `\n * You are an assistant for question-answering tasks. Use\n * the following pieces of retrieved context to answer the\n * question. If you don't know the answer, just say that you\n * don't know. Use three sentences maximum and keep the answer\n * concise.\n * \\n\\n\n * {context}`;\n * const qaPrompt = ChatPromptTemplate.fromMessages([\n * [\"system\", qaSystemPrompt],\n * new MessagesPlaceholder(\"chat_history\"),\n * [\"human\", \"{input}\"],\n * ]);\n *\n * // Below we use createStuffDocuments_chain to feed all retrieved context\n * // into the LLM. Note that we can also use StuffDocumentsChain and other\n * // instances of BaseCombineDocumentsChain.\n * const questionAnswerChain = await createStuffDocumentsChain({\n * llm,\n * prompt: qaPrompt,\n * });\n *\n * const ragChain = await createRetrievalChain({\n * retriever: historyAwareRetriever,\n * combineDocsChain: questionAnswerChain,\n * });\n *\n * // Usage:\n * const chat_history: BaseMessage[] = [];\n * const response = await ragChain.invoke({\n * chat_history,\n * input: \"...\",\n * });\n * ```\n */\nexport class ConversationalRetrievalQAChain\n extends BaseChain\n implements ConversationalRetrievalQAChainInput\n{\n static lc_name() {\n return \"ConversationalRetrievalQAChain\";\n }\n\n inputKey = \"question\";\n\n chatHistoryKey = \"chat_history\";\n\n get inputKeys() {\n return [this.inputKey, this.chatHistoryKey];\n }\n\n get outputKeys() {\n return this.combineDocumentsChain.outputKeys.concat(\n this.returnSourceDocuments ? [\"sourceDocuments\"] : []\n );\n }\n\n retriever: BaseRetrieverInterface;\n\n combineDocumentsChain: BaseChain;\n\n questionGeneratorChain: LLMChain;\n\n returnSourceDocuments = false;\n\n returnGeneratedQuestion = false;\n\n constructor(fields: ConversationalRetrievalQAChainInput) {\n super(fields);\n this.retriever = fields.retriever;\n this.combineDocumentsChain = fields.combineDocumentsChain;\n this.questionGeneratorChain = fields.questionGeneratorChain;\n this.inputKey = fields.inputKey ?? this.inputKey;\n this.returnSourceDocuments =\n fields.returnSourceDocuments ?? this.returnSourceDocuments;\n this.returnGeneratedQuestion =\n fields.returnGeneratedQuestion ?? this.returnGeneratedQuestion;\n }\n\n /**\n * Static method to convert the chat history input into a formatted\n * string.\n * @param chatHistory Chat history input which can be a string, an array of BaseMessage instances, or an array of string arrays.\n * @returns A formatted string representing the chat history.\n */\n static getChatHistoryString(\n chatHistory: string | BaseMessage[] | string[][]\n ) {\n let historyMessages: BaseMessage[];\n if (Array.isArray(chatHistory)) {\n // TODO: Deprecate on a breaking release\n if (\n Array.isArray(chatHistory[0]) &&\n typeof chatHistory[0][0] === \"string\"\n ) {\n console.warn(\n \"Passing chat history as an array of strings is deprecated.\\nPlease see https://js.langchain.com/docs/modules/chains/popular/chat_vector_db#externally-managed-memory for more information.\"\n );\n historyMessages = chatHistory.flat().map((stringMessage, i) => {\n if (i % 2 === 0) {\n return new HumanMessage(stringMessage);\n } else {\n return new AIMessage(stringMessage);\n }\n });\n } else {\n historyMessages = chatHistory as BaseMessage[];\n }\n return historyMessages\n .map((chatMessage) => {\n if (chatMessage._getType() === \"human\") {\n return `Human: ${chatMessage.content}`;\n } else if (chatMessage._getType() === \"ai\") {\n return `Assistant: ${chatMessage.content}`;\n } else {\n return `${chatMessage.content}`;\n }\n })\n .join(\"\\n\");\n }\n return chatHistory;\n }\n\n /** @ignore */\n async _call(\n values: ChainValues,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n if (!(this.inputKey in values)) {\n throw new Error(`Question key ${this.inputKey} not found.`);\n }\n if (!(this.chatHistoryKey in values)) {\n throw new Error(`Chat history key ${this.chatHistoryKey} not found.`);\n }\n const question: string = values[this.inputKey];\n const chatHistory: string =\n ConversationalRetrievalQAChain.getChatHistoryString(\n values[this.chatHistoryKey]\n );\n let newQuestion = question;\n if (chatHistory.length > 0) {\n const result = await this.questionGeneratorChain.call(\n {\n question,\n chat_history: chatHistory,\n },\n runManager?.getChild(\"question_generator\")\n );\n const keys = Object.keys(result);\n if (keys.length === 1) {\n newQuestion = result[keys[0]];\n } else {\n throw new Error(\n \"Return from llm chain has multiple values, only single values supported.\"\n );\n }\n }\n const docs = await this.retriever.invoke(\n newQuestion,\n runManager?.getChild(\"retriever\")\n );\n const inputs = {\n question: newQuestion,\n input_documents: docs,\n chat_history: chatHistory,\n };\n let result = await this.combineDocumentsChain.call(\n inputs,\n runManager?.getChild(\"combine_documents\")\n );\n if (this.returnSourceDocuments) {\n result = {\n ...result,\n sourceDocuments: docs,\n };\n }\n if (this.returnGeneratedQuestion) {\n result = {\n ...result,\n generatedQuestion: newQuestion,\n };\n }\n return result;\n }\n\n _chainType(): string {\n return \"conversational_retrieval_chain\";\n }\n\n static async deserialize(\n _data: SerializedChatVectorDBQAChain,\n _values: LoadValues\n ): Promise<ConversationalRetrievalQAChain> {\n throw new Error(\"Not implemented.\");\n }\n\n serialize(): SerializedChatVectorDBQAChain {\n throw new Error(\"Not implemented.\");\n }\n\n /**\n * Static method to create a new ConversationalRetrievalQAChain from a\n * BaseLanguageModel and a BaseRetriever.\n * @param llm {@link BaseLanguageModelInterface} instance used to generate a new question.\n * @param retriever {@link BaseRetrieverInterface} instance used to retrieve relevant documents.\n * @param options.returnSourceDocuments Whether to return source documents in the final output\n * @param options.questionGeneratorChainOptions Options to initialize the standalone question generation chain used as the first internal step\n * @param options.qaChainOptions {@link QAChainParams} used to initialize the QA chain used as the second internal step\n * @returns A new instance of ConversationalRetrievalQAChain.\n */\n static fromLLM(\n llm: BaseLanguageModelInterface,\n retriever: BaseRetrieverInterface,\n options: {\n outputKey?: string; // not used\n returnSourceDocuments?: boolean;\n /** @deprecated Pass in questionGeneratorChainOptions.template instead */\n questionGeneratorTemplate?: string;\n /** @deprecated Pass in qaChainOptions.prompt instead */\n qaTemplate?: string;\n questionGeneratorChainOptions?: {\n llm?: BaseLanguageModelInterface;\n template?: string;\n };\n qaChainOptions?: QAChainParams;\n } & Omit<\n ConversationalRetrievalQAChainInput,\n \"retriever\" | \"combineDocumentsChain\" | \"questionGeneratorChain\"\n > = {}\n ): ConversationalRetrievalQAChain {\n const {\n questionGeneratorTemplate,\n qaTemplate,\n qaChainOptions = {\n type: \"stuff\",\n prompt: qaTemplate\n ? PromptTemplate.fromTemplate(qaTemplate)\n : undefined,\n },\n questionGeneratorChainOptions,\n verbose,\n ...rest\n } = options;\n\n const qaChain = loadQAChain(llm, qaChainOptions);\n\n const questionGeneratorChainPrompt = PromptTemplate.fromTemplate(\n questionGeneratorChainOptions?.template ??\n questionGeneratorTemplate ??\n question_generator_template\n );\n const questionGeneratorChain = new LLMChain({\n prompt: questionGeneratorChainPrompt,\n llm: questionGeneratorChainOptions?.llm ?? llm,\n verbose,\n });\n const instance = new this({\n retriever,\n combineDocumentsChain: qaChain,\n questionGeneratorChain,\n verbose,\n ...rest,\n });\n return instance;\n }\n}\n"],"mappings":";;;;;;AAcA,MAAM,8BAA8B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA6FpC,IAAa,iCAAb,MAAa,uCACH,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,WAAW;CAEX,iBAAiB;CAEjB,IAAI,YAAY;AACd,SAAO,CAAC,KAAK,UAAU,KAAK,eAAe;;CAG7C,IAAI,aAAa;AACf,SAAO,KAAK,sBAAsB,WAAW,OAC3C,KAAK,wBAAwB,CAAC,kBAAkB,GAAG,EAAE,CACtD;;CAGH;CAEA;CAEA;CAEA,wBAAwB;CAExB,0BAA0B;CAE1B,YAAY,QAA6C;AACvD,QAAM,OAAO;AACb,OAAK,YAAY,OAAO;AACxB,OAAK,wBAAwB,OAAO;AACpC,OAAK,yBAAyB,OAAO;AACrC,OAAK,WAAW,OAAO,YAAY,KAAK;AACxC,OAAK,wBACH,OAAO,yBAAyB,KAAK;AACvC,OAAK,0BACH,OAAO,2BAA2B,KAAK;;;;;;;;CAS3C,OAAO,qBACL,aACA;EACA,IAAI;AACJ,MAAI,MAAM,QAAQ,YAAY,EAAE;AAE9B,OACE,MAAM,QAAQ,YAAY,GAAG,IAC7B,OAAO,YAAY,GAAG,OAAO,UAC7B;AACA,YAAQ,KACN,6LACD;AACD,sBAAkB,YAAY,MAAM,CAAC,KAAK,eAAe,MAAM;AAC7D,SAAI,IAAI,MAAM,EACZ,QAAO,IAAI,aAAa,cAAc;SAEtC,QAAO,IAAI,UAAU,cAAc;MAErC;SAEF,mBAAkB;AAEpB,UAAO,gBACJ,KAAK,gBAAgB;AACpB,QAAI,YAAY,UAAU,KAAK,QAC7B,QAAO,UAAU,YAAY;aACpB,YAAY,UAAU,KAAK,KACpC,QAAO,cAAc,YAAY;QAEjC,QAAO,GAAG,YAAY;KAExB,CACD,KAAK,KAAK;;AAEf,SAAO;;;CAIT,MAAM,MACJ,QACA,YACsB;AACtB,MAAI,EAAE,KAAK,YAAY,QACrB,OAAM,IAAI,MAAM,gBAAgB,KAAK,SAAS,aAAa;AAE7D,MAAI,EAAE,KAAK,kBAAkB,QAC3B,OAAM,IAAI,MAAM,oBAAoB,KAAK,eAAe,aAAa;EAEvE,MAAM,WAAmB,OAAO,KAAK;EACrC,MAAM,cACJ,+BAA+B,qBAC7B,OAAO,KAAK,gBACb;EACH,IAAI,cAAc;AAClB,MAAI,YAAY,SAAS,GAAG;GAC1B,MAAM,SAAS,MAAM,KAAK,uBAAuB,KAC/C;IACE;IACA,cAAc;IACf,EACD,YAAY,SAAS,qBAAqB,CAC3C;GACD,MAAM,OAAO,OAAO,KAAK,OAAO;AAChC,OAAI,KAAK,WAAW,EAClB,eAAc,OAAO,KAAK;OAE1B,OAAM,IAAI,MACR,2EACD;;EAGL,MAAM,OAAO,MAAM,KAAK,UAAU,OAChC,aACA,YAAY,SAAS,YAAY,CAClC;EACD,MAAM,SAAS;GACb,UAAU;GACV,iBAAiB;GACjB,cAAc;GACf;EACD,IAAI,SAAS,MAAM,KAAK,sBAAsB,KAC5C,QACA,YAAY,SAAS,oBAAoB,CAC1C;AACD,MAAI,KAAK,sBACP,UAAS;GACP,GAAG;GACH,iBAAiB;GAClB;AAEH,MAAI,KAAK,wBACP,UAAS;GACP,GAAG;GACH,mBAAmB;GACpB;AAEH,SAAO;;CAGT,aAAqB;AACnB,SAAO;;CAGT,aAAa,YACX,OACA,SACyC;AACzC,QAAM,IAAI,MAAM,mBAAmB;;CAGrC,YAA2C;AACzC,QAAM,IAAI,MAAM,mBAAmB;;;;;;;;;;;;CAarC,OAAO,QACL,KACA,WACA,UAeI,EAAE,EAC0B;EAChC,MAAM,EACJ,2BACA,YACA,iBAAiB;GACf,MAAM;GACN,QAAQ,aACJ,eAAe,aAAa,WAAW,GACvC,KAAA;GACL,EACD,+BACA,SACA,GAAG,SACD;EAEJ,MAAM,UAAU,YAAY,KAAK,eAAe;EAOhD,MAAM,yBAAyB,IAAI,SAAS;GAC1C,QANmC,eAAe,aAClD,+BAA+B,YAC7B,6BACA,4BACH;GAGC,KAAK,+BAA+B,OAAO;GAC3C;GACD,CAAC;AAQF,SAPiB,IAAI,KAAK;GACxB;GACA,uBAAuB;GACvB;GACA;GACA,GAAG;GACJ,CAAC"}
|
|
1
|
+
{"version":3,"file":"conversational_retrieval_chain.js","names":[],"sources":["../../src/chains/conversational_retrieval_chain.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport type { BaseRetrieverInterface } from \"@langchain/core/retrievers\";\nimport { PromptTemplate } from \"@langchain/core/prompts\";\nimport { BaseMessage, HumanMessage, AIMessage } from \"@langchain/core/messages\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { SerializedChatVectorDBQAChain } from \"./serde.js\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { LLMChain } from \"./llm_chain.js\";\nimport { QAChainParams, loadQAChain } from \"./question_answering/load.js\";\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\nexport type LoadValues = Record<string, any>;\n\nconst question_generator_template = `Given the following conversation and a follow up question, rephrase the follow up question to be a standalone question.\n\nChat History:\n{chat_history}\nFollow Up Input: {question}\nStandalone question:`;\n\n/**\n * Interface for the input parameters of the\n * ConversationalRetrievalQAChain class.\n */\nexport interface ConversationalRetrievalQAChainInput extends ChainInputs {\n retriever: BaseRetrieverInterface;\n combineDocumentsChain: BaseChain;\n questionGeneratorChain: LLMChain;\n returnSourceDocuments?: boolean;\n returnGeneratedQuestion?: boolean;\n inputKey?: string;\n}\n\n/**\n * Class for conducting conversational question-answering tasks with a\n * retrieval component. Extends the BaseChain class and implements the\n * ConversationalRetrievalQAChainInput interface.\n * @example\n * ```typescript\n * import { ChatAnthropic } from \"@langchain/anthropic\";\n * import {\n * ChatPromptTemplate,\n * MessagesPlaceholder,\n * } from \"@langchain/core/prompts\";\n * import { BaseMessage } from \"@langchain/core/messages\";\n * import { createStuffDocumentsChain } from \"@langchain/classic/chains/combine_documents\";\n * import { createHistoryAwareRetriever } from \"@langchain/classic/chains/history_aware_retriever\";\n * import { createRetrievalChain } from \"@langchain/classic/chains/retrieval\";\n *\n * const retriever = ...your retriever;\n * const llm = new ChatAnthropic();\n *\n * // Contextualize question\n * const contextualizeQSystemPrompt = `\n * Given a chat history and the latest user question\n * which might reference context in the chat history,\n * formulate a standalone question which can be understood\n * without the chat history. Do NOT answer the question, just\n * reformulate it if needed and otherwise return it as is.`;\n * const contextualizeQPrompt = ChatPromptTemplate.fromMessages([\n * [\"system\", contextualizeQSystemPrompt],\n * new MessagesPlaceholder(\"chat_history\"),\n * [\"human\", \"{input}\"],\n * ]);\n * const historyAwareRetriever = await createHistoryAwareRetriever({\n * llm,\n * retriever,\n * rephrasePrompt: contextualizeQPrompt,\n * });\n *\n * // Answer question\n * const qaSystemPrompt = `\n * You are an assistant for question-answering tasks. Use\n * the following pieces of retrieved context to answer the\n * question. If you don't know the answer, just say that you\n * don't know. Use three sentences maximum and keep the answer\n * concise.\n * \\n\\n\n * {context}`;\n * const qaPrompt = ChatPromptTemplate.fromMessages([\n * [\"system\", qaSystemPrompt],\n * new MessagesPlaceholder(\"chat_history\"),\n * [\"human\", \"{input}\"],\n * ]);\n *\n * // Below we use createStuffDocuments_chain to feed all retrieved context\n * // into the LLM. Note that we can also use StuffDocumentsChain and other\n * // instances of BaseCombineDocumentsChain.\n * const questionAnswerChain = await createStuffDocumentsChain({\n * llm,\n * prompt: qaPrompt,\n * });\n *\n * const ragChain = await createRetrievalChain({\n * retriever: historyAwareRetriever,\n * combineDocsChain: questionAnswerChain,\n * });\n *\n * // Usage:\n * const chat_history: BaseMessage[] = [];\n * const response = await ragChain.invoke({\n * chat_history,\n * input: \"...\",\n * });\n * ```\n */\nexport class ConversationalRetrievalQAChain\n extends BaseChain\n implements ConversationalRetrievalQAChainInput\n{\n static lc_name() {\n return \"ConversationalRetrievalQAChain\";\n }\n\n inputKey = \"question\";\n\n chatHistoryKey = \"chat_history\";\n\n get inputKeys() {\n return [this.inputKey, this.chatHistoryKey];\n }\n\n get outputKeys() {\n return this.combineDocumentsChain.outputKeys.concat(\n this.returnSourceDocuments ? [\"sourceDocuments\"] : []\n );\n }\n\n retriever: BaseRetrieverInterface;\n\n combineDocumentsChain: BaseChain;\n\n questionGeneratorChain: LLMChain;\n\n returnSourceDocuments = false;\n\n returnGeneratedQuestion = false;\n\n constructor(fields: ConversationalRetrievalQAChainInput) {\n super(fields);\n this.retriever = fields.retriever;\n this.combineDocumentsChain = fields.combineDocumentsChain;\n this.questionGeneratorChain = fields.questionGeneratorChain;\n this.inputKey = fields.inputKey ?? this.inputKey;\n this.returnSourceDocuments =\n fields.returnSourceDocuments ?? this.returnSourceDocuments;\n this.returnGeneratedQuestion =\n fields.returnGeneratedQuestion ?? this.returnGeneratedQuestion;\n }\n\n /**\n * Static method to convert the chat history input into a formatted\n * string.\n * @param chatHistory Chat history input which can be a string, an array of BaseMessage instances, or an array of string arrays.\n * @returns A formatted string representing the chat history.\n */\n static getChatHistoryString(\n chatHistory: string | BaseMessage[] | string[][]\n ) {\n let historyMessages: BaseMessage[];\n if (Array.isArray(chatHistory)) {\n // TODO: Deprecate on a breaking release\n if (\n Array.isArray(chatHistory[0]) &&\n typeof chatHistory[0][0] === \"string\"\n ) {\n console.warn(\n \"Passing chat history as an array of strings is deprecated.\\nPlease see https://js.langchain.com/docs/modules/chains/popular/chat_vector_db#externally-managed-memory for more information.\"\n );\n historyMessages = chatHistory.flat().map((stringMessage, i) => {\n if (i % 2 === 0) {\n return new HumanMessage(stringMessage);\n } else {\n return new AIMessage(stringMessage);\n }\n });\n } else {\n historyMessages = chatHistory as BaseMessage[];\n }\n return historyMessages\n .map((chatMessage) => {\n if (chatMessage._getType() === \"human\") {\n return `Human: ${chatMessage.content}`;\n } else if (chatMessage._getType() === \"ai\") {\n return `Assistant: ${chatMessage.content}`;\n } else {\n return `${chatMessage.content}`;\n }\n })\n .join(\"\\n\");\n }\n return chatHistory;\n }\n\n /** @ignore */\n async _call(\n values: ChainValues,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n if (!(this.inputKey in values)) {\n throw new Error(`Question key ${this.inputKey} not found.`);\n }\n if (!(this.chatHistoryKey in values)) {\n throw new Error(`Chat history key ${this.chatHistoryKey} not found.`);\n }\n const question: string = values[this.inputKey];\n const chatHistory: string =\n ConversationalRetrievalQAChain.getChatHistoryString(\n values[this.chatHistoryKey]\n );\n let newQuestion = question;\n if (chatHistory.length > 0) {\n const result = await this.questionGeneratorChain.call(\n {\n question,\n chat_history: chatHistory,\n },\n runManager?.getChild(\"question_generator\")\n );\n const keys = Object.keys(result);\n if (keys.length === 1) {\n newQuestion = result[keys[0]];\n } else {\n throw new Error(\n \"Return from llm chain has multiple values, only single values supported.\"\n );\n }\n }\n const docs = await this.retriever.invoke(\n newQuestion,\n runManager?.getChild(\"retriever\")\n );\n const inputs = {\n question: newQuestion,\n input_documents: docs,\n chat_history: chatHistory,\n };\n let result = await this.combineDocumentsChain.call(\n inputs,\n runManager?.getChild(\"combine_documents\")\n );\n if (this.returnSourceDocuments) {\n result = {\n ...result,\n sourceDocuments: docs,\n };\n }\n if (this.returnGeneratedQuestion) {\n result = {\n ...result,\n generatedQuestion: newQuestion,\n };\n }\n return result;\n }\n\n _chainType(): string {\n return \"conversational_retrieval_chain\";\n }\n\n static async deserialize(\n _data: SerializedChatVectorDBQAChain,\n _values: LoadValues\n ): Promise<ConversationalRetrievalQAChain> {\n throw new Error(\"Not implemented.\");\n }\n\n serialize(): SerializedChatVectorDBQAChain {\n throw new Error(\"Not implemented.\");\n }\n\n /**\n * Static method to create a new ConversationalRetrievalQAChain from a\n * BaseLanguageModel and a BaseRetriever.\n * @param llm {@link BaseLanguageModelInterface} instance used to generate a new question.\n * @param retriever {@link BaseRetrieverInterface} instance used to retrieve relevant documents.\n * @param options.returnSourceDocuments Whether to return source documents in the final output\n * @param options.questionGeneratorChainOptions Options to initialize the standalone question generation chain used as the first internal step\n * @param options.qaChainOptions {@link QAChainParams} used to initialize the QA chain used as the second internal step\n * @returns A new instance of ConversationalRetrievalQAChain.\n */\n static fromLLM(\n llm: BaseLanguageModelInterface,\n retriever: BaseRetrieverInterface,\n options: {\n outputKey?: string; // not used\n returnSourceDocuments?: boolean;\n /** @deprecated Pass in questionGeneratorChainOptions.template instead */\n questionGeneratorTemplate?: string;\n /** @deprecated Pass in qaChainOptions.prompt instead */\n qaTemplate?: string;\n questionGeneratorChainOptions?: {\n llm?: BaseLanguageModelInterface;\n template?: string;\n };\n qaChainOptions?: QAChainParams;\n } & Omit<\n ConversationalRetrievalQAChainInput,\n \"retriever\" | \"combineDocumentsChain\" | \"questionGeneratorChain\"\n > = {}\n ): ConversationalRetrievalQAChain {\n const {\n questionGeneratorTemplate,\n qaTemplate,\n qaChainOptions = {\n type: \"stuff\",\n prompt: qaTemplate\n ? PromptTemplate.fromTemplate(qaTemplate)\n : undefined,\n },\n questionGeneratorChainOptions,\n verbose,\n ...rest\n } = options;\n\n const qaChain = loadQAChain(llm, qaChainOptions);\n\n const questionGeneratorChainPrompt = PromptTemplate.fromTemplate(\n questionGeneratorChainOptions?.template ??\n questionGeneratorTemplate ??\n question_generator_template\n );\n const questionGeneratorChain = new LLMChain({\n prompt: questionGeneratorChainPrompt,\n llm: questionGeneratorChainOptions?.llm ?? llm,\n verbose,\n });\n const instance = new this({\n retriever,\n combineDocumentsChain: qaChain,\n questionGeneratorChain,\n verbose,\n ...rest,\n });\n return instance;\n }\n}\n"],"mappings":";;;;;;AAcA,MAAM,8BAA8B;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA6FpC,IAAa,iCAAb,MAAa,uCACH,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,WAAW;CAEX,iBAAiB;CAEjB,IAAI,YAAY;AACd,SAAO,CAAC,KAAK,UAAU,KAAK,eAAe;;CAG7C,IAAI,aAAa;AACf,SAAO,KAAK,sBAAsB,WAAW,OAC3C,KAAK,wBAAwB,CAAC,kBAAkB,GAAG,EAAE,CACtD;;CAGH;CAEA;CAEA;CAEA,wBAAwB;CAExB,0BAA0B;CAE1B,YAAY,QAA6C;AACvD,QAAM,OAAO;AACb,OAAK,YAAY,OAAO;AACxB,OAAK,wBAAwB,OAAO;AACpC,OAAK,yBAAyB,OAAO;AACrC,OAAK,WAAW,OAAO,YAAY,KAAK;AACxC,OAAK,wBACH,OAAO,yBAAyB,KAAK;AACvC,OAAK,0BACH,OAAO,2BAA2B,KAAK;;;;;;;;CAS3C,OAAO,qBACL,aACA;EACA,IAAI;AACJ,MAAI,MAAM,QAAQ,YAAY,EAAE;AAE9B,OACE,MAAM,QAAQ,YAAY,GAAG,IAC7B,OAAO,YAAY,GAAG,OAAO,UAC7B;AACA,YAAQ,KACN,6LACD;AACD,sBAAkB,YAAY,MAAM,CAAC,KAAK,eAAe,MAAM;AAC7D,SAAI,IAAI,MAAM,EACZ,QAAO,IAAI,aAAa,cAAc;SAEtC,QAAO,IAAI,UAAU,cAAc;MAErC;SAEF,mBAAkB;AAEpB,UAAO,gBACJ,KAAK,gBAAgB;AACpB,QAAI,YAAY,UAAU,KAAK,QAC7B,QAAO,UAAU,YAAY;aACpB,YAAY,UAAU,KAAK,KACpC,QAAO,cAAc,YAAY;QAEjC,QAAO,GAAG,YAAY;KAExB,CACD,KAAK,KAAK;;AAEf,SAAO;;;CAIT,MAAM,MACJ,QACA,YACsB;AACtB,MAAI,EAAE,KAAK,YAAY,QACrB,OAAM,IAAI,MAAM,gBAAgB,KAAK,SAAS,aAAa;AAE7D,MAAI,EAAE,KAAK,kBAAkB,QAC3B,OAAM,IAAI,MAAM,oBAAoB,KAAK,eAAe,aAAa;EAEvE,MAAM,WAAmB,OAAO,KAAK;EACrC,MAAM,cACJ,+BAA+B,qBAC7B,OAAO,KAAK,gBACb;EACH,IAAI,cAAc;AAClB,MAAI,YAAY,SAAS,GAAG;GAC1B,MAAM,SAAS,MAAM,KAAK,uBAAuB,KAC/C;IACE;IACA,cAAc;IACf,EACD,YAAY,SAAS,qBAAqB,CAC3C;GACD,MAAM,OAAO,OAAO,KAAK,OAAO;AAChC,OAAI,KAAK,WAAW,EAClB,eAAc,OAAO,KAAK;OAE1B,OAAM,IAAI,MACR,2EACD;;EAGL,MAAM,OAAO,MAAM,KAAK,UAAU,OAChC,aACA,YAAY,SAAS,YAAY,CAClC;EACD,MAAM,SAAS;GACb,UAAU;GACV,iBAAiB;GACjB,cAAc;GACf;EACD,IAAI,SAAS,MAAM,KAAK,sBAAsB,KAC5C,QACA,YAAY,SAAS,oBAAoB,CAC1C;AACD,MAAI,KAAK,sBACP,UAAS;GACP,GAAG;GACH,iBAAiB;GAClB;AAEH,MAAI,KAAK,wBACP,UAAS;GACP,GAAG;GACH,mBAAmB;GACpB;AAEH,SAAO;;CAGT,aAAqB;AACnB,SAAO;;CAGT,aAAa,YACX,OACA,SACyC;AACzC,QAAM,IAAI,MAAM,mBAAmB;;CAGrC,YAA2C;AACzC,QAAM,IAAI,MAAM,mBAAmB;;;;;;;;;;;;CAarC,OAAO,QACL,KACA,WACA,UAeI,EAAE,EAC0B;EAChC,MAAM,EACJ,2BACA,YACA,iBAAiB;GACf,MAAM;GACN,QAAQ,aACJ,eAAe,aAAa,WAAW,GACvC,KAAA;GACL,EACD,+BACA,SACA,GAAG,SACD;EAEJ,MAAM,UAAU,YAAY,KAAK,eAAe;EAOhD,MAAM,yBAAyB,IAAI,SAAS;GAC1C,QANmC,eAAe,aAClD,+BAA+B,YAC7B,6BACA,4BACH;GAGC,KAAK,+BAA+B,OAAO;GAC3C;GACD,CAAC;AAQF,SAPiB,IAAI,KAAK;GACxB;GACA,uBAAuB;GACvB;GACA;GACA,GAAG;GACJ,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"cypher.cjs","names":["BaseChain","CYPHER_QA_PROMPT","CYPHER_GENERATION_PROMPT","LLMChain"],"sources":["../../../src/chains/graph_qa/cypher.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { LLMChain } from \"../llm_chain.js\";\nimport { BaseChain, ChainInputs } from \"../base.js\";\nimport { CYPHER_GENERATION_PROMPT, CYPHER_QA_PROMPT } from \"./prompts.js\";\nimport { logVersion020MigrationWarning } from \"../../util/entrypoint_deprecation.js\";\n\n/* #__PURE__ */ logVersion020MigrationWarning({\n oldEntrypointName: \"chains/graph_qa/cypher\",\n newPackageName: \"@langchain/community\",\n});\n\nexport const INTERMEDIATE_STEPS_KEY = \"intermediateSteps\";\n\nexport interface GraphCypherQAChainInput extends ChainInputs {\n //
|
|
1
|
+
{"version":3,"file":"cypher.cjs","names":["BaseChain","CYPHER_QA_PROMPT","CYPHER_GENERATION_PROMPT","LLMChain"],"sources":["../../../src/chains/graph_qa/cypher.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { LLMChain } from \"../llm_chain.js\";\nimport { BaseChain, ChainInputs } from \"../base.js\";\nimport { CYPHER_GENERATION_PROMPT, CYPHER_QA_PROMPT } from \"./prompts.js\";\nimport { logVersion020MigrationWarning } from \"../../util/entrypoint_deprecation.js\";\n\n/* #__PURE__ */ logVersion020MigrationWarning({\n oldEntrypointName: \"chains/graph_qa/cypher\",\n newPackageName: \"@langchain/community\",\n});\n\nexport const INTERMEDIATE_STEPS_KEY = \"intermediateSteps\";\n\nexport interface GraphCypherQAChainInput extends ChainInputs {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n graph: any;\n cypherGenerationChain: LLMChain;\n qaChain: LLMChain;\n inputKey?: string;\n outputKey?: string;\n topK?: number;\n returnIntermediateSteps?: boolean;\n returnDirect?: boolean;\n}\n\nexport interface FromLLMInput {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n graph: any;\n llm?: BaseLanguageModelInterface;\n cypherLLM?: BaseLanguageModelInterface;\n qaLLM?: BaseLanguageModelInterface;\n qaPrompt?: BasePromptTemplate;\n cypherPrompt?: BasePromptTemplate;\n returnIntermediateSteps?: boolean;\n returnDirect?: boolean;\n}\n\n/**\n * Chain for question-answering against a graph by generating Cypher statements.\n *\n * @example\n * ```typescript\n * const chain = new GraphCypherQAChain({\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\", temperature: 0 }),\n * graph: new Neo4jGraph(),\n * });\n * const res = await chain.invoke(\"Who played in Pulp Fiction?\");\n * ```\n *\n * @security\n * This chain will execute Cypher statements against the provided database.\n * Make sure that the database connection uses credentials\n * that are narrowly-scoped to only include necessary permissions.\n * Failure to do so may result in data corruption or loss, since the calling code\n * may attempt commands that would result in deletion, mutation of data\n * if appropriately prompted or reading sensitive data if such data is present in the database.\n * The best way to guard against such negative outcomes is to (as appropriate) limit the\n * permissions granted to the credentials used with this tool.\n *\n * See https://js.langchain.com/docs/security for more information.\n */\nexport class GraphCypherQAChain extends BaseChain {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n private graph: any;\n\n private cypherGenerationChain: LLMChain;\n\n private qaChain: LLMChain;\n\n private inputKey = \"query\";\n\n private outputKey = \"result\";\n\n private topK = 10;\n\n private returnDirect = false;\n\n private returnIntermediateSteps = false;\n\n constructor(props: GraphCypherQAChainInput) {\n super(props);\n const {\n graph,\n cypherGenerationChain,\n qaChain,\n inputKey,\n outputKey,\n topK,\n returnIntermediateSteps,\n returnDirect,\n } = props;\n\n this.graph = graph;\n this.cypherGenerationChain = cypherGenerationChain;\n this.qaChain = qaChain;\n\n if (inputKey) {\n this.inputKey = inputKey;\n }\n if (outputKey) {\n this.outputKey = outputKey;\n }\n if (topK) {\n this.topK = topK;\n }\n if (returnIntermediateSteps) {\n this.returnIntermediateSteps = returnIntermediateSteps;\n }\n if (returnDirect) {\n this.returnDirect = returnDirect;\n }\n }\n\n _chainType() {\n return \"graph_cypher_chain\" as const;\n }\n\n get inputKeys(): string[] {\n return [this.inputKey];\n }\n\n get outputKeys(): string[] {\n return [this.outputKey];\n }\n\n static fromLLM(props: FromLLMInput): GraphCypherQAChain {\n const {\n graph,\n qaPrompt = CYPHER_QA_PROMPT,\n cypherPrompt = CYPHER_GENERATION_PROMPT,\n llm,\n cypherLLM,\n qaLLM,\n returnIntermediateSteps = false,\n returnDirect = false,\n } = props;\n\n if (!cypherLLM && !llm) {\n throw new Error(\n \"Either 'llm' or 'cypherLLM' parameters must be provided\"\n );\n }\n\n if (!qaLLM && !llm) {\n throw new Error(\"Either 'llm' or 'qaLLM' parameters must be provided\");\n }\n\n if (cypherLLM && qaLLM && llm) {\n throw new Error(\n \"You can specify up to two of 'cypherLLM', 'qaLLM', and 'llm', but not all three simultaneously.\"\n );\n }\n\n const qaChain = new LLMChain({\n llm: (qaLLM || llm) as BaseLanguageModelInterface,\n prompt: qaPrompt,\n });\n\n const cypherGenerationChain = new LLMChain({\n llm: (cypherLLM || llm) as BaseLanguageModelInterface,\n prompt: cypherPrompt,\n });\n\n return new GraphCypherQAChain({\n cypherGenerationChain,\n qaChain,\n graph,\n returnIntermediateSteps,\n returnDirect,\n });\n }\n\n private extractCypher(text: string): string {\n const pattern = /```(.*?)```/s;\n const matches = text.match(pattern);\n return matches ? matches[1] : text;\n }\n\n async _call(\n values: ChainValues,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const callbacks = runManager?.getChild();\n const question = values[this.inputKey];\n\n const intermediateSteps = [];\n\n const generatedCypher = await this.cypherGenerationChain.call(\n { question, schema: this.graph.getSchema() },\n callbacks\n );\n\n const extractedCypher = this.extractCypher(generatedCypher.text);\n\n await runManager?.handleText(`Generated Cypher:\\n`);\n await runManager?.handleText(`${extractedCypher} green\\n`);\n\n intermediateSteps.push({ query: extractedCypher });\n\n let chainResult: ChainValues;\n const context = await this.graph.query(extractedCypher, {\n topK: this.topK,\n });\n\n if (this.returnDirect) {\n chainResult = { [this.outputKey]: context };\n } else {\n await runManager?.handleText(\"Full Context:\\n\");\n await runManager?.handleText(`${context} green\\n`);\n\n intermediateSteps.push({ context });\n\n const result = await this.qaChain.call(\n { question, context: JSON.stringify(context) },\n callbacks\n );\n\n chainResult = {\n [this.outputKey]: result[this.qaChain.outputKey],\n };\n }\n\n if (this.returnIntermediateSteps) {\n chainResult[INTERMEDIATE_STEPS_KEY] = intermediateSteps;\n }\n\n return chainResult;\n }\n}\n"],"mappings":";;;;;;;;;;;AAcA,MAAa,yBAAyB;;;;;;;;;;;;;;;;;;;;;;;;;AAkDtC,IAAa,qBAAb,MAAa,2BAA2BA,aAAAA,UAAU;CAEhD;CAEA;CAEA;CAEA,WAAmB;CAEnB,YAAoB;CAEpB,OAAe;CAEf,eAAuB;CAEvB,0BAAkC;CAElC,YAAY,OAAgC;AAC1C,QAAM,MAAM;EACZ,MAAM,EACJ,OACA,uBACA,SACA,UACA,WACA,MACA,yBACA,iBACE;AAEJ,OAAK,QAAQ;AACb,OAAK,wBAAwB;AAC7B,OAAK,UAAU;AAEf,MAAI,SACF,MAAK,WAAW;AAElB,MAAI,UACF,MAAK,YAAY;AAEnB,MAAI,KACF,MAAK,OAAO;AAEd,MAAI,wBACF,MAAK,0BAA0B;AAEjC,MAAI,aACF,MAAK,eAAe;;CAIxB,aAAa;AACX,SAAO;;CAGT,IAAI,YAAsB;AACxB,SAAO,CAAC,KAAK,SAAS;;CAGxB,IAAI,aAAuB;AACzB,SAAO,CAAC,KAAK,UAAU;;CAGzB,OAAO,QAAQ,OAAyC;EACtD,MAAM,EACJ,OACA,WAAWC,gBAAAA,kBACX,eAAeC,gBAAAA,0BACf,KACA,WACA,OACA,0BAA0B,OAC1B,eAAe,UACb;AAEJ,MAAI,CAAC,aAAa,CAAC,IACjB,OAAM,IAAI,MACR,0DACD;AAGH,MAAI,CAAC,SAAS,CAAC,IACb,OAAM,IAAI,MAAM,sDAAsD;AAGxE,MAAI,aAAa,SAAS,IACxB,OAAM,IAAI,MACR,kGACD;EAGH,MAAM,UAAU,IAAIC,kBAAAA,SAAS;GAC3B,KAAM,SAAS;GACf,QAAQ;GACT,CAAC;AAOF,SAAO,IAAI,mBAAmB;GAC5B,uBAN4B,IAAIA,kBAAAA,SAAS;IACzC,KAAM,aAAa;IACnB,QAAQ;IACT,CAAC;GAIA;GACA;GACA;GACA;GACD,CAAC;;CAGJ,cAAsB,MAAsB;EAE1C,MAAM,UAAU,KAAK,MADL,eACmB;AACnC,SAAO,UAAU,QAAQ,KAAK;;CAGhC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,YAAY,YAAY,UAAU;EACxC,MAAM,WAAW,OAAO,KAAK;EAE7B,MAAM,oBAAoB,EAAE;EAE5B,MAAM,kBAAkB,MAAM,KAAK,sBAAsB,KACvD;GAAE;GAAU,QAAQ,KAAK,MAAM,WAAW;GAAE,EAC5C,UACD;EAED,MAAM,kBAAkB,KAAK,cAAc,gBAAgB,KAAK;AAEhE,QAAM,YAAY,WAAW,sBAAsB;AACnD,QAAM,YAAY,WAAW,GAAG,gBAAgB,UAAU;AAE1D,oBAAkB,KAAK,EAAE,OAAO,iBAAiB,CAAC;EAElD,IAAI;EACJ,MAAM,UAAU,MAAM,KAAK,MAAM,MAAM,iBAAiB,EACtD,MAAM,KAAK,MACZ,CAAC;AAEF,MAAI,KAAK,aACP,eAAc,GAAG,KAAK,YAAY,SAAS;OACtC;AACL,SAAM,YAAY,WAAW,kBAAkB;AAC/C,SAAM,YAAY,WAAW,GAAG,QAAQ,UAAU;AAElD,qBAAkB,KAAK,EAAE,SAAS,CAAC;GAEnC,MAAM,SAAS,MAAM,KAAK,QAAQ,KAChC;IAAE;IAAU,SAAS,KAAK,UAAU,QAAQ;IAAE,EAC9C,UACD;AAED,iBAAc,GACX,KAAK,YAAY,OAAO,KAAK,QAAQ,YACvC;;AAGH,MAAI,KAAK,wBACP,aAAY,0BAA0B;AAGxC,SAAO"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"cypher.js","names":[],"sources":["../../../src/chains/graph_qa/cypher.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { LLMChain } from \"../llm_chain.js\";\nimport { BaseChain, ChainInputs } from \"../base.js\";\nimport { CYPHER_GENERATION_PROMPT, CYPHER_QA_PROMPT } from \"./prompts.js\";\nimport { logVersion020MigrationWarning } from \"../../util/entrypoint_deprecation.js\";\n\n/* #__PURE__ */ logVersion020MigrationWarning({\n oldEntrypointName: \"chains/graph_qa/cypher\",\n newPackageName: \"@langchain/community\",\n});\n\nexport const INTERMEDIATE_STEPS_KEY = \"intermediateSteps\";\n\nexport interface GraphCypherQAChainInput extends ChainInputs {\n //
|
|
1
|
+
{"version":3,"file":"cypher.js","names":[],"sources":["../../../src/chains/graph_qa/cypher.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { LLMChain } from \"../llm_chain.js\";\nimport { BaseChain, ChainInputs } from \"../base.js\";\nimport { CYPHER_GENERATION_PROMPT, CYPHER_QA_PROMPT } from \"./prompts.js\";\nimport { logVersion020MigrationWarning } from \"../../util/entrypoint_deprecation.js\";\n\n/* #__PURE__ */ logVersion020MigrationWarning({\n oldEntrypointName: \"chains/graph_qa/cypher\",\n newPackageName: \"@langchain/community\",\n});\n\nexport const INTERMEDIATE_STEPS_KEY = \"intermediateSteps\";\n\nexport interface GraphCypherQAChainInput extends ChainInputs {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n graph: any;\n cypherGenerationChain: LLMChain;\n qaChain: LLMChain;\n inputKey?: string;\n outputKey?: string;\n topK?: number;\n returnIntermediateSteps?: boolean;\n returnDirect?: boolean;\n}\n\nexport interface FromLLMInput {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n graph: any;\n llm?: BaseLanguageModelInterface;\n cypherLLM?: BaseLanguageModelInterface;\n qaLLM?: BaseLanguageModelInterface;\n qaPrompt?: BasePromptTemplate;\n cypherPrompt?: BasePromptTemplate;\n returnIntermediateSteps?: boolean;\n returnDirect?: boolean;\n}\n\n/**\n * Chain for question-answering against a graph by generating Cypher statements.\n *\n * @example\n * ```typescript\n * const chain = new GraphCypherQAChain({\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\", temperature: 0 }),\n * graph: new Neo4jGraph(),\n * });\n * const res = await chain.invoke(\"Who played in Pulp Fiction?\");\n * ```\n *\n * @security\n * This chain will execute Cypher statements against the provided database.\n * Make sure that the database connection uses credentials\n * that are narrowly-scoped to only include necessary permissions.\n * Failure to do so may result in data corruption or loss, since the calling code\n * may attempt commands that would result in deletion, mutation of data\n * if appropriately prompted or reading sensitive data if such data is present in the database.\n * The best way to guard against such negative outcomes is to (as appropriate) limit the\n * permissions granted to the credentials used with this tool.\n *\n * See https://js.langchain.com/docs/security for more information.\n */\nexport class GraphCypherQAChain extends BaseChain {\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n private graph: any;\n\n private cypherGenerationChain: LLMChain;\n\n private qaChain: LLMChain;\n\n private inputKey = \"query\";\n\n private outputKey = \"result\";\n\n private topK = 10;\n\n private returnDirect = false;\n\n private returnIntermediateSteps = false;\n\n constructor(props: GraphCypherQAChainInput) {\n super(props);\n const {\n graph,\n cypherGenerationChain,\n qaChain,\n inputKey,\n outputKey,\n topK,\n returnIntermediateSteps,\n returnDirect,\n } = props;\n\n this.graph = graph;\n this.cypherGenerationChain = cypherGenerationChain;\n this.qaChain = qaChain;\n\n if (inputKey) {\n this.inputKey = inputKey;\n }\n if (outputKey) {\n this.outputKey = outputKey;\n }\n if (topK) {\n this.topK = topK;\n }\n if (returnIntermediateSteps) {\n this.returnIntermediateSteps = returnIntermediateSteps;\n }\n if (returnDirect) {\n this.returnDirect = returnDirect;\n }\n }\n\n _chainType() {\n return \"graph_cypher_chain\" as const;\n }\n\n get inputKeys(): string[] {\n return [this.inputKey];\n }\n\n get outputKeys(): string[] {\n return [this.outputKey];\n }\n\n static fromLLM(props: FromLLMInput): GraphCypherQAChain {\n const {\n graph,\n qaPrompt = CYPHER_QA_PROMPT,\n cypherPrompt = CYPHER_GENERATION_PROMPT,\n llm,\n cypherLLM,\n qaLLM,\n returnIntermediateSteps = false,\n returnDirect = false,\n } = props;\n\n if (!cypherLLM && !llm) {\n throw new Error(\n \"Either 'llm' or 'cypherLLM' parameters must be provided\"\n );\n }\n\n if (!qaLLM && !llm) {\n throw new Error(\"Either 'llm' or 'qaLLM' parameters must be provided\");\n }\n\n if (cypherLLM && qaLLM && llm) {\n throw new Error(\n \"You can specify up to two of 'cypherLLM', 'qaLLM', and 'llm', but not all three simultaneously.\"\n );\n }\n\n const qaChain = new LLMChain({\n llm: (qaLLM || llm) as BaseLanguageModelInterface,\n prompt: qaPrompt,\n });\n\n const cypherGenerationChain = new LLMChain({\n llm: (cypherLLM || llm) as BaseLanguageModelInterface,\n prompt: cypherPrompt,\n });\n\n return new GraphCypherQAChain({\n cypherGenerationChain,\n qaChain,\n graph,\n returnIntermediateSteps,\n returnDirect,\n });\n }\n\n private extractCypher(text: string): string {\n const pattern = /```(.*?)```/s;\n const matches = text.match(pattern);\n return matches ? matches[1] : text;\n }\n\n async _call(\n values: ChainValues,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const callbacks = runManager?.getChild();\n const question = values[this.inputKey];\n\n const intermediateSteps = [];\n\n const generatedCypher = await this.cypherGenerationChain.call(\n { question, schema: this.graph.getSchema() },\n callbacks\n );\n\n const extractedCypher = this.extractCypher(generatedCypher.text);\n\n await runManager?.handleText(`Generated Cypher:\\n`);\n await runManager?.handleText(`${extractedCypher} green\\n`);\n\n intermediateSteps.push({ query: extractedCypher });\n\n let chainResult: ChainValues;\n const context = await this.graph.query(extractedCypher, {\n topK: this.topK,\n });\n\n if (this.returnDirect) {\n chainResult = { [this.outputKey]: context };\n } else {\n await runManager?.handleText(\"Full Context:\\n\");\n await runManager?.handleText(`${context} green\\n`);\n\n intermediateSteps.push({ context });\n\n const result = await this.qaChain.call(\n { question, context: JSON.stringify(context) },\n callbacks\n );\n\n chainResult = {\n [this.outputKey]: result[this.qaChain.outputKey],\n };\n }\n\n if (this.returnIntermediateSteps) {\n chainResult[INTERMEDIATE_STEPS_KEY] = intermediateSteps;\n }\n\n return chainResult;\n }\n}\n"],"mappings":";;;;;;;;;;AAcA,MAAa,yBAAyB;;;;;;;;;;;;;;;;;;;;;;;;;AAkDtC,IAAa,qBAAb,MAAa,2BAA2B,UAAU;CAEhD;CAEA;CAEA;CAEA,WAAmB;CAEnB,YAAoB;CAEpB,OAAe;CAEf,eAAuB;CAEvB,0BAAkC;CAElC,YAAY,OAAgC;AAC1C,QAAM,MAAM;EACZ,MAAM,EACJ,OACA,uBACA,SACA,UACA,WACA,MACA,yBACA,iBACE;AAEJ,OAAK,QAAQ;AACb,OAAK,wBAAwB;AAC7B,OAAK,UAAU;AAEf,MAAI,SACF,MAAK,WAAW;AAElB,MAAI,UACF,MAAK,YAAY;AAEnB,MAAI,KACF,MAAK,OAAO;AAEd,MAAI,wBACF,MAAK,0BAA0B;AAEjC,MAAI,aACF,MAAK,eAAe;;CAIxB,aAAa;AACX,SAAO;;CAGT,IAAI,YAAsB;AACxB,SAAO,CAAC,KAAK,SAAS;;CAGxB,IAAI,aAAuB;AACzB,SAAO,CAAC,KAAK,UAAU;;CAGzB,OAAO,QAAQ,OAAyC;EACtD,MAAM,EACJ,OACA,WAAW,kBACX,eAAe,0BACf,KACA,WACA,OACA,0BAA0B,OAC1B,eAAe,UACb;AAEJ,MAAI,CAAC,aAAa,CAAC,IACjB,OAAM,IAAI,MACR,0DACD;AAGH,MAAI,CAAC,SAAS,CAAC,IACb,OAAM,IAAI,MAAM,sDAAsD;AAGxE,MAAI,aAAa,SAAS,IACxB,OAAM,IAAI,MACR,kGACD;EAGH,MAAM,UAAU,IAAI,SAAS;GAC3B,KAAM,SAAS;GACf,QAAQ;GACT,CAAC;AAOF,SAAO,IAAI,mBAAmB;GAC5B,uBAN4B,IAAI,SAAS;IACzC,KAAM,aAAa;IACnB,QAAQ;IACT,CAAC;GAIA;GACA;GACA;GACA;GACD,CAAC;;CAGJ,cAAsB,MAAsB;EAE1C,MAAM,UAAU,KAAK,MADL,eACmB;AACnC,SAAO,UAAU,QAAQ,KAAK;;CAGhC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,YAAY,YAAY,UAAU;EACxC,MAAM,WAAW,OAAO,KAAK;EAE7B,MAAM,oBAAoB,EAAE;EAE5B,MAAM,kBAAkB,MAAM,KAAK,sBAAsB,KACvD;GAAE;GAAU,QAAQ,KAAK,MAAM,WAAW;GAAE,EAC5C,UACD;EAED,MAAM,kBAAkB,KAAK,cAAc,gBAAgB,KAAK;AAEhE,QAAM,YAAY,WAAW,sBAAsB;AACnD,QAAM,YAAY,WAAW,GAAG,gBAAgB,UAAU;AAE1D,oBAAkB,KAAK,EAAE,OAAO,iBAAiB,CAAC;EAElD,IAAI;EACJ,MAAM,UAAU,MAAM,KAAK,MAAM,MAAM,iBAAiB,EACtD,MAAM,KAAK,MACZ,CAAC;AAEF,MAAI,KAAK,aACP,eAAc,GAAG,KAAK,YAAY,SAAS;OACtC;AACL,SAAM,YAAY,WAAW,kBAAkB;AAC/C,SAAM,YAAY,WAAW,GAAG,QAAQ,UAAU;AAElD,qBAAkB,KAAK,EAAE,SAAS,CAAC;GAEnC,MAAM,SAAS,MAAM,KAAK,QAAQ,KAChC;IAAE;IAAU,SAAS,KAAK,UAAU,QAAQ;IAAE,EAC9C,UACD;AAED,iBAAc,GACX,KAAK,YAAY,OAAO,KAAK,QAAQ,YACvC;;AAGH,MAAI,KAAK,wBACP,aAAY,0BAA0B;AAGxC,SAAO"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"llm_chain.cjs","names":["Runnable","BaseChain","NoOpOutputParser","BaseLanguageModel","BasePromptTemplate"],"sources":["../../src/chains/llm_chain.ts"],"sourcesContent":["import {\n BaseLanguageModel,\n BaseLanguageModelInterface,\n BaseLanguageModelInput,\n} from \"@langchain/core/language_models/base\";\nimport type { ChainValues } from \"@langchain/core/utils/types\";\nimport type { Generation } from \"@langchain/core/outputs\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport type { BasePromptValueInterface } from \"@langchain/core/prompt_values\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n BaseLLMOutputParser,\n BaseOutputParser,\n} from \"@langchain/core/output_parsers\";\nimport {\n CallbackManager,\n BaseCallbackConfig,\n CallbackManagerForChainRun,\n Callbacks,\n} from \"@langchain/core/callbacks/manager\";\nimport { Runnable, type RunnableInterface } from \"@langchain/core/runnables\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { SerializedLLMChain } from \"./serde.js\";\nimport { NoOpOutputParser } from \"../output_parsers/noop.js\";\n\ntype LLMType =\n | BaseLanguageModelInterface\n | Runnable<BaseLanguageModelInput, string>\n | Runnable<BaseLanguageModelInput, BaseMessage>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype CallOptionsIfAvailable<T> = T extends { CallOptions: infer CO } ? CO : any;\n/**\n * Interface for the input parameters of the LLMChain class.\n */\nexport interface LLMChainInput<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n> extends ChainInputs {\n /** Prompt object to use */\n prompt: BasePromptTemplate;\n /** LLM Wrapper to use */\n llm: Model;\n /** Kwargs to pass to LLM */\n llmKwargs?: CallOptionsIfAvailable<Model>;\n /** OutputParser to use */\n outputParser?: BaseLLMOutputParser<T>;\n /** Key to use for output, defaults to `text` */\n outputKey?: string;\n}\n\nfunction isBaseLanguageModel(llmLike: unknown): llmLike is BaseLanguageModel {\n return typeof (llmLike as BaseLanguageModelInterface)._llmType === \"function\";\n}\n\nfunction _getLanguageModel(llmLike: RunnableInterface): BaseLanguageModel {\n if (isBaseLanguageModel(llmLike)) {\n return llmLike;\n } else if (\"bound\" in llmLike && Runnable.isRunnable(llmLike.bound)) {\n return _getLanguageModel(llmLike.bound);\n } else if (\n \"runnable\" in llmLike &&\n \"fallbacks\" in llmLike &&\n Runnable.isRunnable(llmLike.runnable)\n ) {\n return _getLanguageModel(llmLike.runnable);\n } else if (\"default\" in llmLike && Runnable.isRunnable(llmLike.default)) {\n return _getLanguageModel(llmLike.default);\n } else {\n throw new Error(\"Unable to extract BaseLanguageModel from llmLike object.\");\n }\n}\n\n/**\n * Chain to run queries against LLMs.\n *\n * @example\n * ```ts\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n *\n * const prompt = ChatPromptTemplate.fromTemplate(\"Tell me a {adjective} joke\");\n * const llm = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const chain = prompt.pipe(llm);\n *\n * const response = await chain.invoke({ adjective: \"funny\" });\n * ```\n */\nexport class LLMChain<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n>\n extends BaseChain\n implements LLMChainInput<T>\n{\n static lc_name() {\n return \"LLMChain\";\n }\n\n lc_serializable = true;\n\n prompt: BasePromptTemplate;\n\n llm: Model;\n\n llmKwargs?: CallOptionsIfAvailable<Model>;\n\n outputKey = \"text\";\n\n outputParser?: BaseLLMOutputParser<T>;\n\n get inputKeys() {\n return this.prompt.inputVariables;\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n constructor(fields: LLMChainInput<T, Model>) {\n super(fields);\n this.prompt = fields.prompt;\n this.llm = fields.llm;\n this.llmKwargs = fields.llmKwargs;\n this.outputKey = fields.outputKey ?? this.outputKey;\n this.outputParser =\n fields.outputParser ?? (new NoOpOutputParser() as BaseOutputParser<T>);\n if (this.prompt.outputParser) {\n if (fields.outputParser) {\n throw new Error(\"Cannot set both outputParser and prompt.outputParser\");\n }\n this.outputParser = this.prompt.outputParser as BaseOutputParser<T>;\n }\n }\n\n private getCallKeys(): string[] {\n const callKeys = \"callKeys\" in this.llm ? this.llm.callKeys : [];\n return callKeys;\n }\n\n /** @ignore */\n _selectMemoryInputs(values: ChainValues): ChainValues {\n const valuesForMemory = super._selectMemoryInputs(values);\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n delete valuesForMemory[key];\n }\n }\n return valuesForMemory;\n }\n\n /** @ignore */\n async _getFinalOutput(\n generations: Generation[],\n promptValue: BasePromptValueInterface,\n runManager?: CallbackManagerForChainRun\n ): Promise<unknown> {\n let finalCompletion: unknown;\n if (this.outputParser) {\n finalCompletion = await this.outputParser.parseResultWithPrompt(\n generations,\n promptValue,\n runManager?.getChild()\n );\n } else {\n finalCompletion = generations[0].text;\n }\n return finalCompletion;\n }\n\n /**\n * Run the core logic of this chain and add to output if desired.\n *\n * Wraps _call and handles memory.\n */\n call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n config?: Callbacks | BaseCallbackConfig\n ): Promise<ChainValues> {\n return super.call(values, config);\n }\n\n /** @ignore */\n async _call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const valuesForPrompt = { ...values };\n const valuesForLLM = {\n ...this.llmKwargs,\n } as CallOptionsIfAvailable<Model>;\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n if (valuesForLLM) {\n valuesForLLM[key as keyof CallOptionsIfAvailable<Model>] =\n values[key];\n delete valuesForPrompt[key];\n }\n }\n }\n const promptValue = await this.prompt.formatPromptValue(valuesForPrompt);\n if (\"generatePrompt\" in this.llm) {\n const { generations } = await this.llm.generatePrompt(\n [promptValue],\n valuesForLLM,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: await this._getFinalOutput(\n generations[0],\n promptValue,\n runManager\n ),\n };\n }\n\n const modelWithParser = this.outputParser\n ? this.llm.pipe(this.outputParser)\n : this.llm;\n const response = await modelWithParser.invoke(\n promptValue,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: response,\n };\n }\n\n /**\n * Format prompt with values and pass to LLM\n *\n * @param values - keys to pass to prompt template\n * @param callbackManager - CallbackManager to use\n * @returns Completion from LLM.\n *\n * @example\n * ```ts\n * llm.predict({ adjective: \"funny\" })\n * ```\n */\n async predict(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n callbackManager?: CallbackManager\n ): Promise<T> {\n const output = await this.call(values, callbackManager);\n return output[this.outputKey];\n }\n\n _chainType() {\n return \"llm\" as const;\n }\n\n static async deserialize(data: SerializedLLMChain): Promise<LLMChain> {\n const { llm, prompt } = data;\n if (!llm) {\n throw new Error(\"LLMChain must have llm\");\n }\n if (!prompt) {\n throw new Error(\"LLMChain must have prompt\");\n }\n\n return new LLMChain({\n llm: await BaseLanguageModel.deserialize(llm),\n prompt: await BasePromptTemplate.deserialize(prompt),\n });\n }\n\n /** @deprecated */\n serialize(): SerializedLLMChain {\n const serialize =\n \"serialize\" in this.llm ? this.llm.serialize() : undefined;\n return {\n _type: `${this._chainType()}_chain`,\n llm: serialize,\n prompt: this.prompt.serialize(),\n };\n }\n\n _getNumTokens(text: string): Promise<number> {\n return _getLanguageModel(this.llm).getNumTokens(text);\n }\n}\n"],"mappings":";;;;;;;AAmDA,SAAS,oBAAoB,SAAgD;AAC3E,QAAO,OAAQ,QAAuC,aAAa;;AAGrE,SAAS,kBAAkB,SAA+C;AACxE,KAAI,oBAAoB,QAAQ,CAC9B,QAAO;UACE,WAAW,WAAWA,0BAAAA,SAAS,WAAW,QAAQ,MAAM,CACjE,QAAO,kBAAkB,QAAQ,MAAM;UAEvC,cAAc,WACd,eAAe,WACfA,0BAAAA,SAAS,WAAW,QAAQ,SAAS,CAErC,QAAO,kBAAkB,QAAQ,SAAS;UACjC,aAAa,WAAWA,0BAAAA,SAAS,WAAW,QAAQ,QAAQ,CACrE,QAAO,kBAAkB,QAAQ,QAAQ;KAEzC,OAAM,IAAI,MAAM,2DAA2D;;;;;;;;;;;;;;;;;AAmB/E,IAAa,WAAb,MAAa,iBAIHC,aAAAA,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,kBAAkB;CAElB;CAEA;CAEA;CAEA,YAAY;CAEZ;CAEA,IAAI,YAAY;AACd,SAAO,KAAK,OAAO;;CAGrB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,YAAY,QAAiC;AAC3C,QAAM,OAAO;AACb,OAAK,SAAS,OAAO;AACrB,OAAK,MAAM,OAAO;AAClB,OAAK,YAAY,OAAO;AACxB,OAAK,YAAY,OAAO,aAAa,KAAK;AAC1C,OAAK,eACH,OAAO,gBAAiB,IAAIC,aAAAA,kBAAkB;AAChD,MAAI,KAAK,OAAO,cAAc;AAC5B,OAAI,OAAO,aACT,OAAM,IAAI,MAAM,uDAAuD;AAEzE,QAAK,eAAe,KAAK,OAAO;;;CAIpC,cAAgC;AAE9B,SADiB,cAAc,KAAK,MAAM,KAAK,IAAI,WAAW,EAAE;;;CAKlE,oBAAoB,QAAkC;EACpD,MAAM,kBAAkB,MAAM,oBAAoB,OAAO;EACzD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO,OACT,QAAO,gBAAgB;AAG3B,SAAO;;;CAIT,MAAM,gBACJ,aACA,aACA,YACkB;EAClB,IAAI;AACJ,MAAI,KAAK,aACP,mBAAkB,MAAM,KAAK,aAAa,sBACxC,aACA,aACA,YAAY,UAAU,CACvB;MAED,mBAAkB,YAAY,GAAG;AAEnC,SAAO;;;;;;;CAQT,KACE,QACA,QACsB;AACtB,SAAO,MAAM,KAAK,QAAQ,OAAO;;;CAInC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,kBAAkB,EAAE,GAAG,QAAQ;EACrC,MAAM,eAAe,EACnB,GAAG,KAAK,WACT;EACD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO;OACL,cAAc;AAChB,iBAAa,OACX,OAAO;AACT,WAAO,gBAAgB;;;EAI7B,MAAM,cAAc,MAAM,KAAK,OAAO,kBAAkB,gBAAgB;AACxE,MAAI,oBAAoB,KAAK,KAAK;GAChC,MAAM,EAAE,gBAAgB,MAAM,KAAK,IAAI,eACrC,CAAC,YAAY,EACb,cACA,YAAY,UAAU,CACvB;AACD,UAAO,GACJ,KAAK,YAAY,MAAM,KAAK,gBAC3B,YAAY,IACZ,aACA,WACD,EACF;;EAMH,MAAM,WAAW,OAHO,KAAK,eACzB,KAAK,IAAI,KAAK,KAAK,aAAa,GAChC,KAAK,KAC8B,OACrC,aACA,YAAY,UAAU,CACvB;AACD,SAAO,GACJ,KAAK,YAAY,UACnB;;;;;;;;;;;;;;CAeH,MAAM,QACJ,QACA,iBACY;AAEZ,UADe,MAAM,KAAK,KAAK,QAAQ,gBAAgB,EACzC,KAAK;;CAGrB,aAAa;AACX,SAAO;;CAGT,aAAa,YAAY,MAA6C;EACpE,MAAM,EAAE,KAAK,WAAW;AACxB,MAAI,CAAC,IACH,OAAM,IAAI,MAAM,yBAAyB;AAE3C,MAAI,CAAC,OACH,OAAM,IAAI,MAAM,4BAA4B;AAG9C,SAAO,IAAI,SAAS;GAClB,KAAK,MAAMC,qCAAAA,kBAAkB,YAAY,IAAI;GAC7C,QAAQ,MAAMC,wBAAAA,mBAAmB,YAAY,OAAO;GACrD,CAAC;;;CAIJ,YAAgC;EAC9B,MAAM,YACJ,eAAe,KAAK,MAAM,KAAK,IAAI,WAAW,GAAG,KAAA;AACnD,SAAO;GACL,OAAO,GAAG,KAAK,YAAY,CAAC;GAC5B,KAAK;GACL,QAAQ,KAAK,OAAO,WAAW;GAChC;;CAGH,cAAc,MAA+B;AAC3C,SAAO,kBAAkB,KAAK,IAAI,CAAC,aAAa,KAAK"}
|
|
1
|
+
{"version":3,"file":"llm_chain.cjs","names":["Runnable","BaseChain","NoOpOutputParser","BaseLanguageModel","BasePromptTemplate"],"sources":["../../src/chains/llm_chain.ts"],"sourcesContent":["import {\n BaseLanguageModel,\n BaseLanguageModelInterface,\n BaseLanguageModelInput,\n} from \"@langchain/core/language_models/base\";\nimport type { ChainValues } from \"@langchain/core/utils/types\";\nimport type { Generation } from \"@langchain/core/outputs\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport type { BasePromptValueInterface } from \"@langchain/core/prompt_values\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n BaseLLMOutputParser,\n BaseOutputParser,\n} from \"@langchain/core/output_parsers\";\nimport {\n CallbackManager,\n BaseCallbackConfig,\n CallbackManagerForChainRun,\n Callbacks,\n} from \"@langchain/core/callbacks/manager\";\nimport { Runnable, type RunnableInterface } from \"@langchain/core/runnables\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { SerializedLLMChain } from \"./serde.js\";\nimport { NoOpOutputParser } from \"../output_parsers/noop.js\";\n\ntype LLMType =\n | BaseLanguageModelInterface\n | Runnable<BaseLanguageModelInput, string>\n | Runnable<BaseLanguageModelInput, BaseMessage>;\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\ntype CallOptionsIfAvailable<T> = T extends { CallOptions: infer CO } ? CO : any;\n/**\n * Interface for the input parameters of the LLMChain class.\n */\nexport interface LLMChainInput<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n> extends ChainInputs {\n /** Prompt object to use */\n prompt: BasePromptTemplate;\n /** LLM Wrapper to use */\n llm: Model;\n /** Kwargs to pass to LLM */\n llmKwargs?: CallOptionsIfAvailable<Model>;\n /** OutputParser to use */\n outputParser?: BaseLLMOutputParser<T>;\n /** Key to use for output, defaults to `text` */\n outputKey?: string;\n}\n\nfunction isBaseLanguageModel(llmLike: unknown): llmLike is BaseLanguageModel {\n return typeof (llmLike as BaseLanguageModelInterface)._llmType === \"function\";\n}\n\nfunction _getLanguageModel(llmLike: RunnableInterface): BaseLanguageModel {\n if (isBaseLanguageModel(llmLike)) {\n return llmLike;\n } else if (\"bound\" in llmLike && Runnable.isRunnable(llmLike.bound)) {\n return _getLanguageModel(llmLike.bound);\n } else if (\n \"runnable\" in llmLike &&\n \"fallbacks\" in llmLike &&\n Runnable.isRunnable(llmLike.runnable)\n ) {\n return _getLanguageModel(llmLike.runnable);\n } else if (\"default\" in llmLike && Runnable.isRunnable(llmLike.default)) {\n return _getLanguageModel(llmLike.default);\n } else {\n throw new Error(\"Unable to extract BaseLanguageModel from llmLike object.\");\n }\n}\n\n/**\n * Chain to run queries against LLMs.\n *\n * @example\n * ```ts\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n *\n * const prompt = ChatPromptTemplate.fromTemplate(\"Tell me a {adjective} joke\");\n * const llm = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const chain = prompt.pipe(llm);\n *\n * const response = await chain.invoke({ adjective: \"funny\" });\n * ```\n */\nexport class LLMChain<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n>\n extends BaseChain\n implements LLMChainInput<T>\n{\n static lc_name() {\n return \"LLMChain\";\n }\n\n lc_serializable = true;\n\n prompt: BasePromptTemplate;\n\n llm: Model;\n\n llmKwargs?: CallOptionsIfAvailable<Model>;\n\n outputKey = \"text\";\n\n outputParser?: BaseLLMOutputParser<T>;\n\n get inputKeys() {\n return this.prompt.inputVariables;\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n constructor(fields: LLMChainInput<T, Model>) {\n super(fields);\n this.prompt = fields.prompt;\n this.llm = fields.llm;\n this.llmKwargs = fields.llmKwargs;\n this.outputKey = fields.outputKey ?? this.outputKey;\n this.outputParser =\n fields.outputParser ?? (new NoOpOutputParser() as BaseOutputParser<T>);\n if (this.prompt.outputParser) {\n if (fields.outputParser) {\n throw new Error(\"Cannot set both outputParser and prompt.outputParser\");\n }\n this.outputParser = this.prompt.outputParser as BaseOutputParser<T>;\n }\n }\n\n private getCallKeys(): string[] {\n const callKeys = \"callKeys\" in this.llm ? this.llm.callKeys : [];\n return callKeys;\n }\n\n /** @ignore */\n _selectMemoryInputs(values: ChainValues): ChainValues {\n const valuesForMemory = super._selectMemoryInputs(values);\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n delete valuesForMemory[key];\n }\n }\n return valuesForMemory;\n }\n\n /** @ignore */\n async _getFinalOutput(\n generations: Generation[],\n promptValue: BasePromptValueInterface,\n runManager?: CallbackManagerForChainRun\n ): Promise<unknown> {\n let finalCompletion: unknown;\n if (this.outputParser) {\n finalCompletion = await this.outputParser.parseResultWithPrompt(\n generations,\n promptValue,\n runManager?.getChild()\n );\n } else {\n finalCompletion = generations[0].text;\n }\n return finalCompletion;\n }\n\n /**\n * Run the core logic of this chain and add to output if desired.\n *\n * Wraps _call and handles memory.\n */\n call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n config?: Callbacks | BaseCallbackConfig\n ): Promise<ChainValues> {\n return super.call(values, config);\n }\n\n /** @ignore */\n async _call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const valuesForPrompt = { ...values };\n const valuesForLLM = {\n ...this.llmKwargs,\n } as CallOptionsIfAvailable<Model>;\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n if (valuesForLLM) {\n valuesForLLM[key as keyof CallOptionsIfAvailable<Model>] =\n values[key];\n delete valuesForPrompt[key];\n }\n }\n }\n const promptValue = await this.prompt.formatPromptValue(valuesForPrompt);\n if (\"generatePrompt\" in this.llm) {\n const { generations } = await this.llm.generatePrompt(\n [promptValue],\n valuesForLLM,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: await this._getFinalOutput(\n generations[0],\n promptValue,\n runManager\n ),\n };\n }\n\n const modelWithParser = this.outputParser\n ? this.llm.pipe(this.outputParser)\n : this.llm;\n const response = await modelWithParser.invoke(\n promptValue,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: response,\n };\n }\n\n /**\n * Format prompt with values and pass to LLM\n *\n * @param values - keys to pass to prompt template\n * @param callbackManager - CallbackManager to use\n * @returns Completion from LLM.\n *\n * @example\n * ```ts\n * llm.predict({ adjective: \"funny\" })\n * ```\n */\n async predict(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n callbackManager?: CallbackManager\n ): Promise<T> {\n const output = await this.call(values, callbackManager);\n return output[this.outputKey];\n }\n\n _chainType() {\n return \"llm\" as const;\n }\n\n static async deserialize(data: SerializedLLMChain): Promise<LLMChain> {\n const { llm, prompt } = data;\n if (!llm) {\n throw new Error(\"LLMChain must have llm\");\n }\n if (!prompt) {\n throw new Error(\"LLMChain must have prompt\");\n }\n\n return new LLMChain({\n llm: await BaseLanguageModel.deserialize(llm),\n prompt: await BasePromptTemplate.deserialize(prompt),\n });\n }\n\n /** @deprecated */\n serialize(): SerializedLLMChain {\n const serialize =\n \"serialize\" in this.llm ? this.llm.serialize() : undefined;\n return {\n _type: `${this._chainType()}_chain`,\n llm: serialize,\n prompt: this.prompt.serialize(),\n };\n }\n\n _getNumTokens(text: string): Promise<number> {\n return _getLanguageModel(this.llm).getNumTokens(text);\n }\n}\n"],"mappings":";;;;;;;AAmDA,SAAS,oBAAoB,SAAgD;AAC3E,QAAO,OAAQ,QAAuC,aAAa;;AAGrE,SAAS,kBAAkB,SAA+C;AACxE,KAAI,oBAAoB,QAAQ,CAC9B,QAAO;UACE,WAAW,WAAWA,0BAAAA,SAAS,WAAW,QAAQ,MAAM,CACjE,QAAO,kBAAkB,QAAQ,MAAM;UAEvC,cAAc,WACd,eAAe,WACfA,0BAAAA,SAAS,WAAW,QAAQ,SAAS,CAErC,QAAO,kBAAkB,QAAQ,SAAS;UACjC,aAAa,WAAWA,0BAAAA,SAAS,WAAW,QAAQ,QAAQ,CACrE,QAAO,kBAAkB,QAAQ,QAAQ;KAEzC,OAAM,IAAI,MAAM,2DAA2D;;;;;;;;;;;;;;;;;AAmB/E,IAAa,WAAb,MAAa,iBAIHC,aAAAA,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,kBAAkB;CAElB;CAEA;CAEA;CAEA,YAAY;CAEZ;CAEA,IAAI,YAAY;AACd,SAAO,KAAK,OAAO;;CAGrB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,YAAY,QAAiC;AAC3C,QAAM,OAAO;AACb,OAAK,SAAS,OAAO;AACrB,OAAK,MAAM,OAAO;AAClB,OAAK,YAAY,OAAO;AACxB,OAAK,YAAY,OAAO,aAAa,KAAK;AAC1C,OAAK,eACH,OAAO,gBAAiB,IAAIC,aAAAA,kBAAkB;AAChD,MAAI,KAAK,OAAO,cAAc;AAC5B,OAAI,OAAO,aACT,OAAM,IAAI,MAAM,uDAAuD;AAEzE,QAAK,eAAe,KAAK,OAAO;;;CAIpC,cAAgC;AAE9B,SADiB,cAAc,KAAK,MAAM,KAAK,IAAI,WAAW,EAAE;;;CAKlE,oBAAoB,QAAkC;EACpD,MAAM,kBAAkB,MAAM,oBAAoB,OAAO;EACzD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO,OACT,QAAO,gBAAgB;AAG3B,SAAO;;;CAIT,MAAM,gBACJ,aACA,aACA,YACkB;EAClB,IAAI;AACJ,MAAI,KAAK,aACP,mBAAkB,MAAM,KAAK,aAAa,sBACxC,aACA,aACA,YAAY,UAAU,CACvB;MAED,mBAAkB,YAAY,GAAG;AAEnC,SAAO;;;;;;;CAQT,KACE,QACA,QACsB;AACtB,SAAO,MAAM,KAAK,QAAQ,OAAO;;;CAInC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,kBAAkB,EAAE,GAAG,QAAQ;EACrC,MAAM,eAAe,EACnB,GAAG,KAAK,WACT;EACD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO;OACL,cAAc;AAChB,iBAAa,OACX,OAAO;AACT,WAAO,gBAAgB;;;EAI7B,MAAM,cAAc,MAAM,KAAK,OAAO,kBAAkB,gBAAgB;AACxE,MAAI,oBAAoB,KAAK,KAAK;GAChC,MAAM,EAAE,gBAAgB,MAAM,KAAK,IAAI,eACrC,CAAC,YAAY,EACb,cACA,YAAY,UAAU,CACvB;AACD,UAAO,GACJ,KAAK,YAAY,MAAM,KAAK,gBAC3B,YAAY,IACZ,aACA,WACD,EACF;;EAMH,MAAM,WAAW,OAHO,KAAK,eACzB,KAAK,IAAI,KAAK,KAAK,aAAa,GAChC,KAAK,KAC8B,OACrC,aACA,YAAY,UAAU,CACvB;AACD,SAAO,GACJ,KAAK,YAAY,UACnB;;;;;;;;;;;;;;CAeH,MAAM,QACJ,QACA,iBACY;AAEZ,UADe,MAAM,KAAK,KAAK,QAAQ,gBAAgB,EACzC,KAAK;;CAGrB,aAAa;AACX,SAAO;;CAGT,aAAa,YAAY,MAA6C;EACpE,MAAM,EAAE,KAAK,WAAW;AACxB,MAAI,CAAC,IACH,OAAM,IAAI,MAAM,yBAAyB;AAE3C,MAAI,CAAC,OACH,OAAM,IAAI,MAAM,4BAA4B;AAG9C,SAAO,IAAI,SAAS;GAClB,KAAK,MAAMC,qCAAAA,kBAAkB,YAAY,IAAI;GAC7C,QAAQ,MAAMC,wBAAAA,mBAAmB,YAAY,OAAO;GACrD,CAAC;;;CAIJ,YAAgC;EAC9B,MAAM,YACJ,eAAe,KAAK,MAAM,KAAK,IAAI,WAAW,GAAG,KAAA;AACnD,SAAO;GACL,OAAO,GAAG,KAAK,YAAY,CAAC;GAC5B,KAAK;GACL,QAAQ,KAAK,OAAO,WAAW;GAChC;;CAGH,cAAc,MAA+B;AAC3C,SAAO,kBAAkB,KAAK,IAAI,CAAC,aAAa,KAAK"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"llm_chain.js","names":[],"sources":["../../src/chains/llm_chain.ts"],"sourcesContent":["import {\n BaseLanguageModel,\n BaseLanguageModelInterface,\n BaseLanguageModelInput,\n} from \"@langchain/core/language_models/base\";\nimport type { ChainValues } from \"@langchain/core/utils/types\";\nimport type { Generation } from \"@langchain/core/outputs\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport type { BasePromptValueInterface } from \"@langchain/core/prompt_values\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n BaseLLMOutputParser,\n BaseOutputParser,\n} from \"@langchain/core/output_parsers\";\nimport {\n CallbackManager,\n BaseCallbackConfig,\n CallbackManagerForChainRun,\n Callbacks,\n} from \"@langchain/core/callbacks/manager\";\nimport { Runnable, type RunnableInterface } from \"@langchain/core/runnables\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { SerializedLLMChain } from \"./serde.js\";\nimport { NoOpOutputParser } from \"../output_parsers/noop.js\";\n\ntype LLMType =\n | BaseLanguageModelInterface\n | Runnable<BaseLanguageModelInput, string>\n | Runnable<BaseLanguageModelInput, BaseMessage>;\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype CallOptionsIfAvailable<T> = T extends { CallOptions: infer CO } ? CO : any;\n/**\n * Interface for the input parameters of the LLMChain class.\n */\nexport interface LLMChainInput<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n> extends ChainInputs {\n /** Prompt object to use */\n prompt: BasePromptTemplate;\n /** LLM Wrapper to use */\n llm: Model;\n /** Kwargs to pass to LLM */\n llmKwargs?: CallOptionsIfAvailable<Model>;\n /** OutputParser to use */\n outputParser?: BaseLLMOutputParser<T>;\n /** Key to use for output, defaults to `text` */\n outputKey?: string;\n}\n\nfunction isBaseLanguageModel(llmLike: unknown): llmLike is BaseLanguageModel {\n return typeof (llmLike as BaseLanguageModelInterface)._llmType === \"function\";\n}\n\nfunction _getLanguageModel(llmLike: RunnableInterface): BaseLanguageModel {\n if (isBaseLanguageModel(llmLike)) {\n return llmLike;\n } else if (\"bound\" in llmLike && Runnable.isRunnable(llmLike.bound)) {\n return _getLanguageModel(llmLike.bound);\n } else if (\n \"runnable\" in llmLike &&\n \"fallbacks\" in llmLike &&\n Runnable.isRunnable(llmLike.runnable)\n ) {\n return _getLanguageModel(llmLike.runnable);\n } else if (\"default\" in llmLike && Runnable.isRunnable(llmLike.default)) {\n return _getLanguageModel(llmLike.default);\n } else {\n throw new Error(\"Unable to extract BaseLanguageModel from llmLike object.\");\n }\n}\n\n/**\n * Chain to run queries against LLMs.\n *\n * @example\n * ```ts\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n *\n * const prompt = ChatPromptTemplate.fromTemplate(\"Tell me a {adjective} joke\");\n * const llm = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const chain = prompt.pipe(llm);\n *\n * const response = await chain.invoke({ adjective: \"funny\" });\n * ```\n */\nexport class LLMChain<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n>\n extends BaseChain\n implements LLMChainInput<T>\n{\n static lc_name() {\n return \"LLMChain\";\n }\n\n lc_serializable = true;\n\n prompt: BasePromptTemplate;\n\n llm: Model;\n\n llmKwargs?: CallOptionsIfAvailable<Model>;\n\n outputKey = \"text\";\n\n outputParser?: BaseLLMOutputParser<T>;\n\n get inputKeys() {\n return this.prompt.inputVariables;\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n constructor(fields: LLMChainInput<T, Model>) {\n super(fields);\n this.prompt = fields.prompt;\n this.llm = fields.llm;\n this.llmKwargs = fields.llmKwargs;\n this.outputKey = fields.outputKey ?? this.outputKey;\n this.outputParser =\n fields.outputParser ?? (new NoOpOutputParser() as BaseOutputParser<T>);\n if (this.prompt.outputParser) {\n if (fields.outputParser) {\n throw new Error(\"Cannot set both outputParser and prompt.outputParser\");\n }\n this.outputParser = this.prompt.outputParser as BaseOutputParser<T>;\n }\n }\n\n private getCallKeys(): string[] {\n const callKeys = \"callKeys\" in this.llm ? this.llm.callKeys : [];\n return callKeys;\n }\n\n /** @ignore */\n _selectMemoryInputs(values: ChainValues): ChainValues {\n const valuesForMemory = super._selectMemoryInputs(values);\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n delete valuesForMemory[key];\n }\n }\n return valuesForMemory;\n }\n\n /** @ignore */\n async _getFinalOutput(\n generations: Generation[],\n promptValue: BasePromptValueInterface,\n runManager?: CallbackManagerForChainRun\n ): Promise<unknown> {\n let finalCompletion: unknown;\n if (this.outputParser) {\n finalCompletion = await this.outputParser.parseResultWithPrompt(\n generations,\n promptValue,\n runManager?.getChild()\n );\n } else {\n finalCompletion = generations[0].text;\n }\n return finalCompletion;\n }\n\n /**\n * Run the core logic of this chain and add to output if desired.\n *\n * Wraps _call and handles memory.\n */\n call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n config?: Callbacks | BaseCallbackConfig\n ): Promise<ChainValues> {\n return super.call(values, config);\n }\n\n /** @ignore */\n async _call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const valuesForPrompt = { ...values };\n const valuesForLLM = {\n ...this.llmKwargs,\n } as CallOptionsIfAvailable<Model>;\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n if (valuesForLLM) {\n valuesForLLM[key as keyof CallOptionsIfAvailable<Model>] =\n values[key];\n delete valuesForPrompt[key];\n }\n }\n }\n const promptValue = await this.prompt.formatPromptValue(valuesForPrompt);\n if (\"generatePrompt\" in this.llm) {\n const { generations } = await this.llm.generatePrompt(\n [promptValue],\n valuesForLLM,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: await this._getFinalOutput(\n generations[0],\n promptValue,\n runManager\n ),\n };\n }\n\n const modelWithParser = this.outputParser\n ? this.llm.pipe(this.outputParser)\n : this.llm;\n const response = await modelWithParser.invoke(\n promptValue,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: response,\n };\n }\n\n /**\n * Format prompt with values and pass to LLM\n *\n * @param values - keys to pass to prompt template\n * @param callbackManager - CallbackManager to use\n * @returns Completion from LLM.\n *\n * @example\n * ```ts\n * llm.predict({ adjective: \"funny\" })\n * ```\n */\n async predict(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n callbackManager?: CallbackManager\n ): Promise<T> {\n const output = await this.call(values, callbackManager);\n return output[this.outputKey];\n }\n\n _chainType() {\n return \"llm\" as const;\n }\n\n static async deserialize(data: SerializedLLMChain): Promise<LLMChain> {\n const { llm, prompt } = data;\n if (!llm) {\n throw new Error(\"LLMChain must have llm\");\n }\n if (!prompt) {\n throw new Error(\"LLMChain must have prompt\");\n }\n\n return new LLMChain({\n llm: await BaseLanguageModel.deserialize(llm),\n prompt: await BasePromptTemplate.deserialize(prompt),\n });\n }\n\n /** @deprecated */\n serialize(): SerializedLLMChain {\n const serialize =\n \"serialize\" in this.llm ? this.llm.serialize() : undefined;\n return {\n _type: `${this._chainType()}_chain`,\n llm: serialize,\n prompt: this.prompt.serialize(),\n };\n }\n\n _getNumTokens(text: string): Promise<number> {\n return _getLanguageModel(this.llm).getNumTokens(text);\n }\n}\n"],"mappings":";;;;;;AAmDA,SAAS,oBAAoB,SAAgD;AAC3E,QAAO,OAAQ,QAAuC,aAAa;;AAGrE,SAAS,kBAAkB,SAA+C;AACxE,KAAI,oBAAoB,QAAQ,CAC9B,QAAO;UACE,WAAW,WAAW,SAAS,WAAW,QAAQ,MAAM,CACjE,QAAO,kBAAkB,QAAQ,MAAM;UAEvC,cAAc,WACd,eAAe,WACf,SAAS,WAAW,QAAQ,SAAS,CAErC,QAAO,kBAAkB,QAAQ,SAAS;UACjC,aAAa,WAAW,SAAS,WAAW,QAAQ,QAAQ,CACrE,QAAO,kBAAkB,QAAQ,QAAQ;KAEzC,OAAM,IAAI,MAAM,2DAA2D;;;;;;;;;;;;;;;;;AAmB/E,IAAa,WAAb,MAAa,iBAIH,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,kBAAkB;CAElB;CAEA;CAEA;CAEA,YAAY;CAEZ;CAEA,IAAI,YAAY;AACd,SAAO,KAAK,OAAO;;CAGrB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,YAAY,QAAiC;AAC3C,QAAM,OAAO;AACb,OAAK,SAAS,OAAO;AACrB,OAAK,MAAM,OAAO;AAClB,OAAK,YAAY,OAAO;AACxB,OAAK,YAAY,OAAO,aAAa,KAAK;AAC1C,OAAK,eACH,OAAO,gBAAiB,IAAI,kBAAkB;AAChD,MAAI,KAAK,OAAO,cAAc;AAC5B,OAAI,OAAO,aACT,OAAM,IAAI,MAAM,uDAAuD;AAEzE,QAAK,eAAe,KAAK,OAAO;;;CAIpC,cAAgC;AAE9B,SADiB,cAAc,KAAK,MAAM,KAAK,IAAI,WAAW,EAAE;;;CAKlE,oBAAoB,QAAkC;EACpD,MAAM,kBAAkB,MAAM,oBAAoB,OAAO;EACzD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO,OACT,QAAO,gBAAgB;AAG3B,SAAO;;;CAIT,MAAM,gBACJ,aACA,aACA,YACkB;EAClB,IAAI;AACJ,MAAI,KAAK,aACP,mBAAkB,MAAM,KAAK,aAAa,sBACxC,aACA,aACA,YAAY,UAAU,CACvB;MAED,mBAAkB,YAAY,GAAG;AAEnC,SAAO;;;;;;;CAQT,KACE,QACA,QACsB;AACtB,SAAO,MAAM,KAAK,QAAQ,OAAO;;;CAInC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,kBAAkB,EAAE,GAAG,QAAQ;EACrC,MAAM,eAAe,EACnB,GAAG,KAAK,WACT;EACD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO;OACL,cAAc;AAChB,iBAAa,OACX,OAAO;AACT,WAAO,gBAAgB;;;EAI7B,MAAM,cAAc,MAAM,KAAK,OAAO,kBAAkB,gBAAgB;AACxE,MAAI,oBAAoB,KAAK,KAAK;GAChC,MAAM,EAAE,gBAAgB,MAAM,KAAK,IAAI,eACrC,CAAC,YAAY,EACb,cACA,YAAY,UAAU,CACvB;AACD,UAAO,GACJ,KAAK,YAAY,MAAM,KAAK,gBAC3B,YAAY,IACZ,aACA,WACD,EACF;;EAMH,MAAM,WAAW,OAHO,KAAK,eACzB,KAAK,IAAI,KAAK,KAAK,aAAa,GAChC,KAAK,KAC8B,OACrC,aACA,YAAY,UAAU,CACvB;AACD,SAAO,GACJ,KAAK,YAAY,UACnB;;;;;;;;;;;;;;CAeH,MAAM,QACJ,QACA,iBACY;AAEZ,UADe,MAAM,KAAK,KAAK,QAAQ,gBAAgB,EACzC,KAAK;;CAGrB,aAAa;AACX,SAAO;;CAGT,aAAa,YAAY,MAA6C;EACpE,MAAM,EAAE,KAAK,WAAW;AACxB,MAAI,CAAC,IACH,OAAM,IAAI,MAAM,yBAAyB;AAE3C,MAAI,CAAC,OACH,OAAM,IAAI,MAAM,4BAA4B;AAG9C,SAAO,IAAI,SAAS;GAClB,KAAK,MAAM,kBAAkB,YAAY,IAAI;GAC7C,QAAQ,MAAM,mBAAmB,YAAY,OAAO;GACrD,CAAC;;;CAIJ,YAAgC;EAC9B,MAAM,YACJ,eAAe,KAAK,MAAM,KAAK,IAAI,WAAW,GAAG,KAAA;AACnD,SAAO;GACL,OAAO,GAAG,KAAK,YAAY,CAAC;GAC5B,KAAK;GACL,QAAQ,KAAK,OAAO,WAAW;GAChC;;CAGH,cAAc,MAA+B;AAC3C,SAAO,kBAAkB,KAAK,IAAI,CAAC,aAAa,KAAK"}
|
|
1
|
+
{"version":3,"file":"llm_chain.js","names":[],"sources":["../../src/chains/llm_chain.ts"],"sourcesContent":["import {\n BaseLanguageModel,\n BaseLanguageModelInterface,\n BaseLanguageModelInput,\n} from \"@langchain/core/language_models/base\";\nimport type { ChainValues } from \"@langchain/core/utils/types\";\nimport type { Generation } from \"@langchain/core/outputs\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport type { BasePromptValueInterface } from \"@langchain/core/prompt_values\";\nimport { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n BaseLLMOutputParser,\n BaseOutputParser,\n} from \"@langchain/core/output_parsers\";\nimport {\n CallbackManager,\n BaseCallbackConfig,\n CallbackManagerForChainRun,\n Callbacks,\n} from \"@langchain/core/callbacks/manager\";\nimport { Runnable, type RunnableInterface } from \"@langchain/core/runnables\";\nimport { BaseChain, ChainInputs } from \"./base.js\";\nimport { SerializedLLMChain } from \"./serde.js\";\nimport { NoOpOutputParser } from \"../output_parsers/noop.js\";\n\ntype LLMType =\n | BaseLanguageModelInterface\n | Runnable<BaseLanguageModelInput, string>\n | Runnable<BaseLanguageModelInput, BaseMessage>;\n\n// oxlint-disable-next-line @typescript-eslint/no-explicit-any\ntype CallOptionsIfAvailable<T> = T extends { CallOptions: infer CO } ? CO : any;\n/**\n * Interface for the input parameters of the LLMChain class.\n */\nexport interface LLMChainInput<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n> extends ChainInputs {\n /** Prompt object to use */\n prompt: BasePromptTemplate;\n /** LLM Wrapper to use */\n llm: Model;\n /** Kwargs to pass to LLM */\n llmKwargs?: CallOptionsIfAvailable<Model>;\n /** OutputParser to use */\n outputParser?: BaseLLMOutputParser<T>;\n /** Key to use for output, defaults to `text` */\n outputKey?: string;\n}\n\nfunction isBaseLanguageModel(llmLike: unknown): llmLike is BaseLanguageModel {\n return typeof (llmLike as BaseLanguageModelInterface)._llmType === \"function\";\n}\n\nfunction _getLanguageModel(llmLike: RunnableInterface): BaseLanguageModel {\n if (isBaseLanguageModel(llmLike)) {\n return llmLike;\n } else if (\"bound\" in llmLike && Runnable.isRunnable(llmLike.bound)) {\n return _getLanguageModel(llmLike.bound);\n } else if (\n \"runnable\" in llmLike &&\n \"fallbacks\" in llmLike &&\n Runnable.isRunnable(llmLike.runnable)\n ) {\n return _getLanguageModel(llmLike.runnable);\n } else if (\"default\" in llmLike && Runnable.isRunnable(llmLike.default)) {\n return _getLanguageModel(llmLike.default);\n } else {\n throw new Error(\"Unable to extract BaseLanguageModel from llmLike object.\");\n }\n}\n\n/**\n * Chain to run queries against LLMs.\n *\n * @example\n * ```ts\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n *\n * const prompt = ChatPromptTemplate.fromTemplate(\"Tell me a {adjective} joke\");\n * const llm = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const chain = prompt.pipe(llm);\n *\n * const response = await chain.invoke({ adjective: \"funny\" });\n * ```\n */\nexport class LLMChain<\n T extends string | object = string,\n Model extends LLMType = LLMType,\n>\n extends BaseChain\n implements LLMChainInput<T>\n{\n static lc_name() {\n return \"LLMChain\";\n }\n\n lc_serializable = true;\n\n prompt: BasePromptTemplate;\n\n llm: Model;\n\n llmKwargs?: CallOptionsIfAvailable<Model>;\n\n outputKey = \"text\";\n\n outputParser?: BaseLLMOutputParser<T>;\n\n get inputKeys() {\n return this.prompt.inputVariables;\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n constructor(fields: LLMChainInput<T, Model>) {\n super(fields);\n this.prompt = fields.prompt;\n this.llm = fields.llm;\n this.llmKwargs = fields.llmKwargs;\n this.outputKey = fields.outputKey ?? this.outputKey;\n this.outputParser =\n fields.outputParser ?? (new NoOpOutputParser() as BaseOutputParser<T>);\n if (this.prompt.outputParser) {\n if (fields.outputParser) {\n throw new Error(\"Cannot set both outputParser and prompt.outputParser\");\n }\n this.outputParser = this.prompt.outputParser as BaseOutputParser<T>;\n }\n }\n\n private getCallKeys(): string[] {\n const callKeys = \"callKeys\" in this.llm ? this.llm.callKeys : [];\n return callKeys;\n }\n\n /** @ignore */\n _selectMemoryInputs(values: ChainValues): ChainValues {\n const valuesForMemory = super._selectMemoryInputs(values);\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n delete valuesForMemory[key];\n }\n }\n return valuesForMemory;\n }\n\n /** @ignore */\n async _getFinalOutput(\n generations: Generation[],\n promptValue: BasePromptValueInterface,\n runManager?: CallbackManagerForChainRun\n ): Promise<unknown> {\n let finalCompletion: unknown;\n if (this.outputParser) {\n finalCompletion = await this.outputParser.parseResultWithPrompt(\n generations,\n promptValue,\n runManager?.getChild()\n );\n } else {\n finalCompletion = generations[0].text;\n }\n return finalCompletion;\n }\n\n /**\n * Run the core logic of this chain and add to output if desired.\n *\n * Wraps _call and handles memory.\n */\n call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n config?: Callbacks | BaseCallbackConfig\n ): Promise<ChainValues> {\n return super.call(values, config);\n }\n\n /** @ignore */\n async _call(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const valuesForPrompt = { ...values };\n const valuesForLLM = {\n ...this.llmKwargs,\n } as CallOptionsIfAvailable<Model>;\n const callKeys = this.getCallKeys();\n for (const key of callKeys) {\n if (key in values) {\n if (valuesForLLM) {\n valuesForLLM[key as keyof CallOptionsIfAvailable<Model>] =\n values[key];\n delete valuesForPrompt[key];\n }\n }\n }\n const promptValue = await this.prompt.formatPromptValue(valuesForPrompt);\n if (\"generatePrompt\" in this.llm) {\n const { generations } = await this.llm.generatePrompt(\n [promptValue],\n valuesForLLM,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: await this._getFinalOutput(\n generations[0],\n promptValue,\n runManager\n ),\n };\n }\n\n const modelWithParser = this.outputParser\n ? this.llm.pipe(this.outputParser)\n : this.llm;\n const response = await modelWithParser.invoke(\n promptValue,\n runManager?.getChild()\n );\n return {\n [this.outputKey]: response,\n };\n }\n\n /**\n * Format prompt with values and pass to LLM\n *\n * @param values - keys to pass to prompt template\n * @param callbackManager - CallbackManager to use\n * @returns Completion from LLM.\n *\n * @example\n * ```ts\n * llm.predict({ adjective: \"funny\" })\n * ```\n */\n async predict(\n values: ChainValues & CallOptionsIfAvailable<Model>,\n callbackManager?: CallbackManager\n ): Promise<T> {\n const output = await this.call(values, callbackManager);\n return output[this.outputKey];\n }\n\n _chainType() {\n return \"llm\" as const;\n }\n\n static async deserialize(data: SerializedLLMChain): Promise<LLMChain> {\n const { llm, prompt } = data;\n if (!llm) {\n throw new Error(\"LLMChain must have llm\");\n }\n if (!prompt) {\n throw new Error(\"LLMChain must have prompt\");\n }\n\n return new LLMChain({\n llm: await BaseLanguageModel.deserialize(llm),\n prompt: await BasePromptTemplate.deserialize(prompt),\n });\n }\n\n /** @deprecated */\n serialize(): SerializedLLMChain {\n const serialize =\n \"serialize\" in this.llm ? this.llm.serialize() : undefined;\n return {\n _type: `${this._chainType()}_chain`,\n llm: serialize,\n prompt: this.prompt.serialize(),\n };\n }\n\n _getNumTokens(text: string): Promise<number> {\n return _getLanguageModel(this.llm).getNumTokens(text);\n }\n}\n"],"mappings":";;;;;;AAmDA,SAAS,oBAAoB,SAAgD;AAC3E,QAAO,OAAQ,QAAuC,aAAa;;AAGrE,SAAS,kBAAkB,SAA+C;AACxE,KAAI,oBAAoB,QAAQ,CAC9B,QAAO;UACE,WAAW,WAAW,SAAS,WAAW,QAAQ,MAAM,CACjE,QAAO,kBAAkB,QAAQ,MAAM;UAEvC,cAAc,WACd,eAAe,WACf,SAAS,WAAW,QAAQ,SAAS,CAErC,QAAO,kBAAkB,QAAQ,SAAS;UACjC,aAAa,WAAW,SAAS,WAAW,QAAQ,QAAQ,CACrE,QAAO,kBAAkB,QAAQ,QAAQ;KAEzC,OAAM,IAAI,MAAM,2DAA2D;;;;;;;;;;;;;;;;;AAmB/E,IAAa,WAAb,MAAa,iBAIH,UAEV;CACE,OAAO,UAAU;AACf,SAAO;;CAGT,kBAAkB;CAElB;CAEA;CAEA;CAEA,YAAY;CAEZ;CAEA,IAAI,YAAY;AACd,SAAO,KAAK,OAAO;;CAGrB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,YAAY,QAAiC;AAC3C,QAAM,OAAO;AACb,OAAK,SAAS,OAAO;AACrB,OAAK,MAAM,OAAO;AAClB,OAAK,YAAY,OAAO;AACxB,OAAK,YAAY,OAAO,aAAa,KAAK;AAC1C,OAAK,eACH,OAAO,gBAAiB,IAAI,kBAAkB;AAChD,MAAI,KAAK,OAAO,cAAc;AAC5B,OAAI,OAAO,aACT,OAAM,IAAI,MAAM,uDAAuD;AAEzE,QAAK,eAAe,KAAK,OAAO;;;CAIpC,cAAgC;AAE9B,SADiB,cAAc,KAAK,MAAM,KAAK,IAAI,WAAW,EAAE;;;CAKlE,oBAAoB,QAAkC;EACpD,MAAM,kBAAkB,MAAM,oBAAoB,OAAO;EACzD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO,OACT,QAAO,gBAAgB;AAG3B,SAAO;;;CAIT,MAAM,gBACJ,aACA,aACA,YACkB;EAClB,IAAI;AACJ,MAAI,KAAK,aACP,mBAAkB,MAAM,KAAK,aAAa,sBACxC,aACA,aACA,YAAY,UAAU,CACvB;MAED,mBAAkB,YAAY,GAAG;AAEnC,SAAO;;;;;;;CAQT,KACE,QACA,QACsB;AACtB,SAAO,MAAM,KAAK,QAAQ,OAAO;;;CAInC,MAAM,MACJ,QACA,YACsB;EACtB,MAAM,kBAAkB,EAAE,GAAG,QAAQ;EACrC,MAAM,eAAe,EACnB,GAAG,KAAK,WACT;EACD,MAAM,WAAW,KAAK,aAAa;AACnC,OAAK,MAAM,OAAO,SAChB,KAAI,OAAO;OACL,cAAc;AAChB,iBAAa,OACX,OAAO;AACT,WAAO,gBAAgB;;;EAI7B,MAAM,cAAc,MAAM,KAAK,OAAO,kBAAkB,gBAAgB;AACxE,MAAI,oBAAoB,KAAK,KAAK;GAChC,MAAM,EAAE,gBAAgB,MAAM,KAAK,IAAI,eACrC,CAAC,YAAY,EACb,cACA,YAAY,UAAU,CACvB;AACD,UAAO,GACJ,KAAK,YAAY,MAAM,KAAK,gBAC3B,YAAY,IACZ,aACA,WACD,EACF;;EAMH,MAAM,WAAW,OAHO,KAAK,eACzB,KAAK,IAAI,KAAK,KAAK,aAAa,GAChC,KAAK,KAC8B,OACrC,aACA,YAAY,UAAU,CACvB;AACD,SAAO,GACJ,KAAK,YAAY,UACnB;;;;;;;;;;;;;;CAeH,MAAM,QACJ,QACA,iBACY;AAEZ,UADe,MAAM,KAAK,KAAK,QAAQ,gBAAgB,EACzC,KAAK;;CAGrB,aAAa;AACX,SAAO;;CAGT,aAAa,YAAY,MAA6C;EACpE,MAAM,EAAE,KAAK,WAAW;AACxB,MAAI,CAAC,IACH,OAAM,IAAI,MAAM,yBAAyB;AAE3C,MAAI,CAAC,OACH,OAAM,IAAI,MAAM,4BAA4B;AAG9C,SAAO,IAAI,SAAS;GAClB,KAAK,MAAM,kBAAkB,YAAY,IAAI;GAC7C,QAAQ,MAAM,mBAAmB,YAAY,OAAO;GACrD,CAAC;;;CAIJ,YAAgC;EAC9B,MAAM,YACJ,eAAe,KAAK,MAAM,KAAK,IAAI,WAAW,GAAG,KAAA;AACnD,SAAO;GACL,OAAO,GAAG,KAAK,YAAY,CAAC;GAC5B,KAAK;GACL,QAAQ,KAAK,OAAO,WAAW;GAChC;;CAGH,cAAc,MAA+B;AAC3C,SAAO,kBAAkB,KAAK,IAAI,CAAC,aAAa,KAAK"}
|
package/dist/chains/load.d.cts
CHANGED
|
@@ -1,6 +1,5 @@
|
|
|
1
1
|
import { BaseChain } from "./base.cjs";
|
|
2
2
|
import { LoadValues } from "../util/load.cjs";
|
|
3
|
-
import * as _langchain_core_utils_types0 from "@langchain/core/utils/types";
|
|
4
3
|
|
|
5
4
|
//#region src/chains/load.d.ts
|
|
6
5
|
/**
|
|
@@ -21,7 +20,7 @@ import * as _langchain_core_utils_types0 from "@langchain/core/utils/types";
|
|
|
21
20
|
* const chain = await loadChain("/path/to/chain.json");
|
|
22
21
|
* ```
|
|
23
22
|
*/
|
|
24
|
-
declare const loadChain: (uri: string, values?: LoadValues) => Promise<BaseChain
|
|
23
|
+
declare const loadChain: (uri: string, values?: LoadValues) => Promise<BaseChain>;
|
|
25
24
|
//#endregion
|
|
26
25
|
export { loadChain };
|
|
27
26
|
//# sourceMappingURL=load.d.cts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"load.d.cts","names":[],"sources":["../../src/chains/load.ts"],"mappings":"
|
|
1
|
+
{"version":3,"file":"load.d.cts","names":[],"sources":["../../src/chains/load.ts"],"mappings":";;;;;;AAgCA;;;;;;;;;;;;;;;;cAAa,SAAA,GAAS,GAAA,UACT,MAAA,GACH,UAAA,KACP,OAAA,CAAQ,SAAA"}
|
package/dist/chains/load.d.ts
CHANGED
|
@@ -1,6 +1,5 @@
|
|
|
1
1
|
import { BaseChain } from "./base.js";
|
|
2
2
|
import { LoadValues } from "../util/load.js";
|
|
3
|
-
import * as _langchain_core_utils_types0 from "@langchain/core/utils/types";
|
|
4
3
|
|
|
5
4
|
//#region src/chains/load.d.ts
|
|
6
5
|
/**
|
|
@@ -21,7 +20,7 @@ import * as _langchain_core_utils_types0 from "@langchain/core/utils/types";
|
|
|
21
20
|
* const chain = await loadChain("/path/to/chain.json");
|
|
22
21
|
* ```
|
|
23
22
|
*/
|
|
24
|
-
declare const loadChain: (uri: string, values?: LoadValues) => Promise<BaseChain
|
|
23
|
+
declare const loadChain: (uri: string, values?: LoadValues) => Promise<BaseChain>;
|
|
25
24
|
//#endregion
|
|
26
25
|
export { loadChain };
|
|
27
26
|
//# sourceMappingURL=load.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"load.d.ts","names":[],"sources":["../../src/chains/load.ts"],"mappings":"
|
|
1
|
+
{"version":3,"file":"load.d.ts","names":[],"sources":["../../src/chains/load.ts"],"mappings":";;;;;;AAgCA;;;;;;;;;;;;;;;;cAAa,SAAA,GAAS,GAAA,UACT,MAAA,GACH,UAAA,KACP,OAAA,CAAQ,SAAA"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"base.cjs","names":["JsonOutputFunctionsParser"],"sources":["../../../src/chains/openai_functions/base.ts"],"sourcesContent":["import type { BaseOutputParser } from \"@langchain/core/output_parsers\";\nimport type { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport type { Runnable, RunnableInterface } from \"@langchain/core/runnables\";\nimport type {\n BaseFunctionCallOptions,\n BaseLanguageModelInput,\n FunctionDefinition,\n} from \"@langchain/core/language_models/base\";\nimport {\n isInteropZodSchema,\n type InputValues,\n InteropZodObject,\n} from \"@langchain/core/utils/types\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Configuration params for the createOpenAIFnRunnable method.\n */\nexport type CreateOpenAIFnRunnableConfig<\n //
|
|
1
|
+
{"version":3,"file":"base.cjs","names":["JsonOutputFunctionsParser"],"sources":["../../../src/chains/openai_functions/base.ts"],"sourcesContent":["import type { BaseOutputParser } from \"@langchain/core/output_parsers\";\nimport type { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport type { Runnable, RunnableInterface } from \"@langchain/core/runnables\";\nimport type {\n BaseFunctionCallOptions,\n BaseLanguageModelInput,\n FunctionDefinition,\n} from \"@langchain/core/language_models/base\";\nimport {\n isInteropZodSchema,\n type InputValues,\n InteropZodObject,\n} from \"@langchain/core/utils/types\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Configuration params for the createOpenAIFnRunnable method.\n */\nexport type CreateOpenAIFnRunnableConfig<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any>,\n RunOutput,\n> = {\n functions: FunctionDefinition[];\n /** Language model to use, assumed to support the OpenAI function-calling API. */\n llm: RunnableInterface<\n BaseLanguageModelInput,\n BaseMessage,\n BaseFunctionCallOptions\n >;\n /** BasePromptTemplate to pass to the model. */\n prompt: BasePromptTemplate<InputValues<Extract<keyof RunInput, string>>>;\n /**\n * Only used if a single function is passed in. If `true`, then the model will be\n * forced to use the given function. If `false`, then the model will be given the\n * option to use the given function or not.\n */\n enforceSingleFunctionUsage?: boolean;\n /**\n * BaseLLMOutputParser to use for parsing model outputs.\n * By default will be inferred from the function types.\n */\n outputParser?: BaseOutputParser<RunOutput>;\n};\n\n/**\n * Creates a runnable sequence that calls OpenAI functions.\n * @param config - The parameters required to create the runnable.\n * @returns A runnable sequence that will pass the given functions to the model when run.\n *\n * @example\n * ```typescript\n * const openAIFunction = {\n * name: \"get_person_details\",\n * description: \"Get details about a person\",\n * parameters: {\n * title: \"Person\",\n * description: \"Identifying information about a person.\",\n * type: \"object\",\n * properties: {\n * name: { title: \"Name\", description: \"The person's name\", type: \"string\" },\n * age: { title: \"Age\", description: \"The person's age\", type: \"integer\" },\n * fav_food: {\n * title: \"Fav Food\",\n * description: \"The person's favorite food\",\n * type: \"string\",\n * },\n * },\n * required: [\"name\", \"age\"],\n * },\n * };\n *\n * const model = new ChatOpenAI();\n * const prompt = ChatPromptTemplate.fromMessages([\n * [\"human\", \"Human description: {description}\"],\n * ]);\n * const outputParser = new JsonOutputFunctionsParser();\n *\n * const runnable = createOpenAIFnRunnable({\n * functions: [openAIFunction],\n * llm: model,\n * prompt,\n * enforceSingleFunctionUsage: true, // Default is true\n * outputParser\n * });\n * const response = await runnable.invoke({\n * description:\n * \"My name's John Doe and I'm 30 years old. My favorite kind of food are chocolate chip cookies.\",\n * });\n *\n * console.log(response);\n *\n * // { name: 'John Doe', age: 30, fav_food: 'chocolate chip cookies' }\n * ```\n */\nexport function createOpenAIFnRunnable<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any> = Record<string, any>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>,\n>(\n config: CreateOpenAIFnRunnableConfig<RunInput, RunOutput>\n): Runnable<RunInput, RunOutput> {\n const {\n functions,\n llm,\n prompt,\n enforceSingleFunctionUsage = true,\n outputParser = new JsonOutputFunctionsParser<RunOutput>(),\n } = config;\n const llmKwargs: BaseFunctionCallOptions = {\n functions,\n };\n\n if (functions.length === 1 && enforceSingleFunctionUsage) {\n llmKwargs.function_call = {\n name: functions[0].name,\n };\n }\n\n const llmWithKwargs = (llm as Runnable).withConfig(llmKwargs);\n return prompt.pipe(llmWithKwargs).pipe(outputParser);\n}\n\n/**\n * Configuration params for the createStructuredOutputRunnable method.\n */\nexport type CreateStructuredOutputRunnableConfig<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any>,\n RunOutput,\n> = {\n /**\n * Schema to output. Must be either valid JSONSchema or a Zod schema.\n */\n outputSchema: InteropZodObject | JsonSchema7Type;\n /**\n * Language model to use, assumed to support the OpenAI function-calling API.\n */\n llm: RunnableInterface<\n BaseLanguageModelInput,\n BaseMessage,\n BaseFunctionCallOptions\n >;\n /** BasePromptTemplate to pass to the model. */\n prompt: BasePromptTemplate<InputValues<Extract<keyof RunInput, string>>>;\n /**\n * BaseLLMOutputParser to use for parsing model outputs.\n */\n outputParser?: BaseOutputParser<RunOutput>;\n};\n\n/**\n * @deprecated Prefer the `.withStructuredOutput` method on chat model classes.\n *\n * Create a runnable that uses an OpenAI function to get a structured output.\n * @param config Params required to create the runnable.\n * @returns A runnable sequence that will pass the given function to the model when run.\n *\n * @example\n * ```typescript\n * import { createStructuredOutputRunnable } from \"@langchain/classic/chains/openai_functions\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { JsonOutputFunctionsParser } from \"@langchain/classic/output_parsers\";\n *\n * const jsonSchema = {\n * title: \"Person\",\n * description: \"Identifying information about a person.\",\n * type: \"object\",\n * properties: {\n * name: { title: \"Name\", description: \"The person's name\", type: \"string\" },\n * age: { title: \"Age\", description: \"The person's age\", type: \"integer\" },\n * fav_food: {\n * title: \"Fav Food\",\n * description: \"The person's favorite food\",\n * type: \"string\",\n * },\n * },\n * required: [\"name\", \"age\"],\n * };\n *\n * const model = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const prompt = ChatPromptTemplate.fromMessages([\n * [\"human\", \"Human description: {description}\"],\n * ]);\n *\n * const outputParser = new JsonOutputFunctionsParser();\n *\n * // Also works with Zod schema\n * const runnable = createStructuredOutputRunnable({\n * outputSchema: jsonSchema,\n * llm: model,\n * prompt,\n * outputParser\n * });\n *\n * const response = await runnable.invoke({\n * description:\n * \"My name's John Doe and I'm 30 years old. My favorite kind of food are chocolate chip cookies.\",\n * });\n *\n * console.log(response);\n *\n * // { name: 'John Doe', age: 30, fav_food: 'chocolate chip cookies' }\n * ```\n */\nexport function createStructuredOutputRunnable<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any> = Record<string, any>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>,\n>(\n config: CreateStructuredOutputRunnableConfig<RunInput, RunOutput>\n): Runnable<RunInput, RunOutput> {\n const { outputSchema, llm, prompt, outputParser } = config;\n const jsonSchema = isInteropZodSchema(outputSchema)\n ? toJsonSchema(outputSchema)\n : outputSchema;\n const oaiFunction: FunctionDefinition = {\n name: \"outputFormatter\",\n description:\n \"Output formatter. Should always be used to format your response to the user\",\n parameters: jsonSchema,\n };\n\n return createOpenAIFnRunnable({\n functions: [oaiFunction],\n llm,\n prompt,\n enforceSingleFunctionUsage: true,\n outputParser,\n });\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAoGA,SAAgB,uBAMd,QAC+B;CAC/B,MAAM,EACJ,WACA,KACA,QACA,6BAA6B,MAC7B,eAAe,IAAIA,yBAAAA,2BAAsC,KACvD;CACJ,MAAM,YAAqC,EACzC,WACD;AAED,KAAI,UAAU,WAAW,KAAK,2BAC5B,WAAU,gBAAgB,EACxB,MAAM,UAAU,GAAG,MACpB;CAGH,MAAM,gBAAiB,IAAiB,WAAW,UAAU;AAC7D,QAAO,OAAO,KAAK,cAAc,CAAC,KAAK,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsFtD,SAAgB,+BAMd,QAC+B;CAC/B,MAAM,EAAE,cAAc,KAAK,QAAQ,iBAAiB;AAWpD,QAAO,uBAAuB;EAC5B,WAAW,CAR2B;GACtC,MAAM;GACN,aACE;GACF,aAAA,GAAA,4BAAA,oBAPoC,aAAa,IAAA,GAAA,kCAAA,cAClC,aAAa,GAC1B;GAMH,CAGyB;EACxB;EACA;EACA,4BAA4B;EAC5B;EACD,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"base.js","names":[],"sources":["../../../src/chains/openai_functions/base.ts"],"sourcesContent":["import type { BaseOutputParser } from \"@langchain/core/output_parsers\";\nimport type { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport type { Runnable, RunnableInterface } from \"@langchain/core/runnables\";\nimport type {\n BaseFunctionCallOptions,\n BaseLanguageModelInput,\n FunctionDefinition,\n} from \"@langchain/core/language_models/base\";\nimport {\n isInteropZodSchema,\n type InputValues,\n InteropZodObject,\n} from \"@langchain/core/utils/types\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Configuration params for the createOpenAIFnRunnable method.\n */\nexport type CreateOpenAIFnRunnableConfig<\n //
|
|
1
|
+
{"version":3,"file":"base.js","names":[],"sources":["../../../src/chains/openai_functions/base.ts"],"sourcesContent":["import type { BaseOutputParser } from \"@langchain/core/output_parsers\";\nimport type { BasePromptTemplate } from \"@langchain/core/prompts\";\nimport type { Runnable, RunnableInterface } from \"@langchain/core/runnables\";\nimport type {\n BaseFunctionCallOptions,\n BaseLanguageModelInput,\n FunctionDefinition,\n} from \"@langchain/core/language_models/base\";\nimport {\n isInteropZodSchema,\n type InputValues,\n InteropZodObject,\n} from \"@langchain/core/utils/types\";\nimport type { BaseMessage } from \"@langchain/core/messages\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Configuration params for the createOpenAIFnRunnable method.\n */\nexport type CreateOpenAIFnRunnableConfig<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any>,\n RunOutput,\n> = {\n functions: FunctionDefinition[];\n /** Language model to use, assumed to support the OpenAI function-calling API. */\n llm: RunnableInterface<\n BaseLanguageModelInput,\n BaseMessage,\n BaseFunctionCallOptions\n >;\n /** BasePromptTemplate to pass to the model. */\n prompt: BasePromptTemplate<InputValues<Extract<keyof RunInput, string>>>;\n /**\n * Only used if a single function is passed in. If `true`, then the model will be\n * forced to use the given function. If `false`, then the model will be given the\n * option to use the given function or not.\n */\n enforceSingleFunctionUsage?: boolean;\n /**\n * BaseLLMOutputParser to use for parsing model outputs.\n * By default will be inferred from the function types.\n */\n outputParser?: BaseOutputParser<RunOutput>;\n};\n\n/**\n * Creates a runnable sequence that calls OpenAI functions.\n * @param config - The parameters required to create the runnable.\n * @returns A runnable sequence that will pass the given functions to the model when run.\n *\n * @example\n * ```typescript\n * const openAIFunction = {\n * name: \"get_person_details\",\n * description: \"Get details about a person\",\n * parameters: {\n * title: \"Person\",\n * description: \"Identifying information about a person.\",\n * type: \"object\",\n * properties: {\n * name: { title: \"Name\", description: \"The person's name\", type: \"string\" },\n * age: { title: \"Age\", description: \"The person's age\", type: \"integer\" },\n * fav_food: {\n * title: \"Fav Food\",\n * description: \"The person's favorite food\",\n * type: \"string\",\n * },\n * },\n * required: [\"name\", \"age\"],\n * },\n * };\n *\n * const model = new ChatOpenAI();\n * const prompt = ChatPromptTemplate.fromMessages([\n * [\"human\", \"Human description: {description}\"],\n * ]);\n * const outputParser = new JsonOutputFunctionsParser();\n *\n * const runnable = createOpenAIFnRunnable({\n * functions: [openAIFunction],\n * llm: model,\n * prompt,\n * enforceSingleFunctionUsage: true, // Default is true\n * outputParser\n * });\n * const response = await runnable.invoke({\n * description:\n * \"My name's John Doe and I'm 30 years old. My favorite kind of food are chocolate chip cookies.\",\n * });\n *\n * console.log(response);\n *\n * // { name: 'John Doe', age: 30, fav_food: 'chocolate chip cookies' }\n * ```\n */\nexport function createOpenAIFnRunnable<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any> = Record<string, any>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>,\n>(\n config: CreateOpenAIFnRunnableConfig<RunInput, RunOutput>\n): Runnable<RunInput, RunOutput> {\n const {\n functions,\n llm,\n prompt,\n enforceSingleFunctionUsage = true,\n outputParser = new JsonOutputFunctionsParser<RunOutput>(),\n } = config;\n const llmKwargs: BaseFunctionCallOptions = {\n functions,\n };\n\n if (functions.length === 1 && enforceSingleFunctionUsage) {\n llmKwargs.function_call = {\n name: functions[0].name,\n };\n }\n\n const llmWithKwargs = (llm as Runnable).withConfig(llmKwargs);\n return prompt.pipe(llmWithKwargs).pipe(outputParser);\n}\n\n/**\n * Configuration params for the createStructuredOutputRunnable method.\n */\nexport type CreateStructuredOutputRunnableConfig<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any>,\n RunOutput,\n> = {\n /**\n * Schema to output. Must be either valid JSONSchema or a Zod schema.\n */\n outputSchema: InteropZodObject | JsonSchema7Type;\n /**\n * Language model to use, assumed to support the OpenAI function-calling API.\n */\n llm: RunnableInterface<\n BaseLanguageModelInput,\n BaseMessage,\n BaseFunctionCallOptions\n >;\n /** BasePromptTemplate to pass to the model. */\n prompt: BasePromptTemplate<InputValues<Extract<keyof RunInput, string>>>;\n /**\n * BaseLLMOutputParser to use for parsing model outputs.\n */\n outputParser?: BaseOutputParser<RunOutput>;\n};\n\n/**\n * @deprecated Prefer the `.withStructuredOutput` method on chat model classes.\n *\n * Create a runnable that uses an OpenAI function to get a structured output.\n * @param config Params required to create the runnable.\n * @returns A runnable sequence that will pass the given function to the model when run.\n *\n * @example\n * ```typescript\n * import { createStructuredOutputRunnable } from \"@langchain/classic/chains/openai_functions\";\n * import { ChatOpenAI } from \"@langchain/openai\";\n * import { ChatPromptTemplate } from \"@langchain/core/prompts\";\n * import { JsonOutputFunctionsParser } from \"@langchain/classic/output_parsers\";\n *\n * const jsonSchema = {\n * title: \"Person\",\n * description: \"Identifying information about a person.\",\n * type: \"object\",\n * properties: {\n * name: { title: \"Name\", description: \"The person's name\", type: \"string\" },\n * age: { title: \"Age\", description: \"The person's age\", type: \"integer\" },\n * fav_food: {\n * title: \"Fav Food\",\n * description: \"The person's favorite food\",\n * type: \"string\",\n * },\n * },\n * required: [\"name\", \"age\"],\n * };\n *\n * const model = new ChatOpenAI({ model: \"gpt-4o-mini\" });\n * const prompt = ChatPromptTemplate.fromMessages([\n * [\"human\", \"Human description: {description}\"],\n * ]);\n *\n * const outputParser = new JsonOutputFunctionsParser();\n *\n * // Also works with Zod schema\n * const runnable = createStructuredOutputRunnable({\n * outputSchema: jsonSchema,\n * llm: model,\n * prompt,\n * outputParser\n * });\n *\n * const response = await runnable.invoke({\n * description:\n * \"My name's John Doe and I'm 30 years old. My favorite kind of food are chocolate chip cookies.\",\n * });\n *\n * console.log(response);\n *\n * // { name: 'John Doe', age: 30, fav_food: 'chocolate chip cookies' }\n * ```\n */\nexport function createStructuredOutputRunnable<\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunInput extends Record<string, any> = Record<string, any>,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n RunOutput extends Record<string, any> = Record<string, any>,\n>(\n config: CreateStructuredOutputRunnableConfig<RunInput, RunOutput>\n): Runnable<RunInput, RunOutput> {\n const { outputSchema, llm, prompt, outputParser } = config;\n const jsonSchema = isInteropZodSchema(outputSchema)\n ? toJsonSchema(outputSchema)\n : outputSchema;\n const oaiFunction: FunctionDefinition = {\n name: \"outputFormatter\",\n description:\n \"Output formatter. Should always be used to format your response to the user\",\n parameters: jsonSchema,\n };\n\n return createOpenAIFnRunnable({\n functions: [oaiFunction],\n llm,\n prompt,\n enforceSingleFunctionUsage: true,\n outputParser,\n });\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAoGA,SAAgB,uBAMd,QAC+B;CAC/B,MAAM,EACJ,WACA,KACA,QACA,6BAA6B,MAC7B,eAAe,IAAI,2BAAsC,KACvD;CACJ,MAAM,YAAqC,EACzC,WACD;AAED,KAAI,UAAU,WAAW,KAAK,2BAC5B,WAAU,gBAAgB,EACxB,MAAM,UAAU,GAAG,MACpB;CAGH,MAAM,gBAAiB,IAAiB,WAAW,UAAU;AAC7D,QAAO,OAAO,KAAK,cAAc,CAAC,KAAK,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAsFtD,SAAgB,+BAMd,QAC+B;CAC/B,MAAM,EAAE,cAAc,KAAK,QAAQ,iBAAiB;AAWpD,QAAO,uBAAuB;EAC5B,WAAW,CAR2B;GACtC,MAAM;GACN,aACE;GACF,YAPiB,mBAAmB,aAAa,GAC/C,aAAa,aAAa,GAC1B;GAMH,CAGyB;EACxB;EACA;EACA,4BAA4B;EAC5B;EACD,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"openapi.cjs","names":["OpenAPISpec","BaseChain","ChatPromptTemplate","HumanMessagePromptTemplate","LLMChain","JsonOutputFunctionsParser","SequentialChain"],"sources":["../../../src/chains/openai_functions/openapi.ts"],"sourcesContent":["import type { OpenAIClient } from \"@langchain/openai\";\nimport {\n type JsonSchema7ObjectType,\n type JsonSchema7ArrayType,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport type { OpenAPIV3_1 } from \"openapi-types\";\n\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BaseChatModel } from \"@langchain/core/language_models/chat_models\";\nimport { BaseFunctionCallOptions } from \"@langchain/core/language_models/base\";\nimport {\n ChatPromptTemplate,\n HumanMessagePromptTemplate,\n BasePromptTemplate,\n} from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { OpenAPISpec } from \"../../util/openapi.js\";\nimport { BaseChain } from \"../base.js\";\nimport { LLMChain, LLMChainInput } from \"../llm_chain.js\";\nimport { SequentialChain } from \"../sequential_chain.js\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Type representing a function for executing OpenAPI requests.\n */\ntype OpenAPIExecutionMethod = (\n name: string,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>,\n options?: {\n headers?: Record<string, string>;\n params?: Record<string, string>;\n }\n) => Promise<string>;\n\n/**\n * Type representing the composition types of a schema.\n */\ntype CompositionType = \"anyOf\" | \"allOf\" | \"oneOf\";\n\n/**\n * Gets the composition type of a schema if it exists.\n * @param schema\n * @returns The composition type of the schema if it exists.\n */\nfunction getCompositionType(\n schema: OpenAPIV3_1.SchemaObject\n): CompositionType | undefined {\n if (schema.anyOf !== undefined) {\n return \"anyOf\";\n }\n if (schema.allOf !== undefined) {\n return \"allOf\";\n }\n if (schema.oneOf !== undefined) {\n return \"oneOf\";\n }\n return undefined;\n}\n\n/**\n * Formats a URL by replacing path parameters with their corresponding\n * values.\n * @param url The URL to format.\n * @param pathParams The path parameters to replace in the URL.\n * @returns The formatted URL.\n */\nfunction formatURL(url: string, pathParams: Record<string, string>): string {\n const expectedPathParamNames = [...url.matchAll(/{(.*?)}/g)].map(\n (match) => match[1]\n );\n const newParams: Record<string, string> = {};\n for (const paramName of expectedPathParamNames) {\n const cleanParamName = paramName.replace(/^\\.;/, \"\").replace(/\\*$/, \"\");\n const value = pathParams[cleanParamName];\n let formattedValue;\n if (Array.isArray(value)) {\n if (paramName.startsWith(\".\")) {\n const separator = paramName.endsWith(\"*\") ? \".\" : \",\";\n formattedValue = `.${value.join(separator)}`;\n } else if (paramName.startsWith(\",\")) {\n const separator = paramName.endsWith(\"*\") ? `${cleanParamName}=` : \",\";\n formattedValue = `${cleanParamName}=${value.join(separator)}`;\n } else {\n formattedValue = value.join(\",\");\n }\n } else if (typeof value === \"object\") {\n const kvSeparator = paramName.endsWith(\"*\") ? \"=\" : \",\";\n const kvStrings = Object.entries(value).map(\n ([k, v]) => k + kvSeparator + v\n );\n let entrySeparator;\n if (paramName.startsWith(\".\")) {\n entrySeparator = \".\";\n formattedValue = \".\";\n } else if (paramName.startsWith(\";\")) {\n entrySeparator = \";\";\n formattedValue = \";\";\n } else {\n entrySeparator = \",\";\n formattedValue = \"\";\n }\n formattedValue += kvStrings.join(entrySeparator);\n } else {\n if (paramName.startsWith(\".\")) {\n formattedValue = `.${value}`;\n } else if (paramName.startsWith(\";\")) {\n formattedValue = `;${cleanParamName}=${value}`;\n } else {\n formattedValue = value;\n }\n }\n newParams[paramName] = formattedValue;\n }\n let formattedUrl = url;\n for (const [key, newValue] of Object.entries(newParams)) {\n formattedUrl = formattedUrl.replace(`{${key}}`, newValue);\n }\n return formattedUrl;\n}\n\n/**\n * Converts OpenAPI parameters to JSON schema format.\n * @param params The OpenAPI parameters to convert.\n * @param spec The OpenAPI specification that contains the parameters.\n * @returns The JSON schema representation of the OpenAPI parameters.\n */\nfunction convertOpenAPIParamsToJSONSchema(\n params: OpenAPIV3_1.ParameterObject[],\n spec: OpenAPISpec\n) {\n return params.reduce(\n (jsonSchema: JsonSchema7ObjectType, param) => {\n let schema;\n if (param.schema) {\n schema = spec.getSchema(param.schema);\n jsonSchema.properties[param.name] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n );\n } else if (param.content) {\n const mediaTypeSchema = Object.values(param.content)[0].schema;\n if (mediaTypeSchema) {\n schema = spec.getSchema(mediaTypeSchema);\n }\n if (!schema) {\n return jsonSchema;\n }\n if (schema.description === undefined) {\n schema.description = param.description ?? \"\";\n }\n jsonSchema.properties[param.name] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n );\n } else {\n return jsonSchema;\n }\n if (param.required && Array.isArray(jsonSchema.required)) {\n jsonSchema.required.push(param.name);\n }\n return jsonSchema;\n },\n {\n type: \"object\",\n properties: {},\n required: [],\n additionalProperties: {},\n }\n );\n}\n\n// OpenAI throws errors on extraneous schema properties, e.g. if \"required\" is set on individual ones\n/**\n * Converts OpenAPI schemas to JSON schema format.\n * @param schema The OpenAPI schema to convert.\n * @param spec The OpenAPI specification that contains the schema.\n * @returns The JSON schema representation of the OpenAPI schema.\n */\nexport function convertOpenAPISchemaToJSONSchema(\n schema: OpenAPIV3_1.SchemaObject,\n spec: OpenAPISpec\n): JsonSchema7Type {\n const compositionType = getCompositionType(schema);\n if (compositionType !== undefined && schema[compositionType] !== undefined) {\n return {\n [compositionType]: schema[compositionType].map((s) =>\n convertOpenAPISchemaToJSONSchema(spec.getSchema(s), spec)\n ),\n } as JsonSchema7Type;\n }\n\n if (schema.type === \"object\") {\n return Object.keys(schema.properties ?? {}).reduce(\n (jsonSchema: JsonSchema7ObjectType, propertyName) => {\n if (!schema.properties) {\n return jsonSchema;\n }\n const openAPIProperty = spec.getSchema(schema.properties[propertyName]);\n if (openAPIProperty.type === undefined) {\n return jsonSchema;\n }\n jsonSchema.properties[propertyName] = convertOpenAPISchemaToJSONSchema(\n openAPIProperty,\n spec\n );\n if (\n schema.required?.includes(propertyName) &&\n jsonSchema.required !== undefined\n ) {\n jsonSchema.required.push(propertyName);\n }\n return jsonSchema;\n },\n {\n type: \"object\",\n properties: {},\n required: [],\n additionalProperties: {},\n }\n );\n }\n\n if (schema.type === \"array\") {\n const openAPIItems = spec.getSchema(schema.items ?? {});\n return {\n type: \"array\",\n items: convertOpenAPISchemaToJSONSchema(openAPIItems, spec),\n minItems: schema.minItems,\n maxItems: schema.maxItems,\n } as JsonSchema7ArrayType;\n }\n return {\n type: schema.type ?? \"string\",\n } as JsonSchema7Type;\n}\n\n/**\n * Converts an OpenAPI specification to OpenAI functions.\n * @param spec The OpenAPI specification to convert.\n * @returns An object containing the OpenAI functions derived from the OpenAPI specification and a default execution method.\n */\nexport function convertOpenAPISpecToOpenAIFunctions(spec: OpenAPISpec): {\n openAIFunctions: OpenAIClient.Chat.ChatCompletionCreateParams.Function[];\n defaultExecutionMethod?: OpenAPIExecutionMethod;\n} {\n if (!spec.document.paths) {\n return { openAIFunctions: [] };\n }\n const openAIFunctions = [];\n const nameToCallMap: Record<string, { method: string; url: string }> = {};\n for (const path of Object.keys(spec.document.paths)) {\n const pathParameters = spec.getParametersForPath(path);\n for (const method of spec.getMethodsForPath(path)) {\n const operation = spec.getOperation(path, method);\n if (!operation) {\n return { openAIFunctions: [] };\n }\n const operationParametersByLocation = pathParameters\n .concat(spec.getParametersForOperation(operation))\n .reduce(\n (\n operationParams: Record<string, OpenAPIV3_1.ParameterObject[]>,\n param\n ) => {\n if (!operationParams[param.in]) {\n operationParams[param.in] = [];\n }\n operationParams[param.in].push(param);\n return operationParams;\n },\n {}\n );\n const paramLocationToRequestArgNameMap: Record<string, string> = {\n query: \"params\",\n header: \"headers\",\n cookie: \"cookies\",\n path: \"path_params\",\n };\n const requestArgsSchema: Record<string, JsonSchema7ObjectType> & {\n data?:\n | JsonSchema7ObjectType\n | {\n anyOf?: JsonSchema7ObjectType[];\n };\n } = {};\n for (const paramLocation of Object.keys(\n paramLocationToRequestArgNameMap\n )) {\n if (operationParametersByLocation[paramLocation]) {\n requestArgsSchema[paramLocationToRequestArgNameMap[paramLocation]] =\n convertOpenAPIParamsToJSONSchema(\n operationParametersByLocation[paramLocation],\n spec\n );\n }\n }\n const requestBody = spec.getRequestBodyForOperation(operation);\n if (requestBody?.content !== undefined) {\n const requestBodySchemas: Record<string, JsonSchema7ObjectType> = {};\n for (const [mediaType, mediaTypeObject] of Object.entries(\n requestBody.content\n )) {\n if (mediaTypeObject.schema !== undefined) {\n const schema = spec.getSchema(mediaTypeObject.schema);\n requestBodySchemas[mediaType] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n ) as JsonSchema7ObjectType;\n }\n }\n const mediaTypes = Object.keys(requestBodySchemas);\n if (mediaTypes.length === 1) {\n requestArgsSchema.data = requestBodySchemas[mediaTypes[0]];\n } else if (mediaTypes.length > 1) {\n requestArgsSchema.data = {\n anyOf: Object.values(requestBodySchemas),\n };\n }\n }\n const openAIFunction: OpenAIClient.Chat.ChatCompletionCreateParams.Function =\n {\n name: OpenAPISpec.getCleanedOperationId(operation, path, method),\n description: operation.description ?? operation.summary ?? \"\",\n parameters: {\n type: \"object\",\n properties: requestArgsSchema,\n // All remaining top-level parameters are required\n required: Object.keys(requestArgsSchema),\n },\n };\n\n openAIFunctions.push(openAIFunction);\n const baseUrl = (spec.baseUrl ?? \"\").endsWith(\"/\")\n ? (spec.baseUrl ?? \"\").slice(0, -1)\n : (spec.baseUrl ?? \"\");\n nameToCallMap[openAIFunction.name] = {\n method,\n url: baseUrl + path,\n };\n }\n }\n return {\n openAIFunctions,\n defaultExecutionMethod: async (\n name: string,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>,\n options?: {\n headers?: Record<string, string>;\n params?: Record<string, string>;\n }\n ) => {\n const {\n headers: customHeaders,\n params: customParams,\n ...rest\n } = options ?? {};\n const { method, url } = nameToCallMap[name];\n const requestParams = requestArgs.params ?? {};\n const nonEmptyParams = Object.keys(requestParams).reduce(\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n (filteredArgs: Record<string, any>, argName) => {\n if (\n requestParams[argName] !== \"\" &&\n requestParams[argName] !== null &&\n requestParams[argName] !== undefined\n ) {\n filteredArgs[argName] = requestParams[argName];\n }\n return filteredArgs;\n },\n {}\n );\n const queryString = new URLSearchParams({\n ...nonEmptyParams,\n ...customParams,\n }).toString();\n const pathParams = requestArgs.path_params;\n const formattedUrl =\n formatURL(url, pathParams) +\n (queryString.length ? `?${queryString}` : \"\");\n const headers: Record<string, string> = {};\n let body;\n if (requestArgs.data !== undefined) {\n let contentType = \"text/plain\";\n if (typeof requestArgs.data !== \"string\") {\n if (typeof requestArgs.data === \"object\") {\n contentType = \"application/json\";\n }\n body = JSON.stringify(requestArgs.data);\n } else {\n body = requestArgs.data;\n }\n headers[\"content-type\"] = contentType;\n }\n const response = await fetch(formattedUrl, {\n ...requestArgs,\n method,\n headers: {\n ...headers,\n ...requestArgs.headers,\n ...customHeaders,\n },\n body,\n ...rest,\n });\n let output;\n if (response.status < 200 || response.status > 299) {\n output = `${response.status}: ${\n response.statusText\n } for ${name} called with ${JSON.stringify(queryString)}`;\n } else {\n output = await response.text();\n }\n return output;\n },\n };\n}\n\n/**\n * Type representing a function for executing simple requests.\n */\ntype SimpleRequestChainExecutionMethod = (\n name: string,\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>\n) => Promise<string>;\n\n/**\n * A chain for making simple API requests.\n */\nclass SimpleRequestChain extends BaseChain {\n static lc_name() {\n return \"SimpleRequestChain\";\n }\n\n private requestMethod: SimpleRequestChainExecutionMethod;\n\n inputKey = \"function\";\n\n outputKey = \"response\";\n\n constructor(config: { requestMethod: SimpleRequestChainExecutionMethod }) {\n super();\n this.requestMethod = config.requestMethod;\n }\n\n get inputKeys() {\n return [this.inputKey];\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n _chainType() {\n return \"simple_request_chain\" as const;\n }\n\n /** @ignore */\n async _call(\n values: ChainValues,\n _runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const inputKeyValue = values[this.inputKey];\n const methodName = inputKeyValue.name;\n const args = inputKeyValue.arguments;\n const response = await this.requestMethod(methodName, args);\n return { [this.outputKey]: response };\n }\n}\n\n/**\n * Type representing the options for creating an OpenAPI chain.\n */\nexport type OpenAPIChainOptions = {\n llm?: BaseChatModel<BaseFunctionCallOptions>;\n prompt?: BasePromptTemplate;\n requestChain?: BaseChain;\n llmChainInputs?: LLMChainInput;\n headers?: Record<string, string>;\n params?: Record<string, string>;\n verbose?: boolean;\n};\n\n/**\n * Create a chain for querying an API from a OpenAPI spec.\n * @param spec OpenAPISpec or url/file/text string corresponding to one.\n * @param options Custom options passed into the chain\n * @returns OpenAPIChain\n */\nexport async function createOpenAPIChain(\n spec: OpenAPIV3_1.Document | string,\n options: OpenAPIChainOptions = {}\n) {\n let convertedSpec;\n if (typeof spec === \"string\") {\n try {\n convertedSpec = await OpenAPISpec.fromURL(spec);\n } catch {\n try {\n convertedSpec = OpenAPISpec.fromString(spec);\n } catch {\n throw new Error(`Unable to parse spec from source ${spec}.`);\n }\n }\n } else {\n convertedSpec = OpenAPISpec.fromObject(spec);\n }\n const { openAIFunctions, defaultExecutionMethod } =\n convertOpenAPISpecToOpenAIFunctions(convertedSpec);\n if (defaultExecutionMethod === undefined) {\n throw new Error(\n `Could not parse any valid operations from the provided spec.`\n );\n }\n\n if (!options.llm) {\n throw new Error(\"`llm` option is required\");\n }\n\n const {\n llm = options.llm,\n prompt = ChatPromptTemplate.fromMessages([\n HumanMessagePromptTemplate.fromTemplate(\n \"Use the provided API's to respond to this user query:\\n\\n{query}\"\n ),\n ]),\n requestChain = new SimpleRequestChain({\n requestMethod: async (name, args) =>\n defaultExecutionMethod(name, args, {\n headers: options.headers,\n params: options.params,\n }),\n }),\n llmChainInputs = {},\n verbose,\n ...rest\n } = options;\n const formatChain = new LLMChain({\n llm,\n prompt,\n outputParser: new JsonOutputFunctionsParser({ argsOnly: false }),\n outputKey: \"function\",\n llmKwargs: { functions: openAIFunctions },\n ...llmChainInputs,\n });\n return new SequentialChain({\n chains: [formatChain, requestChain],\n outputVariables: [\"response\"],\n inputVariables: formatChain.inputKeys,\n verbose,\n ...rest,\n });\n}\n"],"mappings":";;;;;;;;;;;;;AA8CA,SAAS,mBACP,QAC6B;AAC7B,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;AAET,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;AAET,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;;;;;;;;;AAYX,SAAS,UAAU,KAAa,YAA4C;CAC1E,MAAM,yBAAyB,CAAC,GAAG,IAAI,SAAS,WAAW,CAAC,CAAC,KAC1D,UAAU,MAAM,GAClB;CACD,MAAM,YAAoC,EAAE;AAC5C,MAAK,MAAM,aAAa,wBAAwB;EAC9C,MAAM,iBAAiB,UAAU,QAAQ,QAAQ,GAAG,CAAC,QAAQ,OAAO,GAAG;EACvE,MAAM,QAAQ,WAAW;EACzB,IAAI;AACJ,MAAI,MAAM,QAAQ,MAAM,CACtB,KAAI,UAAU,WAAW,IAAI,EAAE;GAC7B,MAAM,YAAY,UAAU,SAAS,IAAI,GAAG,MAAM;AAClD,oBAAiB,IAAI,MAAM,KAAK,UAAU;aACjC,UAAU,WAAW,IAAI,EAAE;GACpC,MAAM,YAAY,UAAU,SAAS,IAAI,GAAG,GAAG,eAAe,KAAK;AACnE,oBAAiB,GAAG,eAAe,GAAG,MAAM,KAAK,UAAU;QAE3D,kBAAiB,MAAM,KAAK,IAAI;WAEzB,OAAO,UAAU,UAAU;GACpC,MAAM,cAAc,UAAU,SAAS,IAAI,GAAG,MAAM;GACpD,MAAM,YAAY,OAAO,QAAQ,MAAM,CAAC,KACrC,CAAC,GAAG,OAAO,IAAI,cAAc,EAC/B;GACD,IAAI;AACJ,OAAI,UAAU,WAAW,IAAI,EAAE;AAC7B,qBAAiB;AACjB,qBAAiB;cACR,UAAU,WAAW,IAAI,EAAE;AACpC,qBAAiB;AACjB,qBAAiB;UACZ;AACL,qBAAiB;AACjB,qBAAiB;;AAEnB,qBAAkB,UAAU,KAAK,eAAe;aAE5C,UAAU,WAAW,IAAI,CAC3B,kBAAiB,IAAI;WACZ,UAAU,WAAW,IAAI,CAClC,kBAAiB,IAAI,eAAe,GAAG;MAEvC,kBAAiB;AAGrB,YAAU,aAAa;;CAEzB,IAAI,eAAe;AACnB,MAAK,MAAM,CAAC,KAAK,aAAa,OAAO,QAAQ,UAAU,CACrD,gBAAe,aAAa,QAAQ,IAAI,IAAI,IAAI,SAAS;AAE3D,QAAO;;;;;;;;AAST,SAAS,iCACP,QACA,MACA;AACA,QAAO,OAAO,QACX,YAAmC,UAAU;EAC5C,IAAI;AACJ,MAAI,MAAM,QAAQ;AAChB,YAAS,KAAK,UAAU,MAAM,OAAO;AACrC,cAAW,WAAW,MAAM,QAAQ,iCAClC,QACA,KACD;aACQ,MAAM,SAAS;GACxB,MAAM,kBAAkB,OAAO,OAAO,MAAM,QAAQ,CAAC,GAAG;AACxD,OAAI,gBACF,UAAS,KAAK,UAAU,gBAAgB;AAE1C,OAAI,CAAC,OACH,QAAO;AAET,OAAI,OAAO,gBAAgB,KAAA,EACzB,QAAO,cAAc,MAAM,eAAe;AAE5C,cAAW,WAAW,MAAM,QAAQ,iCAClC,QACA,KACD;QAED,QAAO;AAET,MAAI,MAAM,YAAY,MAAM,QAAQ,WAAW,SAAS,CACtD,YAAW,SAAS,KAAK,MAAM,KAAK;AAEtC,SAAO;IAET;EACE,MAAM;EACN,YAAY,EAAE;EACd,UAAU,EAAE;EACZ,sBAAsB,EAAE;EACzB,CACF;;;;;;;;AAUH,SAAgB,iCACd,QACA,MACiB;CACjB,MAAM,kBAAkB,mBAAmB,OAAO;AAClD,KAAI,oBAAoB,KAAA,KAAa,OAAO,qBAAqB,KAAA,EAC/D,QAAO,GACJ,kBAAkB,OAAO,iBAAiB,KAAK,MAC9C,iCAAiC,KAAK,UAAU,EAAE,EAAE,KAAK,CAC1D,EACF;AAGH,KAAI,OAAO,SAAS,SAClB,QAAO,OAAO,KAAK,OAAO,cAAc,EAAE,CAAC,CAAC,QACzC,YAAmC,iBAAiB;AACnD,MAAI,CAAC,OAAO,WACV,QAAO;EAET,MAAM,kBAAkB,KAAK,UAAU,OAAO,WAAW,cAAc;AACvE,MAAI,gBAAgB,SAAS,KAAA,EAC3B,QAAO;AAET,aAAW,WAAW,gBAAgB,iCACpC,iBACA,KACD;AACD,MACE,OAAO,UAAU,SAAS,aAAa,IACvC,WAAW,aAAa,KAAA,EAExB,YAAW,SAAS,KAAK,aAAa;AAExC,SAAO;IAET;EACE,MAAM;EACN,YAAY,EAAE;EACd,UAAU,EAAE;EACZ,sBAAsB,EAAE;EACzB,CACF;AAGH,KAAI,OAAO,SAAS,QAElB,QAAO;EACL,MAAM;EACN,OAAO,iCAHY,KAAK,UAAU,OAAO,SAAS,EAAE,CAAC,EAGC,KAAK;EAC3D,UAAU,OAAO;EACjB,UAAU,OAAO;EAClB;AAEH,QAAO,EACL,MAAM,OAAO,QAAQ,UACtB;;;;;;;AAQH,SAAgB,oCAAoC,MAGlD;AACA,KAAI,CAAC,KAAK,SAAS,MACjB,QAAO,EAAE,iBAAiB,EAAE,EAAE;CAEhC,MAAM,kBAAkB,EAAE;CAC1B,MAAM,gBAAiE,EAAE;AACzE,MAAK,MAAM,QAAQ,OAAO,KAAK,KAAK,SAAS,MAAM,EAAE;EACnD,MAAM,iBAAiB,KAAK,qBAAqB,KAAK;AACtD,OAAK,MAAM,UAAU,KAAK,kBAAkB,KAAK,EAAE;GACjD,MAAM,YAAY,KAAK,aAAa,MAAM,OAAO;AACjD,OAAI,CAAC,UACH,QAAO,EAAE,iBAAiB,EAAE,EAAE;GAEhC,MAAM,gCAAgC,eACnC,OAAO,KAAK,0BAA0B,UAAU,CAAC,CACjD,QAEG,iBACA,UACG;AACH,QAAI,CAAC,gBAAgB,MAAM,IACzB,iBAAgB,MAAM,MAAM,EAAE;AAEhC,oBAAgB,MAAM,IAAI,KAAK,MAAM;AACrC,WAAO;MAET,EAAE,CACH;GACH,MAAM,mCAA2D;IAC/D,OAAO;IACP,QAAQ;IACR,QAAQ;IACR,MAAM;IACP;GACD,MAAM,oBAMF,EAAE;AACN,QAAK,MAAM,iBAAiB,OAAO,KACjC,iCACD,CACC,KAAI,8BAA8B,eAChC,mBAAkB,iCAAiC,kBACjD,iCACE,8BAA8B,gBAC9B,KACD;GAGP,MAAM,cAAc,KAAK,2BAA2B,UAAU;AAC9D,OAAI,aAAa,YAAY,KAAA,GAAW;IACtC,MAAM,qBAA4D,EAAE;AACpE,SAAK,MAAM,CAAC,WAAW,oBAAoB,OAAO,QAChD,YAAY,QACb,CACC,KAAI,gBAAgB,WAAW,KAAA,EAE7B,oBAAmB,aAAa,iCADjB,KAAK,UAAU,gBAAgB,OAAO,EAGnD,KACD;IAGL,MAAM,aAAa,OAAO,KAAK,mBAAmB;AAClD,QAAI,WAAW,WAAW,EACxB,mBAAkB,OAAO,mBAAmB,WAAW;aAC9C,WAAW,SAAS,EAC7B,mBAAkB,OAAO,EACvB,OAAO,OAAO,OAAO,mBAAmB,EACzC;;GAGL,MAAM,iBACJ;IACE,MAAMA,gBAAAA,YAAY,sBAAsB,WAAW,MAAM,OAAO;IAChE,aAAa,UAAU,eAAe,UAAU,WAAW;IAC3D,YAAY;KACV,MAAM;KACN,YAAY;KAEZ,UAAU,OAAO,KAAK,kBAAkB;KACzC;IACF;AAEH,mBAAgB,KAAK,eAAe;GACpC,MAAM,WAAW,KAAK,WAAW,IAAI,SAAS,IAAI,IAC7C,KAAK,WAAW,IAAI,MAAM,GAAG,GAAG,GAChC,KAAK,WAAW;AACrB,iBAAc,eAAe,QAAQ;IACnC;IACA,KAAK,UAAU;IAChB;;;AAGL,QAAO;EACL;EACA,wBAAwB,OACtB,MAEA,aACA,YAIG;GACH,MAAM,EACJ,SAAS,eACT,QAAQ,cACR,GAAG,SACD,WAAW,EAAE;GACjB,MAAM,EAAE,QAAQ,QAAQ,cAAc;GACtC,MAAM,gBAAgB,YAAY,UAAU,EAAE;GAC9C,MAAM,iBAAiB,OAAO,KAAK,cAAc,CAAC,QAE/C,cAAmC,YAAY;AAC9C,QACE,cAAc,aAAa,MAC3B,cAAc,aAAa,QAC3B,cAAc,aAAa,KAAA,EAE3B,cAAa,WAAW,cAAc;AAExC,WAAO;MAET,EAAE,CACH;GACD,MAAM,cAAc,IAAI,gBAAgB;IACtC,GAAG;IACH,GAAG;IACJ,CAAC,CAAC,UAAU;GACb,MAAM,aAAa,YAAY;GAC/B,MAAM,eACJ,UAAU,KAAK,WAAW,IACzB,YAAY,SAAS,IAAI,gBAAgB;GAC5C,MAAM,UAAkC,EAAE;GAC1C,IAAI;AACJ,OAAI,YAAY,SAAS,KAAA,GAAW;IAClC,IAAI,cAAc;AAClB,QAAI,OAAO,YAAY,SAAS,UAAU;AACxC,SAAI,OAAO,YAAY,SAAS,SAC9B,eAAc;AAEhB,YAAO,KAAK,UAAU,YAAY,KAAK;UAEvC,QAAO,YAAY;AAErB,YAAQ,kBAAkB;;GAE5B,MAAM,WAAW,MAAM,MAAM,cAAc;IACzC,GAAG;IACH;IACA,SAAS;KACP,GAAG;KACH,GAAG,YAAY;KACf,GAAG;KACJ;IACD;IACA,GAAG;IACJ,CAAC;GACF,IAAI;AACJ,OAAI,SAAS,SAAS,OAAO,SAAS,SAAS,IAC7C,UAAS,GAAG,SAAS,OAAO,IAC1B,SAAS,WACV,OAAO,KAAK,eAAe,KAAK,UAAU,YAAY;OAEvD,UAAS,MAAM,SAAS,MAAM;AAEhC,UAAO;;EAEV;;;;;AAeH,IAAM,qBAAN,cAAiCC,aAAAA,UAAU;CACzC,OAAO,UAAU;AACf,SAAO;;CAGT;CAEA,WAAW;CAEX,YAAY;CAEZ,YAAY,QAA8D;AACxE,SAAO;AACP,OAAK,gBAAgB,OAAO;;CAG9B,IAAI,YAAY;AACd,SAAO,CAAC,KAAK,SAAS;;CAGxB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,aAAa;AACX,SAAO;;;CAIT,MAAM,MACJ,QACA,aACsB;EACtB,MAAM,gBAAgB,OAAO,KAAK;EAClC,MAAM,aAAa,cAAc;EACjC,MAAM,OAAO,cAAc;EAC3B,MAAM,WAAW,MAAM,KAAK,cAAc,YAAY,KAAK;AAC3D,SAAO,GAAG,KAAK,YAAY,UAAU;;;;;;;;;AAuBzC,eAAsB,mBACpB,MACA,UAA+B,EAAE,EACjC;CACA,IAAI;AACJ,KAAI,OAAO,SAAS,SAClB,KAAI;AACF,kBAAgB,MAAMD,gBAAAA,YAAY,QAAQ,KAAK;SACzC;AACN,MAAI;AACF,mBAAgBA,gBAAAA,YAAY,WAAW,KAAK;UACtC;AACN,SAAM,IAAI,MAAM,oCAAoC,KAAK,GAAG;;;KAIhE,iBAAgBA,gBAAAA,YAAY,WAAW,KAAK;CAE9C,MAAM,EAAE,iBAAiB,2BACvB,oCAAoC,cAAc;AACpD,KAAI,2BAA2B,KAAA,EAC7B,OAAM,IAAI,MACR,+DACD;AAGH,KAAI,CAAC,QAAQ,IACX,OAAM,IAAI,MAAM,2BAA2B;CAG7C,MAAM,EACJ,MAAM,QAAQ,KACd,SAASE,wBAAAA,mBAAmB,aAAa,CACvCC,wBAAAA,2BAA2B,aACzB,mEACD,CACF,CAAC,EACF,eAAe,IAAI,mBAAmB,EACpC,eAAe,OAAO,MAAM,SAC1B,uBAAuB,MAAM,MAAM;EACjC,SAAS,QAAQ;EACjB,QAAQ,QAAQ;EACjB,CAAC,EACL,CAAC,EACF,iBAAiB,EAAE,EACnB,SACA,GAAG,SACD;CACJ,MAAM,cAAc,IAAIC,kBAAAA,SAAS;EAC/B;EACA;EACA,cAAc,IAAIC,yBAAAA,0BAA0B,EAAE,UAAU,OAAO,CAAC;EAChE,WAAW;EACX,WAAW,EAAE,WAAW,iBAAiB;EACzC,GAAG;EACJ,CAAC;AACF,QAAO,IAAIC,yBAAAA,gBAAgB;EACzB,QAAQ,CAAC,aAAa,aAAa;EACnC,iBAAiB,CAAC,WAAW;EAC7B,gBAAgB,YAAY;EAC5B;EACA,GAAG;EACJ,CAAC"}
|
|
1
|
+
{"version":3,"file":"openapi.cjs","names":["OpenAPISpec","BaseChain","ChatPromptTemplate","HumanMessagePromptTemplate","LLMChain","JsonOutputFunctionsParser","SequentialChain"],"sources":["../../../src/chains/openai_functions/openapi.ts"],"sourcesContent":["import type { OpenAIClient } from \"@langchain/openai\";\nimport {\n type JsonSchema7ObjectType,\n type JsonSchema7ArrayType,\n type JsonSchema7Type,\n} from \"@langchain/core/utils/json_schema\";\nimport type { OpenAPIV3_1 } from \"openapi-types\";\n\nimport { ChainValues } from \"@langchain/core/utils/types\";\nimport { BaseChatModel } from \"@langchain/core/language_models/chat_models\";\nimport { BaseFunctionCallOptions } from \"@langchain/core/language_models/base\";\nimport {\n ChatPromptTemplate,\n HumanMessagePromptTemplate,\n BasePromptTemplate,\n} from \"@langchain/core/prompts\";\nimport { CallbackManagerForChainRun } from \"@langchain/core/callbacks/manager\";\nimport { OpenAPISpec } from \"../../util/openapi.js\";\nimport { BaseChain } from \"../base.js\";\nimport { LLMChain, LLMChainInput } from \"../llm_chain.js\";\nimport { SequentialChain } from \"../sequential_chain.js\";\nimport { JsonOutputFunctionsParser } from \"../../output_parsers/openai_functions.js\";\n\n/**\n * Type representing a function for executing OpenAPI requests.\n */\ntype OpenAPIExecutionMethod = (\n name: string,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>,\n options?: {\n headers?: Record<string, string>;\n params?: Record<string, string>;\n }\n) => Promise<string>;\n\n/**\n * Type representing the composition types of a schema.\n */\ntype CompositionType = \"anyOf\" | \"allOf\" | \"oneOf\";\n\n/**\n * Gets the composition type of a schema if it exists.\n * @param schema\n * @returns The composition type of the schema if it exists.\n */\nfunction getCompositionType(\n schema: OpenAPIV3_1.SchemaObject\n): CompositionType | undefined {\n if (schema.anyOf !== undefined) {\n return \"anyOf\";\n }\n if (schema.allOf !== undefined) {\n return \"allOf\";\n }\n if (schema.oneOf !== undefined) {\n return \"oneOf\";\n }\n return undefined;\n}\n\n/**\n * Formats a URL by replacing path parameters with their corresponding\n * values.\n * @param url The URL to format.\n * @param pathParams The path parameters to replace in the URL.\n * @returns The formatted URL.\n */\nfunction formatURL(url: string, pathParams: Record<string, string>): string {\n const expectedPathParamNames = [...url.matchAll(/{(.*?)}/g)].map(\n (match) => match[1]\n );\n const newParams: Record<string, string> = {};\n for (const paramName of expectedPathParamNames) {\n const cleanParamName = paramName.replace(/^\\.;/, \"\").replace(/\\*$/, \"\");\n const value = pathParams[cleanParamName];\n let formattedValue;\n if (Array.isArray(value)) {\n if (paramName.startsWith(\".\")) {\n const separator = paramName.endsWith(\"*\") ? \".\" : \",\";\n formattedValue = `.${value.join(separator)}`;\n } else if (paramName.startsWith(\",\")) {\n const separator = paramName.endsWith(\"*\") ? `${cleanParamName}=` : \",\";\n formattedValue = `${cleanParamName}=${value.join(separator)}`;\n } else {\n formattedValue = value.join(\",\");\n }\n } else if (typeof value === \"object\") {\n const kvSeparator = paramName.endsWith(\"*\") ? \"=\" : \",\";\n const kvStrings = Object.entries(value).map(\n ([k, v]) => k + kvSeparator + v\n );\n let entrySeparator;\n if (paramName.startsWith(\".\")) {\n entrySeparator = \".\";\n formattedValue = \".\";\n } else if (paramName.startsWith(\";\")) {\n entrySeparator = \";\";\n formattedValue = \";\";\n } else {\n entrySeparator = \",\";\n formattedValue = \"\";\n }\n formattedValue += kvStrings.join(entrySeparator);\n } else {\n if (paramName.startsWith(\".\")) {\n formattedValue = `.${value}`;\n } else if (paramName.startsWith(\";\")) {\n formattedValue = `;${cleanParamName}=${value}`;\n } else {\n formattedValue = value;\n }\n }\n newParams[paramName] = formattedValue;\n }\n let formattedUrl = url;\n for (const [key, newValue] of Object.entries(newParams)) {\n formattedUrl = formattedUrl.replace(`{${key}}`, newValue);\n }\n return formattedUrl;\n}\n\n/**\n * Converts OpenAPI parameters to JSON schema format.\n * @param params The OpenAPI parameters to convert.\n * @param spec The OpenAPI specification that contains the parameters.\n * @returns The JSON schema representation of the OpenAPI parameters.\n */\nfunction convertOpenAPIParamsToJSONSchema(\n params: OpenAPIV3_1.ParameterObject[],\n spec: OpenAPISpec\n) {\n return params.reduce(\n (jsonSchema: JsonSchema7ObjectType, param) => {\n let schema;\n if (param.schema) {\n schema = spec.getSchema(param.schema);\n jsonSchema.properties[param.name] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n );\n } else if (param.content) {\n const mediaTypeSchema = Object.values(param.content)[0].schema;\n if (mediaTypeSchema) {\n schema = spec.getSchema(mediaTypeSchema);\n }\n if (!schema) {\n return jsonSchema;\n }\n if (schema.description === undefined) {\n schema.description = param.description ?? \"\";\n }\n jsonSchema.properties[param.name] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n );\n } else {\n return jsonSchema;\n }\n if (param.required && Array.isArray(jsonSchema.required)) {\n jsonSchema.required.push(param.name);\n }\n return jsonSchema;\n },\n {\n type: \"object\",\n properties: {},\n required: [],\n additionalProperties: {},\n }\n );\n}\n\n// OpenAI throws errors on extraneous schema properties, e.g. if \"required\" is set on individual ones\n/**\n * Converts OpenAPI schemas to JSON schema format.\n * @param schema The OpenAPI schema to convert.\n * @param spec The OpenAPI specification that contains the schema.\n * @returns The JSON schema representation of the OpenAPI schema.\n */\nexport function convertOpenAPISchemaToJSONSchema(\n schema: OpenAPIV3_1.SchemaObject,\n spec: OpenAPISpec\n): JsonSchema7Type {\n const compositionType = getCompositionType(schema);\n if (compositionType !== undefined && schema[compositionType] !== undefined) {\n return {\n [compositionType]: schema[compositionType].map((s) =>\n convertOpenAPISchemaToJSONSchema(spec.getSchema(s), spec)\n ),\n } as JsonSchema7Type;\n }\n\n if (schema.type === \"object\") {\n return Object.keys(schema.properties ?? {}).reduce(\n (jsonSchema: JsonSchema7ObjectType, propertyName) => {\n if (!schema.properties) {\n return jsonSchema;\n }\n const openAPIProperty = spec.getSchema(schema.properties[propertyName]);\n if (openAPIProperty.type === undefined) {\n return jsonSchema;\n }\n jsonSchema.properties[propertyName] = convertOpenAPISchemaToJSONSchema(\n openAPIProperty,\n spec\n );\n if (\n schema.required?.includes(propertyName) &&\n jsonSchema.required !== undefined\n ) {\n jsonSchema.required.push(propertyName);\n }\n return jsonSchema;\n },\n {\n type: \"object\",\n properties: {},\n required: [],\n additionalProperties: {},\n }\n );\n }\n\n if (schema.type === \"array\") {\n const openAPIItems = spec.getSchema(schema.items ?? {});\n return {\n type: \"array\",\n items: convertOpenAPISchemaToJSONSchema(openAPIItems, spec),\n minItems: schema.minItems,\n maxItems: schema.maxItems,\n } as JsonSchema7ArrayType;\n }\n return {\n type: schema.type ?? \"string\",\n } as JsonSchema7Type;\n}\n\n/**\n * Converts an OpenAPI specification to OpenAI functions.\n * @param spec The OpenAPI specification to convert.\n * @returns An object containing the OpenAI functions derived from the OpenAPI specification and a default execution method.\n */\nexport function convertOpenAPISpecToOpenAIFunctions(spec: OpenAPISpec): {\n openAIFunctions: OpenAIClient.Chat.ChatCompletionCreateParams.Function[];\n defaultExecutionMethod?: OpenAPIExecutionMethod;\n} {\n if (!spec.document.paths) {\n return { openAIFunctions: [] };\n }\n const openAIFunctions = [];\n const nameToCallMap: Record<string, { method: string; url: string }> = {};\n for (const path of Object.keys(spec.document.paths)) {\n const pathParameters = spec.getParametersForPath(path);\n for (const method of spec.getMethodsForPath(path)) {\n const operation = spec.getOperation(path, method);\n if (!operation) {\n return { openAIFunctions: [] };\n }\n const operationParametersByLocation = pathParameters\n .concat(spec.getParametersForOperation(operation))\n .reduce(\n (\n operationParams: Record<string, OpenAPIV3_1.ParameterObject[]>,\n param\n ) => {\n if (!operationParams[param.in]) {\n operationParams[param.in] = [];\n }\n operationParams[param.in].push(param);\n return operationParams;\n },\n {}\n );\n const paramLocationToRequestArgNameMap: Record<string, string> = {\n query: \"params\",\n header: \"headers\",\n cookie: \"cookies\",\n path: \"path_params\",\n };\n const requestArgsSchema: Record<string, JsonSchema7ObjectType> & {\n data?:\n | JsonSchema7ObjectType\n | {\n anyOf?: JsonSchema7ObjectType[];\n };\n } = {};\n for (const paramLocation of Object.keys(\n paramLocationToRequestArgNameMap\n )) {\n if (operationParametersByLocation[paramLocation]) {\n requestArgsSchema[paramLocationToRequestArgNameMap[paramLocation]] =\n convertOpenAPIParamsToJSONSchema(\n operationParametersByLocation[paramLocation],\n spec\n );\n }\n }\n const requestBody = spec.getRequestBodyForOperation(operation);\n if (requestBody?.content !== undefined) {\n const requestBodySchemas: Record<string, JsonSchema7ObjectType> = {};\n for (const [mediaType, mediaTypeObject] of Object.entries(\n requestBody.content\n )) {\n if (mediaTypeObject.schema !== undefined) {\n const schema = spec.getSchema(mediaTypeObject.schema);\n requestBodySchemas[mediaType] = convertOpenAPISchemaToJSONSchema(\n schema,\n spec\n ) as JsonSchema7ObjectType;\n }\n }\n const mediaTypes = Object.keys(requestBodySchemas);\n if (mediaTypes.length === 1) {\n requestArgsSchema.data = requestBodySchemas[mediaTypes[0]];\n } else if (mediaTypes.length > 1) {\n requestArgsSchema.data = {\n anyOf: Object.values(requestBodySchemas),\n };\n }\n }\n const openAIFunction: OpenAIClient.Chat.ChatCompletionCreateParams.Function =\n {\n name: OpenAPISpec.getCleanedOperationId(operation, path, method),\n description: operation.description ?? operation.summary ?? \"\",\n parameters: {\n type: \"object\",\n properties: requestArgsSchema,\n // All remaining top-level parameters are required\n required: Object.keys(requestArgsSchema),\n },\n };\n\n openAIFunctions.push(openAIFunction);\n const baseUrl = (spec.baseUrl ?? \"\").endsWith(\"/\")\n ? (spec.baseUrl ?? \"\").slice(0, -1)\n : (spec.baseUrl ?? \"\");\n nameToCallMap[openAIFunction.name] = {\n method,\n url: baseUrl + path,\n };\n }\n }\n return {\n openAIFunctions,\n defaultExecutionMethod: async (\n name: string,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>,\n options?: {\n headers?: Record<string, string>;\n params?: Record<string, string>;\n }\n ) => {\n const {\n headers: customHeaders,\n params: customParams,\n ...rest\n } = options ?? {};\n const { method, url } = nameToCallMap[name];\n const requestParams = requestArgs.params ?? {};\n const nonEmptyParams = Object.keys(requestParams).reduce(\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n (filteredArgs: Record<string, any>, argName) => {\n if (\n requestParams[argName] !== \"\" &&\n requestParams[argName] !== null &&\n requestParams[argName] !== undefined\n ) {\n filteredArgs[argName] = requestParams[argName];\n }\n return filteredArgs;\n },\n {}\n );\n const queryString = new URLSearchParams({\n ...nonEmptyParams,\n ...customParams,\n }).toString();\n const pathParams = requestArgs.path_params;\n const formattedUrl =\n formatURL(url, pathParams) +\n (queryString.length ? `?${queryString}` : \"\");\n const headers: Record<string, string> = {};\n let body;\n if (requestArgs.data !== undefined) {\n let contentType = \"text/plain\";\n if (typeof requestArgs.data !== \"string\") {\n if (typeof requestArgs.data === \"object\") {\n contentType = \"application/json\";\n }\n body = JSON.stringify(requestArgs.data);\n } else {\n body = requestArgs.data;\n }\n headers[\"content-type\"] = contentType;\n }\n const response = await fetch(formattedUrl, {\n ...requestArgs,\n method,\n headers: {\n ...headers,\n ...requestArgs.headers,\n ...customHeaders,\n },\n body,\n ...rest,\n });\n let output;\n if (response.status < 200 || response.status > 299) {\n output = `${response.status}: ${\n response.statusText\n } for ${name} called with ${JSON.stringify(queryString)}`;\n } else {\n output = await response.text();\n }\n return output;\n },\n };\n}\n\n/**\n * Type representing a function for executing simple requests.\n */\ntype SimpleRequestChainExecutionMethod = (\n name: string,\n // oxlint-disable-next-line @typescript-eslint/no-explicit-any\n requestArgs: Record<string, any>\n) => Promise<string>;\n\n/**\n * A chain for making simple API requests.\n */\nclass SimpleRequestChain extends BaseChain {\n static lc_name() {\n return \"SimpleRequestChain\";\n }\n\n private requestMethod: SimpleRequestChainExecutionMethod;\n\n inputKey = \"function\";\n\n outputKey = \"response\";\n\n constructor(config: { requestMethod: SimpleRequestChainExecutionMethod }) {\n super();\n this.requestMethod = config.requestMethod;\n }\n\n get inputKeys() {\n return [this.inputKey];\n }\n\n get outputKeys() {\n return [this.outputKey];\n }\n\n _chainType() {\n return \"simple_request_chain\" as const;\n }\n\n /** @ignore */\n async _call(\n values: ChainValues,\n _runManager?: CallbackManagerForChainRun\n ): Promise<ChainValues> {\n const inputKeyValue = values[this.inputKey];\n const methodName = inputKeyValue.name;\n const args = inputKeyValue.arguments;\n const response = await this.requestMethod(methodName, args);\n return { [this.outputKey]: response };\n }\n}\n\n/**\n * Type representing the options for creating an OpenAPI chain.\n */\nexport type OpenAPIChainOptions = {\n llm?: BaseChatModel<BaseFunctionCallOptions>;\n prompt?: BasePromptTemplate;\n requestChain?: BaseChain;\n llmChainInputs?: LLMChainInput;\n headers?: Record<string, string>;\n params?: Record<string, string>;\n verbose?: boolean;\n};\n\n/**\n * Create a chain for querying an API from a OpenAPI spec.\n * @param spec OpenAPISpec or url/file/text string corresponding to one.\n * @param options Custom options passed into the chain\n * @returns OpenAPIChain\n */\nexport async function createOpenAPIChain(\n spec: OpenAPIV3_1.Document | string,\n options: OpenAPIChainOptions = {}\n) {\n let convertedSpec;\n if (typeof spec === \"string\") {\n try {\n convertedSpec = await OpenAPISpec.fromURL(spec);\n } catch {\n try {\n convertedSpec = OpenAPISpec.fromString(spec);\n } catch {\n throw new Error(`Unable to parse spec from source ${spec}.`);\n }\n }\n } else {\n convertedSpec = OpenAPISpec.fromObject(spec);\n }\n const { openAIFunctions, defaultExecutionMethod } =\n convertOpenAPISpecToOpenAIFunctions(convertedSpec);\n if (defaultExecutionMethod === undefined) {\n throw new Error(\n `Could not parse any valid operations from the provided spec.`\n );\n }\n\n if (!options.llm) {\n throw new Error(\"`llm` option is required\");\n }\n\n const {\n llm = options.llm,\n prompt = ChatPromptTemplate.fromMessages([\n HumanMessagePromptTemplate.fromTemplate(\n \"Use the provided API's to respond to this user query:\\n\\n{query}\"\n ),\n ]),\n requestChain = new SimpleRequestChain({\n requestMethod: async (name, args) =>\n defaultExecutionMethod(name, args, {\n headers: options.headers,\n params: options.params,\n }),\n }),\n llmChainInputs = {},\n verbose,\n ...rest\n } = options;\n const formatChain = new LLMChain({\n llm,\n prompt,\n outputParser: new JsonOutputFunctionsParser({ argsOnly: false }),\n outputKey: \"function\",\n llmKwargs: { functions: openAIFunctions },\n ...llmChainInputs,\n });\n return new SequentialChain({\n chains: [formatChain, requestChain],\n outputVariables: [\"response\"],\n inputVariables: formatChain.inputKeys,\n verbose,\n ...rest,\n });\n}\n"],"mappings":";;;;;;;;;;;;;AA8CA,SAAS,mBACP,QAC6B;AAC7B,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;AAET,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;AAET,KAAI,OAAO,UAAU,KAAA,EACnB,QAAO;;;;;;;;;AAYX,SAAS,UAAU,KAAa,YAA4C;CAC1E,MAAM,yBAAyB,CAAC,GAAG,IAAI,SAAS,WAAW,CAAC,CAAC,KAC1D,UAAU,MAAM,GAClB;CACD,MAAM,YAAoC,EAAE;AAC5C,MAAK,MAAM,aAAa,wBAAwB;EAC9C,MAAM,iBAAiB,UAAU,QAAQ,QAAQ,GAAG,CAAC,QAAQ,OAAO,GAAG;EACvE,MAAM,QAAQ,WAAW;EACzB,IAAI;AACJ,MAAI,MAAM,QAAQ,MAAM,CACtB,KAAI,UAAU,WAAW,IAAI,EAAE;GAC7B,MAAM,YAAY,UAAU,SAAS,IAAI,GAAG,MAAM;AAClD,oBAAiB,IAAI,MAAM,KAAK,UAAU;aACjC,UAAU,WAAW,IAAI,EAAE;GACpC,MAAM,YAAY,UAAU,SAAS,IAAI,GAAG,GAAG,eAAe,KAAK;AACnE,oBAAiB,GAAG,eAAe,GAAG,MAAM,KAAK,UAAU;QAE3D,kBAAiB,MAAM,KAAK,IAAI;WAEzB,OAAO,UAAU,UAAU;GACpC,MAAM,cAAc,UAAU,SAAS,IAAI,GAAG,MAAM;GACpD,MAAM,YAAY,OAAO,QAAQ,MAAM,CAAC,KACrC,CAAC,GAAG,OAAO,IAAI,cAAc,EAC/B;GACD,IAAI;AACJ,OAAI,UAAU,WAAW,IAAI,EAAE;AAC7B,qBAAiB;AACjB,qBAAiB;cACR,UAAU,WAAW,IAAI,EAAE;AACpC,qBAAiB;AACjB,qBAAiB;UACZ;AACL,qBAAiB;AACjB,qBAAiB;;AAEnB,qBAAkB,UAAU,KAAK,eAAe;aAE5C,UAAU,WAAW,IAAI,CAC3B,kBAAiB,IAAI;WACZ,UAAU,WAAW,IAAI,CAClC,kBAAiB,IAAI,eAAe,GAAG;MAEvC,kBAAiB;AAGrB,YAAU,aAAa;;CAEzB,IAAI,eAAe;AACnB,MAAK,MAAM,CAAC,KAAK,aAAa,OAAO,QAAQ,UAAU,CACrD,gBAAe,aAAa,QAAQ,IAAI,IAAI,IAAI,SAAS;AAE3D,QAAO;;;;;;;;AAST,SAAS,iCACP,QACA,MACA;AACA,QAAO,OAAO,QACX,YAAmC,UAAU;EAC5C,IAAI;AACJ,MAAI,MAAM,QAAQ;AAChB,YAAS,KAAK,UAAU,MAAM,OAAO;AACrC,cAAW,WAAW,MAAM,QAAQ,iCAClC,QACA,KACD;aACQ,MAAM,SAAS;GACxB,MAAM,kBAAkB,OAAO,OAAO,MAAM,QAAQ,CAAC,GAAG;AACxD,OAAI,gBACF,UAAS,KAAK,UAAU,gBAAgB;AAE1C,OAAI,CAAC,OACH,QAAO;AAET,OAAI,OAAO,gBAAgB,KAAA,EACzB,QAAO,cAAc,MAAM,eAAe;AAE5C,cAAW,WAAW,MAAM,QAAQ,iCAClC,QACA,KACD;QAED,QAAO;AAET,MAAI,MAAM,YAAY,MAAM,QAAQ,WAAW,SAAS,CACtD,YAAW,SAAS,KAAK,MAAM,KAAK;AAEtC,SAAO;IAET;EACE,MAAM;EACN,YAAY,EAAE;EACd,UAAU,EAAE;EACZ,sBAAsB,EAAE;EACzB,CACF;;;;;;;;AAUH,SAAgB,iCACd,QACA,MACiB;CACjB,MAAM,kBAAkB,mBAAmB,OAAO;AAClD,KAAI,oBAAoB,KAAA,KAAa,OAAO,qBAAqB,KAAA,EAC/D,QAAO,GACJ,kBAAkB,OAAO,iBAAiB,KAAK,MAC9C,iCAAiC,KAAK,UAAU,EAAE,EAAE,KAAK,CAC1D,EACF;AAGH,KAAI,OAAO,SAAS,SAClB,QAAO,OAAO,KAAK,OAAO,cAAc,EAAE,CAAC,CAAC,QACzC,YAAmC,iBAAiB;AACnD,MAAI,CAAC,OAAO,WACV,QAAO;EAET,MAAM,kBAAkB,KAAK,UAAU,OAAO,WAAW,cAAc;AACvE,MAAI,gBAAgB,SAAS,KAAA,EAC3B,QAAO;AAET,aAAW,WAAW,gBAAgB,iCACpC,iBACA,KACD;AACD,MACE,OAAO,UAAU,SAAS,aAAa,IACvC,WAAW,aAAa,KAAA,EAExB,YAAW,SAAS,KAAK,aAAa;AAExC,SAAO;IAET;EACE,MAAM;EACN,YAAY,EAAE;EACd,UAAU,EAAE;EACZ,sBAAsB,EAAE;EACzB,CACF;AAGH,KAAI,OAAO,SAAS,QAElB,QAAO;EACL,MAAM;EACN,OAAO,iCAHY,KAAK,UAAU,OAAO,SAAS,EAAE,CAAC,EAGC,KAAK;EAC3D,UAAU,OAAO;EACjB,UAAU,OAAO;EAClB;AAEH,QAAO,EACL,MAAM,OAAO,QAAQ,UACtB;;;;;;;AAQH,SAAgB,oCAAoC,MAGlD;AACA,KAAI,CAAC,KAAK,SAAS,MACjB,QAAO,EAAE,iBAAiB,EAAE,EAAE;CAEhC,MAAM,kBAAkB,EAAE;CAC1B,MAAM,gBAAiE,EAAE;AACzE,MAAK,MAAM,QAAQ,OAAO,KAAK,KAAK,SAAS,MAAM,EAAE;EACnD,MAAM,iBAAiB,KAAK,qBAAqB,KAAK;AACtD,OAAK,MAAM,UAAU,KAAK,kBAAkB,KAAK,EAAE;GACjD,MAAM,YAAY,KAAK,aAAa,MAAM,OAAO;AACjD,OAAI,CAAC,UACH,QAAO,EAAE,iBAAiB,EAAE,EAAE;GAEhC,MAAM,gCAAgC,eACnC,OAAO,KAAK,0BAA0B,UAAU,CAAC,CACjD,QAEG,iBACA,UACG;AACH,QAAI,CAAC,gBAAgB,MAAM,IACzB,iBAAgB,MAAM,MAAM,EAAE;AAEhC,oBAAgB,MAAM,IAAI,KAAK,MAAM;AACrC,WAAO;MAET,EAAE,CACH;GACH,MAAM,mCAA2D;IAC/D,OAAO;IACP,QAAQ;IACR,QAAQ;IACR,MAAM;IACP;GACD,MAAM,oBAMF,EAAE;AACN,QAAK,MAAM,iBAAiB,OAAO,KACjC,iCACD,CACC,KAAI,8BAA8B,eAChC,mBAAkB,iCAAiC,kBACjD,iCACE,8BAA8B,gBAC9B,KACD;GAGP,MAAM,cAAc,KAAK,2BAA2B,UAAU;AAC9D,OAAI,aAAa,YAAY,KAAA,GAAW;IACtC,MAAM,qBAA4D,EAAE;AACpE,SAAK,MAAM,CAAC,WAAW,oBAAoB,OAAO,QAChD,YAAY,QACb,CACC,KAAI,gBAAgB,WAAW,KAAA,EAE7B,oBAAmB,aAAa,iCADjB,KAAK,UAAU,gBAAgB,OAAO,EAGnD,KACD;IAGL,MAAM,aAAa,OAAO,KAAK,mBAAmB;AAClD,QAAI,WAAW,WAAW,EACxB,mBAAkB,OAAO,mBAAmB,WAAW;aAC9C,WAAW,SAAS,EAC7B,mBAAkB,OAAO,EACvB,OAAO,OAAO,OAAO,mBAAmB,EACzC;;GAGL,MAAM,iBACJ;IACE,MAAMA,gBAAAA,YAAY,sBAAsB,WAAW,MAAM,OAAO;IAChE,aAAa,UAAU,eAAe,UAAU,WAAW;IAC3D,YAAY;KACV,MAAM;KACN,YAAY;KAEZ,UAAU,OAAO,KAAK,kBAAkB;KACzC;IACF;AAEH,mBAAgB,KAAK,eAAe;GACpC,MAAM,WAAW,KAAK,WAAW,IAAI,SAAS,IAAI,IAC7C,KAAK,WAAW,IAAI,MAAM,GAAG,GAAG,GAChC,KAAK,WAAW;AACrB,iBAAc,eAAe,QAAQ;IACnC;IACA,KAAK,UAAU;IAChB;;;AAGL,QAAO;EACL;EACA,wBAAwB,OACtB,MAEA,aACA,YAIG;GACH,MAAM,EACJ,SAAS,eACT,QAAQ,cACR,GAAG,SACD,WAAW,EAAE;GACjB,MAAM,EAAE,QAAQ,QAAQ,cAAc;GACtC,MAAM,gBAAgB,YAAY,UAAU,EAAE;GAC9C,MAAM,iBAAiB,OAAO,KAAK,cAAc,CAAC,QAE/C,cAAmC,YAAY;AAC9C,QACE,cAAc,aAAa,MAC3B,cAAc,aAAa,QAC3B,cAAc,aAAa,KAAA,EAE3B,cAAa,WAAW,cAAc;AAExC,WAAO;MAET,EAAE,CACH;GACD,MAAM,cAAc,IAAI,gBAAgB;IACtC,GAAG;IACH,GAAG;IACJ,CAAC,CAAC,UAAU;GACb,MAAM,aAAa,YAAY;GAC/B,MAAM,eACJ,UAAU,KAAK,WAAW,IACzB,YAAY,SAAS,IAAI,gBAAgB;GAC5C,MAAM,UAAkC,EAAE;GAC1C,IAAI;AACJ,OAAI,YAAY,SAAS,KAAA,GAAW;IAClC,IAAI,cAAc;AAClB,QAAI,OAAO,YAAY,SAAS,UAAU;AACxC,SAAI,OAAO,YAAY,SAAS,SAC9B,eAAc;AAEhB,YAAO,KAAK,UAAU,YAAY,KAAK;UAEvC,QAAO,YAAY;AAErB,YAAQ,kBAAkB;;GAE5B,MAAM,WAAW,MAAM,MAAM,cAAc;IACzC,GAAG;IACH;IACA,SAAS;KACP,GAAG;KACH,GAAG,YAAY;KACf,GAAG;KACJ;IACD;IACA,GAAG;IACJ,CAAC;GACF,IAAI;AACJ,OAAI,SAAS,SAAS,OAAO,SAAS,SAAS,IAC7C,UAAS,GAAG,SAAS,OAAO,IAC1B,SAAS,WACV,OAAO,KAAK,eAAe,KAAK,UAAU,YAAY;OAEvD,UAAS,MAAM,SAAS,MAAM;AAEhC,UAAO;;EAEV;;;;;AAeH,IAAM,qBAAN,cAAiCC,aAAAA,UAAU;CACzC,OAAO,UAAU;AACf,SAAO;;CAGT;CAEA,WAAW;CAEX,YAAY;CAEZ,YAAY,QAA8D;AACxE,SAAO;AACP,OAAK,gBAAgB,OAAO;;CAG9B,IAAI,YAAY;AACd,SAAO,CAAC,KAAK,SAAS;;CAGxB,IAAI,aAAa;AACf,SAAO,CAAC,KAAK,UAAU;;CAGzB,aAAa;AACX,SAAO;;;CAIT,MAAM,MACJ,QACA,aACsB;EACtB,MAAM,gBAAgB,OAAO,KAAK;EAClC,MAAM,aAAa,cAAc;EACjC,MAAM,OAAO,cAAc;EAC3B,MAAM,WAAW,MAAM,KAAK,cAAc,YAAY,KAAK;AAC3D,SAAO,GAAG,KAAK,YAAY,UAAU;;;;;;;;;AAuBzC,eAAsB,mBACpB,MACA,UAA+B,EAAE,EACjC;CACA,IAAI;AACJ,KAAI,OAAO,SAAS,SAClB,KAAI;AACF,kBAAgB,MAAMD,gBAAAA,YAAY,QAAQ,KAAK;SACzC;AACN,MAAI;AACF,mBAAgBA,gBAAAA,YAAY,WAAW,KAAK;UACtC;AACN,SAAM,IAAI,MAAM,oCAAoC,KAAK,GAAG;;;KAIhE,iBAAgBA,gBAAAA,YAAY,WAAW,KAAK;CAE9C,MAAM,EAAE,iBAAiB,2BACvB,oCAAoC,cAAc;AACpD,KAAI,2BAA2B,KAAA,EAC7B,OAAM,IAAI,MACR,+DACD;AAGH,KAAI,CAAC,QAAQ,IACX,OAAM,IAAI,MAAM,2BAA2B;CAG7C,MAAM,EACJ,MAAM,QAAQ,KACd,SAASE,wBAAAA,mBAAmB,aAAa,CACvCC,wBAAAA,2BAA2B,aACzB,mEACD,CACF,CAAC,EACF,eAAe,IAAI,mBAAmB,EACpC,eAAe,OAAO,MAAM,SAC1B,uBAAuB,MAAM,MAAM;EACjC,SAAS,QAAQ;EACjB,QAAQ,QAAQ;EACjB,CAAC,EACL,CAAC,EACF,iBAAiB,EAAE,EACnB,SACA,GAAG,SACD;CACJ,MAAM,cAAc,IAAIC,kBAAAA,SAAS;EAC/B;EACA;EACA,cAAc,IAAIC,yBAAAA,0BAA0B,EAAE,UAAU,OAAO,CAAC;EAChE,WAAW;EACX,WAAW,EAAE,WAAW,iBAAiB;EACzC,GAAG;EACJ,CAAC;AACF,QAAO,IAAIC,yBAAAA,gBAAgB;EACzB,QAAQ,CAAC,aAAa,aAAa;EACnC,iBAAiB,CAAC,WAAW;EAC7B,gBAAgB,YAAY;EAC5B;EACA,GAAG;EACJ,CAAC"}
|