@langchain/classic 1.0.7 → 1.0.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +18 -0
- package/dist/agents/agent.cjs.map +1 -1
- package/dist/agents/agent.js.map +1 -1
- package/dist/agents/chat/outputParser.d.cts +2 -2
- package/dist/agents/chat/outputParser.d.cts.map +1 -1
- package/dist/agents/chat/outputParser.d.ts +2 -2
- package/dist/agents/chat/outputParser.d.ts.map +1 -1
- package/dist/agents/helpers.cjs.map +1 -1
- package/dist/agents/helpers.js.map +1 -1
- package/dist/agents/mrkl/outputParser.d.ts +1 -1
- package/dist/agents/mrkl/outputParser.d.ts.map +1 -1
- package/dist/agents/openai_functions/index.cjs.map +1 -1
- package/dist/agents/openai_functions/index.js.map +1 -1
- package/dist/agents/openai_tools/index.d.ts +2 -2
- package/dist/agents/openai_tools/index.d.ts.map +1 -1
- package/dist/agents/structured_chat/index.d.ts +2 -2
- package/dist/agents/structured_chat/index.d.ts.map +1 -1
- package/dist/agents/tool_calling/index.d.ts +2 -2
- package/dist/agents/tool_calling/index.d.ts.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.cts.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts +1 -1
- package/dist/agents/toolkits/conversational_retrieval/token_buffer_memory.d.ts.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/tool.d.cts +1 -1
- package/dist/agents/toolkits/conversational_retrieval/tool.d.cts.map +1 -1
- package/dist/agents/toolkits/conversational_retrieval/tool.d.ts +1 -1
- package/dist/agents/toolkits/conversational_retrieval/tool.d.ts.map +1 -1
- package/dist/agents/types.cjs.map +1 -1
- package/dist/agents/types.js.map +1 -1
- package/dist/chains/base.cjs.map +1 -1
- package/dist/chains/base.js.map +1 -1
- package/dist/chains/combine_docs_chain.cjs.map +1 -1
- package/dist/chains/combine_docs_chain.js.map +1 -1
- package/dist/chains/llm_chain.cjs.map +1 -1
- package/dist/chains/llm_chain.js.map +1 -1
- package/dist/chains/openai_functions/base.cjs.map +1 -1
- package/dist/chains/openai_functions/base.js.map +1 -1
- package/dist/chains/openai_functions/openapi.cjs.map +1 -1
- package/dist/chains/openai_functions/openapi.js.map +1 -1
- package/dist/chains/query_constructor/index.cjs.map +1 -1
- package/dist/chains/query_constructor/index.js.map +1 -1
- package/dist/chains/query_constructor/parser.cjs.map +1 -1
- package/dist/chains/query_constructor/parser.js.map +1 -1
- package/dist/chains/router/multi_prompt.cjs.map +1 -1
- package/dist/chains/router/multi_prompt.js.map +1 -1
- package/dist/chains/transform.cjs.map +1 -1
- package/dist/chains/transform.js.map +1 -1
- package/dist/chat_models/universal.cjs +3 -2
- package/dist/chat_models/universal.cjs.map +1 -1
- package/dist/chat_models/universal.js +3 -2
- package/dist/chat_models/universal.js.map +1 -1
- package/dist/document_loaders/fs/json.cjs.map +1 -1
- package/dist/document_loaders/fs/json.js.map +1 -1
- package/dist/embeddings/cache_backed.cjs.map +1 -1
- package/dist/embeddings/cache_backed.js.map +1 -1
- package/dist/evaluation/agents/trajectory.d.ts +2 -2
- package/dist/evaluation/agents/trajectory.d.ts.map +1 -1
- package/dist/evaluation/base.cjs.map +1 -1
- package/dist/evaluation/base.js.map +1 -1
- package/dist/evaluation/comparison/pairwise.d.ts +3 -3
- package/dist/evaluation/comparison/pairwise.d.ts.map +1 -1
- package/dist/evaluation/criteria/criteria.d.ts +3 -3
- package/dist/evaluation/criteria/criteria.d.ts.map +1 -1
- package/dist/experimental/autogpt/prompt.d.cts +2 -2
- package/dist/experimental/autogpt/prompt.d.cts.map +1 -1
- package/dist/experimental/autogpt/prompt.d.ts +2 -2
- package/dist/experimental/autogpt/prompt.d.ts.map +1 -1
- package/dist/experimental/masking/regex_masking_transformer.cjs.map +1 -1
- package/dist/experimental/masking/regex_masking_transformer.js.map +1 -1
- package/dist/experimental/openai_assistant/index.cjs.map +1 -1
- package/dist/experimental/openai_assistant/index.js.map +1 -1
- package/dist/experimental/plan_and_execute/agent_executor.cjs.map +1 -1
- package/dist/experimental/plan_and_execute/agent_executor.js.map +1 -1
- package/dist/experimental/prompts/custom_format.cjs.map +1 -1
- package/dist/experimental/prompts/custom_format.d.cts.map +1 -1
- package/dist/experimental/prompts/custom_format.js.map +1 -1
- package/dist/experimental/prompts/handlebars.cjs.map +1 -1
- package/dist/experimental/prompts/handlebars.d.cts.map +1 -1
- package/dist/experimental/prompts/handlebars.js.map +1 -1
- package/dist/hub/index.cjs +8 -1
- package/dist/hub/index.cjs.map +1 -1
- package/dist/hub/index.d.cts +9 -0
- package/dist/hub/index.d.cts.map +1 -1
- package/dist/hub/index.d.ts +9 -0
- package/dist/hub/index.d.ts.map +1 -1
- package/dist/hub/index.js +8 -1
- package/dist/hub/index.js.map +1 -1
- package/dist/hub/node.cjs +30 -3
- package/dist/hub/node.cjs.map +1 -1
- package/dist/hub/node.d.cts +9 -0
- package/dist/hub/node.d.cts.map +1 -1
- package/dist/hub/node.d.ts +9 -0
- package/dist/hub/node.d.ts.map +1 -1
- package/dist/hub/node.js +30 -3
- package/dist/hub/node.js.map +1 -1
- package/dist/load/import_constants.cjs +24 -24
- package/dist/load/import_constants.cjs.map +1 -1
- package/dist/load/import_constants.js +24 -24
- package/dist/load/import_constants.js.map +1 -1
- package/dist/load/import_map.cjs +44 -44
- package/dist/load/import_map.cjs.map +1 -1
- package/dist/load/import_map.js +52 -52
- package/dist/load/import_map.js.map +1 -1
- package/dist/load/index.cjs +3 -2
- package/dist/load/index.cjs.map +1 -1
- package/dist/load/index.d.cts +1 -1
- package/dist/load/index.d.cts.map +1 -1
- package/dist/load/index.d.ts +1 -1
- package/dist/load/index.d.ts.map +1 -1
- package/dist/load/index.js +3 -2
- package/dist/load/index.js.map +1 -1
- package/dist/output_parsers/openai_functions.cjs.map +1 -1
- package/dist/output_parsers/openai_functions.js.map +1 -1
- package/dist/output_parsers/regex.cjs.map +1 -1
- package/dist/output_parsers/regex.js.map +1 -1
- package/dist/output_parsers/router.cjs.map +1 -1
- package/dist/output_parsers/router.js.map +1 -1
- package/dist/output_parsers/structured.cjs.map +1 -1
- package/dist/output_parsers/structured.js.map +1 -1
- package/dist/retrievers/document_compressors/index.cjs.map +1 -1
- package/dist/retrievers/document_compressors/index.js.map +1 -1
- package/dist/retrievers/hyde.cjs +1 -1
- package/dist/retrievers/hyde.cjs.map +1 -1
- package/dist/retrievers/hyde.js +1 -1
- package/dist/retrievers/hyde.js.map +1 -1
- package/dist/retrievers/matryoshka_retriever.cjs +1 -1
- package/dist/retrievers/matryoshka_retriever.cjs.map +1 -1
- package/dist/retrievers/matryoshka_retriever.js +1 -1
- package/dist/retrievers/matryoshka_retriever.js.map +1 -1
- package/dist/retrievers/score_threshold.cjs.map +1 -1
- package/dist/retrievers/score_threshold.js.map +1 -1
- package/dist/retrievers/self_query/index.cjs +1 -1
- package/dist/retrievers/self_query/index.d.ts +1 -1
- package/dist/retrievers/self_query/index.js +1 -1
- package/dist/smith/config.cjs.map +1 -1
- package/dist/smith/config.js.map +1 -1
- package/dist/smith/runner_utils.cjs.map +1 -1
- package/dist/smith/runner_utils.js.map +1 -1
- package/dist/tools/retriever.d.cts +1 -1
- package/dist/tools/retriever.d.cts.map +1 -1
- package/dist/tools/retriever.d.ts +1 -1
- package/dist/tools/retriever.d.ts.map +1 -1
- package/dist/util/sql_utils.cjs.map +1 -1
- package/dist/util/sql_utils.js.map +1 -1
- package/dist/vectorstores/memory.cjs +1 -1
- package/dist/vectorstores/memory.js +1 -1
- package/package.json +17 -18
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"openai_functions.cjs","names":["BaseLLMOutputParser","config?: { argsOnly?: boolean }","generations: Generation[] | ChatGeneration[]","BaseCumulativeTransformOutputParser","config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput","prev: unknown | undefined","next: unknown","generations: ChatGeneration[]","text: string","OutputParserException","fields: { attrName: string }"],"sources":["../../src/output_parsers/openai_functions.ts"],"sourcesContent":["import { type JsonSchema7ObjectType } from \"@langchain/core/utils/json_schema\";\nimport {\n compare,\n type Operation as JSONPatchOperation,\n} from \"@langchain/core/utils/json_patch\";\n\nimport { ChatGeneration, Generation } from \"@langchain/core/outputs\";\nimport {\n BaseCumulativeTransformOutputParser,\n type BaseCumulativeTransformOutputParserInput,\n BaseLLMOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport { parsePartialJson } from \"@langchain/core/output_parsers\";\nimport { Optional } from \"../types/type-utils.js\";\n\n/**\n * Represents optional parameters for a function in a JSON Schema.\n */\nexport type FunctionParameters = Optional<\n JsonSchema7ObjectType,\n \"additionalProperties\"\n>;\n\n/**\n * Class for parsing the output of an LLM. Can be configured to return\n * only the arguments of the function call in the output.\n */\nexport class OutputFunctionsParser extends BaseLLMOutputParser<string> {\n static lc_name() {\n return \"OutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n argsOnly = true;\n\n constructor(config?: { argsOnly?: boolean }) {\n super();\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n }\n\n /**\n * Parses the output and returns a string representation of the function\n * call or its arguments.\n * @param generations The output of the LLM to parse.\n * @returns A string representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<string> {\n if (\"message\" in generations[0]) {\n const gen = generations[0] as ChatGeneration;\n const functionCall = gen.message.additional_kwargs.function_call;\n if (!functionCall) {\n throw new Error(\n `No function_call in message ${JSON.stringify(generations)}`\n );\n }\n if (!functionCall.arguments) {\n throw new Error(\n `No arguments in function_call ${JSON.stringify(generations)}`\n );\n }\n if (this.argsOnly) {\n return functionCall.arguments;\n }\n return JSON.stringify(functionCall);\n } else {\n throw new Error(\n `No message in generations ${JSON.stringify(generations)}`\n );\n }\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object. Uses an\n * instance of `OutputFunctionsParser` to parse the output.\n */\nexport class JsonOutputFunctionsParser<\n Output extends object = object
|
|
1
|
+
{"version":3,"file":"openai_functions.cjs","names":["BaseLLMOutputParser","config?: { argsOnly?: boolean }","generations: Generation[] | ChatGeneration[]","BaseCumulativeTransformOutputParser","config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput","prev: unknown | undefined","next: unknown","generations: ChatGeneration[]","text: string","OutputParserException","fields: { attrName: string }"],"sources":["../../src/output_parsers/openai_functions.ts"],"sourcesContent":["import { type JsonSchema7ObjectType } from \"@langchain/core/utils/json_schema\";\nimport {\n compare,\n type Operation as JSONPatchOperation,\n} from \"@langchain/core/utils/json_patch\";\n\nimport { ChatGeneration, Generation } from \"@langchain/core/outputs\";\nimport {\n BaseCumulativeTransformOutputParser,\n type BaseCumulativeTransformOutputParserInput,\n BaseLLMOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport { parsePartialJson } from \"@langchain/core/output_parsers\";\nimport { Optional } from \"../types/type-utils.js\";\n\n/**\n * Represents optional parameters for a function in a JSON Schema.\n */\nexport type FunctionParameters = Optional<\n JsonSchema7ObjectType,\n \"additionalProperties\"\n>;\n\n/**\n * Class for parsing the output of an LLM. Can be configured to return\n * only the arguments of the function call in the output.\n */\nexport class OutputFunctionsParser extends BaseLLMOutputParser<string> {\n static lc_name() {\n return \"OutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n argsOnly = true;\n\n constructor(config?: { argsOnly?: boolean }) {\n super();\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n }\n\n /**\n * Parses the output and returns a string representation of the function\n * call or its arguments.\n * @param generations The output of the LLM to parse.\n * @returns A string representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<string> {\n if (\"message\" in generations[0]) {\n const gen = generations[0] as ChatGeneration;\n const functionCall = gen.message.additional_kwargs.function_call;\n if (!functionCall) {\n throw new Error(\n `No function_call in message ${JSON.stringify(generations)}`\n );\n }\n if (!functionCall.arguments) {\n throw new Error(\n `No arguments in function_call ${JSON.stringify(generations)}`\n );\n }\n if (this.argsOnly) {\n return functionCall.arguments;\n }\n return JSON.stringify(functionCall);\n } else {\n throw new Error(\n `No message in generations ${JSON.stringify(generations)}`\n );\n }\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object. Uses an\n * instance of `OutputFunctionsParser` to parse the output.\n */\nexport class JsonOutputFunctionsParser<\n Output extends object = object,\n> extends BaseCumulativeTransformOutputParser<Output> {\n static lc_name() {\n return \"JsonOutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n outputParser: OutputFunctionsParser;\n\n argsOnly = true;\n\n constructor(\n config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput\n ) {\n super(config);\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n this.outputParser = new OutputFunctionsParser(config);\n }\n\n protected _diff(\n prev: unknown | undefined,\n next: unknown\n ): JSONPatchOperation[] | undefined {\n if (!next) {\n return undefined;\n }\n const ops = compare(prev ?? {}, next);\n return ops;\n }\n\n async parsePartialResult(\n generations: ChatGeneration[]\n ): Promise<Output | undefined> {\n const generation = generations[0];\n if (!generation.message) {\n return undefined;\n }\n const { message } = generation;\n const functionCall = message.additional_kwargs.function_call;\n if (!functionCall) {\n return undefined;\n }\n if (this.argsOnly) {\n return parsePartialJson(functionCall.arguments);\n }\n\n return {\n ...functionCall,\n arguments: parsePartialJson(functionCall.arguments),\n } as Output;\n }\n\n /**\n * Parses the output and returns a JSON object. If `argsOnly` is true,\n * only the arguments of the function call are returned.\n * @param generations The output of the LLM to parse.\n * @returns A JSON object representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<Output> {\n const result = await this.outputParser.parseResult(generations);\n if (!result) {\n throw new Error(\n `No result from \"OutputFunctionsParser\" ${JSON.stringify(generations)}`\n );\n }\n return this.parse(result);\n }\n\n async parse(text: string): Promise<Output> {\n try {\n const parsedResult = JSON.parse(text);\n if (this.argsOnly) {\n return parsedResult;\n }\n parsedResult.arguments = JSON.parse(parsedResult.arguments);\n return parsedResult;\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`\n );\n }\n }\n\n getFormatInstructions(): string {\n return \"\";\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object and returning\n * a specific attribute. Uses an instance of `JsonOutputFunctionsParser`\n * to parse the output.\n */\nexport class JsonKeyOutputFunctionsParser<\n T = object,\n> extends BaseLLMOutputParser<T> {\n static lc_name() {\n return \"JsonKeyOutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n outputParser = new JsonOutputFunctionsParser();\n\n attrName: string;\n\n get lc_aliases() {\n return {\n attrName: \"key_name\",\n };\n }\n\n constructor(fields: { attrName: string }) {\n super(fields);\n this.attrName = fields.attrName;\n }\n\n /**\n * Parses the output and returns a specific attribute of the parsed JSON\n * object.\n * @param generations The output of the LLM to parse.\n * @returns The value of a specific attribute of the parsed JSON object.\n */\n async parseResult(generations: Generation[] | ChatGeneration[]): Promise<T> {\n const result = await this.outputParser.parseResult(generations);\n return result[this.attrName as keyof typeof result] as T;\n }\n}\n"],"mappings":";;;;;;;;;AA4BA,IAAa,wBAAb,cAA2CA,oDAA4B;CACrE,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB,WAAW;CAEX,YAAYC,QAAiC;EAC3C,OAAO;EACP,KAAK,WAAW,QAAQ,YAAY,KAAK;CAC1C;;;;;;;CAQD,MAAM,YACJC,aACiB;AACjB,MAAI,aAAa,YAAY,IAAI;GAC/B,MAAM,MAAM,YAAY;GACxB,MAAM,eAAe,IAAI,QAAQ,kBAAkB;AACnD,OAAI,CAAC,aACH,OAAM,IAAI,MACR,CAAC,4BAA4B,EAAE,KAAK,UAAU,YAAY,EAAE;AAGhE,OAAI,CAAC,aAAa,UAChB,OAAM,IAAI,MACR,CAAC,8BAA8B,EAAE,KAAK,UAAU,YAAY,EAAE;AAGlE,OAAI,KAAK,SACP,QAAO,aAAa;AAEtB,UAAO,KAAK,UAAU,aAAa;EACpC,MACC,OAAM,IAAI,MACR,CAAC,0BAA0B,EAAE,KAAK,UAAU,YAAY,EAAE;CAG/D;AACF;;;;;AAMD,IAAa,4BAAb,cAEUC,oEAA4C;CACpD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB;CAEA,WAAW;CAEX,YACEC,QACA;EACA,MAAM,OAAO;EACb,KAAK,WAAW,QAAQ,YAAY,KAAK;EACzC,KAAK,eAAe,IAAI,sBAAsB;CAC/C;CAED,AAAU,MACRC,MACAC,MACkC;AAClC,MAAI,CAAC,KACH,QAAO;EAET,MAAM,qDAAc,QAAQ,CAAE,GAAE,KAAK;AACrC,SAAO;CACR;CAED,MAAM,mBACJC,aAC6B;EAC7B,MAAM,aAAa,YAAY;AAC/B,MAAI,CAAC,WAAW,QACd,QAAO;EAET,MAAM,EAAE,SAAS,GAAG;EACpB,MAAM,eAAe,QAAQ,kBAAkB;AAC/C,MAAI,CAAC,aACH,QAAO;AAET,MAAI,KAAK,SACP,8DAAwB,aAAa,UAAU;AAGjD,SAAO;GACL,GAAG;GACH,iEAA4B,aAAa,UAAU;EACpD;CACF;;;;;;;CAQD,MAAM,YACJL,aACiB;EACjB,MAAM,SAAS,MAAM,KAAK,aAAa,YAAY,YAAY;AAC/D,MAAI,CAAC,OACH,OAAM,IAAI,MACR,CAAC,uCAAuC,EAAE,KAAK,UAAU,YAAY,EAAE;AAG3E,SAAO,KAAK,MAAM,OAAO;CAC1B;CAED,MAAM,MAAMM,MAA+B;AACzC,MAAI;GACF,MAAM,eAAe,KAAK,MAAM,KAAK;AACrC,OAAI,KAAK,SACP,QAAO;GAET,aAAa,YAAY,KAAK,MAAM,aAAa,UAAU;AAC3D,UAAO;EACR,SAAQ,GAAG;AACV,SAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG;EAElD;CACF;CAED,wBAAgC;AAC9B,SAAO;CACR;AACF;;;;;;AAOD,IAAa,+BAAb,cAEUT,oDAAuB;CAC/B,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB,eAAe,IAAI;CAEnB;CAEA,IAAI,aAAa;AACf,SAAO,EACL,UAAU,WACX;CACF;CAED,YAAYU,QAA8B;EACxC,MAAM,OAAO;EACb,KAAK,WAAW,OAAO;CACxB;;;;;;;CAQD,MAAM,YAAYR,aAA0D;EAC1E,MAAM,SAAS,MAAM,KAAK,aAAa,YAAY,YAAY;AAC/D,SAAO,OAAO,KAAK;CACpB;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"openai_functions.js","names":["config?: { argsOnly?: boolean }","generations: Generation[] | ChatGeneration[]","config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput","prev: unknown | undefined","next: unknown","generations: ChatGeneration[]","text: string","fields: { attrName: string }"],"sources":["../../src/output_parsers/openai_functions.ts"],"sourcesContent":["import { type JsonSchema7ObjectType } from \"@langchain/core/utils/json_schema\";\nimport {\n compare,\n type Operation as JSONPatchOperation,\n} from \"@langchain/core/utils/json_patch\";\n\nimport { ChatGeneration, Generation } from \"@langchain/core/outputs\";\nimport {\n BaseCumulativeTransformOutputParser,\n type BaseCumulativeTransformOutputParserInput,\n BaseLLMOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport { parsePartialJson } from \"@langchain/core/output_parsers\";\nimport { Optional } from \"../types/type-utils.js\";\n\n/**\n * Represents optional parameters for a function in a JSON Schema.\n */\nexport type FunctionParameters = Optional<\n JsonSchema7ObjectType,\n \"additionalProperties\"\n>;\n\n/**\n * Class for parsing the output of an LLM. Can be configured to return\n * only the arguments of the function call in the output.\n */\nexport class OutputFunctionsParser extends BaseLLMOutputParser<string> {\n static lc_name() {\n return \"OutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n argsOnly = true;\n\n constructor(config?: { argsOnly?: boolean }) {\n super();\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n }\n\n /**\n * Parses the output and returns a string representation of the function\n * call or its arguments.\n * @param generations The output of the LLM to parse.\n * @returns A string representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<string> {\n if (\"message\" in generations[0]) {\n const gen = generations[0] as ChatGeneration;\n const functionCall = gen.message.additional_kwargs.function_call;\n if (!functionCall) {\n throw new Error(\n `No function_call in message ${JSON.stringify(generations)}`\n );\n }\n if (!functionCall.arguments) {\n throw new Error(\n `No arguments in function_call ${JSON.stringify(generations)}`\n );\n }\n if (this.argsOnly) {\n return functionCall.arguments;\n }\n return JSON.stringify(functionCall);\n } else {\n throw new Error(\n `No message in generations ${JSON.stringify(generations)}`\n );\n }\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object. Uses an\n * instance of `OutputFunctionsParser` to parse the output.\n */\nexport class JsonOutputFunctionsParser<\n Output extends object = object
|
|
1
|
+
{"version":3,"file":"openai_functions.js","names":["config?: { argsOnly?: boolean }","generations: Generation[] | ChatGeneration[]","config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput","prev: unknown | undefined","next: unknown","generations: ChatGeneration[]","text: string","fields: { attrName: string }"],"sources":["../../src/output_parsers/openai_functions.ts"],"sourcesContent":["import { type JsonSchema7ObjectType } from \"@langchain/core/utils/json_schema\";\nimport {\n compare,\n type Operation as JSONPatchOperation,\n} from \"@langchain/core/utils/json_patch\";\n\nimport { ChatGeneration, Generation } from \"@langchain/core/outputs\";\nimport {\n BaseCumulativeTransformOutputParser,\n type BaseCumulativeTransformOutputParserInput,\n BaseLLMOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport { parsePartialJson } from \"@langchain/core/output_parsers\";\nimport { Optional } from \"../types/type-utils.js\";\n\n/**\n * Represents optional parameters for a function in a JSON Schema.\n */\nexport type FunctionParameters = Optional<\n JsonSchema7ObjectType,\n \"additionalProperties\"\n>;\n\n/**\n * Class for parsing the output of an LLM. Can be configured to return\n * only the arguments of the function call in the output.\n */\nexport class OutputFunctionsParser extends BaseLLMOutputParser<string> {\n static lc_name() {\n return \"OutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n argsOnly = true;\n\n constructor(config?: { argsOnly?: boolean }) {\n super();\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n }\n\n /**\n * Parses the output and returns a string representation of the function\n * call or its arguments.\n * @param generations The output of the LLM to parse.\n * @returns A string representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<string> {\n if (\"message\" in generations[0]) {\n const gen = generations[0] as ChatGeneration;\n const functionCall = gen.message.additional_kwargs.function_call;\n if (!functionCall) {\n throw new Error(\n `No function_call in message ${JSON.stringify(generations)}`\n );\n }\n if (!functionCall.arguments) {\n throw new Error(\n `No arguments in function_call ${JSON.stringify(generations)}`\n );\n }\n if (this.argsOnly) {\n return functionCall.arguments;\n }\n return JSON.stringify(functionCall);\n } else {\n throw new Error(\n `No message in generations ${JSON.stringify(generations)}`\n );\n }\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object. Uses an\n * instance of `OutputFunctionsParser` to parse the output.\n */\nexport class JsonOutputFunctionsParser<\n Output extends object = object,\n> extends BaseCumulativeTransformOutputParser<Output> {\n static lc_name() {\n return \"JsonOutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n outputParser: OutputFunctionsParser;\n\n argsOnly = true;\n\n constructor(\n config?: { argsOnly?: boolean } & BaseCumulativeTransformOutputParserInput\n ) {\n super(config);\n this.argsOnly = config?.argsOnly ?? this.argsOnly;\n this.outputParser = new OutputFunctionsParser(config);\n }\n\n protected _diff(\n prev: unknown | undefined,\n next: unknown\n ): JSONPatchOperation[] | undefined {\n if (!next) {\n return undefined;\n }\n const ops = compare(prev ?? {}, next);\n return ops;\n }\n\n async parsePartialResult(\n generations: ChatGeneration[]\n ): Promise<Output | undefined> {\n const generation = generations[0];\n if (!generation.message) {\n return undefined;\n }\n const { message } = generation;\n const functionCall = message.additional_kwargs.function_call;\n if (!functionCall) {\n return undefined;\n }\n if (this.argsOnly) {\n return parsePartialJson(functionCall.arguments);\n }\n\n return {\n ...functionCall,\n arguments: parsePartialJson(functionCall.arguments),\n } as Output;\n }\n\n /**\n * Parses the output and returns a JSON object. If `argsOnly` is true,\n * only the arguments of the function call are returned.\n * @param generations The output of the LLM to parse.\n * @returns A JSON object representation of the function call or its arguments.\n */\n async parseResult(\n generations: Generation[] | ChatGeneration[]\n ): Promise<Output> {\n const result = await this.outputParser.parseResult(generations);\n if (!result) {\n throw new Error(\n `No result from \"OutputFunctionsParser\" ${JSON.stringify(generations)}`\n );\n }\n return this.parse(result);\n }\n\n async parse(text: string): Promise<Output> {\n try {\n const parsedResult = JSON.parse(text);\n if (this.argsOnly) {\n return parsedResult;\n }\n parsedResult.arguments = JSON.parse(parsedResult.arguments);\n return parsedResult;\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`\n );\n }\n }\n\n getFormatInstructions(): string {\n return \"\";\n }\n}\n\n/**\n * Class for parsing the output of an LLM into a JSON object and returning\n * a specific attribute. Uses an instance of `JsonOutputFunctionsParser`\n * to parse the output.\n */\nexport class JsonKeyOutputFunctionsParser<\n T = object,\n> extends BaseLLMOutputParser<T> {\n static lc_name() {\n return \"JsonKeyOutputFunctionsParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"openai_functions\"];\n\n lc_serializable = true;\n\n outputParser = new JsonOutputFunctionsParser();\n\n attrName: string;\n\n get lc_aliases() {\n return {\n attrName: \"key_name\",\n };\n }\n\n constructor(fields: { attrName: string }) {\n super(fields);\n this.attrName = fields.attrName;\n }\n\n /**\n * Parses the output and returns a specific attribute of the parsed JSON\n * object.\n * @param generations The output of the LLM to parse.\n * @returns The value of a specific attribute of the parsed JSON object.\n */\n async parseResult(generations: Generation[] | ChatGeneration[]): Promise<T> {\n const result = await this.outputParser.parseResult(generations);\n return result[this.attrName as keyof typeof result] as T;\n }\n}\n"],"mappings":";;;;;;;;AA4BA,IAAa,wBAAb,cAA2C,oBAA4B;CACrE,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB,WAAW;CAEX,YAAYA,QAAiC;EAC3C,OAAO;EACP,KAAK,WAAW,QAAQ,YAAY,KAAK;CAC1C;;;;;;;CAQD,MAAM,YACJC,aACiB;AACjB,MAAI,aAAa,YAAY,IAAI;GAC/B,MAAM,MAAM,YAAY;GACxB,MAAM,eAAe,IAAI,QAAQ,kBAAkB;AACnD,OAAI,CAAC,aACH,OAAM,IAAI,MACR,CAAC,4BAA4B,EAAE,KAAK,UAAU,YAAY,EAAE;AAGhE,OAAI,CAAC,aAAa,UAChB,OAAM,IAAI,MACR,CAAC,8BAA8B,EAAE,KAAK,UAAU,YAAY,EAAE;AAGlE,OAAI,KAAK,SACP,QAAO,aAAa;AAEtB,UAAO,KAAK,UAAU,aAAa;EACpC,MACC,OAAM,IAAI,MACR,CAAC,0BAA0B,EAAE,KAAK,UAAU,YAAY,EAAE;CAG/D;AACF;;;;;AAMD,IAAa,4BAAb,cAEU,oCAA4C;CACpD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB;CAEA,WAAW;CAEX,YACEC,QACA;EACA,MAAM,OAAO;EACb,KAAK,WAAW,QAAQ,YAAY,KAAK;EACzC,KAAK,eAAe,IAAI,sBAAsB;CAC/C;CAED,AAAU,MACRC,MACAC,MACkC;AAClC,MAAI,CAAC,KACH,QAAO;EAET,MAAM,MAAM,QAAQ,QAAQ,CAAE,GAAE,KAAK;AACrC,SAAO;CACR;CAED,MAAM,mBACJC,aAC6B;EAC7B,MAAM,aAAa,YAAY;AAC/B,MAAI,CAAC,WAAW,QACd,QAAO;EAET,MAAM,EAAE,SAAS,GAAG;EACpB,MAAM,eAAe,QAAQ,kBAAkB;AAC/C,MAAI,CAAC,aACH,QAAO;AAET,MAAI,KAAK,SACP,QAAO,iBAAiB,aAAa,UAAU;AAGjD,SAAO;GACL,GAAG;GACH,WAAW,iBAAiB,aAAa,UAAU;EACpD;CACF;;;;;;;CAQD,MAAM,YACJJ,aACiB;EACjB,MAAM,SAAS,MAAM,KAAK,aAAa,YAAY,YAAY;AAC/D,MAAI,CAAC,OACH,OAAM,IAAI,MACR,CAAC,uCAAuC,EAAE,KAAK,UAAU,YAAY,EAAE;AAG3E,SAAO,KAAK,MAAM,OAAO;CAC1B;CAED,MAAM,MAAMK,MAA+B;AACzC,MAAI;GACF,MAAM,eAAe,KAAK,MAAM,KAAK;AACrC,OAAI,KAAK,SACP,QAAO;GAET,aAAa,YAAY,KAAK,MAAM,aAAa,UAAU;AAC3D,UAAO;EACR,SAAQ,GAAG;AACV,SAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG;EAElD;CACF;CAED,wBAAgC;AAC9B,SAAO;CACR;AACF;;;;;;AAOD,IAAa,+BAAb,cAEU,oBAAuB;CAC/B,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAmB;CAElE,kBAAkB;CAElB,eAAe,IAAI;CAEnB;CAEA,IAAI,aAAa;AACf,SAAO,EACL,UAAU,WACX;CACF;CAED,YAAYC,QAA8B;EACxC,MAAM,OAAO;EACb,KAAK,WAAW,OAAO;CACxB;;;;;;;CAQD,MAAM,YAAYN,aAA0D;EAC1E,MAAM,SAAS,MAAM,KAAK,aAAa,YAAY,YAAY;AAC/D,SAAO,OAAO,KAAK;CACpB;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"regex.cjs","names":["BaseOutputParser","fields: string | RegExp | RegexParserFields","outputKeys?: string[]","defaultOutputKey?: string","text: string","OutputParserException"],"sources":["../../src/output_parsers/regex.ts"],"sourcesContent":["import {\n BaseOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport type { SerializedFields } from \"../load/map_keys.js\";\n\nexport interface RegExpFields {\n pattern: string;\n flags?: string;\n}\n\n/**\n * Interface for the fields required to create a RegexParser instance.\n */\nexport interface RegexParserFields {\n regex: string | RegExp | RegExpFields;\n outputKeys: string[];\n defaultOutputKey?: string;\n}\n\n/**\n * Class to parse the output of an LLM call into a dictionary.\n * @augments BaseOutputParser\n */\nexport class RegexParser extends BaseOutputParser<Record<string, string>> {\n static lc_name() {\n return \"RegexParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"regex\"];\n\n lc_serializable = true;\n\n get lc_attributes(): SerializedFields | undefined {\n return {\n regex: this.lc_kwargs.regex,\n };\n }\n\n regex: string | RegExp;\n\n outputKeys: string[];\n\n defaultOutputKey?: string;\n\n constructor(fields: RegexParserFields);\n\n constructor(\n regex: string | RegExp,\n outputKeys: string[],\n defaultOutputKey?: string\n );\n\n constructor(\n fields: string | RegExp | RegexParserFields,\n outputKeys?: string[],\n defaultOutputKey?: string\n ) {\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (typeof fields === \"string\" || fields instanceof RegExp) {\n // eslint-disable-next-line no-param-reassign\n fields = { regex: fields, outputKeys: outputKeys!, defaultOutputKey };\n }\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (fields.regex instanceof RegExp) {\n fields.regex = {\n pattern: fields.regex.source,\n flags: fields.regex.flags,\n };\n }\n super(fields);\n this.regex =\n typeof fields.regex === \"string\"\n ? new RegExp(fields.regex)\n : \"pattern\" in fields.regex\n
|
|
1
|
+
{"version":3,"file":"regex.cjs","names":["BaseOutputParser","fields: string | RegExp | RegexParserFields","outputKeys?: string[]","defaultOutputKey?: string","text: string","OutputParserException"],"sources":["../../src/output_parsers/regex.ts"],"sourcesContent":["import {\n BaseOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport type { SerializedFields } from \"../load/map_keys.js\";\n\nexport interface RegExpFields {\n pattern: string;\n flags?: string;\n}\n\n/**\n * Interface for the fields required to create a RegexParser instance.\n */\nexport interface RegexParserFields {\n regex: string | RegExp | RegExpFields;\n outputKeys: string[];\n defaultOutputKey?: string;\n}\n\n/**\n * Class to parse the output of an LLM call into a dictionary.\n * @augments BaseOutputParser\n */\nexport class RegexParser extends BaseOutputParser<Record<string, string>> {\n static lc_name() {\n return \"RegexParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"regex\"];\n\n lc_serializable = true;\n\n get lc_attributes(): SerializedFields | undefined {\n return {\n regex: this.lc_kwargs.regex,\n };\n }\n\n regex: string | RegExp;\n\n outputKeys: string[];\n\n defaultOutputKey?: string;\n\n constructor(fields: RegexParserFields);\n\n constructor(\n regex: string | RegExp,\n outputKeys: string[],\n defaultOutputKey?: string\n );\n\n constructor(\n fields: string | RegExp | RegexParserFields,\n outputKeys?: string[],\n defaultOutputKey?: string\n ) {\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (typeof fields === \"string\" || fields instanceof RegExp) {\n // eslint-disable-next-line no-param-reassign\n fields = { regex: fields, outputKeys: outputKeys!, defaultOutputKey };\n }\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (fields.regex instanceof RegExp) {\n fields.regex = {\n pattern: fields.regex.source,\n flags: fields.regex.flags,\n };\n }\n super(fields);\n this.regex =\n typeof fields.regex === \"string\"\n ? new RegExp(fields.regex)\n : \"pattern\" in fields.regex\n ? new RegExp(fields.regex.pattern, fields.regex.flags)\n : fields.regex;\n this.outputKeys = fields.outputKeys;\n this.defaultOutputKey = fields.defaultOutputKey;\n }\n\n _type() {\n return \"regex_parser\";\n }\n\n /**\n * Parses the given text using the regex pattern and returns a dictionary\n * with the parsed output. If the regex pattern does not match the text\n * and no defaultOutputKey is provided, throws an OutputParserException.\n * @param text The text to be parsed.\n * @returns A dictionary with the parsed output.\n */\n async parse(text: string): Promise<Record<string, string>> {\n const match = text.match(this.regex);\n if (match) {\n return this.outputKeys.reduce(\n (acc, key, index) => {\n acc[key] = match[index + 1];\n return acc;\n },\n {} as Record<string, string>\n );\n }\n\n if (this.defaultOutputKey === undefined) {\n throw new OutputParserException(`Could not parse output: ${text}`, text);\n }\n\n return this.outputKeys.reduce(\n (acc, key) => {\n acc[key] = key === this.defaultOutputKey ? text : \"\";\n return acc;\n },\n {} as Record<string, string>\n );\n }\n\n /**\n * Returns a string with instructions on how the LLM output should be\n * formatted to match the regex pattern.\n * @returns A string with formatting instructions.\n */\n getFormatInstructions(): string {\n return `Your response should match the following regex: ${this.regex}`;\n }\n}\n"],"mappings":";;;;;;;;AAwBA,IAAa,cAAb,cAAiCA,iDAAyC;CACxE,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAQ;CAEvD,kBAAkB;CAElB,IAAI,gBAA8C;AAChD,SAAO,EACL,OAAO,KAAK,UAAU,MACvB;CACF;CAED;CAEA;CAEA;CAUA,YACEC,QACAC,YACAC,kBACA;AAEA,MAAI,OAAO,WAAW,YAAY,kBAAkB,QAElD,SAAS;GAAE,OAAO;GAAoB;GAAa;EAAkB;AAGvE,MAAI,OAAO,iBAAiB,QAC1B,OAAO,QAAQ;GACb,SAAS,OAAO,MAAM;GACtB,OAAO,OAAO,MAAM;EACrB;EAEH,MAAM,OAAO;EACb,KAAK,QACH,OAAO,OAAO,UAAU,WACpB,IAAI,OAAO,OAAO,SAClB,aAAa,OAAO,QAClB,IAAI,OAAO,OAAO,MAAM,SAAS,OAAO,MAAM,SAC9C,OAAO;EACf,KAAK,aAAa,OAAO;EACzB,KAAK,mBAAmB,OAAO;CAChC;CAED,QAAQ;AACN,SAAO;CACR;;;;;;;;CASD,MAAM,MAAMC,MAA+C;EACzD,MAAM,QAAQ,KAAK,MAAM,KAAK,MAAM;AACpC,MAAI,MACF,QAAO,KAAK,WAAW,OACrB,CAAC,KAAK,KAAK,UAAU;GACnB,IAAI,OAAO,MAAM,QAAQ;AACzB,UAAO;EACR,GACD,CAAE,EACH;AAGH,MAAI,KAAK,qBAAqB,OAC5B,OAAM,IAAIC,sDAAsB,CAAC,wBAAwB,EAAE,MAAM,EAAE;AAGrE,SAAO,KAAK,WAAW,OACrB,CAAC,KAAK,QAAQ;GACZ,IAAI,OAAO,QAAQ,KAAK,mBAAmB,OAAO;AAClD,UAAO;EACR,GACD,CAAE,EACH;CACF;;;;;;CAOD,wBAAgC;AAC9B,SAAO,CAAC,gDAAgD,EAAE,KAAK,OAAO;CACvE;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"regex.js","names":["fields: string | RegExp | RegexParserFields","outputKeys?: string[]","defaultOutputKey?: string","text: string"],"sources":["../../src/output_parsers/regex.ts"],"sourcesContent":["import {\n BaseOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport type { SerializedFields } from \"../load/map_keys.js\";\n\nexport interface RegExpFields {\n pattern: string;\n flags?: string;\n}\n\n/**\n * Interface for the fields required to create a RegexParser instance.\n */\nexport interface RegexParserFields {\n regex: string | RegExp | RegExpFields;\n outputKeys: string[];\n defaultOutputKey?: string;\n}\n\n/**\n * Class to parse the output of an LLM call into a dictionary.\n * @augments BaseOutputParser\n */\nexport class RegexParser extends BaseOutputParser<Record<string, string>> {\n static lc_name() {\n return \"RegexParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"regex\"];\n\n lc_serializable = true;\n\n get lc_attributes(): SerializedFields | undefined {\n return {\n regex: this.lc_kwargs.regex,\n };\n }\n\n regex: string | RegExp;\n\n outputKeys: string[];\n\n defaultOutputKey?: string;\n\n constructor(fields: RegexParserFields);\n\n constructor(\n regex: string | RegExp,\n outputKeys: string[],\n defaultOutputKey?: string\n );\n\n constructor(\n fields: string | RegExp | RegexParserFields,\n outputKeys?: string[],\n defaultOutputKey?: string\n ) {\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (typeof fields === \"string\" || fields instanceof RegExp) {\n // eslint-disable-next-line no-param-reassign\n fields = { regex: fields, outputKeys: outputKeys!, defaultOutputKey };\n }\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (fields.regex instanceof RegExp) {\n fields.regex = {\n pattern: fields.regex.source,\n flags: fields.regex.flags,\n };\n }\n super(fields);\n this.regex =\n typeof fields.regex === \"string\"\n ? new RegExp(fields.regex)\n : \"pattern\" in fields.regex\n
|
|
1
|
+
{"version":3,"file":"regex.js","names":["fields: string | RegExp | RegexParserFields","outputKeys?: string[]","defaultOutputKey?: string","text: string"],"sources":["../../src/output_parsers/regex.ts"],"sourcesContent":["import {\n BaseOutputParser,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport type { SerializedFields } from \"../load/map_keys.js\";\n\nexport interface RegExpFields {\n pattern: string;\n flags?: string;\n}\n\n/**\n * Interface for the fields required to create a RegexParser instance.\n */\nexport interface RegexParserFields {\n regex: string | RegExp | RegExpFields;\n outputKeys: string[];\n defaultOutputKey?: string;\n}\n\n/**\n * Class to parse the output of an LLM call into a dictionary.\n * @augments BaseOutputParser\n */\nexport class RegexParser extends BaseOutputParser<Record<string, string>> {\n static lc_name() {\n return \"RegexParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"regex\"];\n\n lc_serializable = true;\n\n get lc_attributes(): SerializedFields | undefined {\n return {\n regex: this.lc_kwargs.regex,\n };\n }\n\n regex: string | RegExp;\n\n outputKeys: string[];\n\n defaultOutputKey?: string;\n\n constructor(fields: RegexParserFields);\n\n constructor(\n regex: string | RegExp,\n outputKeys: string[],\n defaultOutputKey?: string\n );\n\n constructor(\n fields: string | RegExp | RegexParserFields,\n outputKeys?: string[],\n defaultOutputKey?: string\n ) {\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (typeof fields === \"string\" || fields instanceof RegExp) {\n // eslint-disable-next-line no-param-reassign\n fields = { regex: fields, outputKeys: outputKeys!, defaultOutputKey };\n }\n // eslint-disable-next-line no-instanceof/no-instanceof\n if (fields.regex instanceof RegExp) {\n fields.regex = {\n pattern: fields.regex.source,\n flags: fields.regex.flags,\n };\n }\n super(fields);\n this.regex =\n typeof fields.regex === \"string\"\n ? new RegExp(fields.regex)\n : \"pattern\" in fields.regex\n ? new RegExp(fields.regex.pattern, fields.regex.flags)\n : fields.regex;\n this.outputKeys = fields.outputKeys;\n this.defaultOutputKey = fields.defaultOutputKey;\n }\n\n _type() {\n return \"regex_parser\";\n }\n\n /**\n * Parses the given text using the regex pattern and returns a dictionary\n * with the parsed output. If the regex pattern does not match the text\n * and no defaultOutputKey is provided, throws an OutputParserException.\n * @param text The text to be parsed.\n * @returns A dictionary with the parsed output.\n */\n async parse(text: string): Promise<Record<string, string>> {\n const match = text.match(this.regex);\n if (match) {\n return this.outputKeys.reduce(\n (acc, key, index) => {\n acc[key] = match[index + 1];\n return acc;\n },\n {} as Record<string, string>\n );\n }\n\n if (this.defaultOutputKey === undefined) {\n throw new OutputParserException(`Could not parse output: ${text}`, text);\n }\n\n return this.outputKeys.reduce(\n (acc, key) => {\n acc[key] = key === this.defaultOutputKey ? text : \"\";\n return acc;\n },\n {} as Record<string, string>\n );\n }\n\n /**\n * Returns a string with instructions on how the LLM output should be\n * formatted to match the regex pattern.\n * @returns A string with formatting instructions.\n */\n getFormatInstructions(): string {\n return `Your response should match the following regex: ${this.regex}`;\n }\n}\n"],"mappings":";;;;;;;AAwBA,IAAa,cAAb,cAAiC,iBAAyC;CACxE,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAQ;CAEvD,kBAAkB;CAElB,IAAI,gBAA8C;AAChD,SAAO,EACL,OAAO,KAAK,UAAU,MACvB;CACF;CAED;CAEA;CAEA;CAUA,YACEA,QACAC,YACAC,kBACA;AAEA,MAAI,OAAO,WAAW,YAAY,kBAAkB,QAElD,SAAS;GAAE,OAAO;GAAoB;GAAa;EAAkB;AAGvE,MAAI,OAAO,iBAAiB,QAC1B,OAAO,QAAQ;GACb,SAAS,OAAO,MAAM;GACtB,OAAO,OAAO,MAAM;EACrB;EAEH,MAAM,OAAO;EACb,KAAK,QACH,OAAO,OAAO,UAAU,WACpB,IAAI,OAAO,OAAO,SAClB,aAAa,OAAO,QAClB,IAAI,OAAO,OAAO,MAAM,SAAS,OAAO,MAAM,SAC9C,OAAO;EACf,KAAK,aAAa,OAAO;EACzB,KAAK,mBAAmB,OAAO;CAChC;CAED,QAAQ;AACN,SAAO;CACR;;;;;;;;CASD,MAAM,MAAMC,MAA+C;EACzD,MAAM,QAAQ,KAAK,MAAM,KAAK,MAAM;AACpC,MAAI,MACF,QAAO,KAAK,WAAW,OACrB,CAAC,KAAK,KAAK,UAAU;GACnB,IAAI,OAAO,MAAM,QAAQ;AACzB,UAAO;EACR,GACD,CAAE,EACH;AAGH,MAAI,KAAK,qBAAqB,OAC5B,OAAM,IAAI,sBAAsB,CAAC,wBAAwB,EAAE,MAAM,EAAE;AAGrE,SAAO,KAAK,WAAW,OACrB,CAAC,KAAK,QAAQ;GACZ,IAAI,OAAO,QAAQ,KAAK,mBAAmB,OAAO;AAClD,UAAO;EACR,GACD,CAAE,EACH;CACF;;;;;;CAOD,wBAAgC;AAC9B,SAAO,CAAC,gDAAgD,EAAE,KAAK,OAAO;CACvE;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"router.cjs","names":["JsonMarkdownStructuredOutputParser","schema: Y","options?: RouterOutputParserInput","text: string","OutputParserException"],"sources":["../../src/output_parsers/router.ts"],"sourcesContent":["import { OutputParserException } from \"@langchain/core/output_parsers\";\nimport {\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\nimport { JsonMarkdownStructuredOutputParser } from \"./structured.js\";\n\n/**\n * Defines the input parameters for the RouterOutputParser class. It can\n * include a default destination and an interpolation depth.\n */\nexport type RouterOutputParserInput = {\n defaultDestination?: string;\n interpolationDepth?: number;\n};\n\n/**\n * A type of output parser that extends the\n * JsonMarkdownStructuredOutputParser. It is used to parse the output of a\n * router in LangChain. The class takes a schema and an optional\n * RouterOutputParserInput object as parameters.\n */\nexport class RouterOutputParser<\n Y extends InteropZodType
|
|
1
|
+
{"version":3,"file":"router.cjs","names":["JsonMarkdownStructuredOutputParser","schema: Y","options?: RouterOutputParserInput","text: string","OutputParserException"],"sources":["../../src/output_parsers/router.ts"],"sourcesContent":["import { OutputParserException } from \"@langchain/core/output_parsers\";\nimport {\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\nimport { JsonMarkdownStructuredOutputParser } from \"./structured.js\";\n\n/**\n * Defines the input parameters for the RouterOutputParser class. It can\n * include a default destination and an interpolation depth.\n */\nexport type RouterOutputParserInput = {\n defaultDestination?: string;\n interpolationDepth?: number;\n};\n\n/**\n * A type of output parser that extends the\n * JsonMarkdownStructuredOutputParser. It is used to parse the output of a\n * router in LangChain. The class takes a schema and an optional\n * RouterOutputParserInput object as parameters.\n */\nexport class RouterOutputParser<\n Y extends InteropZodType,\n> extends JsonMarkdownStructuredOutputParser<Y> {\n defaultDestination = \"DEFAULT\";\n\n constructor(schema: Y, options?: RouterOutputParserInput) {\n super(schema);\n this.defaultDestination =\n options?.defaultDestination ?? this.defaultDestination;\n }\n\n /**\n * Overrides the parse method from JsonMarkdownStructuredOutputParser.\n * This method takes a string as input, attempts to parse it, and returns\n * the parsed text. If the destination of the parsed text matches the\n * defaultDestination, the destination is set to null. If the parsing\n * fails, an OutputParserException is thrown.\n * @param text The text to be parsed.\n * @returns The parsed text as a Promise.\n */\n async parse(text: string): Promise<InferInteropZodOutput<Y>> {\n try {\n const parsedText = await super.parse(text);\n if (\n parsedText.destination?.toLowerCase() ===\n this.defaultDestination.toLowerCase()\n ) {\n parsedText.destination = null;\n }\n\n return parsedText;\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n }\n}\n"],"mappings":";;;;;;;;;;;AAsBA,IAAa,qBAAb,cAEUA,sDAAsC;CAC9C,qBAAqB;CAErB,YAAYC,QAAWC,SAAmC;EACxD,MAAM,OAAO;EACb,KAAK,qBACH,SAAS,sBAAsB,KAAK;CACvC;;;;;;;;;;CAWD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,aAAa,MAAM,MAAM,MAAM,KAAK;AAC1C,OACE,WAAW,aAAa,aAAa,KACrC,KAAK,mBAAmB,aAAa,EAErC,WAAW,cAAc;AAG3B,UAAO;EACR,SAAQ,GAAG;AACV,SAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;CACF;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"router.js","names":["schema: Y","options?: RouterOutputParserInput","text: string"],"sources":["../../src/output_parsers/router.ts"],"sourcesContent":["import { OutputParserException } from \"@langchain/core/output_parsers\";\nimport {\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\nimport { JsonMarkdownStructuredOutputParser } from \"./structured.js\";\n\n/**\n * Defines the input parameters for the RouterOutputParser class. It can\n * include a default destination and an interpolation depth.\n */\nexport type RouterOutputParserInput = {\n defaultDestination?: string;\n interpolationDepth?: number;\n};\n\n/**\n * A type of output parser that extends the\n * JsonMarkdownStructuredOutputParser. It is used to parse the output of a\n * router in LangChain. The class takes a schema and an optional\n * RouterOutputParserInput object as parameters.\n */\nexport class RouterOutputParser<\n Y extends InteropZodType
|
|
1
|
+
{"version":3,"file":"router.js","names":["schema: Y","options?: RouterOutputParserInput","text: string"],"sources":["../../src/output_parsers/router.ts"],"sourcesContent":["import { OutputParserException } from \"@langchain/core/output_parsers\";\nimport {\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\nimport { JsonMarkdownStructuredOutputParser } from \"./structured.js\";\n\n/**\n * Defines the input parameters for the RouterOutputParser class. It can\n * include a default destination and an interpolation depth.\n */\nexport type RouterOutputParserInput = {\n defaultDestination?: string;\n interpolationDepth?: number;\n};\n\n/**\n * A type of output parser that extends the\n * JsonMarkdownStructuredOutputParser. It is used to parse the output of a\n * router in LangChain. The class takes a schema and an optional\n * RouterOutputParserInput object as parameters.\n */\nexport class RouterOutputParser<\n Y extends InteropZodType,\n> extends JsonMarkdownStructuredOutputParser<Y> {\n defaultDestination = \"DEFAULT\";\n\n constructor(schema: Y, options?: RouterOutputParserInput) {\n super(schema);\n this.defaultDestination =\n options?.defaultDestination ?? this.defaultDestination;\n }\n\n /**\n * Overrides the parse method from JsonMarkdownStructuredOutputParser.\n * This method takes a string as input, attempts to parse it, and returns\n * the parsed text. If the destination of the parsed text matches the\n * defaultDestination, the destination is set to null. If the parsing\n * fails, an OutputParserException is thrown.\n * @param text The text to be parsed.\n * @returns The parsed text as a Promise.\n */\n async parse(text: string): Promise<InferInteropZodOutput<Y>> {\n try {\n const parsedText = await super.parse(text);\n if (\n parsedText.destination?.toLowerCase() ===\n this.defaultDestination.toLowerCase()\n ) {\n parsedText.destination = null;\n }\n\n return parsedText;\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n }\n}\n"],"mappings":";;;;;;;;;;AAsBA,IAAa,qBAAb,cAEU,mCAAsC;CAC9C,qBAAqB;CAErB,YAAYA,QAAWC,SAAmC;EACxD,MAAM,OAAO;EACb,KAAK,qBACH,SAAS,sBAAsB,KAAK;CACvC;;;;;;;;;;CAWD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,aAAa,MAAM,MAAM,MAAM,KAAK;AAC1C,OACE,WAAW,aAAa,aAAa,KACrC,KAAK,mBAAmB,aAAa,EAErC,WAAW,cAAc;AAG3B,UAAO;EACR,SAAQ,GAAG;AACV,SAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;CACF;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"structured.cjs","names":["BaseOutputParser","schema: T","schemas: S","z","text: string","OutputParserException","options?: JsonMarkdownFormatInstructionsOptions","schemaInput: JsonSchema7Type","type: string","type","description"],"sources":["../../src/output_parsers/structured.ts"],"sourcesContent":["import { z } from \"zod/v3\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n type JsonSchema7ArrayType,\n type JsonSchema7ObjectType,\n type JsonSchema7StringType,\n type JsonSchema7NumberType,\n type JsonSchema7NullableType,\n} from \"@langchain/core/utils/json_schema\";\nimport {\n BaseOutputParser,\n FormatInstructionsOptions,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport {\n interopParseAsync,\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\n\nexport type JsonMarkdownStructuredOutputParserInput = {\n interpolationDepth?: number;\n};\n\nexport interface JsonMarkdownFormatInstructionsOptions\n extends FormatInstructionsOptions {\n interpolationDepth?: number;\n}\n\nexport class StructuredOutputParser<\n T extends InteropZodType\n> extends BaseOutputParser<InferInteropZodOutput<T>> {\n static lc_name() {\n return \"StructuredOutputParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"structured\"];\n\n toJSON() {\n return this.toJSONNotImplemented();\n }\n\n constructor(public schema: T) {\n super(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a Zod schema.\n * @param schema The Zod schema which the output should match\n * @returns A new instance of StructuredOutputParser.\n */\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a set of names and\n * descriptions.\n * @param schemas An object where each key is a name and each value is a description\n * @returns A new instance of StructuredOutputParser.\n */\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this(zodSchema);\n }\n\n /**\n * Returns a markdown code snippet with a JSON object formatted according\n * to the schema.\n * @param options Optional. The options for formatting the instructions\n * @returns A markdown code snippet with a JSON object formatted according to the schema.\n */\n getFormatInstructions(): string {\n return `You must format your output as a JSON value that adheres to a given \"JSON Schema\" instance.\n\n\"JSON Schema\" is a declarative language that allows you to annotate and validate JSON documents.\n\nFor example, the example \"JSON Schema\" instance {{\"properties\": {{\"foo\": {{\"description\": \"a list of test words\", \"type\": \"array\", \"items\": {{\"type\": \"string\"}}}}}}, \"required\": [\"foo\"]}}}}\nwould match an object with one required property, \"foo\". The \"type\" property specifies \"foo\" must be an \"array\", and the \"description\" property semantically describes it as \"a list of test words\". The items within \"foo\" must be strings.\nThus, the object {{\"foo\": [\"bar\", \"baz\"]}} is a well-formatted instance of this example \"JSON Schema\". The object {{\"properties\": {{\"foo\": [\"bar\", \"baz\"]}}}} is not well-formatted.\n\nYour output will be parsed and type-checked according to the provided schema instance, so make sure all fields in your output match the schema exactly and there are no trailing commas!\n\nHere is the JSON Schema instance your output must adhere to. Include the enclosing markdown codeblock:\n\\`\\`\\`json\n${JSON.stringify(toJsonSchema(this.schema))}\n\\`\\`\\`\n`;\n }\n\n /**\n * Parses the given text according to the schema.\n * @param text The text to parse\n * @returns The parsed output.\n */\n async parse(text: string): Promise<InferInteropZodOutput<T>> {\n try {\n const json = text.includes(\"```\")\n ? text.trim().split(/```(?:json)?/)[1]\n : text.trim();\n return await interopParseAsync(this.schema, JSON.parse(json));\n } catch {\n try {\n return await interopParseAsync(this.schema, JSON.parse(text.trim()));\n } catch (e2) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e2}`,\n text\n );\n }\n }\n }\n}\n\n/**\n * A specific type of `StructuredOutputParser` that parses JSON data\n * formatted as a markdown code snippet.\n */\nexport class JsonMarkdownStructuredOutputParser<\n T extends InteropZodType\n> extends StructuredOutputParser<T> {\n static lc_name() {\n return \"JsonMarkdownStructuredOutputParser\";\n }\n\n getFormatInstructions(\n options?: JsonMarkdownFormatInstructionsOptions\n ): string {\n const interpolationDepth = options?.interpolationDepth ?? 1;\n if (interpolationDepth < 1) {\n throw new Error(\"f string interpolation depth must be at least 1\");\n }\n\n return `Return a markdown code snippet with a JSON object formatted to look like:\\n\\`\\`\\`json\\n${this._schemaToInstruction(\n toJsonSchema(this.schema)\n )\n .replaceAll(\"{\", \"{\".repeat(interpolationDepth))\n .replaceAll(\"}\", \"}\".repeat(interpolationDepth))}\\n\\`\\`\\``;\n }\n\n private _schemaToInstruction(\n schemaInput: JsonSchema7Type,\n indent = 2\n ): string {\n const schema = schemaInput as Extract<\n JsonSchema7Type,\n | JsonSchema7ObjectType\n | JsonSchema7ArrayType\n | JsonSchema7StringType\n | JsonSchema7NumberType\n | JsonSchema7NullableType\n >;\n\n if (\"type\" in schema) {\n let nullable = false;\n let type: string;\n if (Array.isArray(schema.type)) {\n const nullIdx = schema.type.findIndex((type) => type === \"null\");\n if (nullIdx !== -1) {\n nullable = true;\n schema.type.splice(nullIdx, 1);\n }\n type = schema.type.join(\" | \") as string;\n } else {\n type = schema.type;\n }\n\n if (schema.type === \"object\" && schema.properties) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n const properties = Object.entries(schema.properties)\n .map(([key, value]) => {\n const isOptional = schema.required?.includes(key)\n ? \"\"\n : \" (optional)\";\n return `${\" \".repeat(indent)}\"${key}\": ${this._schemaToInstruction(\n value,\n indent + 2\n )}${isOptional}`;\n })\n .join(\"\\n\");\n return `{\\n${properties}\\n${\" \".repeat(indent - 2)}}${description}`;\n }\n if (schema.type === \"array\" && schema.items) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n return `array[\\n${\" \".repeat(indent)}${this._schemaToInstruction(\n schema.items,\n indent + 2\n )}\\n${\" \".repeat(indent - 2)}] ${description}`;\n }\n const isNullable = nullable ? \" (nullable)\" : \"\";\n const description = schema.description ? ` // ${schema.description}` : \"\";\n return `${type}${description}${isNullable}`;\n }\n\n if (\"anyOf\" in schema) {\n return schema.anyOf\n .map((s) => this._schemaToInstruction(s, indent))\n .join(`\\n${\" \".repeat(indent - 2)}`);\n }\n\n throw new Error(\"unsupported schema type\");\n }\n\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this<T>(schema);\n }\n\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this<typeof zodSchema>(zodSchema);\n }\n}\n\nexport interface AsymmetricStructuredOutputParserFields<\n T extends InteropZodType\n> {\n inputSchema: T;\n}\n\n/**\n * A type of `StructuredOutputParser` that handles asymmetric input and\n * output schemas.\n */\nexport abstract class AsymmetricStructuredOutputParser<\n T extends InteropZodType,\n Y = unknown\n> extends BaseOutputParser<Y> {\n private structuredInputParser: JsonMarkdownStructuredOutputParser<T>;\n\n constructor({ inputSchema }: AsymmetricStructuredOutputParserFields<T>) {\n super(...arguments);\n this.structuredInputParser = new JsonMarkdownStructuredOutputParser(\n inputSchema\n );\n }\n\n /**\n * Processes the parsed input into the desired output format. Must be\n * implemented by subclasses.\n * @param input The parsed input\n * @returns The processed output.\n */\n abstract outputProcessor(input: InferInteropZodOutput<T>): Promise<Y>;\n\n async parse(text: string): Promise<Y> {\n let parsedInput;\n try {\n parsedInput = await this.structuredInputParser.parse(text);\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n\n return this.outputProcessor(parsedInput);\n }\n\n getFormatInstructions(): string {\n return this.structuredInputParser.getFormatInstructions();\n }\n}\n"],"mappings":";;;;;;;AA8BA,IAAa,yBAAb,cAEUA,iDAA2C;CACnD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAa;CAE5D,SAAS;AACP,SAAO,KAAK,sBAAsB;CACnC;CAED,YAAmBC,QAAW;EAC5B,MAAM,OAAO;EADI;CAElB;;;;;;CAOD,OAAO,cAAwCA,QAAW;AACxD,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,OAAO,yBACLC,SACA;EACA,MAAM,YAAYC,SAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAMA,SAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,wBAAgC;AAC9B,SAAO,CAAC;;;;;;;;;;;;AAYZ,EAAE,KAAK,+DAAuB,KAAK,OAAO,CAAC,CAAC;;AAE5C,CAAC;CACE;;;;;;CAOD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,OAAO,KAAK,SAAS,MAAM,GAC7B,KAAK,MAAM,CAAC,MAAM,eAAe,CAAC,KAClC,KAAK,MAAM;AACf,UAAO,0DAAwB,KAAK,QAAQ,KAAK,MAAM,KAAK,CAAC;EAC9D,QAAO;AACN,OAAI;AACF,WAAO,0DAAwB,KAAK,QAAQ,KAAK,MAAM,KAAK,MAAM,CAAC,CAAC;GACrE,SAAQ,IAAI;AACX,UAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,IAAI,EAChD;GAEH;EACF;CACF;AACF;;;;;AAMD,IAAa,qCAAb,cAEU,uBAA0B;CAClC,OAAO,UAAU;AACf,SAAO;CACR;CAED,sBACEC,SACQ;EACR,MAAM,qBAAqB,SAAS,sBAAsB;AAC1D,MAAI,qBAAqB,EACvB,OAAM,IAAI,MAAM;AAGlB,SAAO,CAAC,uFAAuF,EAAE,KAAK,0EACvF,KAAK,OAAO,CAC1B,CACE,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAC/C,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAAC,QAAQ,CAAC;CAC7D;CAED,AAAQ,qBACNC,aACA,SAAS,GACD;EACR,MAAM,SAAS;AASf,MAAI,UAAU,QAAQ;GACpB,IAAI,WAAW;GACf,IAAIC;AACJ,OAAI,MAAM,QAAQ,OAAO,KAAK,EAAE;IAC9B,MAAM,UAAU,OAAO,KAAK,UAAU,CAACC,WAASA,WAAS,OAAO;AAChE,QAAI,YAAY,IAAI;KAClB,WAAW;KACX,OAAO,KAAK,OAAO,SAAS,EAAE;IAC/B;IACD,OAAO,OAAO,KAAK,KAAK,MAAM;GAC/B,OACC,OAAO,OAAO;AAGhB,OAAI,OAAO,SAAS,YAAY,OAAO,YAAY;IACjD,MAAMC,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;IACJ,MAAM,aAAa,OAAO,QAAQ,OAAO,WAAW,CACjD,IAAI,CAAC,CAAC,KAAK,MAAM,KAAK;KACrB,MAAM,aAAa,OAAO,UAAU,SAAS,IAAI,GAC7C,KACA;AACJ,YAAO,GAAG,IAAI,OAAO,OAAO,CAAC,CAAC,EAAE,IAAI,GAAG,EAAE,KAAK,qBAC5C,OACA,SAAS,EACV,GAAG,YAAY;IACjB,EAAC,CACD,KAAK,KAAK;AACb,WAAO,CAAC,GAAG,EAAE,WAAW,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,CAAC,EAAEA,eAAa;GACpE;AACD,OAAI,OAAO,SAAS,WAAW,OAAO,OAAO;IAC3C,MAAMA,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;AACJ,WAAO,CAAC,QAAQ,EAAE,IAAI,OAAO,OAAO,GAAG,KAAK,qBAC1C,OAAO,OACP,SAAS,EACV,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,EAAE,EAAEA,eAAa;GAC/C;GACD,MAAM,aAAa,WAAW,gBAAgB;GAC9C,MAAM,cAAc,OAAO,cAAc,CAAC,IAAI,EAAE,OAAO,aAAa,GAAG;AACvE,UAAO,GAAG,OAAO,cAAc,YAAY;EAC5C;AAED,MAAI,WAAW,OACb,QAAO,OAAO,MACX,IAAI,CAAC,MAAM,KAAK,qBAAqB,GAAG,OAAO,CAAC,CAChD,KAAK,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,EAAE,CAAC;AAGxC,QAAM,IAAI,MAAM;CACjB;CAED,OAAO,cAAwCT,QAAW;AACxD,SAAO,IAAI,KAAQ;CACpB;CAED,OAAO,yBACLC,SACA;EACA,MAAM,YAAYC,SAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAMA,SAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAuB;CACnC;AACF;;;;;AAYD,IAAsB,mCAAtB,cAGUH,iDAAoB;CAC5B,AAAQ;CAER,YAAY,EAAE,aAAwD,EAAE;EACtE,MAAM,GAAG,UAAU;EACnB,KAAK,wBAAwB,IAAI,mCAC/B;CAEH;CAUD,MAAM,MAAMI,MAA0B;EACpC,IAAI;AACJ,MAAI;GACF,cAAc,MAAM,KAAK,sBAAsB,MAAM,KAAK;EAC3D,SAAQ,GAAG;AACV,SAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;AAED,SAAO,KAAK,gBAAgB,YAAY;CACzC;CAED,wBAAgC;AAC9B,SAAO,KAAK,sBAAsB,uBAAuB;CAC1D;AACF"}
|
|
1
|
+
{"version":3,"file":"structured.cjs","names":["BaseOutputParser","schema: T","schemas: S","z","text: string","OutputParserException","options?: JsonMarkdownFormatInstructionsOptions","schemaInput: JsonSchema7Type","type: string","type","description"],"sources":["../../src/output_parsers/structured.ts"],"sourcesContent":["import { z } from \"zod/v3\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n type JsonSchema7ArrayType,\n type JsonSchema7ObjectType,\n type JsonSchema7StringType,\n type JsonSchema7NumberType,\n type JsonSchema7NullableType,\n} from \"@langchain/core/utils/json_schema\";\nimport {\n BaseOutputParser,\n FormatInstructionsOptions,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport {\n interopParseAsync,\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\n\nexport type JsonMarkdownStructuredOutputParserInput = {\n interpolationDepth?: number;\n};\n\nexport interface JsonMarkdownFormatInstructionsOptions\n extends FormatInstructionsOptions {\n interpolationDepth?: number;\n}\n\nexport class StructuredOutputParser<\n T extends InteropZodType,\n> extends BaseOutputParser<InferInteropZodOutput<T>> {\n static lc_name() {\n return \"StructuredOutputParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"structured\"];\n\n toJSON() {\n return this.toJSONNotImplemented();\n }\n\n constructor(public schema: T) {\n super(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a Zod schema.\n * @param schema The Zod schema which the output should match\n * @returns A new instance of StructuredOutputParser.\n */\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a set of names and\n * descriptions.\n * @param schemas An object where each key is a name and each value is a description\n * @returns A new instance of StructuredOutputParser.\n */\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this(zodSchema);\n }\n\n /**\n * Returns a markdown code snippet with a JSON object formatted according\n * to the schema.\n * @param options Optional. The options for formatting the instructions\n * @returns A markdown code snippet with a JSON object formatted according to the schema.\n */\n getFormatInstructions(): string {\n return `You must format your output as a JSON value that adheres to a given \"JSON Schema\" instance.\n\n\"JSON Schema\" is a declarative language that allows you to annotate and validate JSON documents.\n\nFor example, the example \"JSON Schema\" instance {{\"properties\": {{\"foo\": {{\"description\": \"a list of test words\", \"type\": \"array\", \"items\": {{\"type\": \"string\"}}}}}}, \"required\": [\"foo\"]}}}}\nwould match an object with one required property, \"foo\". The \"type\" property specifies \"foo\" must be an \"array\", and the \"description\" property semantically describes it as \"a list of test words\". The items within \"foo\" must be strings.\nThus, the object {{\"foo\": [\"bar\", \"baz\"]}} is a well-formatted instance of this example \"JSON Schema\". The object {{\"properties\": {{\"foo\": [\"bar\", \"baz\"]}}}} is not well-formatted.\n\nYour output will be parsed and type-checked according to the provided schema instance, so make sure all fields in your output match the schema exactly and there are no trailing commas!\n\nHere is the JSON Schema instance your output must adhere to. Include the enclosing markdown codeblock:\n\\`\\`\\`json\n${JSON.stringify(toJsonSchema(this.schema))}\n\\`\\`\\`\n`;\n }\n\n /**\n * Parses the given text according to the schema.\n * @param text The text to parse\n * @returns The parsed output.\n */\n async parse(text: string): Promise<InferInteropZodOutput<T>> {\n try {\n const json = text.includes(\"```\")\n ? text.trim().split(/```(?:json)?/)[1]\n : text.trim();\n return await interopParseAsync(this.schema, JSON.parse(json));\n } catch {\n try {\n return await interopParseAsync(this.schema, JSON.parse(text.trim()));\n } catch (e2) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e2}`,\n text\n );\n }\n }\n }\n}\n\n/**\n * A specific type of `StructuredOutputParser` that parses JSON data\n * formatted as a markdown code snippet.\n */\nexport class JsonMarkdownStructuredOutputParser<\n T extends InteropZodType,\n> extends StructuredOutputParser<T> {\n static lc_name() {\n return \"JsonMarkdownStructuredOutputParser\";\n }\n\n getFormatInstructions(\n options?: JsonMarkdownFormatInstructionsOptions\n ): string {\n const interpolationDepth = options?.interpolationDepth ?? 1;\n if (interpolationDepth < 1) {\n throw new Error(\"f string interpolation depth must be at least 1\");\n }\n\n return `Return a markdown code snippet with a JSON object formatted to look like:\\n\\`\\`\\`json\\n${this._schemaToInstruction(\n toJsonSchema(this.schema)\n )\n .replaceAll(\"{\", \"{\".repeat(interpolationDepth))\n .replaceAll(\"}\", \"}\".repeat(interpolationDepth))}\\n\\`\\`\\``;\n }\n\n private _schemaToInstruction(\n schemaInput: JsonSchema7Type,\n indent = 2\n ): string {\n const schema = schemaInput as Extract<\n JsonSchema7Type,\n | JsonSchema7ObjectType\n | JsonSchema7ArrayType\n | JsonSchema7StringType\n | JsonSchema7NumberType\n | JsonSchema7NullableType\n >;\n\n if (\"type\" in schema) {\n let nullable = false;\n let type: string;\n if (Array.isArray(schema.type)) {\n const nullIdx = schema.type.findIndex((type) => type === \"null\");\n if (nullIdx !== -1) {\n nullable = true;\n schema.type.splice(nullIdx, 1);\n }\n type = schema.type.join(\" | \") as string;\n } else {\n type = schema.type;\n }\n\n if (schema.type === \"object\" && schema.properties) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n const properties = Object.entries(schema.properties)\n .map(([key, value]) => {\n const isOptional = schema.required?.includes(key)\n ? \"\"\n : \" (optional)\";\n return `${\" \".repeat(indent)}\"${key}\": ${this._schemaToInstruction(\n value,\n indent + 2\n )}${isOptional}`;\n })\n .join(\"\\n\");\n return `{\\n${properties}\\n${\" \".repeat(indent - 2)}}${description}`;\n }\n if (schema.type === \"array\" && schema.items) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n return `array[\\n${\" \".repeat(indent)}${this._schemaToInstruction(\n schema.items,\n indent + 2\n )}\\n${\" \".repeat(indent - 2)}] ${description}`;\n }\n const isNullable = nullable ? \" (nullable)\" : \"\";\n const description = schema.description ? ` // ${schema.description}` : \"\";\n return `${type}${description}${isNullable}`;\n }\n\n if (\"anyOf\" in schema) {\n return schema.anyOf\n .map((s) => this._schemaToInstruction(s, indent))\n .join(`\\n${\" \".repeat(indent - 2)}`);\n }\n\n throw new Error(\"unsupported schema type\");\n }\n\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this<T>(schema);\n }\n\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this<typeof zodSchema>(zodSchema);\n }\n}\n\nexport interface AsymmetricStructuredOutputParserFields<\n T extends InteropZodType,\n> {\n inputSchema: T;\n}\n\n/**\n * A type of `StructuredOutputParser` that handles asymmetric input and\n * output schemas.\n */\nexport abstract class AsymmetricStructuredOutputParser<\n T extends InteropZodType,\n Y = unknown,\n> extends BaseOutputParser<Y> {\n private structuredInputParser: JsonMarkdownStructuredOutputParser<T>;\n\n constructor({ inputSchema }: AsymmetricStructuredOutputParserFields<T>) {\n super(...arguments);\n this.structuredInputParser = new JsonMarkdownStructuredOutputParser(\n inputSchema\n );\n }\n\n /**\n * Processes the parsed input into the desired output format. Must be\n * implemented by subclasses.\n * @param input The parsed input\n * @returns The processed output.\n */\n abstract outputProcessor(input: InferInteropZodOutput<T>): Promise<Y>;\n\n async parse(text: string): Promise<Y> {\n let parsedInput;\n try {\n parsedInput = await this.structuredInputParser.parse(text);\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n\n return this.outputProcessor(parsedInput);\n }\n\n getFormatInstructions(): string {\n return this.structuredInputParser.getFormatInstructions();\n }\n}\n"],"mappings":";;;;;;;AA8BA,IAAa,yBAAb,cAEUA,iDAA2C;CACnD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAa;CAE5D,SAAS;AACP,SAAO,KAAK,sBAAsB;CACnC;CAED,YAAmBC,QAAW;EAC5B,MAAM,OAAO;EADI;CAElB;;;;;;CAOD,OAAO,cAAwCA,QAAW;AACxD,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,OAAO,yBACLC,SACA;EACA,MAAM,YAAYC,SAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAMA,SAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,wBAAgC;AAC9B,SAAO,CAAC;;;;;;;;;;;;AAYZ,EAAE,KAAK,+DAAuB,KAAK,OAAO,CAAC,CAAC;;AAE5C,CAAC;CACE;;;;;;CAOD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,OAAO,KAAK,SAAS,MAAM,GAC7B,KAAK,MAAM,CAAC,MAAM,eAAe,CAAC,KAClC,KAAK,MAAM;AACf,UAAO,0DAAwB,KAAK,QAAQ,KAAK,MAAM,KAAK,CAAC;EAC9D,QAAO;AACN,OAAI;AACF,WAAO,0DAAwB,KAAK,QAAQ,KAAK,MAAM,KAAK,MAAM,CAAC,CAAC;GACrE,SAAQ,IAAI;AACX,UAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,IAAI,EAChD;GAEH;EACF;CACF;AACF;;;;;AAMD,IAAa,qCAAb,cAEU,uBAA0B;CAClC,OAAO,UAAU;AACf,SAAO;CACR;CAED,sBACEC,SACQ;EACR,MAAM,qBAAqB,SAAS,sBAAsB;AAC1D,MAAI,qBAAqB,EACvB,OAAM,IAAI,MAAM;AAGlB,SAAO,CAAC,uFAAuF,EAAE,KAAK,0EACvF,KAAK,OAAO,CAC1B,CACE,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAC/C,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAAC,QAAQ,CAAC;CAC7D;CAED,AAAQ,qBACNC,aACA,SAAS,GACD;EACR,MAAM,SAAS;AASf,MAAI,UAAU,QAAQ;GACpB,IAAI,WAAW;GACf,IAAIC;AACJ,OAAI,MAAM,QAAQ,OAAO,KAAK,EAAE;IAC9B,MAAM,UAAU,OAAO,KAAK,UAAU,CAACC,WAASA,WAAS,OAAO;AAChE,QAAI,YAAY,IAAI;KAClB,WAAW;KACX,OAAO,KAAK,OAAO,SAAS,EAAE;IAC/B;IACD,OAAO,OAAO,KAAK,KAAK,MAAM;GAC/B,OACC,OAAO,OAAO;AAGhB,OAAI,OAAO,SAAS,YAAY,OAAO,YAAY;IACjD,MAAMC,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;IACJ,MAAM,aAAa,OAAO,QAAQ,OAAO,WAAW,CACjD,IAAI,CAAC,CAAC,KAAK,MAAM,KAAK;KACrB,MAAM,aAAa,OAAO,UAAU,SAAS,IAAI,GAC7C,KACA;AACJ,YAAO,GAAG,IAAI,OAAO,OAAO,CAAC,CAAC,EAAE,IAAI,GAAG,EAAE,KAAK,qBAC5C,OACA,SAAS,EACV,GAAG,YAAY;IACjB,EAAC,CACD,KAAK,KAAK;AACb,WAAO,CAAC,GAAG,EAAE,WAAW,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,CAAC,EAAEA,eAAa;GACpE;AACD,OAAI,OAAO,SAAS,WAAW,OAAO,OAAO;IAC3C,MAAMA,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;AACJ,WAAO,CAAC,QAAQ,EAAE,IAAI,OAAO,OAAO,GAAG,KAAK,qBAC1C,OAAO,OACP,SAAS,EACV,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,EAAE,EAAEA,eAAa;GAC/C;GACD,MAAM,aAAa,WAAW,gBAAgB;GAC9C,MAAM,cAAc,OAAO,cAAc,CAAC,IAAI,EAAE,OAAO,aAAa,GAAG;AACvE,UAAO,GAAG,OAAO,cAAc,YAAY;EAC5C;AAED,MAAI,WAAW,OACb,QAAO,OAAO,MACX,IAAI,CAAC,MAAM,KAAK,qBAAqB,GAAG,OAAO,CAAC,CAChD,KAAK,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,EAAE,CAAC;AAGxC,QAAM,IAAI,MAAM;CACjB;CAED,OAAO,cAAwCT,QAAW;AACxD,SAAO,IAAI,KAAQ;CACpB;CAED,OAAO,yBACLC,SACA;EACA,MAAM,YAAYC,SAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAMA,SAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAuB;CACnC;AACF;;;;;AAYD,IAAsB,mCAAtB,cAGUH,iDAAoB;CAC5B,AAAQ;CAER,YAAY,EAAE,aAAwD,EAAE;EACtE,MAAM,GAAG,UAAU;EACnB,KAAK,wBAAwB,IAAI,mCAC/B;CAEH;CAUD,MAAM,MAAMI,MAA0B;EACpC,IAAI;AACJ,MAAI;GACF,cAAc,MAAM,KAAK,sBAAsB,MAAM,KAAK;EAC3D,SAAQ,GAAG;AACV,SAAM,IAAIC,sDACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;AAED,SAAO,KAAK,gBAAgB,YAAY;CACzC;CAED,wBAAgC;AAC9B,SAAO,KAAK,sBAAsB,uBAAuB;CAC1D;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"structured.js","names":["schema: T","schemas: S","text: string","options?: JsonMarkdownFormatInstructionsOptions","schemaInput: JsonSchema7Type","type: string","type","description"],"sources":["../../src/output_parsers/structured.ts"],"sourcesContent":["import { z } from \"zod/v3\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n type JsonSchema7ArrayType,\n type JsonSchema7ObjectType,\n type JsonSchema7StringType,\n type JsonSchema7NumberType,\n type JsonSchema7NullableType,\n} from \"@langchain/core/utils/json_schema\";\nimport {\n BaseOutputParser,\n FormatInstructionsOptions,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport {\n interopParseAsync,\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\n\nexport type JsonMarkdownStructuredOutputParserInput = {\n interpolationDepth?: number;\n};\n\nexport interface JsonMarkdownFormatInstructionsOptions\n extends FormatInstructionsOptions {\n interpolationDepth?: number;\n}\n\nexport class StructuredOutputParser<\n T extends InteropZodType\n> extends BaseOutputParser<InferInteropZodOutput<T>> {\n static lc_name() {\n return \"StructuredOutputParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"structured\"];\n\n toJSON() {\n return this.toJSONNotImplemented();\n }\n\n constructor(public schema: T) {\n super(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a Zod schema.\n * @param schema The Zod schema which the output should match\n * @returns A new instance of StructuredOutputParser.\n */\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a set of names and\n * descriptions.\n * @param schemas An object where each key is a name and each value is a description\n * @returns A new instance of StructuredOutputParser.\n */\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this(zodSchema);\n }\n\n /**\n * Returns a markdown code snippet with a JSON object formatted according\n * to the schema.\n * @param options Optional. The options for formatting the instructions\n * @returns A markdown code snippet with a JSON object formatted according to the schema.\n */\n getFormatInstructions(): string {\n return `You must format your output as a JSON value that adheres to a given \"JSON Schema\" instance.\n\n\"JSON Schema\" is a declarative language that allows you to annotate and validate JSON documents.\n\nFor example, the example \"JSON Schema\" instance {{\"properties\": {{\"foo\": {{\"description\": \"a list of test words\", \"type\": \"array\", \"items\": {{\"type\": \"string\"}}}}}}, \"required\": [\"foo\"]}}}}\nwould match an object with one required property, \"foo\". The \"type\" property specifies \"foo\" must be an \"array\", and the \"description\" property semantically describes it as \"a list of test words\". The items within \"foo\" must be strings.\nThus, the object {{\"foo\": [\"bar\", \"baz\"]}} is a well-formatted instance of this example \"JSON Schema\". The object {{\"properties\": {{\"foo\": [\"bar\", \"baz\"]}}}} is not well-formatted.\n\nYour output will be parsed and type-checked according to the provided schema instance, so make sure all fields in your output match the schema exactly and there are no trailing commas!\n\nHere is the JSON Schema instance your output must adhere to. Include the enclosing markdown codeblock:\n\\`\\`\\`json\n${JSON.stringify(toJsonSchema(this.schema))}\n\\`\\`\\`\n`;\n }\n\n /**\n * Parses the given text according to the schema.\n * @param text The text to parse\n * @returns The parsed output.\n */\n async parse(text: string): Promise<InferInteropZodOutput<T>> {\n try {\n const json = text.includes(\"```\")\n ? text.trim().split(/```(?:json)?/)[1]\n : text.trim();\n return await interopParseAsync(this.schema, JSON.parse(json));\n } catch {\n try {\n return await interopParseAsync(this.schema, JSON.parse(text.trim()));\n } catch (e2) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e2}`,\n text\n );\n }\n }\n }\n}\n\n/**\n * A specific type of `StructuredOutputParser` that parses JSON data\n * formatted as a markdown code snippet.\n */\nexport class JsonMarkdownStructuredOutputParser<\n T extends InteropZodType\n> extends StructuredOutputParser<T> {\n static lc_name() {\n return \"JsonMarkdownStructuredOutputParser\";\n }\n\n getFormatInstructions(\n options?: JsonMarkdownFormatInstructionsOptions\n ): string {\n const interpolationDepth = options?.interpolationDepth ?? 1;\n if (interpolationDepth < 1) {\n throw new Error(\"f string interpolation depth must be at least 1\");\n }\n\n return `Return a markdown code snippet with a JSON object formatted to look like:\\n\\`\\`\\`json\\n${this._schemaToInstruction(\n toJsonSchema(this.schema)\n )\n .replaceAll(\"{\", \"{\".repeat(interpolationDepth))\n .replaceAll(\"}\", \"}\".repeat(interpolationDepth))}\\n\\`\\`\\``;\n }\n\n private _schemaToInstruction(\n schemaInput: JsonSchema7Type,\n indent = 2\n ): string {\n const schema = schemaInput as Extract<\n JsonSchema7Type,\n | JsonSchema7ObjectType\n | JsonSchema7ArrayType\n | JsonSchema7StringType\n | JsonSchema7NumberType\n | JsonSchema7NullableType\n >;\n\n if (\"type\" in schema) {\n let nullable = false;\n let type: string;\n if (Array.isArray(schema.type)) {\n const nullIdx = schema.type.findIndex((type) => type === \"null\");\n if (nullIdx !== -1) {\n nullable = true;\n schema.type.splice(nullIdx, 1);\n }\n type = schema.type.join(\" | \") as string;\n } else {\n type = schema.type;\n }\n\n if (schema.type === \"object\" && schema.properties) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n const properties = Object.entries(schema.properties)\n .map(([key, value]) => {\n const isOptional = schema.required?.includes(key)\n ? \"\"\n : \" (optional)\";\n return `${\" \".repeat(indent)}\"${key}\": ${this._schemaToInstruction(\n value,\n indent + 2\n )}${isOptional}`;\n })\n .join(\"\\n\");\n return `{\\n${properties}\\n${\" \".repeat(indent - 2)}}${description}`;\n }\n if (schema.type === \"array\" && schema.items) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n return `array[\\n${\" \".repeat(indent)}${this._schemaToInstruction(\n schema.items,\n indent + 2\n )}\\n${\" \".repeat(indent - 2)}] ${description}`;\n }\n const isNullable = nullable ? \" (nullable)\" : \"\";\n const description = schema.description ? ` // ${schema.description}` : \"\";\n return `${type}${description}${isNullable}`;\n }\n\n if (\"anyOf\" in schema) {\n return schema.anyOf\n .map((s) => this._schemaToInstruction(s, indent))\n .join(`\\n${\" \".repeat(indent - 2)}`);\n }\n\n throw new Error(\"unsupported schema type\");\n }\n\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this<T>(schema);\n }\n\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this<typeof zodSchema>(zodSchema);\n }\n}\n\nexport interface AsymmetricStructuredOutputParserFields<\n T extends InteropZodType\n> {\n inputSchema: T;\n}\n\n/**\n * A type of `StructuredOutputParser` that handles asymmetric input and\n * output schemas.\n */\nexport abstract class AsymmetricStructuredOutputParser<\n T extends InteropZodType,\n Y = unknown\n> extends BaseOutputParser<Y> {\n private structuredInputParser: JsonMarkdownStructuredOutputParser<T>;\n\n constructor({ inputSchema }: AsymmetricStructuredOutputParserFields<T>) {\n super(...arguments);\n this.structuredInputParser = new JsonMarkdownStructuredOutputParser(\n inputSchema\n );\n }\n\n /**\n * Processes the parsed input into the desired output format. Must be\n * implemented by subclasses.\n * @param input The parsed input\n * @returns The processed output.\n */\n abstract outputProcessor(input: InferInteropZodOutput<T>): Promise<Y>;\n\n async parse(text: string): Promise<Y> {\n let parsedInput;\n try {\n parsedInput = await this.structuredInputParser.parse(text);\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n\n return this.outputProcessor(parsedInput);\n }\n\n getFormatInstructions(): string {\n return this.structuredInputParser.getFormatInstructions();\n }\n}\n"],"mappings":";;;;;;AA8BA,IAAa,yBAAb,cAEU,iBAA2C;CACnD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAa;CAE5D,SAAS;AACP,SAAO,KAAK,sBAAsB;CACnC;CAED,YAAmBA,QAAW;EAC5B,MAAM,OAAO;EADI;CAElB;;;;;;CAOD,OAAO,cAAwCA,QAAW;AACxD,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,OAAO,yBACLC,SACA;EACA,MAAM,YAAY,EAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAM,EAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,wBAAgC;AAC9B,SAAO,CAAC;;;;;;;;;;;;AAYZ,EAAE,KAAK,UAAU,aAAa,KAAK,OAAO,CAAC,CAAC;;AAE5C,CAAC;CACE;;;;;;CAOD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,OAAO,KAAK,SAAS,MAAM,GAC7B,KAAK,MAAM,CAAC,MAAM,eAAe,CAAC,KAClC,KAAK,MAAM;AACf,UAAO,MAAM,kBAAkB,KAAK,QAAQ,KAAK,MAAM,KAAK,CAAC;EAC9D,QAAO;AACN,OAAI;AACF,WAAO,MAAM,kBAAkB,KAAK,QAAQ,KAAK,MAAM,KAAK,MAAM,CAAC,CAAC;GACrE,SAAQ,IAAI;AACX,UAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,IAAI,EAChD;GAEH;EACF;CACF;AACF;;;;;AAMD,IAAa,qCAAb,cAEU,uBAA0B;CAClC,OAAO,UAAU;AACf,SAAO;CACR;CAED,sBACEC,SACQ;EACR,MAAM,qBAAqB,SAAS,sBAAsB;AAC1D,MAAI,qBAAqB,EACvB,OAAM,IAAI,MAAM;AAGlB,SAAO,CAAC,uFAAuF,EAAE,KAAK,qBACpG,aAAa,KAAK,OAAO,CAC1B,CACE,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAC/C,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAAC,QAAQ,CAAC;CAC7D;CAED,AAAQ,qBACNC,aACA,SAAS,GACD;EACR,MAAM,SAAS;AASf,MAAI,UAAU,QAAQ;GACpB,IAAI,WAAW;GACf,IAAIC;AACJ,OAAI,MAAM,QAAQ,OAAO,KAAK,EAAE;IAC9B,MAAM,UAAU,OAAO,KAAK,UAAU,CAACC,WAASA,WAAS,OAAO;AAChE,QAAI,YAAY,IAAI;KAClB,WAAW;KACX,OAAO,KAAK,OAAO,SAAS,EAAE;IAC/B;IACD,OAAO,OAAO,KAAK,KAAK,MAAM;GAC/B,OACC,OAAO,OAAO;AAGhB,OAAI,OAAO,SAAS,YAAY,OAAO,YAAY;IACjD,MAAMC,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;IACJ,MAAM,aAAa,OAAO,QAAQ,OAAO,WAAW,CACjD,IAAI,CAAC,CAAC,KAAK,MAAM,KAAK;KACrB,MAAM,aAAa,OAAO,UAAU,SAAS,IAAI,GAC7C,KACA;AACJ,YAAO,GAAG,IAAI,OAAO,OAAO,CAAC,CAAC,EAAE,IAAI,GAAG,EAAE,KAAK,qBAC5C,OACA,SAAS,EACV,GAAG,YAAY;IACjB,EAAC,CACD,KAAK,KAAK;AACb,WAAO,CAAC,GAAG,EAAE,WAAW,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,CAAC,EAAEA,eAAa;GACpE;AACD,OAAI,OAAO,SAAS,WAAW,OAAO,OAAO;IAC3C,MAAMA,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;AACJ,WAAO,CAAC,QAAQ,EAAE,IAAI,OAAO,OAAO,GAAG,KAAK,qBAC1C,OAAO,OACP,SAAS,EACV,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,EAAE,EAAEA,eAAa;GAC/C;GACD,MAAM,aAAa,WAAW,gBAAgB;GAC9C,MAAM,cAAc,OAAO,cAAc,CAAC,IAAI,EAAE,OAAO,aAAa,GAAG;AACvE,UAAO,GAAG,OAAO,cAAc,YAAY;EAC5C;AAED,MAAI,WAAW,OACb,QAAO,OAAO,MACX,IAAI,CAAC,MAAM,KAAK,qBAAqB,GAAG,OAAO,CAAC,CAChD,KAAK,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,EAAE,CAAC;AAGxC,QAAM,IAAI,MAAM;CACjB;CAED,OAAO,cAAwCP,QAAW;AACxD,SAAO,IAAI,KAAQ;CACpB;CAED,OAAO,yBACLC,SACA;EACA,MAAM,YAAY,EAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAM,EAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAuB;CACnC;AACF;;;;;AAYD,IAAsB,mCAAtB,cAGU,iBAAoB;CAC5B,AAAQ;CAER,YAAY,EAAE,aAAwD,EAAE;EACtE,MAAM,GAAG,UAAU;EACnB,KAAK,wBAAwB,IAAI,mCAC/B;CAEH;CAUD,MAAM,MAAMC,MAA0B;EACpC,IAAI;AACJ,MAAI;GACF,cAAc,MAAM,KAAK,sBAAsB,MAAM,KAAK;EAC3D,SAAQ,GAAG;AACV,SAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;AAED,SAAO,KAAK,gBAAgB,YAAY;CACzC;CAED,wBAAgC;AAC9B,SAAO,KAAK,sBAAsB,uBAAuB;CAC1D;AACF"}
|
|
1
|
+
{"version":3,"file":"structured.js","names":["schema: T","schemas: S","text: string","options?: JsonMarkdownFormatInstructionsOptions","schemaInput: JsonSchema7Type","type: string","type","description"],"sources":["../../src/output_parsers/structured.ts"],"sourcesContent":["import { z } from \"zod/v3\";\nimport {\n toJsonSchema,\n type JsonSchema7Type,\n type JsonSchema7ArrayType,\n type JsonSchema7ObjectType,\n type JsonSchema7StringType,\n type JsonSchema7NumberType,\n type JsonSchema7NullableType,\n} from \"@langchain/core/utils/json_schema\";\nimport {\n BaseOutputParser,\n FormatInstructionsOptions,\n OutputParserException,\n} from \"@langchain/core/output_parsers\";\nimport {\n interopParseAsync,\n type InferInteropZodOutput,\n type InteropZodType,\n} from \"@langchain/core/utils/types\";\n\nexport type JsonMarkdownStructuredOutputParserInput = {\n interpolationDepth?: number;\n};\n\nexport interface JsonMarkdownFormatInstructionsOptions\n extends FormatInstructionsOptions {\n interpolationDepth?: number;\n}\n\nexport class StructuredOutputParser<\n T extends InteropZodType,\n> extends BaseOutputParser<InferInteropZodOutput<T>> {\n static lc_name() {\n return \"StructuredOutputParser\";\n }\n\n lc_namespace = [\"langchain\", \"output_parsers\", \"structured\"];\n\n toJSON() {\n return this.toJSONNotImplemented();\n }\n\n constructor(public schema: T) {\n super(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a Zod schema.\n * @param schema The Zod schema which the output should match\n * @returns A new instance of StructuredOutputParser.\n */\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this(schema);\n }\n\n /**\n * Creates a new StructuredOutputParser from a set of names and\n * descriptions.\n * @param schemas An object where each key is a name and each value is a description\n * @returns A new instance of StructuredOutputParser.\n */\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this(zodSchema);\n }\n\n /**\n * Returns a markdown code snippet with a JSON object formatted according\n * to the schema.\n * @param options Optional. The options for formatting the instructions\n * @returns A markdown code snippet with a JSON object formatted according to the schema.\n */\n getFormatInstructions(): string {\n return `You must format your output as a JSON value that adheres to a given \"JSON Schema\" instance.\n\n\"JSON Schema\" is a declarative language that allows you to annotate and validate JSON documents.\n\nFor example, the example \"JSON Schema\" instance {{\"properties\": {{\"foo\": {{\"description\": \"a list of test words\", \"type\": \"array\", \"items\": {{\"type\": \"string\"}}}}}}, \"required\": [\"foo\"]}}}}\nwould match an object with one required property, \"foo\". The \"type\" property specifies \"foo\" must be an \"array\", and the \"description\" property semantically describes it as \"a list of test words\". The items within \"foo\" must be strings.\nThus, the object {{\"foo\": [\"bar\", \"baz\"]}} is a well-formatted instance of this example \"JSON Schema\". The object {{\"properties\": {{\"foo\": [\"bar\", \"baz\"]}}}} is not well-formatted.\n\nYour output will be parsed and type-checked according to the provided schema instance, so make sure all fields in your output match the schema exactly and there are no trailing commas!\n\nHere is the JSON Schema instance your output must adhere to. Include the enclosing markdown codeblock:\n\\`\\`\\`json\n${JSON.stringify(toJsonSchema(this.schema))}\n\\`\\`\\`\n`;\n }\n\n /**\n * Parses the given text according to the schema.\n * @param text The text to parse\n * @returns The parsed output.\n */\n async parse(text: string): Promise<InferInteropZodOutput<T>> {\n try {\n const json = text.includes(\"```\")\n ? text.trim().split(/```(?:json)?/)[1]\n : text.trim();\n return await interopParseAsync(this.schema, JSON.parse(json));\n } catch {\n try {\n return await interopParseAsync(this.schema, JSON.parse(text.trim()));\n } catch (e2) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e2}`,\n text\n );\n }\n }\n }\n}\n\n/**\n * A specific type of `StructuredOutputParser` that parses JSON data\n * formatted as a markdown code snippet.\n */\nexport class JsonMarkdownStructuredOutputParser<\n T extends InteropZodType,\n> extends StructuredOutputParser<T> {\n static lc_name() {\n return \"JsonMarkdownStructuredOutputParser\";\n }\n\n getFormatInstructions(\n options?: JsonMarkdownFormatInstructionsOptions\n ): string {\n const interpolationDepth = options?.interpolationDepth ?? 1;\n if (interpolationDepth < 1) {\n throw new Error(\"f string interpolation depth must be at least 1\");\n }\n\n return `Return a markdown code snippet with a JSON object formatted to look like:\\n\\`\\`\\`json\\n${this._schemaToInstruction(\n toJsonSchema(this.schema)\n )\n .replaceAll(\"{\", \"{\".repeat(interpolationDepth))\n .replaceAll(\"}\", \"}\".repeat(interpolationDepth))}\\n\\`\\`\\``;\n }\n\n private _schemaToInstruction(\n schemaInput: JsonSchema7Type,\n indent = 2\n ): string {\n const schema = schemaInput as Extract<\n JsonSchema7Type,\n | JsonSchema7ObjectType\n | JsonSchema7ArrayType\n | JsonSchema7StringType\n | JsonSchema7NumberType\n | JsonSchema7NullableType\n >;\n\n if (\"type\" in schema) {\n let nullable = false;\n let type: string;\n if (Array.isArray(schema.type)) {\n const nullIdx = schema.type.findIndex((type) => type === \"null\");\n if (nullIdx !== -1) {\n nullable = true;\n schema.type.splice(nullIdx, 1);\n }\n type = schema.type.join(\" | \") as string;\n } else {\n type = schema.type;\n }\n\n if (schema.type === \"object\" && schema.properties) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n const properties = Object.entries(schema.properties)\n .map(([key, value]) => {\n const isOptional = schema.required?.includes(key)\n ? \"\"\n : \" (optional)\";\n return `${\" \".repeat(indent)}\"${key}\": ${this._schemaToInstruction(\n value,\n indent + 2\n )}${isOptional}`;\n })\n .join(\"\\n\");\n return `{\\n${properties}\\n${\" \".repeat(indent - 2)}}${description}`;\n }\n if (schema.type === \"array\" && schema.items) {\n const description = schema.description\n ? ` // ${schema.description}`\n : \"\";\n return `array[\\n${\" \".repeat(indent)}${this._schemaToInstruction(\n schema.items,\n indent + 2\n )}\\n${\" \".repeat(indent - 2)}] ${description}`;\n }\n const isNullable = nullable ? \" (nullable)\" : \"\";\n const description = schema.description ? ` // ${schema.description}` : \"\";\n return `${type}${description}${isNullable}`;\n }\n\n if (\"anyOf\" in schema) {\n return schema.anyOf\n .map((s) => this._schemaToInstruction(s, indent))\n .join(`\\n${\" \".repeat(indent - 2)}`);\n }\n\n throw new Error(\"unsupported schema type\");\n }\n\n static fromZodSchema<T extends InteropZodType>(schema: T) {\n return new this<T>(schema);\n }\n\n static fromNamesAndDescriptions<S extends { [key: string]: string }>(\n schemas: S\n ) {\n const zodSchema = z.object(\n Object.fromEntries(\n Object.entries(schemas).map(\n ([name, description]) =>\n [name, z.string().describe(description)] as const\n )\n )\n );\n\n return new this<typeof zodSchema>(zodSchema);\n }\n}\n\nexport interface AsymmetricStructuredOutputParserFields<\n T extends InteropZodType,\n> {\n inputSchema: T;\n}\n\n/**\n * A type of `StructuredOutputParser` that handles asymmetric input and\n * output schemas.\n */\nexport abstract class AsymmetricStructuredOutputParser<\n T extends InteropZodType,\n Y = unknown,\n> extends BaseOutputParser<Y> {\n private structuredInputParser: JsonMarkdownStructuredOutputParser<T>;\n\n constructor({ inputSchema }: AsymmetricStructuredOutputParserFields<T>) {\n super(...arguments);\n this.structuredInputParser = new JsonMarkdownStructuredOutputParser(\n inputSchema\n );\n }\n\n /**\n * Processes the parsed input into the desired output format. Must be\n * implemented by subclasses.\n * @param input The parsed input\n * @returns The processed output.\n */\n abstract outputProcessor(input: InferInteropZodOutput<T>): Promise<Y>;\n\n async parse(text: string): Promise<Y> {\n let parsedInput;\n try {\n parsedInput = await this.structuredInputParser.parse(text);\n } catch (e) {\n throw new OutputParserException(\n `Failed to parse. Text: \"${text}\". Error: ${e}`,\n text\n );\n }\n\n return this.outputProcessor(parsedInput);\n }\n\n getFormatInstructions(): string {\n return this.structuredInputParser.getFormatInstructions();\n }\n}\n"],"mappings":";;;;;;AA8BA,IAAa,yBAAb,cAEU,iBAA2C;CACnD,OAAO,UAAU;AACf,SAAO;CACR;CAED,eAAe;EAAC;EAAa;EAAkB;CAAa;CAE5D,SAAS;AACP,SAAO,KAAK,sBAAsB;CACnC;CAED,YAAmBA,QAAW;EAC5B,MAAM,OAAO;EADI;CAElB;;;;;;CAOD,OAAO,cAAwCA,QAAW;AACxD,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,OAAO,yBACLC,SACA;EACA,MAAM,YAAY,EAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAM,EAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAK;CACjB;;;;;;;CAQD,wBAAgC;AAC9B,SAAO,CAAC;;;;;;;;;;;;AAYZ,EAAE,KAAK,UAAU,aAAa,KAAK,OAAO,CAAC,CAAC;;AAE5C,CAAC;CACE;;;;;;CAOD,MAAM,MAAMC,MAAiD;AAC3D,MAAI;GACF,MAAM,OAAO,KAAK,SAAS,MAAM,GAC7B,KAAK,MAAM,CAAC,MAAM,eAAe,CAAC,KAClC,KAAK,MAAM;AACf,UAAO,MAAM,kBAAkB,KAAK,QAAQ,KAAK,MAAM,KAAK,CAAC;EAC9D,QAAO;AACN,OAAI;AACF,WAAO,MAAM,kBAAkB,KAAK,QAAQ,KAAK,MAAM,KAAK,MAAM,CAAC,CAAC;GACrE,SAAQ,IAAI;AACX,UAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,IAAI,EAChD;GAEH;EACF;CACF;AACF;;;;;AAMD,IAAa,qCAAb,cAEU,uBAA0B;CAClC,OAAO,UAAU;AACf,SAAO;CACR;CAED,sBACEC,SACQ;EACR,MAAM,qBAAqB,SAAS,sBAAsB;AAC1D,MAAI,qBAAqB,EACvB,OAAM,IAAI,MAAM;AAGlB,SAAO,CAAC,uFAAuF,EAAE,KAAK,qBACpG,aAAa,KAAK,OAAO,CAC1B,CACE,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAC/C,WAAW,KAAK,IAAI,OAAO,mBAAmB,CAAC,CAAC,QAAQ,CAAC;CAC7D;CAED,AAAQ,qBACNC,aACA,SAAS,GACD;EACR,MAAM,SAAS;AASf,MAAI,UAAU,QAAQ;GACpB,IAAI,WAAW;GACf,IAAIC;AACJ,OAAI,MAAM,QAAQ,OAAO,KAAK,EAAE;IAC9B,MAAM,UAAU,OAAO,KAAK,UAAU,CAACC,WAASA,WAAS,OAAO;AAChE,QAAI,YAAY,IAAI;KAClB,WAAW;KACX,OAAO,KAAK,OAAO,SAAS,EAAE;IAC/B;IACD,OAAO,OAAO,KAAK,KAAK,MAAM;GAC/B,OACC,OAAO,OAAO;AAGhB,OAAI,OAAO,SAAS,YAAY,OAAO,YAAY;IACjD,MAAMC,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;IACJ,MAAM,aAAa,OAAO,QAAQ,OAAO,WAAW,CACjD,IAAI,CAAC,CAAC,KAAK,MAAM,KAAK;KACrB,MAAM,aAAa,OAAO,UAAU,SAAS,IAAI,GAC7C,KACA;AACJ,YAAO,GAAG,IAAI,OAAO,OAAO,CAAC,CAAC,EAAE,IAAI,GAAG,EAAE,KAAK,qBAC5C,OACA,SAAS,EACV,GAAG,YAAY;IACjB,EAAC,CACD,KAAK,KAAK;AACb,WAAO,CAAC,GAAG,EAAE,WAAW,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,CAAC,EAAEA,eAAa;GACpE;AACD,OAAI,OAAO,SAAS,WAAW,OAAO,OAAO;IAC3C,MAAMA,gBAAc,OAAO,cACvB,CAAC,IAAI,EAAE,OAAO,aAAa,GAC3B;AACJ,WAAO,CAAC,QAAQ,EAAE,IAAI,OAAO,OAAO,GAAG,KAAK,qBAC1C,OAAO,OACP,SAAS,EACV,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,CAAC,EAAE,EAAEA,eAAa;GAC/C;GACD,MAAM,aAAa,WAAW,gBAAgB;GAC9C,MAAM,cAAc,OAAO,cAAc,CAAC,IAAI,EAAE,OAAO,aAAa,GAAG;AACvE,UAAO,GAAG,OAAO,cAAc,YAAY;EAC5C;AAED,MAAI,WAAW,OACb,QAAO,OAAO,MACX,IAAI,CAAC,MAAM,KAAK,qBAAqB,GAAG,OAAO,CAAC,CAChD,KAAK,CAAC,EAAE,EAAE,IAAI,OAAO,SAAS,EAAE,EAAE,CAAC;AAGxC,QAAM,IAAI,MAAM;CACjB;CAED,OAAO,cAAwCP,QAAW;AACxD,SAAO,IAAI,KAAQ;CACpB;CAED,OAAO,yBACLC,SACA;EACA,MAAM,YAAY,EAAE,OAClB,OAAO,YACL,OAAO,QAAQ,QAAQ,CAAC,IACtB,CAAC,CAAC,MAAM,YAAY,KAClB,CAAC,MAAM,EAAE,QAAQ,CAAC,SAAS,YAAY,AAAC,EAC3C,CACF,CACF;AAED,SAAO,IAAI,KAAuB;CACnC;AACF;;;;;AAYD,IAAsB,mCAAtB,cAGU,iBAAoB;CAC5B,AAAQ;CAER,YAAY,EAAE,aAAwD,EAAE;EACtE,MAAM,GAAG,UAAU;EACnB,KAAK,wBAAwB,IAAI,mCAC/B;CAEH;CAUD,MAAM,MAAMC,MAA0B;EACpC,IAAI;AACJ,MAAI;GACF,cAAc,MAAM,KAAK,sBAAsB,MAAM,KAAK;EAC3D,SAAQ,GAAG;AACV,SAAM,IAAI,sBACR,CAAC,wBAAwB,EAAE,KAAK,UAAU,EAAE,GAAG,EAC/C;EAEH;AAED,SAAO,KAAK,gBAAgB,YAAY;CACzC;CAED,wBAAgC;AAC9B,SAAO,KAAK,sBAAsB,uBAAuB;CAC1D;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.cjs","names":["x: any","fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }","documents: DocumentInterface[]","query: string","callbacks?: Callbacks"],"sources":["../../../src/retrievers/document_compressors/index.ts"],"sourcesContent":["import type { DocumentInterface } from \"@langchain/core/documents\";\nimport { BaseDocumentTransformer } from \"@langchain/core/documents\";\nimport { Callbacks } from \"@langchain/core/callbacks/manager\";\n\n/**\n * Base Document Compression class. All compressors should extend this class.\n */\nexport abstract class BaseDocumentCompressor {\n /**\n * Abstract method that must be implemented by any class that extends\n * `BaseDocumentCompressor`. This method takes an array of `Document`\n * objects and a query string as parameters and returns a Promise that\n * resolves with an array of compressed `Document` objects.\n * @param documents An array of `Document` objects to be compressed.\n * @param query A query string.\n * @returns A Promise that resolves with an array of compressed `Document` objects.\n */\n abstract compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]>;\n\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n static isBaseDocumentCompressor(x: any): x is BaseDocumentCompressor {\n return x?.compressDocuments !== undefined;\n }\n}\n\n/**\n * Document compressor that uses a pipeline of Transformers.\n * @example\n * ```typescript\n * const compressorPipeline = new DocumentCompressorPipeline({\n * transformers: [\n * new RecursiveCharacterTextSplitter({\n * chunkSize: 200,\n * chunkOverlap: 0,\n * }),\n * new EmbeddingsFilter({\n * embeddings: new OpenAIEmbeddings(),\n * similarityThreshold: 0.8,\n * k: 5,\n * }),\n * ],\n * });\n * const retriever = new ContextualCompressionRetriever({\n * baseCompressor: compressorPipeline,\n * baseRetriever: new TavilySearchAPIRetriever({\n * includeRawContent: true,\n * }),\n * });\n * const retrievedDocs = await retriever.invoke(\n * \"What did the speaker say about Justice Breyer in the 2022 State of the Union?\",\n * );\n * console.log({ retrievedDocs });\n * ```\n */\nexport class DocumentCompressorPipeline extends BaseDocumentCompressor {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n\n constructor(fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }) {\n super();\n this.transformers = fields.transformers;\n }\n\n async compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]> {\n let transformedDocuments = documents;\n for (const transformer of this.transformers) {\n if (BaseDocumentCompressor.isBaseDocumentCompressor(transformer)) {\n transformedDocuments = await transformer.compressDocuments(\n transformedDocuments,\n query,\n callbacks\n );\n } else {\n transformedDocuments
|
|
1
|
+
{"version":3,"file":"index.cjs","names":["x: any","fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }","documents: DocumentInterface[]","query: string","callbacks?: Callbacks"],"sources":["../../../src/retrievers/document_compressors/index.ts"],"sourcesContent":["import type { DocumentInterface } from \"@langchain/core/documents\";\nimport { BaseDocumentTransformer } from \"@langchain/core/documents\";\nimport { Callbacks } from \"@langchain/core/callbacks/manager\";\n\n/**\n * Base Document Compression class. All compressors should extend this class.\n */\nexport abstract class BaseDocumentCompressor {\n /**\n * Abstract method that must be implemented by any class that extends\n * `BaseDocumentCompressor`. This method takes an array of `Document`\n * objects and a query string as parameters and returns a Promise that\n * resolves with an array of compressed `Document` objects.\n * @param documents An array of `Document` objects to be compressed.\n * @param query A query string.\n * @returns A Promise that resolves with an array of compressed `Document` objects.\n */\n abstract compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]>;\n\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n static isBaseDocumentCompressor(x: any): x is BaseDocumentCompressor {\n return x?.compressDocuments !== undefined;\n }\n}\n\n/**\n * Document compressor that uses a pipeline of Transformers.\n * @example\n * ```typescript\n * const compressorPipeline = new DocumentCompressorPipeline({\n * transformers: [\n * new RecursiveCharacterTextSplitter({\n * chunkSize: 200,\n * chunkOverlap: 0,\n * }),\n * new EmbeddingsFilter({\n * embeddings: new OpenAIEmbeddings(),\n * similarityThreshold: 0.8,\n * k: 5,\n * }),\n * ],\n * });\n * const retriever = new ContextualCompressionRetriever({\n * baseCompressor: compressorPipeline,\n * baseRetriever: new TavilySearchAPIRetriever({\n * includeRawContent: true,\n * }),\n * });\n * const retrievedDocs = await retriever.invoke(\n * \"What did the speaker say about Justice Breyer in the 2022 State of the Union?\",\n * );\n * console.log({ retrievedDocs });\n * ```\n */\nexport class DocumentCompressorPipeline extends BaseDocumentCompressor {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n\n constructor(fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }) {\n super();\n this.transformers = fields.transformers;\n }\n\n async compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]> {\n let transformedDocuments = documents;\n for (const transformer of this.transformers) {\n if (BaseDocumentCompressor.isBaseDocumentCompressor(transformer)) {\n transformedDocuments = await transformer.compressDocuments(\n transformedDocuments,\n query,\n callbacks\n );\n } else {\n transformedDocuments =\n await transformer.transformDocuments(transformedDocuments);\n }\n }\n return transformedDocuments;\n }\n}\n"],"mappings":";;;;;;;;;;;AAOA,IAAsB,yBAAtB,MAA6C;CAiB3C,OAAO,yBAAyBA,GAAqC;AACnE,SAAO,GAAG,sBAAsB;CACjC;AACF;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA+BD,IAAa,6BAAb,cAAgD,uBAAuB;CACrE;CAEA,YAAYC,QAET;EACD,OAAO;EACP,KAAK,eAAe,OAAO;CAC5B;CAED,MAAM,kBACJC,WACAC,OACAC,WAC8B;EAC9B,IAAI,uBAAuB;AAC3B,OAAK,MAAM,eAAe,KAAK,aAC7B,KAAI,uBAAuB,yBAAyB,YAAY,EAC9D,uBAAuB,MAAM,YAAY,kBACvC,sBACA,OACA,UACD;OAED,uBACE,MAAM,YAAY,mBAAmB,qBAAqB;AAGhE,SAAO;CACR;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","names":["x: any","fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }","documents: DocumentInterface[]","query: string","callbacks?: Callbacks"],"sources":["../../../src/retrievers/document_compressors/index.ts"],"sourcesContent":["import type { DocumentInterface } from \"@langchain/core/documents\";\nimport { BaseDocumentTransformer } from \"@langchain/core/documents\";\nimport { Callbacks } from \"@langchain/core/callbacks/manager\";\n\n/**\n * Base Document Compression class. All compressors should extend this class.\n */\nexport abstract class BaseDocumentCompressor {\n /**\n * Abstract method that must be implemented by any class that extends\n * `BaseDocumentCompressor`. This method takes an array of `Document`\n * objects and a query string as parameters and returns a Promise that\n * resolves with an array of compressed `Document` objects.\n * @param documents An array of `Document` objects to be compressed.\n * @param query A query string.\n * @returns A Promise that resolves with an array of compressed `Document` objects.\n */\n abstract compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]>;\n\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n static isBaseDocumentCompressor(x: any): x is BaseDocumentCompressor {\n return x?.compressDocuments !== undefined;\n }\n}\n\n/**\n * Document compressor that uses a pipeline of Transformers.\n * @example\n * ```typescript\n * const compressorPipeline = new DocumentCompressorPipeline({\n * transformers: [\n * new RecursiveCharacterTextSplitter({\n * chunkSize: 200,\n * chunkOverlap: 0,\n * }),\n * new EmbeddingsFilter({\n * embeddings: new OpenAIEmbeddings(),\n * similarityThreshold: 0.8,\n * k: 5,\n * }),\n * ],\n * });\n * const retriever = new ContextualCompressionRetriever({\n * baseCompressor: compressorPipeline,\n * baseRetriever: new TavilySearchAPIRetriever({\n * includeRawContent: true,\n * }),\n * });\n * const retrievedDocs = await retriever.invoke(\n * \"What did the speaker say about Justice Breyer in the 2022 State of the Union?\",\n * );\n * console.log({ retrievedDocs });\n * ```\n */\nexport class DocumentCompressorPipeline extends BaseDocumentCompressor {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n\n constructor(fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }) {\n super();\n this.transformers = fields.transformers;\n }\n\n async compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]> {\n let transformedDocuments = documents;\n for (const transformer of this.transformers) {\n if (BaseDocumentCompressor.isBaseDocumentCompressor(transformer)) {\n transformedDocuments = await transformer.compressDocuments(\n transformedDocuments,\n query,\n callbacks\n );\n } else {\n transformedDocuments
|
|
1
|
+
{"version":3,"file":"index.js","names":["x: any","fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }","documents: DocumentInterface[]","query: string","callbacks?: Callbacks"],"sources":["../../../src/retrievers/document_compressors/index.ts"],"sourcesContent":["import type { DocumentInterface } from \"@langchain/core/documents\";\nimport { BaseDocumentTransformer } from \"@langchain/core/documents\";\nimport { Callbacks } from \"@langchain/core/callbacks/manager\";\n\n/**\n * Base Document Compression class. All compressors should extend this class.\n */\nexport abstract class BaseDocumentCompressor {\n /**\n * Abstract method that must be implemented by any class that extends\n * `BaseDocumentCompressor`. This method takes an array of `Document`\n * objects and a query string as parameters and returns a Promise that\n * resolves with an array of compressed `Document` objects.\n * @param documents An array of `Document` objects to be compressed.\n * @param query A query string.\n * @returns A Promise that resolves with an array of compressed `Document` objects.\n */\n abstract compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]>;\n\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n static isBaseDocumentCompressor(x: any): x is BaseDocumentCompressor {\n return x?.compressDocuments !== undefined;\n }\n}\n\n/**\n * Document compressor that uses a pipeline of Transformers.\n * @example\n * ```typescript\n * const compressorPipeline = new DocumentCompressorPipeline({\n * transformers: [\n * new RecursiveCharacterTextSplitter({\n * chunkSize: 200,\n * chunkOverlap: 0,\n * }),\n * new EmbeddingsFilter({\n * embeddings: new OpenAIEmbeddings(),\n * similarityThreshold: 0.8,\n * k: 5,\n * }),\n * ],\n * });\n * const retriever = new ContextualCompressionRetriever({\n * baseCompressor: compressorPipeline,\n * baseRetriever: new TavilySearchAPIRetriever({\n * includeRawContent: true,\n * }),\n * });\n * const retrievedDocs = await retriever.invoke(\n * \"What did the speaker say about Justice Breyer in the 2022 State of the Union?\",\n * );\n * console.log({ retrievedDocs });\n * ```\n */\nexport class DocumentCompressorPipeline extends BaseDocumentCompressor {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n\n constructor(fields: {\n transformers: (BaseDocumentTransformer | BaseDocumentCompressor)[];\n }) {\n super();\n this.transformers = fields.transformers;\n }\n\n async compressDocuments(\n documents: DocumentInterface[],\n query: string,\n callbacks?: Callbacks\n ): Promise<DocumentInterface[]> {\n let transformedDocuments = documents;\n for (const transformer of this.transformers) {\n if (BaseDocumentCompressor.isBaseDocumentCompressor(transformer)) {\n transformedDocuments = await transformer.compressDocuments(\n transformedDocuments,\n query,\n callbacks\n );\n } else {\n transformedDocuments =\n await transformer.transformDocuments(transformedDocuments);\n }\n }\n return transformedDocuments;\n }\n}\n"],"mappings":";;;;;;;;;;;AAOA,IAAsB,yBAAtB,MAA6C;CAiB3C,OAAO,yBAAyBA,GAAqC;AACnE,SAAO,GAAG,sBAAsB;CACjC;AACF;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA+BD,IAAa,6BAAb,cAAgD,uBAAuB;CACrE;CAEA,YAAYC,QAET;EACD,OAAO;EACP,KAAK,eAAe,OAAO;CAC5B;CAED,MAAM,kBACJC,WACAC,OACAC,WAC8B;EAC9B,IAAI,uBAAuB;AAC3B,OAAK,MAAM,eAAe,KAAK,aAC7B,KAAI,uBAAuB,yBAAyB,YAAY,EAC9D,uBAAuB,MAAM,YAAY,kBACvC,sBACA,OACA,UACD;OAED,uBACE,MAAM,YAAY,mBAAmB,qBAAqB;AAGhE,SAAO;CACR;AACF"}
|
package/dist/retrievers/hyde.cjs
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
const require_rolldown_runtime = require('../_virtual/rolldown_runtime.cjs');
|
|
2
2
|
const __langchain_core_prompts = require_rolldown_runtime.__toESM(require("@langchain/core/prompts"));
|
|
3
|
-
const __langchain_core_vectorstores = require_rolldown_runtime.__toESM(require("@langchain/core/vectorstores"));
|
|
4
3
|
const __langchain_core_prompt_values = require_rolldown_runtime.__toESM(require("@langchain/core/prompt_values"));
|
|
4
|
+
const __langchain_core_vectorstores = require_rolldown_runtime.__toESM(require("@langchain/core/vectorstores"));
|
|
5
5
|
|
|
6
6
|
//#region src/retrievers/hyde.ts
|
|
7
7
|
var hyde_exports = {};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"hyde.cjs","names":["VectorStoreRetriever","fields: HydeRetrieverOptions<V>","query: string","runManager?: CallbackManagerForRetrieverRun","value: BasePromptValue","StringPromptValue","key: PromptKey","template: string","PromptTemplate"],"sources":["../../src/retrievers/hyde.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { Document } from \"@langchain/core/documents\";\nimport { PromptTemplate, BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n StringPromptValue,\n BasePromptValue,\n} from \"@langchain/core/prompt_values\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\nimport { CallbackManagerForRetrieverRun } from \"@langchain/core/callbacks/manager\";\n\n/**\n * A string that corresponds to a specific prompt template.\n */\nexport type PromptKey =\n | \"websearch\"\n | \"scifact\"\n | \"arguana\"\n | \"trec-covid\"\n | \"fiqa\"\n | \"dbpedia-entity\"\n | \"trec-news\"\n | \"mr-tydi\";\n\n/**\n * Options for the HydeRetriever class, which includes a BaseLanguageModel\n * instance, a VectorStore instance, and an optional promptTemplate which\n * can either be a BasePromptTemplate instance or a PromptKey.\n */\nexport type HydeRetrieverOptions<V extends VectorStore> =\n VectorStoreRetrieverInput<V> & {\n llm: BaseLanguageModelInterface;\n promptTemplate?: BasePromptTemplate | PromptKey;\n };\n\n/**\n * A class for retrieving relevant documents based on a given query. It\n * extends the VectorStoreRetriever class and uses a BaseLanguageModel to\n * generate a hypothetical answer to the query, which is then used to\n * retrieve relevant documents.\n * @example\n * ```typescript\n * const retriever = new HydeRetriever({\n * vectorStore: new MemoryVectorStore(new OpenAIEmbeddings()),\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\" }),\n * k: 1,\n * });\n * await vectorStore.addDocuments(\n * [\n * \"My name is John.\",\n * \"My name is Bob.\",\n * \"My favourite food is pizza.\",\n * \"My favourite food is pasta.\",\n * ].map((pageContent) => new Document({ pageContent })),\n * );\n * const results = await retriever.invoke(\n * \"What is my favourite food?\",\n * );\n * ```\n */\nexport class HydeRetriever<\n V extends VectorStore = VectorStore
|
|
1
|
+
{"version":3,"file":"hyde.cjs","names":["VectorStoreRetriever","fields: HydeRetrieverOptions<V>","query: string","runManager?: CallbackManagerForRetrieverRun","value: BasePromptValue","StringPromptValue","key: PromptKey","template: string","PromptTemplate"],"sources":["../../src/retrievers/hyde.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { Document } from \"@langchain/core/documents\";\nimport { PromptTemplate, BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n StringPromptValue,\n BasePromptValue,\n} from \"@langchain/core/prompt_values\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\nimport { CallbackManagerForRetrieverRun } from \"@langchain/core/callbacks/manager\";\n\n/**\n * A string that corresponds to a specific prompt template.\n */\nexport type PromptKey =\n | \"websearch\"\n | \"scifact\"\n | \"arguana\"\n | \"trec-covid\"\n | \"fiqa\"\n | \"dbpedia-entity\"\n | \"trec-news\"\n | \"mr-tydi\";\n\n/**\n * Options for the HydeRetriever class, which includes a BaseLanguageModel\n * instance, a VectorStore instance, and an optional promptTemplate which\n * can either be a BasePromptTemplate instance or a PromptKey.\n */\nexport type HydeRetrieverOptions<V extends VectorStore> =\n VectorStoreRetrieverInput<V> & {\n llm: BaseLanguageModelInterface;\n promptTemplate?: BasePromptTemplate | PromptKey;\n };\n\n/**\n * A class for retrieving relevant documents based on a given query. It\n * extends the VectorStoreRetriever class and uses a BaseLanguageModel to\n * generate a hypothetical answer to the query, which is then used to\n * retrieve relevant documents.\n * @example\n * ```typescript\n * const retriever = new HydeRetriever({\n * vectorStore: new MemoryVectorStore(new OpenAIEmbeddings()),\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\" }),\n * k: 1,\n * });\n * await vectorStore.addDocuments(\n * [\n * \"My name is John.\",\n * \"My name is Bob.\",\n * \"My favourite food is pizza.\",\n * \"My favourite food is pasta.\",\n * ].map((pageContent) => new Document({ pageContent })),\n * );\n * const results = await retriever.invoke(\n * \"What is my favourite food?\",\n * );\n * ```\n */\nexport class HydeRetriever<\n V extends VectorStore = VectorStore,\n> extends VectorStoreRetriever<V> {\n static lc_name() {\n return \"HydeRetriever\";\n }\n\n get lc_namespace(): string[] {\n return [\"langchain\", \"retrievers\", \"hyde\"];\n }\n\n llm: BaseLanguageModelInterface;\n\n promptTemplate?: BasePromptTemplate;\n\n constructor(fields: HydeRetrieverOptions<V>) {\n super(fields);\n this.llm = fields.llm;\n this.promptTemplate =\n typeof fields.promptTemplate === \"string\"\n ? getPromptTemplateFromKey(fields.promptTemplate)\n : fields.promptTemplate;\n if (this.promptTemplate) {\n const { inputVariables } = this.promptTemplate;\n if (inputVariables.length !== 1 && inputVariables[0] !== \"question\") {\n throw new Error(\n `Prompt template must accept a single input variable 'question'. Invalid input variables for prompt template: ${inputVariables}`\n );\n }\n }\n }\n\n async _getRelevantDocuments(\n query: string,\n runManager?: CallbackManagerForRetrieverRun\n ): Promise<Document[]> {\n let value: BasePromptValue = new StringPromptValue(query);\n\n // Use a custom template if provided\n if (this.promptTemplate) {\n value = await this.promptTemplate.formatPromptValue({ question: query });\n }\n\n // Get a hypothetical answer from the LLM\n const res = await this.llm.generatePrompt([value]);\n const answer = res.generations[0][0].text;\n\n // Retrieve relevant documents based on the hypothetical answer\n const results = await this.vectorStore.similaritySearch(\n answer,\n this.k,\n this.filter,\n runManager?.getChild(\"vectorstore\")\n );\n\n return results;\n }\n}\n\n/**\n * Returns a BasePromptTemplate instance based on a given PromptKey.\n */\nexport function getPromptTemplateFromKey(key: PromptKey): BasePromptTemplate {\n let template: string;\n\n switch (key) {\n case \"websearch\":\n template = `Please write a passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"scifact\":\n template = `Please write a scientific paper passage to support/refute the claim\nClaim: {question}\nPassage:`;\n break;\n case \"arguana\":\n template = `Please write a counter argument for the passage\nPassage: {question}\nCounter Argument:`;\n break;\n case \"trec-covid\":\n template = `Please write a scientific paper passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"fiqa\":\n template = `Please write a financial article passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"dbpedia-entity\":\n template = `Please write a passage to answer the question.\nQuestion: {question}\nPassage:`;\n break;\n case \"trec-news\":\n template = `Please write a news passage about the topic.\nTopic: {question}\nPassage:`;\n break;\n case \"mr-tydi\":\n template = `Please write a passage in Swahili/Korean/Japanese/Bengali to answer the question in detail.\nQuestion: {question}\nPassage:`;\n break;\n default:\n throw new Error(`Invalid prompt key: ${key}`);\n }\n\n return PromptTemplate.fromTemplate(template);\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA+DA,IAAa,gBAAb,cAEUA,mDAAwB;CAChC,OAAO,UAAU;AACf,SAAO;CACR;CAED,IAAI,eAAyB;AAC3B,SAAO;GAAC;GAAa;GAAc;EAAO;CAC3C;CAED;CAEA;CAEA,YAAYC,QAAiC;EAC3C,MAAM,OAAO;EACb,KAAK,MAAM,OAAO;EAClB,KAAK,iBACH,OAAO,OAAO,mBAAmB,WAC7B,yBAAyB,OAAO,eAAe,GAC/C,OAAO;AACb,MAAI,KAAK,gBAAgB;GACvB,MAAM,EAAE,gBAAgB,GAAG,KAAK;AAChC,OAAI,eAAe,WAAW,KAAK,eAAe,OAAO,WACvD,OAAM,IAAI,MACR,CAAC,6GAA6G,EAAE,gBAAgB;EAGrI;CACF;CAED,MAAM,sBACJC,OACAC,YACqB;EACrB,IAAIC,QAAyB,IAAIC,iDAAkB;AAGnD,MAAI,KAAK,gBACP,QAAQ,MAAM,KAAK,eAAe,kBAAkB,EAAE,UAAU,MAAO,EAAC;EAI1E,MAAM,MAAM,MAAM,KAAK,IAAI,eAAe,CAAC,KAAM,EAAC;EAClD,MAAM,SAAS,IAAI,YAAY,GAAG,GAAG;EAGrC,MAAM,UAAU,MAAM,KAAK,YAAY,iBACrC,QACA,KAAK,GACL,KAAK,QACL,YAAY,SAAS,cAAc,CACpC;AAED,SAAO;CACR;AACF;;;;AAKD,SAAgB,yBAAyBC,KAAoC;CAC3E,IAAIC;AAEJ,SAAQ,KAAR;EACE,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;iBAED,CAAC;AACZ;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,QACE,OAAM,IAAI,MAAM,CAAC,oBAAoB,EAAE,KAAK;CAC/C;AAED,QAAOC,wCAAe,aAAa,SAAS;AAC7C"}
|
package/dist/retrievers/hyde.js
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { __export } from "../_virtual/rolldown_runtime.js";
|
|
2
2
|
import { PromptTemplate } from "@langchain/core/prompts";
|
|
3
|
-
import { VectorStoreRetriever } from "@langchain/core/vectorstores";
|
|
4
3
|
import { StringPromptValue } from "@langchain/core/prompt_values";
|
|
4
|
+
import { VectorStoreRetriever } from "@langchain/core/vectorstores";
|
|
5
5
|
|
|
6
6
|
//#region src/retrievers/hyde.ts
|
|
7
7
|
var hyde_exports = {};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"hyde.js","names":["fields: HydeRetrieverOptions<V>","query: string","runManager?: CallbackManagerForRetrieverRun","value: BasePromptValue","key: PromptKey","template: string"],"sources":["../../src/retrievers/hyde.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { Document } from \"@langchain/core/documents\";\nimport { PromptTemplate, BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n StringPromptValue,\n BasePromptValue,\n} from \"@langchain/core/prompt_values\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\nimport { CallbackManagerForRetrieverRun } from \"@langchain/core/callbacks/manager\";\n\n/**\n * A string that corresponds to a specific prompt template.\n */\nexport type PromptKey =\n | \"websearch\"\n | \"scifact\"\n | \"arguana\"\n | \"trec-covid\"\n | \"fiqa\"\n | \"dbpedia-entity\"\n | \"trec-news\"\n | \"mr-tydi\";\n\n/**\n * Options for the HydeRetriever class, which includes a BaseLanguageModel\n * instance, a VectorStore instance, and an optional promptTemplate which\n * can either be a BasePromptTemplate instance or a PromptKey.\n */\nexport type HydeRetrieverOptions<V extends VectorStore> =\n VectorStoreRetrieverInput<V> & {\n llm: BaseLanguageModelInterface;\n promptTemplate?: BasePromptTemplate | PromptKey;\n };\n\n/**\n * A class for retrieving relevant documents based on a given query. It\n * extends the VectorStoreRetriever class and uses a BaseLanguageModel to\n * generate a hypothetical answer to the query, which is then used to\n * retrieve relevant documents.\n * @example\n * ```typescript\n * const retriever = new HydeRetriever({\n * vectorStore: new MemoryVectorStore(new OpenAIEmbeddings()),\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\" }),\n * k: 1,\n * });\n * await vectorStore.addDocuments(\n * [\n * \"My name is John.\",\n * \"My name is Bob.\",\n * \"My favourite food is pizza.\",\n * \"My favourite food is pasta.\",\n * ].map((pageContent) => new Document({ pageContent })),\n * );\n * const results = await retriever.invoke(\n * \"What is my favourite food?\",\n * );\n * ```\n */\nexport class HydeRetriever<\n V extends VectorStore = VectorStore
|
|
1
|
+
{"version":3,"file":"hyde.js","names":["fields: HydeRetrieverOptions<V>","query: string","runManager?: CallbackManagerForRetrieverRun","value: BasePromptValue","key: PromptKey","template: string"],"sources":["../../src/retrievers/hyde.ts"],"sourcesContent":["import type { BaseLanguageModelInterface } from \"@langchain/core/language_models/base\";\nimport { Document } from \"@langchain/core/documents\";\nimport { PromptTemplate, BasePromptTemplate } from \"@langchain/core/prompts\";\nimport {\n StringPromptValue,\n BasePromptValue,\n} from \"@langchain/core/prompt_values\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\nimport { CallbackManagerForRetrieverRun } from \"@langchain/core/callbacks/manager\";\n\n/**\n * A string that corresponds to a specific prompt template.\n */\nexport type PromptKey =\n | \"websearch\"\n | \"scifact\"\n | \"arguana\"\n | \"trec-covid\"\n | \"fiqa\"\n | \"dbpedia-entity\"\n | \"trec-news\"\n | \"mr-tydi\";\n\n/**\n * Options for the HydeRetriever class, which includes a BaseLanguageModel\n * instance, a VectorStore instance, and an optional promptTemplate which\n * can either be a BasePromptTemplate instance or a PromptKey.\n */\nexport type HydeRetrieverOptions<V extends VectorStore> =\n VectorStoreRetrieverInput<V> & {\n llm: BaseLanguageModelInterface;\n promptTemplate?: BasePromptTemplate | PromptKey;\n };\n\n/**\n * A class for retrieving relevant documents based on a given query. It\n * extends the VectorStoreRetriever class and uses a BaseLanguageModel to\n * generate a hypothetical answer to the query, which is then used to\n * retrieve relevant documents.\n * @example\n * ```typescript\n * const retriever = new HydeRetriever({\n * vectorStore: new MemoryVectorStore(new OpenAIEmbeddings()),\n * llm: new ChatOpenAI({ model: \"gpt-4o-mini\" }),\n * k: 1,\n * });\n * await vectorStore.addDocuments(\n * [\n * \"My name is John.\",\n * \"My name is Bob.\",\n * \"My favourite food is pizza.\",\n * \"My favourite food is pasta.\",\n * ].map((pageContent) => new Document({ pageContent })),\n * );\n * const results = await retriever.invoke(\n * \"What is my favourite food?\",\n * );\n * ```\n */\nexport class HydeRetriever<\n V extends VectorStore = VectorStore,\n> extends VectorStoreRetriever<V> {\n static lc_name() {\n return \"HydeRetriever\";\n }\n\n get lc_namespace(): string[] {\n return [\"langchain\", \"retrievers\", \"hyde\"];\n }\n\n llm: BaseLanguageModelInterface;\n\n promptTemplate?: BasePromptTemplate;\n\n constructor(fields: HydeRetrieverOptions<V>) {\n super(fields);\n this.llm = fields.llm;\n this.promptTemplate =\n typeof fields.promptTemplate === \"string\"\n ? getPromptTemplateFromKey(fields.promptTemplate)\n : fields.promptTemplate;\n if (this.promptTemplate) {\n const { inputVariables } = this.promptTemplate;\n if (inputVariables.length !== 1 && inputVariables[0] !== \"question\") {\n throw new Error(\n `Prompt template must accept a single input variable 'question'. Invalid input variables for prompt template: ${inputVariables}`\n );\n }\n }\n }\n\n async _getRelevantDocuments(\n query: string,\n runManager?: CallbackManagerForRetrieverRun\n ): Promise<Document[]> {\n let value: BasePromptValue = new StringPromptValue(query);\n\n // Use a custom template if provided\n if (this.promptTemplate) {\n value = await this.promptTemplate.formatPromptValue({ question: query });\n }\n\n // Get a hypothetical answer from the LLM\n const res = await this.llm.generatePrompt([value]);\n const answer = res.generations[0][0].text;\n\n // Retrieve relevant documents based on the hypothetical answer\n const results = await this.vectorStore.similaritySearch(\n answer,\n this.k,\n this.filter,\n runManager?.getChild(\"vectorstore\")\n );\n\n return results;\n }\n}\n\n/**\n * Returns a BasePromptTemplate instance based on a given PromptKey.\n */\nexport function getPromptTemplateFromKey(key: PromptKey): BasePromptTemplate {\n let template: string;\n\n switch (key) {\n case \"websearch\":\n template = `Please write a passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"scifact\":\n template = `Please write a scientific paper passage to support/refute the claim\nClaim: {question}\nPassage:`;\n break;\n case \"arguana\":\n template = `Please write a counter argument for the passage\nPassage: {question}\nCounter Argument:`;\n break;\n case \"trec-covid\":\n template = `Please write a scientific paper passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"fiqa\":\n template = `Please write a financial article passage to answer the question\nQuestion: {question}\nPassage:`;\n break;\n case \"dbpedia-entity\":\n template = `Please write a passage to answer the question.\nQuestion: {question}\nPassage:`;\n break;\n case \"trec-news\":\n template = `Please write a news passage about the topic.\nTopic: {question}\nPassage:`;\n break;\n case \"mr-tydi\":\n template = `Please write a passage in Swahili/Korean/Japanese/Bengali to answer the question in detail.\nQuestion: {question}\nPassage:`;\n break;\n default:\n throw new Error(`Invalid prompt key: ${key}`);\n }\n\n return PromptTemplate.fromTemplate(template);\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AA+DA,IAAa,gBAAb,cAEU,qBAAwB;CAChC,OAAO,UAAU;AACf,SAAO;CACR;CAED,IAAI,eAAyB;AAC3B,SAAO;GAAC;GAAa;GAAc;EAAO;CAC3C;CAED;CAEA;CAEA,YAAYA,QAAiC;EAC3C,MAAM,OAAO;EACb,KAAK,MAAM,OAAO;EAClB,KAAK,iBACH,OAAO,OAAO,mBAAmB,WAC7B,yBAAyB,OAAO,eAAe,GAC/C,OAAO;AACb,MAAI,KAAK,gBAAgB;GACvB,MAAM,EAAE,gBAAgB,GAAG,KAAK;AAChC,OAAI,eAAe,WAAW,KAAK,eAAe,OAAO,WACvD,OAAM,IAAI,MACR,CAAC,6GAA6G,EAAE,gBAAgB;EAGrI;CACF;CAED,MAAM,sBACJC,OACAC,YACqB;EACrB,IAAIC,QAAyB,IAAI,kBAAkB;AAGnD,MAAI,KAAK,gBACP,QAAQ,MAAM,KAAK,eAAe,kBAAkB,EAAE,UAAU,MAAO,EAAC;EAI1E,MAAM,MAAM,MAAM,KAAK,IAAI,eAAe,CAAC,KAAM,EAAC;EAClD,MAAM,SAAS,IAAI,YAAY,GAAG,GAAG;EAGrC,MAAM,UAAU,MAAM,KAAK,YAAY,iBACrC,QACA,KAAK,GACL,KAAK,QACL,YAAY,SAAS,cAAc,CACpC;AAED,SAAO;CACR;AACF;;;;AAKD,SAAgB,yBAAyBC,KAAoC;CAC3E,IAAIC;AAEJ,SAAQ,KAAR;EACE,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;iBAED,CAAC;AACZ;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,KAAK;GACH,WAAW,CAAC;;QAEV,CAAC;AACH;EACF,QACE,OAAM,IAAI,MAAM,CAAC,oBAAoB,EAAE,KAAK;CAC/C;AAED,QAAO,eAAe,aAAa,SAAS;AAC7C"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
const require_rolldown_runtime = require('../_virtual/rolldown_runtime.cjs');
|
|
2
|
-
const __langchain_core_vectorstores = require_rolldown_runtime.__toESM(require("@langchain/core/vectorstores"));
|
|
3
2
|
const __langchain_core_utils_math = require_rolldown_runtime.__toESM(require("@langchain/core/utils/math"));
|
|
3
|
+
const __langchain_core_vectorstores = require_rolldown_runtime.__toESM(require("@langchain/core/vectorstores"));
|
|
4
4
|
|
|
5
5
|
//#region src/retrievers/matryoshka_retriever.ts
|
|
6
6
|
var matryoshka_retriever_exports = {};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"matryoshka_retriever.cjs","names":["VectorStoreRetriever","fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>","embeddedQuery: number[]","smallResults: DocumentInterface[]","largeEmbeddings: Array<number[]>","func: () => Array<number[]>","query: string","documents: DocumentInterface[]","options?: AddDocumentOptions","newDocuments: Array<DocumentInterface>"],"sources":["../../src/retrievers/matryoshka_retriever.ts"],"sourcesContent":["import { DocumentInterface } from \"@langchain/core/documents\";\nimport { Embeddings } from \"@langchain/core/embeddings\";\nimport {\n cosineSimilarity,\n euclideanDistance,\n innerProduct,\n} from \"@langchain/core/utils/math\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\n\n/**\n * Type for options when adding a document to the VectorStore.\n */\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AddDocumentOptions = Record<string, any>;\n\nexport interface MatryoshkaRetrieverFields {\n /**\n * The number of documents to retrieve from the small store.\n * @default 50\n */\n smallK?: number;\n /**\n * The number of documents to retrieve from the large store.\n * @default 8\n */\n largeK?: number;\n /**\n * The metadata key to store the larger embeddings.\n * @default \"lc_large_embedding\"\n */\n largeEmbeddingKey?: string;\n /**\n * The embedding model to use when generating the large\n * embeddings.\n */\n largeEmbeddingModel: Embeddings;\n /**\n * The type of search to perform using the large embeddings.\n * @default \"cosine\"\n */\n searchType?: \"cosine\" | \"innerProduct\" | \"euclidean\";\n}\n\n/**\n * A retriever that uses two sets of embeddings to perform adaptive retrieval. Based\n * off of the \"Matryoshka embeddings: faster OpenAI vector search using Adaptive Retrieval\"\n * blog post {@link https://supabase.com/blog/matryoshka-embeddings}.\n *\n *\n * This class performs \"Adaptive Retrieval\" for searching text embeddings efficiently using the\n * Matryoshka Representation Learning (MRL) technique. It retrieves documents similar to a query\n * embedding in two steps:\n *\n * First-pass: Uses a lower dimensional sub-vector from the MRL embedding for an initial, fast,\n * but less accurate search.\n *\n * Second-pass: Re-ranks the top results from the first pass using the full, high-dimensional\n * embedding for higher accuracy.\n *\n *\n * This code implements MRL embeddings for efficient vector search by combining faster,\n * lower-dimensional initial search with accurate, high-dimensional re-ranking.\n */\nexport class MatryoshkaRetriever<\n Store extends VectorStore = VectorStore
|
|
1
|
+
{"version":3,"file":"matryoshka_retriever.cjs","names":["VectorStoreRetriever","fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>","embeddedQuery: number[]","smallResults: DocumentInterface[]","largeEmbeddings: Array<number[]>","func: () => Array<number[]>","query: string","documents: DocumentInterface[]","options?: AddDocumentOptions","newDocuments: Array<DocumentInterface>"],"sources":["../../src/retrievers/matryoshka_retriever.ts"],"sourcesContent":["import { DocumentInterface } from \"@langchain/core/documents\";\nimport { Embeddings } from \"@langchain/core/embeddings\";\nimport {\n cosineSimilarity,\n euclideanDistance,\n innerProduct,\n} from \"@langchain/core/utils/math\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\n\n/**\n * Type for options when adding a document to the VectorStore.\n */\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AddDocumentOptions = Record<string, any>;\n\nexport interface MatryoshkaRetrieverFields {\n /**\n * The number of documents to retrieve from the small store.\n * @default 50\n */\n smallK?: number;\n /**\n * The number of documents to retrieve from the large store.\n * @default 8\n */\n largeK?: number;\n /**\n * The metadata key to store the larger embeddings.\n * @default \"lc_large_embedding\"\n */\n largeEmbeddingKey?: string;\n /**\n * The embedding model to use when generating the large\n * embeddings.\n */\n largeEmbeddingModel: Embeddings;\n /**\n * The type of search to perform using the large embeddings.\n * @default \"cosine\"\n */\n searchType?: \"cosine\" | \"innerProduct\" | \"euclidean\";\n}\n\n/**\n * A retriever that uses two sets of embeddings to perform adaptive retrieval. Based\n * off of the \"Matryoshka embeddings: faster OpenAI vector search using Adaptive Retrieval\"\n * blog post {@link https://supabase.com/blog/matryoshka-embeddings}.\n *\n *\n * This class performs \"Adaptive Retrieval\" for searching text embeddings efficiently using the\n * Matryoshka Representation Learning (MRL) technique. It retrieves documents similar to a query\n * embedding in two steps:\n *\n * First-pass: Uses a lower dimensional sub-vector from the MRL embedding for an initial, fast,\n * but less accurate search.\n *\n * Second-pass: Re-ranks the top results from the first pass using the full, high-dimensional\n * embedding for higher accuracy.\n *\n *\n * This code implements MRL embeddings for efficient vector search by combining faster,\n * lower-dimensional initial search with accurate, high-dimensional re-ranking.\n */\nexport class MatryoshkaRetriever<\n Store extends VectorStore = VectorStore,\n> extends VectorStoreRetriever<Store> {\n smallK = 50;\n\n largeK = 8;\n\n largeEmbeddingKey = \"lc_large_embedding\";\n\n largeEmbeddingModel: Embeddings;\n\n searchType: \"cosine\" | \"innerProduct\" | \"euclidean\" = \"cosine\";\n\n constructor(\n fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>\n ) {\n super(fields);\n this.smallK = fields.smallK ?? this.smallK;\n this.largeK = fields.largeK ?? this.largeK;\n this.largeEmbeddingKey = fields.largeEmbeddingKey ?? this.largeEmbeddingKey;\n this.largeEmbeddingModel = fields.largeEmbeddingModel;\n this.searchType = fields.searchType ?? this.searchType;\n }\n\n /**\n * Ranks documents based on their similarity to a query embedding using larger embeddings.\n *\n * This method takes a query embedding and a list of documents (smallResults) as input. Each document\n * in the smallResults array has previously been associated with a large embedding stored in its metadata.\n * Depending on the `searchType` (cosine, innerProduct, or euclidean), it calculates the similarity scores\n * between the query embedding and each document's large embedding. It then ranks the documents based on\n * these similarity scores, from the most similar to the least similar.\n *\n * The method returns a promise that resolves to an array of the top `largeK` documents, where `largeK`\n * is a class property defining the number of documents to return. This subset of documents is determined\n * by sorting the entire list of documents based on their similarity scores and then selecting the top\n * `largeK` documents.\n *\n * @param {number[]} embeddedQuery The embedding of the query, represented as an array of numbers.\n * @param {DocumentInterface[]} smallResults An array of documents, each with metadata that includes a large embedding for similarity comparison.\n * @returns {Promise<DocumentInterface[]>} A promise that resolves to an array of the top `largeK` ranked documents based on their similarity to the query embedding.\n */\n private _rankByLargeEmbeddings(\n embeddedQuery: number[],\n smallResults: DocumentInterface[]\n ): DocumentInterface[] {\n const largeEmbeddings: Array<number[]> = smallResults.map((doc) =>\n JSON.parse(doc.metadata[this.largeEmbeddingKey])\n );\n let func: () => Array<number[]>;\n\n switch (this.searchType) {\n case \"cosine\":\n func = () => cosineSimilarity([embeddedQuery], largeEmbeddings);\n break;\n case \"innerProduct\":\n func = () => innerProduct([embeddedQuery], largeEmbeddings);\n break;\n case \"euclidean\":\n func = () => euclideanDistance([embeddedQuery], largeEmbeddings);\n break;\n default:\n throw new Error(`Unknown search type: ${this.searchType}`);\n }\n\n // Calculate the similarity scores between the query embedding and the large embeddings\n const [similarityScores] = func();\n\n // Create an array of indices from 0 to N-1, where N is the number of documents\n let indices = Array.from(\n { length: smallResults.length },\n (_, index) => index\n );\n\n indices = indices\n .map((v, i) => [similarityScores[i], v])\n .sort(([a], [b]) => b - a)\n .slice(0, this.largeK)\n .map(([, i]) => i);\n\n return indices.map((i) => smallResults[i]);\n }\n\n async _getRelevantDocuments(query: string): Promise<DocumentInterface[]> {\n const [embeddedQuery, smallResults] = await Promise.all([\n this.largeEmbeddingModel.embedQuery(query),\n this.vectorStore.similaritySearch(query, this.smallK, this.filter),\n ]);\n\n return this._rankByLargeEmbeddings(embeddedQuery, smallResults);\n }\n\n /**\n * Override the default `addDocuments` method to embed the documents twice,\n * once using the larger embeddings model, and then again using the default\n * embedding model linked to the vector store.\n *\n * @param {DocumentInterface[]} documents - An array of documents to add to the vector store.\n * @param {AddDocumentOptions} options - An optional object containing additional options for adding documents.\n * @returns {Promise<string[] | void>} A promise that resolves to an array of the document IDs that were added to the vector store.\n */\n override addDocuments = async (\n documents: DocumentInterface[],\n options?: AddDocumentOptions\n ): Promise<string[] | void> => {\n // Insure documents metadata does not contain the large embedding key\n if (documents.some((doc) => this.largeEmbeddingKey in doc.metadata)) {\n throw new Error(\n `All documents must not contain the large embedding key: ${this.largeEmbeddingKey} in their metadata.`\n );\n }\n\n const allDocPageContent = documents.map((doc) => doc.pageContent);\n const allDocLargeEmbeddings =\n await this.largeEmbeddingModel.embedDocuments(allDocPageContent);\n\n const newDocuments: Array<DocumentInterface> = documents.map(\n (doc, idx) => ({\n ...doc,\n metadata: {\n ...doc.metadata,\n [this.largeEmbeddingKey]: JSON.stringify(allDocLargeEmbeddings[idx]),\n },\n })\n );\n\n return this.vectorStore.addDocuments(newDocuments, options);\n };\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;AAmEA,IAAa,sBAAb,cAEUA,mDAA4B;CACpC,SAAS;CAET,SAAS;CAET,oBAAoB;CAEpB;CAEA,aAAsD;CAEtD,YACEC,QACA;EACA,MAAM,OAAO;EACb,KAAK,SAAS,OAAO,UAAU,KAAK;EACpC,KAAK,SAAS,OAAO,UAAU,KAAK;EACpC,KAAK,oBAAoB,OAAO,qBAAqB,KAAK;EAC1D,KAAK,sBAAsB,OAAO;EAClC,KAAK,aAAa,OAAO,cAAc,KAAK;CAC7C;;;;;;;;;;;;;;;;;;;CAoBD,AAAQ,uBACNC,eACAC,cACqB;EACrB,MAAMC,kBAAmC,aAAa,IAAI,CAAC,QACzD,KAAK,MAAM,IAAI,SAAS,KAAK,mBAAmB,CACjD;EACD,IAAIC;AAEJ,UAAQ,KAAK,YAAb;GACE,KAAK;IACH,OAAO,wDAAuB,CAAC,aAAc,GAAE,gBAAgB;AAC/D;GACF,KAAK;IACH,OAAO,oDAAmB,CAAC,aAAc,GAAE,gBAAgB;AAC3D;GACF,KAAK;IACH,OAAO,yDAAwB,CAAC,aAAc,GAAE,gBAAgB;AAChE;GACF,QACE,OAAM,IAAI,MAAM,CAAC,qBAAqB,EAAE,KAAK,YAAY;EAC5D;EAGD,MAAM,CAAC,iBAAiB,GAAG,MAAM;EAGjC,IAAI,UAAU,MAAM,KAClB,EAAE,QAAQ,aAAa,OAAQ,GAC/B,CAAC,GAAG,UAAU,MACf;EAED,UAAU,QACP,IAAI,CAAC,GAAG,MAAM,CAAC,iBAAiB,IAAI,CAAE,EAAC,CACvC,KAAK,CAAC,CAAC,EAAE,EAAE,CAAC,EAAE,KAAK,IAAI,EAAE,CACzB,MAAM,GAAG,KAAK,OAAO,CACrB,IAAI,CAAC,GAAG,EAAE,KAAK,EAAE;AAEpB,SAAO,QAAQ,IAAI,CAAC,MAAM,aAAa,GAAG;CAC3C;CAED,MAAM,sBAAsBC,OAA6C;EACvE,MAAM,CAAC,eAAe,aAAa,GAAG,MAAM,QAAQ,IAAI,CACtD,KAAK,oBAAoB,WAAW,MAAM,EAC1C,KAAK,YAAY,iBAAiB,OAAO,KAAK,QAAQ,KAAK,OAAO,AACnE,EAAC;AAEF,SAAO,KAAK,uBAAuB,eAAe,aAAa;CAChE;;;;;;;;;;CAWD,AAAS,eAAe,OACtBC,WACAC,YAC6B;AAE7B,MAAI,UAAU,KAAK,CAAC,QAAQ,KAAK,qBAAqB,IAAI,SAAS,CACjE,OAAM,IAAI,MACR,CAAC,wDAAwD,EAAE,KAAK,kBAAkB,mBAAmB,CAAC;EAI1G,MAAM,oBAAoB,UAAU,IAAI,CAAC,QAAQ,IAAI,YAAY;EACjE,MAAM,wBACJ,MAAM,KAAK,oBAAoB,eAAe,kBAAkB;EAElE,MAAMC,eAAyC,UAAU,IACvD,CAAC,KAAK,SAAS;GACb,GAAG;GACH,UAAU;IACR,GAAG,IAAI;KACN,KAAK,oBAAoB,KAAK,UAAU,sBAAsB,KAAK;GACrE;EACF,GACF;AAED,SAAO,KAAK,YAAY,aAAa,cAAc,QAAQ;CAC5D;AACF"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { __export } from "../_virtual/rolldown_runtime.js";
|
|
2
|
-
import { VectorStoreRetriever } from "@langchain/core/vectorstores";
|
|
3
2
|
import { cosineSimilarity, euclideanDistance, innerProduct } from "@langchain/core/utils/math";
|
|
3
|
+
import { VectorStoreRetriever } from "@langchain/core/vectorstores";
|
|
4
4
|
|
|
5
5
|
//#region src/retrievers/matryoshka_retriever.ts
|
|
6
6
|
var matryoshka_retriever_exports = {};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"matryoshka_retriever.js","names":["fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>","embeddedQuery: number[]","smallResults: DocumentInterface[]","largeEmbeddings: Array<number[]>","func: () => Array<number[]>","query: string","documents: DocumentInterface[]","options?: AddDocumentOptions","newDocuments: Array<DocumentInterface>"],"sources":["../../src/retrievers/matryoshka_retriever.ts"],"sourcesContent":["import { DocumentInterface } from \"@langchain/core/documents\";\nimport { Embeddings } from \"@langchain/core/embeddings\";\nimport {\n cosineSimilarity,\n euclideanDistance,\n innerProduct,\n} from \"@langchain/core/utils/math\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\n\n/**\n * Type for options when adding a document to the VectorStore.\n */\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AddDocumentOptions = Record<string, any>;\n\nexport interface MatryoshkaRetrieverFields {\n /**\n * The number of documents to retrieve from the small store.\n * @default 50\n */\n smallK?: number;\n /**\n * The number of documents to retrieve from the large store.\n * @default 8\n */\n largeK?: number;\n /**\n * The metadata key to store the larger embeddings.\n * @default \"lc_large_embedding\"\n */\n largeEmbeddingKey?: string;\n /**\n * The embedding model to use when generating the large\n * embeddings.\n */\n largeEmbeddingModel: Embeddings;\n /**\n * The type of search to perform using the large embeddings.\n * @default \"cosine\"\n */\n searchType?: \"cosine\" | \"innerProduct\" | \"euclidean\";\n}\n\n/**\n * A retriever that uses two sets of embeddings to perform adaptive retrieval. Based\n * off of the \"Matryoshka embeddings: faster OpenAI vector search using Adaptive Retrieval\"\n * blog post {@link https://supabase.com/blog/matryoshka-embeddings}.\n *\n *\n * This class performs \"Adaptive Retrieval\" for searching text embeddings efficiently using the\n * Matryoshka Representation Learning (MRL) technique. It retrieves documents similar to a query\n * embedding in two steps:\n *\n * First-pass: Uses a lower dimensional sub-vector from the MRL embedding for an initial, fast,\n * but less accurate search.\n *\n * Second-pass: Re-ranks the top results from the first pass using the full, high-dimensional\n * embedding for higher accuracy.\n *\n *\n * This code implements MRL embeddings for efficient vector search by combining faster,\n * lower-dimensional initial search with accurate, high-dimensional re-ranking.\n */\nexport class MatryoshkaRetriever<\n Store extends VectorStore = VectorStore
|
|
1
|
+
{"version":3,"file":"matryoshka_retriever.js","names":["fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>","embeddedQuery: number[]","smallResults: DocumentInterface[]","largeEmbeddings: Array<number[]>","func: () => Array<number[]>","query: string","documents: DocumentInterface[]","options?: AddDocumentOptions","newDocuments: Array<DocumentInterface>"],"sources":["../../src/retrievers/matryoshka_retriever.ts"],"sourcesContent":["import { DocumentInterface } from \"@langchain/core/documents\";\nimport { Embeddings } from \"@langchain/core/embeddings\";\nimport {\n cosineSimilarity,\n euclideanDistance,\n innerProduct,\n} from \"@langchain/core/utils/math\";\nimport {\n VectorStore,\n VectorStoreRetriever,\n VectorStoreRetrieverInput,\n} from \"@langchain/core/vectorstores\";\n\n/**\n * Type for options when adding a document to the VectorStore.\n */\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\ntype AddDocumentOptions = Record<string, any>;\n\nexport interface MatryoshkaRetrieverFields {\n /**\n * The number of documents to retrieve from the small store.\n * @default 50\n */\n smallK?: number;\n /**\n * The number of documents to retrieve from the large store.\n * @default 8\n */\n largeK?: number;\n /**\n * The metadata key to store the larger embeddings.\n * @default \"lc_large_embedding\"\n */\n largeEmbeddingKey?: string;\n /**\n * The embedding model to use when generating the large\n * embeddings.\n */\n largeEmbeddingModel: Embeddings;\n /**\n * The type of search to perform using the large embeddings.\n * @default \"cosine\"\n */\n searchType?: \"cosine\" | \"innerProduct\" | \"euclidean\";\n}\n\n/**\n * A retriever that uses two sets of embeddings to perform adaptive retrieval. Based\n * off of the \"Matryoshka embeddings: faster OpenAI vector search using Adaptive Retrieval\"\n * blog post {@link https://supabase.com/blog/matryoshka-embeddings}.\n *\n *\n * This class performs \"Adaptive Retrieval\" for searching text embeddings efficiently using the\n * Matryoshka Representation Learning (MRL) technique. It retrieves documents similar to a query\n * embedding in two steps:\n *\n * First-pass: Uses a lower dimensional sub-vector from the MRL embedding for an initial, fast,\n * but less accurate search.\n *\n * Second-pass: Re-ranks the top results from the first pass using the full, high-dimensional\n * embedding for higher accuracy.\n *\n *\n * This code implements MRL embeddings for efficient vector search by combining faster,\n * lower-dimensional initial search with accurate, high-dimensional re-ranking.\n */\nexport class MatryoshkaRetriever<\n Store extends VectorStore = VectorStore,\n> extends VectorStoreRetriever<Store> {\n smallK = 50;\n\n largeK = 8;\n\n largeEmbeddingKey = \"lc_large_embedding\";\n\n largeEmbeddingModel: Embeddings;\n\n searchType: \"cosine\" | \"innerProduct\" | \"euclidean\" = \"cosine\";\n\n constructor(\n fields: MatryoshkaRetrieverFields & VectorStoreRetrieverInput<Store>\n ) {\n super(fields);\n this.smallK = fields.smallK ?? this.smallK;\n this.largeK = fields.largeK ?? this.largeK;\n this.largeEmbeddingKey = fields.largeEmbeddingKey ?? this.largeEmbeddingKey;\n this.largeEmbeddingModel = fields.largeEmbeddingModel;\n this.searchType = fields.searchType ?? this.searchType;\n }\n\n /**\n * Ranks documents based on their similarity to a query embedding using larger embeddings.\n *\n * This method takes a query embedding and a list of documents (smallResults) as input. Each document\n * in the smallResults array has previously been associated with a large embedding stored in its metadata.\n * Depending on the `searchType` (cosine, innerProduct, or euclidean), it calculates the similarity scores\n * between the query embedding and each document's large embedding. It then ranks the documents based on\n * these similarity scores, from the most similar to the least similar.\n *\n * The method returns a promise that resolves to an array of the top `largeK` documents, where `largeK`\n * is a class property defining the number of documents to return. This subset of documents is determined\n * by sorting the entire list of documents based on their similarity scores and then selecting the top\n * `largeK` documents.\n *\n * @param {number[]} embeddedQuery The embedding of the query, represented as an array of numbers.\n * @param {DocumentInterface[]} smallResults An array of documents, each with metadata that includes a large embedding for similarity comparison.\n * @returns {Promise<DocumentInterface[]>} A promise that resolves to an array of the top `largeK` ranked documents based on their similarity to the query embedding.\n */\n private _rankByLargeEmbeddings(\n embeddedQuery: number[],\n smallResults: DocumentInterface[]\n ): DocumentInterface[] {\n const largeEmbeddings: Array<number[]> = smallResults.map((doc) =>\n JSON.parse(doc.metadata[this.largeEmbeddingKey])\n );\n let func: () => Array<number[]>;\n\n switch (this.searchType) {\n case \"cosine\":\n func = () => cosineSimilarity([embeddedQuery], largeEmbeddings);\n break;\n case \"innerProduct\":\n func = () => innerProduct([embeddedQuery], largeEmbeddings);\n break;\n case \"euclidean\":\n func = () => euclideanDistance([embeddedQuery], largeEmbeddings);\n break;\n default:\n throw new Error(`Unknown search type: ${this.searchType}`);\n }\n\n // Calculate the similarity scores between the query embedding and the large embeddings\n const [similarityScores] = func();\n\n // Create an array of indices from 0 to N-1, where N is the number of documents\n let indices = Array.from(\n { length: smallResults.length },\n (_, index) => index\n );\n\n indices = indices\n .map((v, i) => [similarityScores[i], v])\n .sort(([a], [b]) => b - a)\n .slice(0, this.largeK)\n .map(([, i]) => i);\n\n return indices.map((i) => smallResults[i]);\n }\n\n async _getRelevantDocuments(query: string): Promise<DocumentInterface[]> {\n const [embeddedQuery, smallResults] = await Promise.all([\n this.largeEmbeddingModel.embedQuery(query),\n this.vectorStore.similaritySearch(query, this.smallK, this.filter),\n ]);\n\n return this._rankByLargeEmbeddings(embeddedQuery, smallResults);\n }\n\n /**\n * Override the default `addDocuments` method to embed the documents twice,\n * once using the larger embeddings model, and then again using the default\n * embedding model linked to the vector store.\n *\n * @param {DocumentInterface[]} documents - An array of documents to add to the vector store.\n * @param {AddDocumentOptions} options - An optional object containing additional options for adding documents.\n * @returns {Promise<string[] | void>} A promise that resolves to an array of the document IDs that were added to the vector store.\n */\n override addDocuments = async (\n documents: DocumentInterface[],\n options?: AddDocumentOptions\n ): Promise<string[] | void> => {\n // Insure documents metadata does not contain the large embedding key\n if (documents.some((doc) => this.largeEmbeddingKey in doc.metadata)) {\n throw new Error(\n `All documents must not contain the large embedding key: ${this.largeEmbeddingKey} in their metadata.`\n );\n }\n\n const allDocPageContent = documents.map((doc) => doc.pageContent);\n const allDocLargeEmbeddings =\n await this.largeEmbeddingModel.embedDocuments(allDocPageContent);\n\n const newDocuments: Array<DocumentInterface> = documents.map(\n (doc, idx) => ({\n ...doc,\n metadata: {\n ...doc.metadata,\n [this.largeEmbeddingKey]: JSON.stringify(allDocLargeEmbeddings[idx]),\n },\n })\n );\n\n return this.vectorStore.addDocuments(newDocuments, options);\n };\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;AAmEA,IAAa,sBAAb,cAEU,qBAA4B;CACpC,SAAS;CAET,SAAS;CAET,oBAAoB;CAEpB;CAEA,aAAsD;CAEtD,YACEA,QACA;EACA,MAAM,OAAO;EACb,KAAK,SAAS,OAAO,UAAU,KAAK;EACpC,KAAK,SAAS,OAAO,UAAU,KAAK;EACpC,KAAK,oBAAoB,OAAO,qBAAqB,KAAK;EAC1D,KAAK,sBAAsB,OAAO;EAClC,KAAK,aAAa,OAAO,cAAc,KAAK;CAC7C;;;;;;;;;;;;;;;;;;;CAoBD,AAAQ,uBACNC,eACAC,cACqB;EACrB,MAAMC,kBAAmC,aAAa,IAAI,CAAC,QACzD,KAAK,MAAM,IAAI,SAAS,KAAK,mBAAmB,CACjD;EACD,IAAIC;AAEJ,UAAQ,KAAK,YAAb;GACE,KAAK;IACH,OAAO,MAAM,iBAAiB,CAAC,aAAc,GAAE,gBAAgB;AAC/D;GACF,KAAK;IACH,OAAO,MAAM,aAAa,CAAC,aAAc,GAAE,gBAAgB;AAC3D;GACF,KAAK;IACH,OAAO,MAAM,kBAAkB,CAAC,aAAc,GAAE,gBAAgB;AAChE;GACF,QACE,OAAM,IAAI,MAAM,CAAC,qBAAqB,EAAE,KAAK,YAAY;EAC5D;EAGD,MAAM,CAAC,iBAAiB,GAAG,MAAM;EAGjC,IAAI,UAAU,MAAM,KAClB,EAAE,QAAQ,aAAa,OAAQ,GAC/B,CAAC,GAAG,UAAU,MACf;EAED,UAAU,QACP,IAAI,CAAC,GAAG,MAAM,CAAC,iBAAiB,IAAI,CAAE,EAAC,CACvC,KAAK,CAAC,CAAC,EAAE,EAAE,CAAC,EAAE,KAAK,IAAI,EAAE,CACzB,MAAM,GAAG,KAAK,OAAO,CACrB,IAAI,CAAC,GAAG,EAAE,KAAK,EAAE;AAEpB,SAAO,QAAQ,IAAI,CAAC,MAAM,aAAa,GAAG;CAC3C;CAED,MAAM,sBAAsBC,OAA6C;EACvE,MAAM,CAAC,eAAe,aAAa,GAAG,MAAM,QAAQ,IAAI,CACtD,KAAK,oBAAoB,WAAW,MAAM,EAC1C,KAAK,YAAY,iBAAiB,OAAO,KAAK,QAAQ,KAAK,OAAO,AACnE,EAAC;AAEF,SAAO,KAAK,uBAAuB,eAAe,aAAa;CAChE;;;;;;;;;;CAWD,AAAS,eAAe,OACtBC,WACAC,YAC6B;AAE7B,MAAI,UAAU,KAAK,CAAC,QAAQ,KAAK,qBAAqB,IAAI,SAAS,CACjE,OAAM,IAAI,MACR,CAAC,wDAAwD,EAAE,KAAK,kBAAkB,mBAAmB,CAAC;EAI1G,MAAM,oBAAoB,UAAU,IAAI,CAAC,QAAQ,IAAI,YAAY;EACjE,MAAM,wBACJ,MAAM,KAAK,oBAAoB,eAAe,kBAAkB;EAElE,MAAMC,eAAyC,UAAU,IACvD,CAAC,KAAK,SAAS;GACb,GAAG;GACH,UAAU;IACR,GAAG,IAAI;KACN,KAAK,oBAAoB,KAAK,UAAU,sBAAsB,KAAK;GACrE;EACF,GACF;AAED,SAAO,KAAK,YAAY,aAAa,cAAc,QAAQ;CAC5D;AACF"}
|