langchain 0.0.4 → 0.0.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +33 -0
- package/agents.d.ts +1 -0
- package/agents.js +1 -0
- package/chains.d.ts +1 -0
- package/chains.js +1 -0
- package/dist/agents/agent.d.ts +96 -0
- package/dist/agents/agent.js +151 -0
- package/dist/agents/agent.js.map +1 -0
- package/dist/{esm/agents → agents}/executor.d.ts +5 -0
- package/dist/agents/executor.js +99 -0
- package/dist/agents/executor.js.map +1 -0
- package/dist/{cjs/agents → agents}/helpers.d.ts +0 -0
- package/dist/agents/helpers.js +21 -0
- package/dist/agents/helpers.js.map +1 -0
- package/dist/{cjs/agents → agents}/index.d.ts +1 -0
- package/dist/{cjs/agents → agents}/index.js +5 -1
- package/dist/agents/index.js.map +1 -0
- package/dist/agents/initialize.d.ts +4 -0
- package/dist/agents/initialize.js +19 -0
- package/dist/agents/initialize.js.map +1 -0
- package/dist/{cjs/agents → agents}/load.d.ts +0 -0
- package/dist/agents/load.js +19 -0
- package/dist/agents/load.js.map +1 -0
- package/dist/{esm/agents → agents}/mrkl/index.d.ts +18 -1
- package/dist/{cjs/agents → agents}/mrkl/index.js +26 -34
- package/dist/agents/mrkl/index.js.map +1 -0
- package/dist/{cjs/agents → agents}/mrkl/prompt.d.ts +0 -0
- package/dist/{cjs/agents → agents}/mrkl/prompt.js +0 -1
- package/dist/agents/mrkl/prompt.js.map +1 -0
- package/dist/{cjs/agents → agents}/tests/agent.test.d.ts +0 -0
- package/dist/agents/tests/agent.test.js +35 -0
- package/dist/agents/tests/agent.test.js.map +1 -0
- package/dist/{cjs/chains/tests/llm_chain.test.d.ts → agents/tests/calculator.test.d.ts} +0 -0
- package/dist/agents/tests/calculator.test.js +30 -0
- package/dist/agents/tests/calculator.test.js.map +1 -0
- package/dist/agents/tools/base.d.ts +6 -0
- package/dist/agents/tools/base.js +15 -0
- package/dist/agents/tools/base.js.map +1 -0
- package/dist/agents/tools/calculator.d.ts +6 -0
- package/dist/agents/tools/calculator.js +33 -0
- package/dist/agents/tools/calculator.js.map +1 -0
- package/dist/agents/tools/index.d.ts +3 -0
- package/dist/agents/tools/index.js +10 -0
- package/dist/agents/tools/index.js.map +1 -0
- package/dist/{cjs/agents → agents}/tools/llm-math.d.ts +0 -0
- package/dist/{cjs/agents → agents}/tools/llm-math.js +0 -0
- package/dist/{cjs/agents → agents}/tools/llm-math.js.map +1 -1
- package/dist/agents/tools/serpapi.d.ts +18 -0
- package/dist/agents/tools/serpapi.js +93 -0
- package/dist/agents/tools/serpapi.js.map +1 -0
- package/dist/{cjs/agents → agents}/types.d.ts +0 -0
- package/dist/{cjs/agents → agents}/types.js +0 -0
- package/dist/{cjs/agents → agents}/types.js.map +1 -1
- package/dist/{cjs/cache.d.ts → cache.d.ts} +0 -0
- package/dist/{cjs/cache.js → cache.js} +0 -0
- package/dist/cache.js.map +1 -0
- package/dist/chains/base.d.ts +42 -0
- package/dist/chains/base.js +67 -0
- package/dist/chains/base.js.map +1 -0
- package/dist/chains/chat_vector_db_chain.d.ts +41 -0
- package/dist/chains/chat_vector_db_chain.js +146 -0
- package/dist/chains/chat_vector_db_chain.js.map +1 -0
- package/dist/chains/combine_docs_chain.d.ts +35 -0
- package/dist/chains/combine_docs_chain.js +75 -0
- package/dist/chains/combine_docs_chain.js.map +1 -0
- package/dist/chains/conversation.d.ts +12 -0
- package/dist/chains/conversation.js +29 -0
- package/dist/chains/conversation.js.map +1 -0
- package/dist/chains/index.d.ts +7 -0
- package/dist/chains/index.js +19 -0
- package/dist/chains/index.js.map +1 -0
- package/dist/{cjs/chains → chains}/llm_chain.d.ts +36 -1
- package/dist/chains/llm_chain.js +115 -0
- package/dist/chains/llm_chain.js.map +1 -0
- package/dist/chains/load.d.ts +21 -0
- package/dist/chains/load.js +37 -0
- package/dist/chains/load.js.map +1 -0
- package/dist/chains/question_answering/load.d.ts +3 -0
- package/dist/chains/question_answering/load.js +13 -0
- package/dist/chains/question_answering/load.js.map +1 -0
- package/dist/chains/question_answering/stuff_prompts.d.ts +2 -0
- package/dist/chains/question_answering/stuff_prompts.js +10 -0
- package/dist/chains/question_answering/stuff_prompts.js.map +1 -0
- package/dist/{cjs/prompt → chains/question_answering}/tests/load.test.d.ts +0 -0
- package/dist/chains/question_answering/tests/load.test.js +18 -0
- package/dist/chains/question_answering/tests/load.test.js.map +1 -0
- package/dist/{cjs/llms/tests/openai.test.d.ts → chains/tests/chat_vector_db_qa_chain.test.d.ts} +0 -0
- package/dist/chains/tests/chat_vector_db_qa_chain.test.js +35 -0
- package/dist/chains/tests/chat_vector_db_qa_chain.test.js.map +1 -0
- package/dist/{cjs/prompt/tests/template.test.d.ts → chains/tests/combine_docs_chain.test.d.ts} +0 -0
- package/dist/chains/tests/combine_docs_chain.test.js +39 -0
- package/dist/chains/tests/combine_docs_chain.test.js.map +1 -0
- package/dist/{esm/chains → chains}/tests/llm_chain.test.d.ts +0 -0
- package/dist/chains/tests/llm_chain.test.js +29 -0
- package/dist/chains/tests/llm_chain.test.js.map +1 -0
- package/dist/{esm/agents/tests/agent.test.d.ts → chains/tests/vector_db_qa_chain.test.d.ts} +0 -0
- package/dist/chains/tests/vector_db_qa_chain.test.js +46 -0
- package/dist/chains/tests/vector_db_qa_chain.test.js.map +1 -0
- package/dist/chains/vector_db_qa.d.ts +36 -0
- package/dist/chains/vector_db_qa.js +85 -0
- package/dist/chains/vector_db_qa.js.map +1 -0
- package/dist/document.d.ts +13 -0
- package/dist/document.js +28 -0
- package/dist/document.js.map +1 -0
- package/dist/embeddings/base.d.ts +4 -0
- package/dist/embeddings/base.js +7 -0
- package/dist/embeddings/base.js.map +1 -0
- package/dist/embeddings/index.d.ts +1 -0
- package/dist/embeddings/index.js +6 -0
- package/dist/embeddings/index.js.map +1 -0
- package/dist/embeddings/openai.d.ts +20 -0
- package/dist/embeddings/openai.js +86 -0
- package/dist/embeddings/openai.js.map +1 -0
- package/dist/{esm/llms → embeddings}/tests/openai.test.d.ts +0 -0
- package/dist/embeddings/tests/openai.test.js +17 -0
- package/dist/embeddings/tests/openai.test.js.map +1 -0
- package/dist/{cjs/index.d.ts → index.d.ts} +1 -1
- package/dist/{cjs/index.js → index.js} +4 -4
- package/dist/index.js.map +1 -0
- package/dist/llms/base.d.ts +64 -0
- package/dist/llms/base.js +164 -0
- package/dist/llms/base.js.map +1 -0
- package/dist/llms/cohere.d.ts +24 -0
- package/dist/llms/cohere.js +55 -0
- package/dist/llms/cohere.js.map +1 -0
- package/dist/{esm/llms → llms}/index.d.ts +20 -0
- package/dist/{cjs/llms → llms}/index.js +3 -1
- package/dist/llms/index.js.map +1 -0
- package/dist/llms/load.d.ts +11 -0
- package/dist/llms/load.js +17 -0
- package/dist/llms/load.js.map +1 -0
- package/dist/llms/openai.d.ts +125 -0
- package/dist/{cjs/llms → llms}/openai.js +96 -58
- package/dist/llms/openai.js.map +1 -0
- package/dist/{esm/prompt/tests/load.test.d.ts → llms/tests/cohere.test.d.ts} +0 -0
- package/dist/llms/tests/cohere.test.js +10 -0
- package/dist/llms/tests/cohere.test.js.map +1 -0
- package/dist/{esm/prompt/tests/template.test.d.ts → llms/tests/openai.test.d.ts} +0 -0
- package/dist/llms/tests/openai.test.js +10 -0
- package/dist/llms/tests/openai.test.js.map +1 -0
- package/dist/memory/base.d.ts +7 -0
- package/dist/memory/base.js +7 -0
- package/dist/memory/base.js.map +1 -0
- package/dist/memory/buffer_memory.d.ts +15 -0
- package/dist/memory/buffer_memory.js +57 -0
- package/dist/memory/buffer_memory.js.map +1 -0
- package/dist/memory/index.d.ts +2 -0
- package/dist/memory/index.js +8 -0
- package/dist/memory/index.js.map +1 -0
- package/dist/memory/tests/buffer_memory.test.d.ts +1 -0
- package/dist/memory/tests/buffer_memory.test.js +17 -0
- package/dist/memory/tests/buffer_memory.test.js.map +1 -0
- package/dist/prompts/base.d.ts +58 -0
- package/dist/prompts/base.js +53 -0
- package/dist/prompts/base.js.map +1 -0
- package/dist/{esm/prompt → prompts}/few_shot.d.ts +36 -1
- package/dist/{cjs/prompt → prompts}/few_shot.js +27 -33
- package/dist/prompts/few_shot.js.map +1 -0
- package/dist/{cjs/prompt → prompts}/index.d.ts +1 -0
- package/dist/{cjs/prompt → prompts}/index.js +3 -1
- package/dist/prompts/index.js.map +1 -0
- package/dist/prompts/load.d.ts +19 -0
- package/dist/prompts/load.js +33 -0
- package/dist/prompts/load.js.map +1 -0
- package/dist/{cjs/prompt → prompts}/parser.d.ts +30 -0
- package/dist/{cjs/prompt → prompts}/parser.js +21 -0
- package/dist/prompts/parser.js.map +1 -0
- package/dist/prompts/prompt.d.ts +76 -0
- package/dist/{cjs/prompt → prompts}/prompt.js +38 -18
- package/dist/prompts/prompt.js.map +1 -0
- package/dist/{cjs/prompt → prompts}/template.d.ts +0 -0
- package/dist/{cjs/prompt → prompts}/template.js +0 -0
- package/dist/prompts/template.js.map +1 -0
- package/dist/prompts/tests/load.test.d.ts +1 -0
- package/dist/prompts/tests/load.test.js +21 -0
- package/dist/prompts/tests/load.test.js.map +1 -0
- package/dist/prompts/tests/template.test.d.ts +1 -0
- package/dist/{cjs/prompt → prompts}/tests/template.test.js +0 -0
- package/dist/prompts/tests/template.test.js.map +1 -0
- package/dist/text_splitter.d.ts +32 -0
- package/dist/text_splitter.js +167 -0
- package/dist/text_splitter.js.map +1 -0
- package/dist/util/hub.d.ts +2 -0
- package/dist/{cjs/util → util}/hub.js +5 -14
- package/dist/util/hub.js.map +1 -0
- package/dist/{esm/util → util}/index.d.ts +1 -0
- package/dist/{cjs/util → util}/index.js +13 -25
- package/dist/util/index.js.map +1 -0
- package/dist/vectorstores/base.d.ts +18 -0
- package/dist/vectorstores/base.js +38 -0
- package/dist/vectorstores/base.js.map +1 -0
- package/dist/vectorstores/hnswlib.d.ts +19 -0
- package/dist/vectorstores/hnswlib.js +129 -0
- package/dist/vectorstores/hnswlib.js.map +1 -0
- package/dist/vectorstores/index.d.ts +1 -0
- package/dist/vectorstores/index.js +6 -0
- package/dist/vectorstores/index.js.map +1 -0
- package/dist/vectorstores/tests/hnswlib.test.d.ts +1 -0
- package/dist/vectorstores/tests/hnswlib.test.js +44 -0
- package/dist/vectorstores/tests/hnswlib.test.js.map +1 -0
- package/document.d.ts +1 -0
- package/document.js +1 -0
- package/embeddings.d.ts +1 -0
- package/embeddings.js +1 -0
- package/llms.d.ts +1 -0
- package/llms.js +1 -0
- package/memory.d.ts +1 -0
- package/memory.js +1 -0
- package/package.json +53 -44
- package/prompts.d.ts +1 -0
- package/prompts.js +1 -0
- package/text_splitter.d.ts +1 -0
- package/text_splitter.js +1 -0
- package/tools.d.ts +1 -0
- package/tools.js +1 -0
- package/vectorstores.d.ts +1 -0
- package/vectorstores.js +1 -0
- package/dist/cjs/agents/agent.d.ts +0 -42
- package/dist/cjs/agents/agent.js +0 -129
- package/dist/cjs/agents/agent.js.map +0 -1
- package/dist/cjs/agents/executor.d.ts +0 -27
- package/dist/cjs/agents/executor.js +0 -105
- package/dist/cjs/agents/executor.js.map +0 -1
- package/dist/cjs/agents/helpers.js +0 -30
- package/dist/cjs/agents/helpers.js.map +0 -1
- package/dist/cjs/agents/index.js.map +0 -1
- package/dist/cjs/agents/load.js +0 -28
- package/dist/cjs/agents/load.js.map +0 -1
- package/dist/cjs/agents/mrkl/index.d.ts +0 -33
- package/dist/cjs/agents/mrkl/index.js.map +0 -1
- package/dist/cjs/agents/mrkl/prompt.js.map +0 -1
- package/dist/cjs/agents/tests/agent.test.js +0 -31
- package/dist/cjs/agents/tests/agent.test.js.map +0 -1
- package/dist/cjs/agents/tools/index.d.ts +0 -7
- package/dist/cjs/agents/tools/index.js +0 -6
- package/dist/cjs/agents/tools/index.js.map +0 -1
- package/dist/cjs/agents/tools/serpapi.d.ts +0 -3
- package/dist/cjs/agents/tools/serpapi.js +0 -50
- package/dist/cjs/agents/tools/serpapi.js.map +0 -1
- package/dist/cjs/cache.js.map +0 -1
- package/dist/cjs/chains/base.d.ts +0 -12
- package/dist/cjs/chains/base.js +0 -23
- package/dist/cjs/chains/base.js.map +0 -1
- package/dist/cjs/chains/index.d.ts +0 -3
- package/dist/cjs/chains/index.js +0 -10
- package/dist/cjs/chains/index.js.map +0 -1
- package/dist/cjs/chains/llm_chain.js +0 -83
- package/dist/cjs/chains/llm_chain.js.map +0 -1
- package/dist/cjs/chains/llm_math_chain.d.ts +0 -0
- package/dist/cjs/chains/llm_math_chain.js +0 -129
- package/dist/cjs/chains/llm_math_chain.js.map +0 -1
- package/dist/cjs/chains/load.d.ts +0 -2
- package/dist/cjs/chains/load.js +0 -29
- package/dist/cjs/chains/load.js.map +0 -1
- package/dist/cjs/chains/tests/llm_chain.test.js +0 -32
- package/dist/cjs/chains/tests/llm_chain.test.js.map +0 -1
- package/dist/cjs/index.js.map +0 -1
- package/dist/cjs/llms/base.d.ts +0 -23
- package/dist/cjs/llms/base.js +0 -159
- package/dist/cjs/llms/base.js.map +0 -1
- package/dist/cjs/llms/index.d.ts +0 -18
- package/dist/cjs/llms/index.js.map +0 -1
- package/dist/cjs/llms/load.d.ts +0 -3
- package/dist/cjs/llms/load.js +0 -9
- package/dist/cjs/llms/load.js.map +0 -1
- package/dist/cjs/llms/openai.d.ts +0 -63
- package/dist/cjs/llms/openai.js.map +0 -1
- package/dist/cjs/llms/tests/openai.test.js +0 -19
- package/dist/cjs/llms/tests/openai.test.js.map +0 -1
- package/dist/cjs/prompt/base.d.ts +0 -19
- package/dist/cjs/prompt/base.js +0 -54
- package/dist/cjs/prompt/base.js.map +0 -1
- package/dist/cjs/prompt/few_shot.d.ts +0 -47
- package/dist/cjs/prompt/few_shot.js.map +0 -1
- package/dist/cjs/prompt/index.js.map +0 -1
- package/dist/cjs/prompt/load.d.ts +0 -2
- package/dist/cjs/prompt/load.js +0 -25
- package/dist/cjs/prompt/load.js.map +0 -1
- package/dist/cjs/prompt/parser.js.map +0 -1
- package/dist/cjs/prompt/prompt.d.ts +0 -28
- package/dist/cjs/prompt/prompt.js.map +0 -1
- package/dist/cjs/prompt/template.js.map +0 -1
- package/dist/cjs/prompt/tests/load.test.js +0 -30
- package/dist/cjs/prompt/tests/load.test.js.map +0 -1
- package/dist/cjs/prompt/tests/template.test.js.map +0 -1
- package/dist/cjs/util/hub.d.ts +0 -1
- package/dist/cjs/util/hub.js.map +0 -1
- package/dist/cjs/util/index.d.ts +0 -12
- package/dist/cjs/util/index.js.map +0 -1
- package/dist/esm/agents/agent.d.ts +0 -42
- package/dist/esm/agents/agent.js +0 -124
- package/dist/esm/agents/agent.js.map +0 -1
- package/dist/esm/agents/executor.js +0 -101
- package/dist/esm/agents/executor.js.map +0 -1
- package/dist/esm/agents/helpers.d.ts +0 -3
- package/dist/esm/agents/helpers.js +0 -26
- package/dist/esm/agents/helpers.js.map +0 -1
- package/dist/esm/agents/index.d.ts +0 -6
- package/dist/esm/agents/index.js +0 -5
- package/dist/esm/agents/index.js.map +0 -1
- package/dist/esm/agents/load.d.ts +0 -6
- package/dist/esm/agents/load.js +0 -24
- package/dist/esm/agents/load.js.map +0 -1
- package/dist/esm/agents/mrkl/index.js +0 -106
- package/dist/esm/agents/mrkl/index.js.map +0 -1
- package/dist/esm/agents/mrkl/prompt.d.ts +0 -3
- package/dist/esm/agents/mrkl/prompt.js +0 -17
- package/dist/esm/agents/mrkl/prompt.js.map +0 -1
- package/dist/esm/agents/tests/agent.test.js +0 -29
- package/dist/esm/agents/tests/agent.test.js.map +0 -1
- package/dist/esm/agents/tools/index.d.ts +0 -7
- package/dist/esm/agents/tools/index.js +0 -2
- package/dist/esm/agents/tools/index.js.map +0 -1
- package/dist/esm/agents/tools/llm-math.d.ts +0 -0
- package/dist/esm/agents/tools/llm-math.js +0 -2
- package/dist/esm/agents/tools/llm-math.js.map +0 -1
- package/dist/esm/agents/tools/serpapi.d.ts +0 -3
- package/dist/esm/agents/tools/serpapi.js +0 -46
- package/dist/esm/agents/tools/serpapi.js.map +0 -1
- package/dist/esm/agents/types.d.ts +0 -25
- package/dist/esm/agents/types.js +0 -2
- package/dist/esm/agents/types.js.map +0 -1
- package/dist/esm/cache.d.ts +0 -11
- package/dist/esm/cache.js +0 -22
- package/dist/esm/cache.js.map +0 -1
- package/dist/esm/chains/base.d.ts +0 -12
- package/dist/esm/chains/base.js +0 -19
- package/dist/esm/chains/base.js.map +0 -1
- package/dist/esm/chains/index.d.ts +0 -3
- package/dist/esm/chains/index.js +0 -4
- package/dist/esm/chains/index.js.map +0 -1
- package/dist/esm/chains/llm_chain.d.ts +0 -30
- package/dist/esm/chains/llm_chain.js +0 -79
- package/dist/esm/chains/llm_chain.js.map +0 -1
- package/dist/esm/chains/llm_math_chain.d.ts +0 -0
- package/dist/esm/chains/llm_math_chain.js +0 -129
- package/dist/esm/chains/llm_math_chain.js.map +0 -1
- package/dist/esm/chains/load.d.ts +0 -2
- package/dist/esm/chains/load.js +0 -25
- package/dist/esm/chains/load.js.map +0 -1
- package/dist/esm/chains/tests/llm_chain.test.js +0 -30
- package/dist/esm/chains/tests/llm_chain.test.js.map +0 -1
- package/dist/esm/index.d.ts +0 -3
- package/dist/esm/index.js +0 -4
- package/dist/esm/index.js.map +0 -1
- package/dist/esm/llms/base.d.ts +0 -23
- package/dist/esm/llms/base.js +0 -154
- package/dist/esm/llms/base.js.map +0 -1
- package/dist/esm/llms/index.js +0 -4
- package/dist/esm/llms/index.js.map +0 -1
- package/dist/esm/llms/load.d.ts +0 -3
- package/dist/esm/llms/load.js +0 -5
- package/dist/esm/llms/load.js.map +0 -1
- package/dist/esm/llms/openai.d.ts +0 -63
- package/dist/esm/llms/openai.js +0 -198
- package/dist/esm/llms/openai.js.map +0 -1
- package/dist/esm/llms/tests/openai.test.js +0 -17
- package/dist/esm/llms/tests/openai.test.js.map +0 -1
- package/dist/esm/prompt/base.d.ts +0 -19
- package/dist/esm/prompt/base.js +0 -50
- package/dist/esm/prompt/base.js.map +0 -1
- package/dist/esm/prompt/few_shot.js +0 -139
- package/dist/esm/prompt/few_shot.js.map +0 -1
- package/dist/esm/prompt/index.d.ts +0 -3
- package/dist/esm/prompt/index.js +0 -4
- package/dist/esm/prompt/index.js.map +0 -1
- package/dist/esm/prompt/load.d.ts +0 -2
- package/dist/esm/prompt/load.js +0 -21
- package/dist/esm/prompt/load.js.map +0 -1
- package/dist/esm/prompt/parser.d.ts +0 -40
- package/dist/esm/prompt/parser.js +0 -86
- package/dist/esm/prompt/parser.js.map +0 -1
- package/dist/esm/prompt/prompt.d.ts +0 -28
- package/dist/esm/prompt/prompt.js +0 -87
- package/dist/esm/prompt/prompt.js.map +0 -1
- package/dist/esm/prompt/template.d.ts +0 -16
- package/dist/esm/prompt/template.js +0 -80
- package/dist/esm/prompt/template.js.map +0 -1
- package/dist/esm/prompt/tests/load.test.js +0 -25
- package/dist/esm/prompt/tests/load.test.js.map +0 -1
- package/dist/esm/prompt/tests/template.test.js +0 -25
- package/dist/esm/prompt/tests/template.test.js.map +0 -1
- package/dist/esm/util/hub.d.ts +0 -1
- package/dist/esm/util/hub.js +0 -43
- package/dist/esm/util/hub.js.map +0 -1
- package/dist/esm/util/index.js +0 -74
- package/dist/esm/util/index.js.map +0 -1
|
@@ -0,0 +1,64 @@
|
|
|
1
|
+
import { LLMCallbackManager, LLMResult } from "./index";
|
|
2
|
+
export type SerializedLLM = {
|
|
3
|
+
_type: string;
|
|
4
|
+
} & Record<string, any>;
|
|
5
|
+
/**
|
|
6
|
+
* LLM Wrapper. Provides an {@link call} (an {@link generate}) function that takes in a prompt (or prompts) and returns a string.
|
|
7
|
+
*/
|
|
8
|
+
export declare abstract class BaseLLM {
|
|
9
|
+
/**
|
|
10
|
+
* The name of the LLM class
|
|
11
|
+
*/
|
|
12
|
+
name: string;
|
|
13
|
+
cache?: boolean;
|
|
14
|
+
callbackManager: LLMCallbackManager;
|
|
15
|
+
/**
|
|
16
|
+
* Whether to print out response text.
|
|
17
|
+
*/
|
|
18
|
+
verbose?: boolean;
|
|
19
|
+
constructor(callbackManager?: LLMCallbackManager, verbose?: boolean);
|
|
20
|
+
/**
|
|
21
|
+
* Run the LLM on the given prompts and input.
|
|
22
|
+
*/
|
|
23
|
+
abstract _generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
24
|
+
/** @ignore */
|
|
25
|
+
_generateUncached(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
26
|
+
/**
|
|
27
|
+
* Run the LLM on the given propmts an input, handling caching.
|
|
28
|
+
*/
|
|
29
|
+
generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
30
|
+
/**
|
|
31
|
+
* Convenience wrapper for {@link generate} that takes in a single string prompt and returns a single string output.
|
|
32
|
+
*/
|
|
33
|
+
call(prompt: string, stop?: string[]): Promise<string>;
|
|
34
|
+
/**
|
|
35
|
+
* Get the identifying parameters of the LLM.
|
|
36
|
+
*/
|
|
37
|
+
_identifyingParams(): Record<string, any>;
|
|
38
|
+
/**
|
|
39
|
+
* Return the string type key uniquely identifying this class of LLM.
|
|
40
|
+
*/
|
|
41
|
+
abstract _llmType(): string;
|
|
42
|
+
/**
|
|
43
|
+
* Return a json-like object representing this LLM.
|
|
44
|
+
*/
|
|
45
|
+
serialize(): SerializedLLM;
|
|
46
|
+
/**
|
|
47
|
+
* Load an LLM from a json-like object describing it.
|
|
48
|
+
*/
|
|
49
|
+
static deserialize(data: SerializedLLM): Promise<BaseLLM>;
|
|
50
|
+
}
|
|
51
|
+
/**
|
|
52
|
+
* LLM class that provides a simpler interface to subclass than {@link BaseLLM}.
|
|
53
|
+
*
|
|
54
|
+
* Requires only implementing a simpler {@link _call} method instead of {@link _generate}.
|
|
55
|
+
*
|
|
56
|
+
* @augments BaseLLM
|
|
57
|
+
*/
|
|
58
|
+
export declare abstract class LLM extends BaseLLM {
|
|
59
|
+
/**
|
|
60
|
+
* Run the LLM on the given prompt and input.
|
|
61
|
+
*/
|
|
62
|
+
abstract _call(prompt: string, stop?: string[]): Promise<string>;
|
|
63
|
+
_generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
64
|
+
}
|
|
@@ -0,0 +1,164 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.LLM = exports.BaseLLM = void 0;
|
|
4
|
+
const index_1 = require("./index");
|
|
5
|
+
const cache_1 = require("../cache");
|
|
6
|
+
const getCallbackManager = () => ({
|
|
7
|
+
handleStart: (..._args) => {
|
|
8
|
+
// console.log(args);
|
|
9
|
+
},
|
|
10
|
+
handleEnd: (..._args) => {
|
|
11
|
+
// console.log(args);
|
|
12
|
+
},
|
|
13
|
+
handleError: (..._args) => {
|
|
14
|
+
// console.log(args);
|
|
15
|
+
},
|
|
16
|
+
});
|
|
17
|
+
const getVerbosity = () => true;
|
|
18
|
+
const cache = new cache_1.InMemoryCache();
|
|
19
|
+
/**
|
|
20
|
+
* LLM Wrapper. Provides an {@link call} (an {@link generate}) function that takes in a prompt (or prompts) and returns a string.
|
|
21
|
+
*/
|
|
22
|
+
class BaseLLM {
|
|
23
|
+
constructor(callbackManager, verbose) {
|
|
24
|
+
/**
|
|
25
|
+
* The name of the LLM class
|
|
26
|
+
*/
|
|
27
|
+
Object.defineProperty(this, "name", {
|
|
28
|
+
enumerable: true,
|
|
29
|
+
configurable: true,
|
|
30
|
+
writable: true,
|
|
31
|
+
value: void 0
|
|
32
|
+
});
|
|
33
|
+
Object.defineProperty(this, "cache", {
|
|
34
|
+
enumerable: true,
|
|
35
|
+
configurable: true,
|
|
36
|
+
writable: true,
|
|
37
|
+
value: void 0
|
|
38
|
+
});
|
|
39
|
+
Object.defineProperty(this, "callbackManager", {
|
|
40
|
+
enumerable: true,
|
|
41
|
+
configurable: true,
|
|
42
|
+
writable: true,
|
|
43
|
+
value: void 0
|
|
44
|
+
});
|
|
45
|
+
/**
|
|
46
|
+
* Whether to print out response text.
|
|
47
|
+
*/
|
|
48
|
+
Object.defineProperty(this, "verbose", {
|
|
49
|
+
enumerable: true,
|
|
50
|
+
configurable: true,
|
|
51
|
+
writable: true,
|
|
52
|
+
value: false
|
|
53
|
+
});
|
|
54
|
+
this.callbackManager = callbackManager !== null && callbackManager !== void 0 ? callbackManager : getCallbackManager();
|
|
55
|
+
this.verbose = verbose !== null && verbose !== void 0 ? verbose : getVerbosity();
|
|
56
|
+
}
|
|
57
|
+
/** @ignore */
|
|
58
|
+
async _generateUncached(prompts, stop) {
|
|
59
|
+
this.callbackManager.handleStart({ name: this.name }, prompts, this.verbose);
|
|
60
|
+
let output;
|
|
61
|
+
try {
|
|
62
|
+
output = await this._generate(prompts, stop);
|
|
63
|
+
}
|
|
64
|
+
catch (err) {
|
|
65
|
+
this.callbackManager.handleError(`${err}`, this.verbose);
|
|
66
|
+
throw err;
|
|
67
|
+
}
|
|
68
|
+
this.callbackManager.handleEnd(output, this.verbose);
|
|
69
|
+
return output;
|
|
70
|
+
}
|
|
71
|
+
/**
|
|
72
|
+
* Run the LLM on the given propmts an input, handling caching.
|
|
73
|
+
*/
|
|
74
|
+
async generate(prompts, stop) {
|
|
75
|
+
var _a;
|
|
76
|
+
if (!Array.isArray(prompts)) {
|
|
77
|
+
throw new Error("Argument 'prompts' is expected to be a string[]");
|
|
78
|
+
}
|
|
79
|
+
if (this.cache === true && cache === null) {
|
|
80
|
+
throw new Error("Requested cache, but no cache found");
|
|
81
|
+
}
|
|
82
|
+
if (cache === null || this.cache === false) {
|
|
83
|
+
return this._generateUncached(prompts, stop);
|
|
84
|
+
}
|
|
85
|
+
const params = this.serialize();
|
|
86
|
+
params.stop = stop;
|
|
87
|
+
const llmStringKey = `${Object.entries(params).sort()}`;
|
|
88
|
+
const missingPromptIndices = [];
|
|
89
|
+
const generations = prompts.map((prompt, index) => {
|
|
90
|
+
const result = cache.lookup(prompt, llmStringKey);
|
|
91
|
+
if (!result) {
|
|
92
|
+
missingPromptIndices.push(index);
|
|
93
|
+
}
|
|
94
|
+
return result;
|
|
95
|
+
});
|
|
96
|
+
let llmOutput = {};
|
|
97
|
+
if (missingPromptIndices.length > 0) {
|
|
98
|
+
const results = await this._generateUncached(missingPromptIndices.map((i) => prompts[i]), stop);
|
|
99
|
+
results.generations.forEach((generation, index) => {
|
|
100
|
+
const promptIndex = missingPromptIndices[index];
|
|
101
|
+
generations[promptIndex] = generation;
|
|
102
|
+
cache.update(prompts[promptIndex], llmStringKey, generation);
|
|
103
|
+
});
|
|
104
|
+
llmOutput = (_a = results.llmOutput) !== null && _a !== void 0 ? _a : {};
|
|
105
|
+
}
|
|
106
|
+
return { generations, llmOutput };
|
|
107
|
+
}
|
|
108
|
+
/**
|
|
109
|
+
* Convenience wrapper for {@link generate} that takes in a single string prompt and returns a single string output.
|
|
110
|
+
*/
|
|
111
|
+
async call(prompt, stop) {
|
|
112
|
+
const { generations } = await this.generate([prompt], stop);
|
|
113
|
+
return generations[0][0].text;
|
|
114
|
+
}
|
|
115
|
+
/**
|
|
116
|
+
* Get the identifying parameters of the LLM.
|
|
117
|
+
*/
|
|
118
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
119
|
+
_identifyingParams() {
|
|
120
|
+
return {};
|
|
121
|
+
}
|
|
122
|
+
/**
|
|
123
|
+
* Return a json-like object representing this LLM.
|
|
124
|
+
*/
|
|
125
|
+
serialize() {
|
|
126
|
+
return {
|
|
127
|
+
...this._identifyingParams(),
|
|
128
|
+
_type: this._llmType(),
|
|
129
|
+
};
|
|
130
|
+
}
|
|
131
|
+
/**
|
|
132
|
+
* Load an LLM from a json-like object describing it.
|
|
133
|
+
*/
|
|
134
|
+
static async deserialize(data) {
|
|
135
|
+
const { _type, ...rest } = data;
|
|
136
|
+
const Cls = {
|
|
137
|
+
openai: index_1.OpenAI,
|
|
138
|
+
}[_type];
|
|
139
|
+
if (Cls === undefined) {
|
|
140
|
+
throw new Error(`Cannot load LLM with type ${_type}`);
|
|
141
|
+
}
|
|
142
|
+
return new Cls(rest);
|
|
143
|
+
}
|
|
144
|
+
}
|
|
145
|
+
exports.BaseLLM = BaseLLM;
|
|
146
|
+
/**
|
|
147
|
+
* LLM class that provides a simpler interface to subclass than {@link BaseLLM}.
|
|
148
|
+
*
|
|
149
|
+
* Requires only implementing a simpler {@link _call} method instead of {@link _generate}.
|
|
150
|
+
*
|
|
151
|
+
* @augments BaseLLM
|
|
152
|
+
*/
|
|
153
|
+
class LLM extends BaseLLM {
|
|
154
|
+
async _generate(prompts, stop) {
|
|
155
|
+
const generations = [];
|
|
156
|
+
for (let i = 0; i < prompts.length; i += 1) {
|
|
157
|
+
const text = await this._call(prompts[i], stop);
|
|
158
|
+
generations.push([{ text }]);
|
|
159
|
+
}
|
|
160
|
+
return { generations };
|
|
161
|
+
}
|
|
162
|
+
}
|
|
163
|
+
exports.LLM = LLM;
|
|
164
|
+
//# sourceMappingURL=base.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"base.js","sourceRoot":"","sources":["../../llms/base.ts"],"names":[],"mappings":";;;AAAA,mCAAgE;AAChE,oCAAoD;AAEpD,MAAM,kBAAkB,GAAG,GAAuB,EAAE,CAAC,CAAC;IACpD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;IACD,SAAS,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACtB,qBAAqB;IACvB,CAAC;IACD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;CACF,CAAC,CAAC;AAEH,MAAM,YAAY,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC;AAEhC,MAAM,KAAK,GAAc,IAAI,qBAAa,EAAE,CAAC;AAO7C;;GAEG;AACH,MAAsB,OAAO;IAe3B,YAAY,eAAoC,EAAE,OAAiB;QAdnE;;WAEG;QACH;;;;;WAAa;QAEb;;;;;WAAgB;QAEhB;;;;;WAAoC;QAEpC;;WAEG;QACH;;;;mBAAoB,KAAK;WAAC;QAGxB,IAAI,CAAC,eAAe,GAAG,eAAe,aAAf,eAAe,cAAf,eAAe,GAAI,kBAAkB,EAAE,CAAC;QAC/D,IAAI,CAAC,OAAO,GAAG,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,YAAY,EAAE,CAAC;IAC3C,CAAC;IAOD,cAAc;IACd,KAAK,CAAC,iBAAiB,CACrB,OAAiB,EACjB,IAAe;QAEf,IAAI,CAAC,eAAe,CAAC,WAAW,CAC9B,EAAE,IAAI,EAAE,IAAI,CAAC,IAAI,EAAE,EACnB,OAAO,EACP,IAAI,CAAC,OAAO,CACb,CAAC;QACF,IAAI,MAAM,CAAC;QACX,IAAI;YACF,MAAM,GAAG,MAAM,IAAI,CAAC,SAAS,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;SAC9C;QAAC,OAAO,GAAG,EAAE;YACZ,IAAI,CAAC,eAAe,CAAC,WAAW,CAAC,GAAG,GAAG,EAAE,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;YACzD,MAAM,GAAG,CAAC;SACX;QAED,IAAI,CAAC,eAAe,CAAC,SAAS,CAAC,MAAM,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;QACrD,OAAO,MAAM,CAAC;IAChB,CAAC;IAED;;OAEG;IACH,KAAK,CAAC,QAAQ,CAAC,OAAiB,EAAE,IAAe;;QAC/C,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,OAAO,CAAC,EAAE;YAC3B,MAAM,IAAI,KAAK,CAAC,iDAAiD,CAAC,CAAC;SACpE;QAED,IAAI,IAAI,CAAC,KAAK,KAAK,IAAI,IAAI,KAAK,KAAK,IAAI,EAAE;YACzC,MAAM,IAAI,KAAK,CAAC,qCAAqC,CAAC,CAAC;SACxD;QAED,IAAI,KAAK,KAAK,IAAI,IAAI,IAAI,CAAC,KAAK,KAAK,KAAK,EAAE;YAC1C,OAAO,IAAI,CAAC,iBAAiB,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;SAC9C;QAED,MAAM,MAAM,GAAG,IAAI,CAAC,SAAS,EAAE,CAAC;QAChC,MAAM,CAAC,IAAI,GAAG,IAAI,CAAC;QAEnB,MAAM,YAAY,GAAG,GAAG,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAI,EAAE,EAAE,CAAC;QACxD,MAAM,oBAAoB,GAAa,EAAE,CAAC;QAC1C,MAAM,WAAW,GAAG,OAAO,CAAC,GAAG,CAAC,CAAC,MAAM,EAAE,KAAK,EAAE,EAAE;YAChD,MAAM,MAAM,GAAG,KAAK,CAAC,MAAM,CAAC,MAAM,EAAE,YAAY,CAAC,CAAC;YAClD,IAAI,CAAC,MAAM,EAAE;gBACX,oBAAoB,CAAC,IAAI,CAAC,KAAK,CAAC,CAAC;aAClC;YACD,OAAO,MAAM,CAAC;QAChB,CAAC,CAAC,CAAC;QAEH,IAAI,SAAS,GAAG,EAAE,CAAC;QACnB,IAAI,oBAAoB,CAAC,MAAM,GAAG,CAAC,EAAE;YACnC,MAAM,OAAO,GAAG,MAAM,IAAI,CAAC,iBAAiB,CAC1C,oBAAoB,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,EAC3C,IAAI,CACL,CAAC;YACF,OAAO,CAAC,WAAW,CAAC,OAAO,CAAC,CAAC,UAAU,EAAE,KAAK,EAAE,EAAE;gBAChD,MAAM,WAAW,GAAG,oBAAoB,CAAC,KAAK,CAAC,CAAC;gBAChD,WAAW,CAAC,WAAW,CAAC,GAAG,UAAU,CAAC;gBACtC,KAAK,CAAC,MAAM,CAAC,OAAO,CAAC,WAAW,CAAC,EAAE,YAAY,EAAE,UAAU,CAAC,CAAC;YAC/D,CAAC,CAAC,CAAC;YACH,SAAS,GAAG,MAAA,OAAO,CAAC,SAAS,mCAAI,EAAE,CAAC;SACrC;QAED,OAAO,EAAE,WAAW,EAAE,SAAS,EAAe,CAAC;IACjD,CAAC;IAED;;OAEG;IACH,KAAK,CAAC,IAAI,CAAC,MAAc,EAAE,IAAe;QACxC,MAAM,EAAE,WAAW,EAAE,GAAG,MAAM,IAAI,CAAC,QAAQ,CAAC,CAAC,MAAM,CAAC,EAAE,IAAI,CAAC,CAAC;QAC5D,OAAO,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;IAChC,CAAC;IAED;;OAEG;IACH,8DAA8D;IAC9D,kBAAkB;QAChB,OAAO,EAAE,CAAC;IACZ,CAAC;IAOD;;OAEG;IACH,SAAS;QACP,OAAO;YACL,GAAG,IAAI,CAAC,kBAAkB,EAAE;YAC5B,KAAK,EAAE,IAAI,CAAC,QAAQ,EAAE;SACvB,CAAC;IACJ,CAAC;IAED;;OAEG;IACH,MAAM,CAAC,KAAK,CAAC,WAAW,CAAC,IAAmB;QAC1C,MAAM,EAAE,KAAK,EAAE,GAAG,IAAI,EAAE,GAAG,IAAI,CAAC;QAChC,MAAM,GAAG,GAAG;YACV,MAAM,EAAE,cAAM;SACf,CAAC,KAAK,CAAC,CAAC;QACT,IAAI,GAAG,KAAK,SAAS,EAAE;YACrB,MAAM,IAAI,KAAK,CAAC,8BAA8B,KAAK,EAAE,CAAC,CAAC;SACxD;QACD,OAAO,IAAI,GAAG,CAAC,IAAI,CAAC,CAAC;IACvB,CAAC;CAGF;AA3ID,0BA2IC;AAED;;;;;;GAMG;AACH,MAAsB,GAAI,SAAQ,OAAO;IAMvC,KAAK,CAAC,SAAS,CAAC,OAAiB,EAAE,IAAe;QAChD,MAAM,WAAW,GAAG,EAAE,CAAC;QACvB,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,OAAO,CAAC,MAAM,EAAE,CAAC,IAAI,CAAC,EAAE;YAC1C,MAAM,IAAI,GAAG,MAAM,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,CAAC,EAAE,IAAI,CAAC,CAAC;YAChD,WAAW,CAAC,IAAI,CAAC,CAAC,EAAE,IAAI,EAAE,CAAC,CAAC,CAAC;SAC9B;QACD,OAAO,EAAE,WAAW,EAAE,CAAC;IACzB,CAAC;CACF;AAdD,kBAcC"}
|
|
@@ -0,0 +1,24 @@
|
|
|
1
|
+
import { LLM, LLMCallbackManager } from ".";
|
|
2
|
+
interface CohereInput {
|
|
3
|
+
/** Sampling temperature to use */
|
|
4
|
+
temperature: number;
|
|
5
|
+
/**
|
|
6
|
+
* Maximum number of tokens to generate in the completion. -1 returns as many
|
|
7
|
+
* tokens as possible given the prompt and the model's maximum context size.
|
|
8
|
+
*/
|
|
9
|
+
maxTokens: number;
|
|
10
|
+
/** Model to use */
|
|
11
|
+
model: string;
|
|
12
|
+
}
|
|
13
|
+
export declare class Cohere extends LLM implements CohereInput {
|
|
14
|
+
temperature: number;
|
|
15
|
+
maxTokens: number;
|
|
16
|
+
model: string;
|
|
17
|
+
constructor(fields?: Partial<CohereInput> & {
|
|
18
|
+
callbackManager?: LLMCallbackManager;
|
|
19
|
+
verbose?: boolean;
|
|
20
|
+
});
|
|
21
|
+
_llmType(): string;
|
|
22
|
+
_call(prompt: string, _stop?: string[]): Promise<string>;
|
|
23
|
+
}
|
|
24
|
+
export {};
|
|
@@ -0,0 +1,55 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.Cohere = void 0;
|
|
4
|
+
const _1 = require(".");
|
|
5
|
+
class Cohere extends _1.LLM {
|
|
6
|
+
constructor(fields) {
|
|
7
|
+
var _a, _b, _c;
|
|
8
|
+
super(fields === null || fields === void 0 ? void 0 : fields.callbackManager, fields === null || fields === void 0 ? void 0 : fields.verbose);
|
|
9
|
+
Object.defineProperty(this, "temperature", {
|
|
10
|
+
enumerable: true,
|
|
11
|
+
configurable: true,
|
|
12
|
+
writable: true,
|
|
13
|
+
value: 0
|
|
14
|
+
});
|
|
15
|
+
Object.defineProperty(this, "maxTokens", {
|
|
16
|
+
enumerable: true,
|
|
17
|
+
configurable: true,
|
|
18
|
+
writable: true,
|
|
19
|
+
value: 250
|
|
20
|
+
});
|
|
21
|
+
Object.defineProperty(this, "model", {
|
|
22
|
+
enumerable: true,
|
|
23
|
+
configurable: true,
|
|
24
|
+
writable: true,
|
|
25
|
+
value: void 0
|
|
26
|
+
});
|
|
27
|
+
this.maxTokens = (_a = fields === null || fields === void 0 ? void 0 : fields.maxTokens) !== null && _a !== void 0 ? _a : this.maxTokens;
|
|
28
|
+
this.temperature = (_b = fields === null || fields === void 0 ? void 0 : fields.temperature) !== null && _b !== void 0 ? _b : this.temperature;
|
|
29
|
+
this.model = (_c = fields === null || fields === void 0 ? void 0 : fields.model) !== null && _c !== void 0 ? _c : this.model;
|
|
30
|
+
}
|
|
31
|
+
_llmType() {
|
|
32
|
+
return "cohere";
|
|
33
|
+
}
|
|
34
|
+
async _call(prompt, _stop) {
|
|
35
|
+
// eslint-disable-next-line global-require,import/no-extraneous-dependencies,@typescript-eslint/no-var-requires
|
|
36
|
+
const cohere = require("cohere-ai");
|
|
37
|
+
cohere.init(process.env.COHERE_API_KEY);
|
|
38
|
+
// Hit the `generate` endpoint on the `large` model
|
|
39
|
+
const generateResponse = await cohere.generate({
|
|
40
|
+
prompt,
|
|
41
|
+
model: this.model,
|
|
42
|
+
max_tokens: this.maxTokens,
|
|
43
|
+
temperature: this.temperature,
|
|
44
|
+
});
|
|
45
|
+
try {
|
|
46
|
+
return generateResponse.body.generations[0].text;
|
|
47
|
+
}
|
|
48
|
+
catch (_a) {
|
|
49
|
+
console.log(generateResponse);
|
|
50
|
+
throw new Error("Could not parse response.");
|
|
51
|
+
}
|
|
52
|
+
}
|
|
53
|
+
}
|
|
54
|
+
exports.Cohere = Cohere;
|
|
55
|
+
//# sourceMappingURL=cohere.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"cohere.js","sourceRoot":"","sources":["../../llms/cohere.ts"],"names":[],"mappings":";;;AAAA,wBAA4C;AAgB5C,MAAa,MAAO,SAAQ,MAAG;IAO7B,YACE,MAGC;;QAED,KAAK,CAAC,MAAM,aAAN,MAAM,uBAAN,MAAM,CAAE,eAAe,EAAE,MAAM,aAAN,MAAM,uBAAN,MAAM,CAAE,OAAO,CAAC,CAAC;QAZlD;;;;mBAAc,CAAC;WAAC;QAEhB;;;;mBAAY,GAAG;WAAC;QAEhB;;;;;WAAc;QAUZ,IAAI,CAAC,SAAS,GAAG,MAAA,MAAM,aAAN,MAAM,uBAAN,MAAM,CAAE,SAAS,mCAAI,IAAI,CAAC,SAAS,CAAC;QACrD,IAAI,CAAC,WAAW,GAAG,MAAA,MAAM,aAAN,MAAM,uBAAN,MAAM,CAAE,WAAW,mCAAI,IAAI,CAAC,WAAW,CAAC;QAC3D,IAAI,CAAC,KAAK,GAAG,MAAA,MAAM,aAAN,MAAM,uBAAN,MAAM,CAAE,KAAK,mCAAI,IAAI,CAAC,KAAK,CAAC;IAC3C,CAAC;IAED,QAAQ;QACN,OAAO,QAAQ,CAAC;IAClB,CAAC;IAED,KAAK,CAAC,KAAK,CAAC,MAAc,EAAE,KAAgB;QAC1C,+GAA+G;QAC/G,MAAM,MAAM,GAAG,OAAO,CAAC,WAAW,CAAC,CAAC;QAEpC,MAAM,CAAC,IAAI,CAAC,OAAO,CAAC,GAAG,CAAC,cAAc,CAAC,CAAC;QAExC,mDAAmD;QACnD,MAAM,gBAAgB,GAAG,MAAM,MAAM,CAAC,QAAQ,CAAC;YAC7C,MAAM;YACN,KAAK,EAAE,IAAI,CAAC,KAAK;YACjB,UAAU,EAAE,IAAI,CAAC,SAAS;YAC1B,WAAW,EAAE,IAAI,CAAC,WAAW;SAC9B,CAAC,CAAC;QACH,IAAI;YACF,OAAO,gBAAgB,CAAC,IAAI,CAAC,WAAW,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;SAClD;QAAC,WAAM;YACN,OAAO,CAAC,GAAG,CAAC,gBAAgB,CAAC,CAAC;YAC9B,MAAM,IAAI,KAAK,CAAC,2BAA2B,CAAC,CAAC;SAC9C;IACH,CAAC;CACF;AA5CD,wBA4CC"}
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
export { BaseLLM, LLM, SerializedLLM } from "./base";
|
|
2
2
|
export { OpenAI } from "./openai";
|
|
3
|
+
export { Cohere } from "./cohere";
|
|
3
4
|
export { loadLLM } from "./load";
|
|
4
5
|
export type LLMCallbackManager = {
|
|
5
6
|
handleStart: (llm: {
|
|
@@ -8,11 +9,30 @@ export type LLMCallbackManager = {
|
|
|
8
9
|
handleError: (err: string, verbose?: boolean) => void;
|
|
9
10
|
handleEnd: (output: LLMResult, verbose?: boolean) => void;
|
|
10
11
|
};
|
|
12
|
+
/**
|
|
13
|
+
* Output of a single generation.
|
|
14
|
+
*/
|
|
11
15
|
export type Generation = {
|
|
16
|
+
/**
|
|
17
|
+
* Generated text output
|
|
18
|
+
*/
|
|
12
19
|
text: string;
|
|
20
|
+
/**
|
|
21
|
+
* Raw generation info response from the provider.
|
|
22
|
+
* May include things like reason for finishing (e.g. in {@link OpenAI})
|
|
23
|
+
*/
|
|
13
24
|
generationInfo?: Record<string, any>;
|
|
14
25
|
};
|
|
26
|
+
/**
|
|
27
|
+
* Contains all relevant information returned by an LLM.
|
|
28
|
+
*/
|
|
15
29
|
export type LLMResult = {
|
|
30
|
+
/**
|
|
31
|
+
* List of the things generated. Each input could have multiple {@link Generation | generations}, hence this is a list of lists.
|
|
32
|
+
*/
|
|
16
33
|
generations: Generation[][];
|
|
34
|
+
/**
|
|
35
|
+
* Dictionary of arbitrary LLM-provider specific output.
|
|
36
|
+
*/
|
|
17
37
|
llmOutput?: Record<string, any>;
|
|
18
38
|
};
|
|
@@ -1,11 +1,13 @@
|
|
|
1
1
|
"use strict";
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
-
exports.loadLLM = exports.OpenAI = exports.LLM = exports.BaseLLM = void 0;
|
|
3
|
+
exports.loadLLM = exports.Cohere = exports.OpenAI = exports.LLM = exports.BaseLLM = void 0;
|
|
4
4
|
var base_1 = require("./base");
|
|
5
5
|
Object.defineProperty(exports, "BaseLLM", { enumerable: true, get: function () { return base_1.BaseLLM; } });
|
|
6
6
|
Object.defineProperty(exports, "LLM", { enumerable: true, get: function () { return base_1.LLM; } });
|
|
7
7
|
var openai_1 = require("./openai");
|
|
8
8
|
Object.defineProperty(exports, "OpenAI", { enumerable: true, get: function () { return openai_1.OpenAI; } });
|
|
9
|
+
var cohere_1 = require("./cohere");
|
|
10
|
+
Object.defineProperty(exports, "Cohere", { enumerable: true, get: function () { return cohere_1.Cohere; } });
|
|
9
11
|
var load_1 = require("./load");
|
|
10
12
|
Object.defineProperty(exports, "loadLLM", { enumerable: true, get: function () { return load_1.loadLLM; } });
|
|
11
13
|
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../llms/index.ts"],"names":[],"mappings":";;;AAAA,+BAAqD;AAA5C,+FAAA,OAAO,OAAA;AAAE,2FAAA,GAAG,OAAA;AACrB,mCAAkC;AAAzB,gGAAA,MAAM,OAAA;AACf,mCAAkC;AAAzB,gGAAA,MAAM,OAAA;AACf,+BAAiC;AAAxB,+FAAA,OAAO,OAAA"}
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
import { BaseLLM } from "./base";
|
|
2
|
+
/**
|
|
3
|
+
* Load an LLM from a local file.
|
|
4
|
+
*
|
|
5
|
+
* @example
|
|
6
|
+
* ```ts
|
|
7
|
+
* import { loadLLM } from "langchain/llms";
|
|
8
|
+
* const model = await loadLLM("/path/to/llm.json");
|
|
9
|
+
* ```
|
|
10
|
+
*/
|
|
11
|
+
export declare const loadLLM: (file: string) => Promise<BaseLLM>;
|
|
@@ -0,0 +1,17 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.loadLLM = void 0;
|
|
4
|
+
const base_1 = require("./base");
|
|
5
|
+
const util_1 = require("../util");
|
|
6
|
+
/**
|
|
7
|
+
* Load an LLM from a local file.
|
|
8
|
+
*
|
|
9
|
+
* @example
|
|
10
|
+
* ```ts
|
|
11
|
+
* import { loadLLM } from "langchain/llms";
|
|
12
|
+
* const model = await loadLLM("/path/to/llm.json");
|
|
13
|
+
* ```
|
|
14
|
+
*/
|
|
15
|
+
const loadLLM = (file) => base_1.BaseLLM.deserialize((0, util_1.parseFileConfig)(file));
|
|
16
|
+
exports.loadLLM = loadLLM;
|
|
17
|
+
//# sourceMappingURL=load.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"load.js","sourceRoot":"","sources":["../../llms/load.ts"],"names":[],"mappings":";;;AAAA,iCAAiC;AACjC,kCAA0C;AAE1C;;;;;;;;GAQG;AACI,MAAM,OAAO,GAAG,CAAC,IAAY,EAAE,EAAE,CACtC,cAAO,CAAC,WAAW,CAAC,IAAA,sBAAe,EAAC,IAAI,CAAC,CAAC,CAAC;AADhC,QAAA,OAAO,WACyB"}
|
|
@@ -0,0 +1,125 @@
|
|
|
1
|
+
import type { CreateCompletionRequest } from "openai";
|
|
2
|
+
import { BaseLLM, LLMResult, LLMCallbackManager } from ".";
|
|
3
|
+
interface ModelParams {
|
|
4
|
+
/** Sampling temperature to use */
|
|
5
|
+
temperature: number;
|
|
6
|
+
/**
|
|
7
|
+
* Maximum number of tokens to generate in the completion. -1 returns as many
|
|
8
|
+
* tokens as possible given the prompt and the model's maximum context size.
|
|
9
|
+
*/
|
|
10
|
+
maxTokens: number;
|
|
11
|
+
/** Total probability mass of tokens to consider at each step */
|
|
12
|
+
topP: number;
|
|
13
|
+
/** Penalizes repeated tokens according to frequency */
|
|
14
|
+
frequencyPenalty: number;
|
|
15
|
+
/** Penalizes repeated tokens */
|
|
16
|
+
presencePenalty: number;
|
|
17
|
+
/** Number of completions to generate for each prompt */
|
|
18
|
+
n: number;
|
|
19
|
+
/** Generates `bestOf` completions server side and returns the "best" */
|
|
20
|
+
bestOf: number;
|
|
21
|
+
/** Dictionary used to adjust the probability of specific tokens being generated */
|
|
22
|
+
logitBias?: Record<string, number>;
|
|
23
|
+
}
|
|
24
|
+
/**
|
|
25
|
+
* Input to OpenAI class.
|
|
26
|
+
* @augments ModelParams
|
|
27
|
+
*/
|
|
28
|
+
interface OpenAIInput extends ModelParams {
|
|
29
|
+
/** Model name to use */
|
|
30
|
+
modelName: string;
|
|
31
|
+
/** Holds any additional parameters that are valid to pass to {@link
|
|
32
|
+
* https://platform.openai.com/docs/api-reference/completions/create |
|
|
33
|
+
* `openai.createCompletion`} that are not explicitly specified on this class.
|
|
34
|
+
*/
|
|
35
|
+
modelKwargs?: Kwargs;
|
|
36
|
+
/** Batch size to use when passing multiple documents to generate */
|
|
37
|
+
batchSize: number;
|
|
38
|
+
/** Maximum number of retries to make when generating */
|
|
39
|
+
maxRetries: number;
|
|
40
|
+
/** List of stop words to use when generating */
|
|
41
|
+
stop?: string[];
|
|
42
|
+
}
|
|
43
|
+
type Kwargs = Record<string, any>;
|
|
44
|
+
/**
|
|
45
|
+
* Wrapper around OpenAI large language models.
|
|
46
|
+
*
|
|
47
|
+
* To use you should have the `openai` package installed, with the
|
|
48
|
+
* `OPENAI_API_KEY` environment variable set.
|
|
49
|
+
*
|
|
50
|
+
* @remarks
|
|
51
|
+
* Any parameters that are valid to be passed to {@link
|
|
52
|
+
* https://platform.openai.com/docs/api-reference/completions/create |
|
|
53
|
+
* `openai.createCompletion`} can be passed through {@link modelKwargs}, even
|
|
54
|
+
* if not explicitly available on this class.
|
|
55
|
+
*
|
|
56
|
+
* @augments BaseLLM
|
|
57
|
+
* @augments OpenAIInput
|
|
58
|
+
*/
|
|
59
|
+
export declare class OpenAI extends BaseLLM implements OpenAIInput {
|
|
60
|
+
temperature: number;
|
|
61
|
+
maxTokens: number;
|
|
62
|
+
topP: number;
|
|
63
|
+
frequencyPenalty: number;
|
|
64
|
+
presencePenalty: number;
|
|
65
|
+
n: number;
|
|
66
|
+
bestOf: number;
|
|
67
|
+
logitBias?: Record<string, number>;
|
|
68
|
+
modelName: string;
|
|
69
|
+
modelKwargs?: Kwargs;
|
|
70
|
+
batchSize: number;
|
|
71
|
+
maxRetries: number;
|
|
72
|
+
stop?: string[];
|
|
73
|
+
private client;
|
|
74
|
+
constructor(fields?: Partial<OpenAIInput> & {
|
|
75
|
+
callbackManager?: LLMCallbackManager;
|
|
76
|
+
verbose?: boolean;
|
|
77
|
+
openAIApiKey?: string;
|
|
78
|
+
});
|
|
79
|
+
/**
|
|
80
|
+
* Get the parameters used to invoke the model
|
|
81
|
+
*/
|
|
82
|
+
invocationParams(): CreateCompletionRequest & Kwargs;
|
|
83
|
+
/**
|
|
84
|
+
* Get the identifyin parameters for the model
|
|
85
|
+
*/
|
|
86
|
+
identifyingParams(): {
|
|
87
|
+
model: string;
|
|
88
|
+
prompt?: import("openai").CreateCompletionRequestPrompt | null | undefined;
|
|
89
|
+
suffix?: string | null | undefined;
|
|
90
|
+
max_tokens?: number | null | undefined;
|
|
91
|
+
temperature?: number | null | undefined;
|
|
92
|
+
top_p?: number | null | undefined;
|
|
93
|
+
n?: number | null | undefined;
|
|
94
|
+
stream?: boolean | null | undefined;
|
|
95
|
+
logprobs?: number | null | undefined;
|
|
96
|
+
echo?: boolean | null | undefined;
|
|
97
|
+
stop?: import("openai").CreateCompletionRequestStop | null | undefined;
|
|
98
|
+
presence_penalty?: number | null | undefined;
|
|
99
|
+
frequency_penalty?: number | null | undefined;
|
|
100
|
+
best_of?: number | null | undefined;
|
|
101
|
+
logit_bias?: object | null | undefined;
|
|
102
|
+
user?: string | undefined;
|
|
103
|
+
model_name: string;
|
|
104
|
+
};
|
|
105
|
+
/**
|
|
106
|
+
* Call out to OpenAI's endpoint with k unique prompts
|
|
107
|
+
*
|
|
108
|
+
* @param prompts - The prompts to pass into the model.
|
|
109
|
+
* @param [stop] - Optional list of stop words to use when generating.
|
|
110
|
+
*
|
|
111
|
+
* @returns The full LLM output.
|
|
112
|
+
*
|
|
113
|
+
* @example
|
|
114
|
+
* ```ts
|
|
115
|
+
* import { OpenAI } from "langchain/llms";
|
|
116
|
+
* const openai = new OpenAI();
|
|
117
|
+
* const response = await openai.generate(["Tell me a joke."]);
|
|
118
|
+
* ```
|
|
119
|
+
*/
|
|
120
|
+
_generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
121
|
+
/** @ignore */
|
|
122
|
+
completionWithRetry(request: CreateCompletionRequest): Promise<import("axios").AxiosResponse<import("openai").CreateCompletionResponse, any>>;
|
|
123
|
+
_llmType(): string;
|
|
124
|
+
}
|
|
125
|
+
export {};
|