langchain 0.0.192 → 0.0.193
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/agents/agent.cjs +24 -0
- package/dist/agents/agent.d.ts +24 -0
- package/dist/agents/agent.js +24 -0
- package/dist/agents/chat/outputParser.cjs +32 -0
- package/dist/agents/chat/outputParser.d.ts +32 -0
- package/dist/agents/chat/outputParser.js +32 -0
- package/dist/agents/executor.cjs +14 -0
- package/dist/agents/executor.d.ts +14 -0
- package/dist/agents/executor.js +14 -0
- package/dist/agents/mrkl/index.cjs +21 -0
- package/dist/agents/mrkl/index.d.ts +21 -0
- package/dist/agents/mrkl/index.js +21 -0
- package/dist/agents/openai/output_parser.cjs +65 -0
- package/dist/agents/openai/output_parser.d.ts +65 -0
- package/dist/agents/openai/output_parser.js +65 -0
- package/dist/agents/react/output_parser.cjs +16 -0
- package/dist/agents/react/output_parser.d.ts +16 -0
- package/dist/agents/react/output_parser.js +16 -0
- package/dist/agents/structured_chat/outputParser.cjs +12 -0
- package/dist/agents/structured_chat/outputParser.d.ts +12 -0
- package/dist/agents/structured_chat/outputParser.js +12 -0
- package/dist/agents/toolkits/aws_sfn.cjs +20 -0
- package/dist/agents/toolkits/aws_sfn.d.ts +20 -0
- package/dist/agents/toolkits/aws_sfn.js +20 -0
- package/dist/agents/toolkits/json/json.cjs +8 -0
- package/dist/agents/toolkits/json/json.d.ts +8 -0
- package/dist/agents/toolkits/json/json.js +8 -0
- package/dist/agents/toolkits/openapi/openapi.cjs +18 -0
- package/dist/agents/toolkits/openapi/openapi.d.ts +18 -0
- package/dist/agents/toolkits/openapi/openapi.js +18 -0
- package/dist/agents/toolkits/sql/sql.cjs +8 -0
- package/dist/agents/toolkits/sql/sql.d.ts +8 -0
- package/dist/agents/toolkits/sql/sql.js +8 -0
- package/dist/agents/toolkits/vectorstore/vectorstore.cjs +16 -0
- package/dist/agents/toolkits/vectorstore/vectorstore.d.ts +16 -0
- package/dist/agents/toolkits/vectorstore/vectorstore.js +16 -0
- package/dist/agents/toolkits/zapier/zapier.cjs +10 -0
- package/dist/agents/toolkits/zapier/zapier.d.ts +10 -0
- package/dist/agents/toolkits/zapier/zapier.js +10 -0
- package/dist/agents/xml/output_parser.cjs +21 -0
- package/dist/agents/xml/output_parser.d.ts +21 -0
- package/dist/agents/xml/output_parser.js +21 -0
- package/dist/cache/cloudflare_kv.cjs +13 -0
- package/dist/cache/cloudflare_kv.d.ts +13 -0
- package/dist/cache/cloudflare_kv.js +13 -0
- package/dist/cache/ioredis.cjs +13 -0
- package/dist/cache/ioredis.d.ts +13 -0
- package/dist/cache/ioredis.js +13 -0
- package/dist/cache/momento.cjs +19 -0
- package/dist/cache/momento.d.ts +19 -0
- package/dist/cache/momento.js +19 -0
- package/dist/cache/redis.cjs +10 -0
- package/dist/cache/redis.d.ts +10 -0
- package/dist/cache/redis.js +10 -0
- package/dist/cache/upstash_redis.cjs +15 -0
- package/dist/cache/upstash_redis.d.ts +15 -0
- package/dist/cache/upstash_redis.js +15 -0
- package/dist/callbacks/handlers/console.cjs +10 -0
- package/dist/callbacks/handlers/console.d.ts +10 -0
- package/dist/callbacks/handlers/console.js +10 -0
- package/dist/callbacks/manager.cjs +49 -0
- package/dist/callbacks/manager.d.ts +49 -0
- package/dist/callbacks/manager.js +49 -0
- package/dist/chains/analyze_documents_chain.cjs +18 -0
- package/dist/chains/analyze_documents_chain.d.ts +18 -0
- package/dist/chains/analyze_documents_chain.js +18 -0
- package/dist/chains/constitutional_ai/constitutional_chain.cjs +21 -0
- package/dist/chains/constitutional_ai/constitutional_chain.d.ts +21 -0
- package/dist/chains/constitutional_ai/constitutional_chain.js +21 -0
- package/dist/chains/constitutional_ai/constitutional_principle.cjs +21 -0
- package/dist/chains/constitutional_ai/constitutional_principle.d.ts +21 -0
- package/dist/chains/constitutional_ai/constitutional_principle.js +21 -0
- package/dist/chains/conversation.cjs +13 -0
- package/dist/chains/conversation.d.ts +13 -0
- package/dist/chains/conversation.js +13 -0
- package/dist/chains/conversational_retrieval_chain.cjs +29 -0
- package/dist/chains/conversational_retrieval_chain.d.ts +29 -0
- package/dist/chains/conversational_retrieval_chain.js +29 -0
- package/dist/chains/graph_qa/cypher.cjs +10 -0
- package/dist/chains/graph_qa/cypher.d.ts +10 -0
- package/dist/chains/graph_qa/cypher.js +10 -0
- package/dist/chains/openai_moderation.cjs +28 -0
- package/dist/chains/openai_moderation.d.ts +28 -0
- package/dist/chains/openai_moderation.js +28 -0
- package/dist/chains/retrieval_qa.cjs +21 -0
- package/dist/chains/retrieval_qa.d.ts +21 -0
- package/dist/chains/retrieval_qa.js +21 -0
- package/dist/chains/router/multi_prompt.cjs +19 -0
- package/dist/chains/router/multi_prompt.d.ts +19 -0
- package/dist/chains/router/multi_prompt.js +19 -0
- package/dist/chains/router/multi_retrieval_qa.cjs +29 -0
- package/dist/chains/router/multi_retrieval_qa.d.ts +29 -0
- package/dist/chains/router/multi_retrieval_qa.js +29 -0
- package/dist/chains/sequential_chain.cjs +43 -0
- package/dist/chains/sequential_chain.d.ts +43 -0
- package/dist/chains/sequential_chain.js +43 -0
- package/dist/chains/sql_db/sql_db_chain.cjs +9 -0
- package/dist/chains/sql_db/sql_db_chain.d.ts +9 -0
- package/dist/chains/sql_db/sql_db_chain.js +9 -0
- package/dist/chat_models/anthropic.cjs +9 -1
- package/dist/chat_models/anthropic.d.ts +9 -1
- package/dist/chat_models/anthropic.js +9 -1
- package/dist/chat_models/baiduwenxin.cjs +20 -0
- package/dist/chat_models/baiduwenxin.d.ts +20 -0
- package/dist/chat_models/baiduwenxin.js +20 -0
- package/dist/chat_models/base.cjs +2 -0
- package/dist/chat_models/base.d.ts +1 -1
- package/dist/chat_models/base.js +2 -0
- package/dist/chat_models/bedrock/index.cjs +11 -0
- package/dist/chat_models/bedrock/index.d.ts +11 -0
- package/dist/chat_models/bedrock/index.js +11 -0
- package/dist/chat_models/bedrock/web.cjs +9 -0
- package/dist/chat_models/bedrock/web.d.ts +9 -0
- package/dist/chat_models/bedrock/web.js +9 -0
- package/dist/chat_models/cloudflare_workersai.cjs +15 -0
- package/dist/chat_models/cloudflare_workersai.d.ts +15 -0
- package/dist/chat_models/cloudflare_workersai.js +15 -0
- package/dist/chat_models/fake.cjs +16 -0
- package/dist/chat_models/fake.d.ts +16 -0
- package/dist/chat_models/fake.js +16 -0
- package/dist/chat_models/fireworks.cjs +10 -0
- package/dist/chat_models/fireworks.d.ts +10 -0
- package/dist/chat_models/fireworks.js +10 -0
- package/dist/chat_models/googlepalm.cjs +24 -0
- package/dist/chat_models/googlepalm.d.ts +24 -0
- package/dist/chat_models/googlepalm.js +24 -0
- package/dist/chat_models/googlevertexai/index.cjs +7 -0
- package/dist/chat_models/googlevertexai/index.d.ts +7 -0
- package/dist/chat_models/googlevertexai/index.js +7 -0
- package/dist/chat_models/googlevertexai/web.cjs +9 -0
- package/dist/chat_models/googlevertexai/web.d.ts +9 -0
- package/dist/chat_models/googlevertexai/web.js +9 -0
- package/dist/chat_models/iflytek_xinghuo/index.cjs +8 -0
- package/dist/chat_models/iflytek_xinghuo/index.d.ts +8 -0
- package/dist/chat_models/iflytek_xinghuo/index.js +8 -0
- package/dist/chat_models/iflytek_xinghuo/web.cjs +8 -0
- package/dist/chat_models/iflytek_xinghuo/web.d.ts +8 -0
- package/dist/chat_models/iflytek_xinghuo/web.js +8 -0
- package/dist/chat_models/llama_cpp.cjs +17 -0
- package/dist/chat_models/llama_cpp.d.ts +17 -0
- package/dist/chat_models/llama_cpp.js +17 -0
- package/dist/chat_models/minimax.cjs +27 -0
- package/dist/chat_models/minimax.d.ts +27 -0
- package/dist/chat_models/minimax.js +27 -0
- package/dist/chat_models/ollama.cjs +24 -0
- package/dist/chat_models/ollama.d.ts +24 -0
- package/dist/chat_models/ollama.js +24 -0
- package/dist/chat_models/openai.cjs +15 -0
- package/dist/chat_models/openai.d.ts +15 -0
- package/dist/chat_models/openai.js +15 -0
- package/dist/chat_models/yandex.cjs +14 -0
- package/dist/chat_models/yandex.d.ts +14 -0
- package/dist/chat_models/yandex.js +14 -0
- package/dist/document_loaders/fs/directory.cjs +14 -0
- package/dist/document_loaders/fs/directory.d.ts +14 -0
- package/dist/document_loaders/fs/directory.js +14 -0
- package/dist/document_loaders/fs/openai_whisper_audio.cjs +10 -0
- package/dist/document_loaders/fs/openai_whisper_audio.d.ts +10 -0
- package/dist/document_loaders/fs/openai_whisper_audio.js +10 -0
- package/dist/document_loaders/fs/pdf.cjs +6 -0
- package/dist/document_loaders/fs/pdf.d.ts +6 -0
- package/dist/document_loaders/fs/pdf.js +6 -0
- package/dist/document_loaders/fs/srt.cjs +6 -0
- package/dist/document_loaders/fs/srt.d.ts +6 -0
- package/dist/document_loaders/fs/srt.js +6 -0
- package/dist/document_loaders/fs/text.cjs +5 -0
- package/dist/document_loaders/fs/text.d.ts +5 -0
- package/dist/document_loaders/fs/text.js +5 -0
- package/dist/document_loaders/fs/unstructured.cjs +7 -0
- package/dist/document_loaders/fs/unstructured.d.ts +7 -0
- package/dist/document_loaders/fs/unstructured.js +7 -0
- package/dist/document_loaders/web/apify_dataset.cjs +21 -0
- package/dist/document_loaders/web/apify_dataset.d.ts +21 -0
- package/dist/document_loaders/web/apify_dataset.js +21 -0
- package/dist/document_loaders/web/assemblyai.cjs +24 -0
- package/dist/document_loaders/web/assemblyai.d.ts +24 -0
- package/dist/document_loaders/web/assemblyai.js +24 -0
- package/dist/document_loaders/web/azure_blob_storage_file.cjs +11 -0
- package/dist/document_loaders/web/azure_blob_storage_file.d.ts +11 -0
- package/dist/document_loaders/web/azure_blob_storage_file.js +11 -0
- package/dist/document_loaders/web/cheerio.cjs +6 -0
- package/dist/document_loaders/web/cheerio.d.ts +6 -0
- package/dist/document_loaders/web/cheerio.js +6 -0
- package/dist/document_loaders/web/college_confidential.cjs +6 -0
- package/dist/document_loaders/web/college_confidential.d.ts +6 -0
- package/dist/document_loaders/web/college_confidential.js +6 -0
- package/dist/document_loaders/web/confluence.cjs +11 -0
- package/dist/document_loaders/web/confluence.d.ts +11 -0
- package/dist/document_loaders/web/confluence.js +11 -0
- package/dist/document_loaders/web/figma.cjs +9 -0
- package/dist/document_loaders/web/figma.d.ts +9 -0
- package/dist/document_loaders/web/figma.js +9 -0
- package/dist/document_loaders/web/notionapi.cjs +16 -0
- package/dist/document_loaders/web/notionapi.d.ts +16 -0
- package/dist/document_loaders/web/notionapi.js +16 -0
- package/dist/document_loaders/web/notiondb.cjs +11 -1
- package/dist/document_loaders/web/notiondb.d.ts +11 -1
- package/dist/document_loaders/web/notiondb.js +11 -1
- package/dist/document_loaders/web/pdf.cjs +6 -0
- package/dist/document_loaders/web/pdf.d.ts +6 -0
- package/dist/document_loaders/web/pdf.js +6 -0
- package/dist/document_loaders/web/puppeteer.cjs +12 -0
- package/dist/document_loaders/web/puppeteer.d.ts +12 -0
- package/dist/document_loaders/web/puppeteer.js +12 -0
- package/dist/document_loaders/web/s3.cjs +17 -0
- package/dist/document_loaders/web/s3.d.ts +17 -0
- package/dist/document_loaders/web/s3.js +17 -0
- package/dist/document_loaders/web/searchapi.cjs +9 -0
- package/dist/document_loaders/web/searchapi.d.ts +9 -0
- package/dist/document_loaders/web/searchapi.js +9 -0
- package/dist/document_loaders/web/serpapi.cjs +5 -0
- package/dist/document_loaders/web/serpapi.d.ts +5 -0
- package/dist/document_loaders/web/serpapi.js +5 -0
- package/dist/document_loaders/web/sonix_audio.cjs +12 -0
- package/dist/document_loaders/web/sonix_audio.d.ts +12 -0
- package/dist/document_loaders/web/sonix_audio.js +12 -0
- package/dist/document_loaders/web/sort_xyz_blockchain.cjs +22 -0
- package/dist/document_loaders/web/sort_xyz_blockchain.d.ts +22 -0
- package/dist/document_loaders/web/sort_xyz_blockchain.js +22 -0
- package/dist/document_loaders/web/youtube.cjs +9 -0
- package/dist/document_loaders/web/youtube.d.ts +9 -0
- package/dist/document_loaders/web/youtube.js +9 -0
- package/dist/document_transformers/html_to_text.cjs +18 -0
- package/dist/document_transformers/html_to_text.d.ts +18 -0
- package/dist/document_transformers/html_to_text.js +18 -0
- package/dist/document_transformers/mozilla_readability.cjs +18 -0
- package/dist/document_transformers/mozilla_readability.d.ts +18 -0
- package/dist/document_transformers/mozilla_readability.js +18 -0
- package/dist/embeddings/bedrock.cjs +17 -0
- package/dist/embeddings/bedrock.d.ts +17 -0
- package/dist/embeddings/bedrock.js +17 -0
- package/dist/embeddings/cache_backed.cjs +37 -0
- package/dist/embeddings/cache_backed.d.ts +37 -0
- package/dist/embeddings/cache_backed.js +37 -0
- package/dist/embeddings/cohere.cjs +10 -0
- package/dist/embeddings/cohere.d.ts +10 -0
- package/dist/embeddings/cohere.js +10 -0
- package/dist/embeddings/googlepalm.cjs +17 -0
- package/dist/embeddings/googlepalm.d.ts +17 -0
- package/dist/embeddings/googlepalm.js +17 -0
- package/dist/embeddings/googlevertexai.cjs +8 -0
- package/dist/embeddings/googlevertexai.d.ts +8 -0
- package/dist/embeddings/googlevertexai.js +8 -0
- package/dist/embeddings/hf_transformers.cjs +18 -0
- package/dist/embeddings/hf_transformers.d.ts +18 -0
- package/dist/embeddings/hf_transformers.js +18 -0
- package/dist/embeddings/llama_cpp.cjs +16 -0
- package/dist/embeddings/llama_cpp.d.ts +16 -0
- package/dist/embeddings/llama_cpp.js +16 -0
- package/dist/embeddings/minimax.cjs +15 -0
- package/dist/embeddings/minimax.d.ts +15 -0
- package/dist/embeddings/minimax.js +15 -0
- package/dist/embeddings/openai.cjs +10 -0
- package/dist/embeddings/openai.d.ts +10 -0
- package/dist/embeddings/openai.js +10 -0
- package/dist/embeddings/tensorflow.cjs +16 -0
- package/dist/embeddings/tensorflow.d.ts +16 -0
- package/dist/embeddings/tensorflow.js +16 -0
- package/dist/experimental/autogpt/agent.cjs +21 -0
- package/dist/experimental/autogpt/agent.d.ts +21 -0
- package/dist/experimental/autogpt/agent.js +21 -0
- package/dist/experimental/babyagi/agent.cjs +12 -0
- package/dist/experimental/babyagi/agent.d.ts +12 -0
- package/dist/experimental/babyagi/agent.js +12 -0
- package/dist/experimental/chat_models/ollama_functions.cjs +12 -15
- package/dist/experimental/chat_models/ollama_functions.d.ts +3 -4
- package/dist/experimental/chat_models/ollama_functions.js +13 -16
- package/dist/experimental/generative_agents/generative_agent.cjs +30 -0
- package/dist/experimental/generative_agents/generative_agent.d.ts +30 -0
- package/dist/experimental/generative_agents/generative_agent.js +30 -0
- package/dist/experimental/generative_agents/generative_agent_memory.cjs +18 -0
- package/dist/experimental/generative_agents/generative_agent_memory.d.ts +18 -0
- package/dist/experimental/generative_agents/generative_agent_memory.js +18 -0
- package/dist/llms/base.cjs +3 -1
- package/dist/llms/base.d.ts +1 -1
- package/dist/llms/base.js +3 -1
- package/dist/llms/cohere.cjs +13 -0
- package/dist/llms/cohere.d.ts +13 -0
- package/dist/llms/cohere.js +13 -0
- package/dist/llms/googlevertexai/index.cjs +12 -0
- package/dist/llms/googlevertexai/index.d.ts +12 -0
- package/dist/llms/googlevertexai/index.js +12 -0
- package/dist/llms/googlevertexai/web.cjs +12 -0
- package/dist/llms/googlevertexai/web.d.ts +12 -0
- package/dist/llms/googlevertexai/web.js +12 -0
- package/dist/llms/hf.cjs +13 -0
- package/dist/llms/hf.d.ts +13 -0
- package/dist/llms/hf.js +13 -0
- package/dist/llms/ollama.cjs +19 -0
- package/dist/llms/ollama.d.ts +19 -0
- package/dist/llms/ollama.js +19 -0
- package/dist/llms/openai-chat.cjs +17 -0
- package/dist/llms/openai-chat.d.ts +17 -0
- package/dist/llms/openai-chat.js +17 -0
- package/dist/llms/openai.cjs +14 -0
- package/dist/llms/openai.d.ts +14 -0
- package/dist/llms/openai.js +14 -0
- package/dist/llms/portkey.cjs +24 -0
- package/dist/llms/portkey.d.ts +24 -0
- package/dist/llms/portkey.js +24 -0
- package/dist/llms/replicate.cjs +11 -0
- package/dist/llms/replicate.d.ts +11 -0
- package/dist/llms/replicate.js +11 -0
- package/dist/memory/buffer_memory.cjs +24 -0
- package/dist/memory/buffer_memory.d.ts +24 -0
- package/dist/memory/buffer_memory.js +24 -0
- package/dist/memory/buffer_token_memory.cjs +15 -0
- package/dist/memory/buffer_token_memory.d.ts +15 -0
- package/dist/memory/buffer_token_memory.js +15 -0
- package/dist/memory/buffer_window_memory.cjs +23 -0
- package/dist/memory/buffer_window_memory.d.ts +23 -0
- package/dist/memory/buffer_window_memory.js +23 -0
- package/dist/memory/entity_memory.cjs +29 -0
- package/dist/memory/entity_memory.d.ts +29 -0
- package/dist/memory/entity_memory.js +29 -0
- package/dist/memory/summary.cjs +24 -0
- package/dist/memory/summary.d.ts +24 -0
- package/dist/memory/summary.js +24 -0
- package/dist/memory/summary_buffer.cjs +32 -0
- package/dist/memory/summary_buffer.d.ts +32 -0
- package/dist/memory/summary_buffer.js +32 -0
- package/dist/memory/vector_store.cjs +24 -0
- package/dist/memory/vector_store.d.ts +24 -0
- package/dist/memory/vector_store.js +24 -0
- package/dist/memory/zep.cjs +34 -0
- package/dist/memory/zep.d.ts +34 -0
- package/dist/memory/zep.js +34 -0
- package/dist/prompts/chat.cjs +31 -0
- package/dist/prompts/chat.d.ts +31 -0
- package/dist/prompts/chat.js +31 -0
- package/dist/prompts/few_shot.cjs +31 -0
- package/dist/prompts/few_shot.d.ts +31 -0
- package/dist/prompts/few_shot.js +31 -0
- package/dist/prompts/pipeline.cjs +39 -0
- package/dist/prompts/pipeline.d.ts +39 -0
- package/dist/prompts/pipeline.js +39 -0
- package/dist/prompts/selectors/LengthBasedExampleSelector.cjs +36 -0
- package/dist/prompts/selectors/LengthBasedExampleSelector.d.ts +36 -0
- package/dist/prompts/selectors/LengthBasedExampleSelector.js +36 -0
- package/dist/prompts/selectors/SemanticSimilarityExampleSelector.cjs +25 -0
- package/dist/prompts/selectors/SemanticSimilarityExampleSelector.d.ts +25 -0
- package/dist/prompts/selectors/SemanticSimilarityExampleSelector.js +25 -0
- package/dist/retrievers/amazon_kendra.cjs +16 -0
- package/dist/retrievers/amazon_kendra.d.ts +16 -0
- package/dist/retrievers/amazon_kendra.js +16 -0
- package/dist/retrievers/chaindesk.cjs +11 -0
- package/dist/retrievers/chaindesk.d.ts +11 -0
- package/dist/retrievers/chaindesk.js +11 -0
- package/dist/retrievers/contextual_compression.cjs +10 -0
- package/dist/retrievers/contextual_compression.d.ts +10 -0
- package/dist/retrievers/contextual_compression.js +10 -0
- package/dist/retrievers/document_compressors/embeddings_filter.cjs +13 -0
- package/dist/retrievers/document_compressors/embeddings_filter.d.ts +13 -0
- package/dist/retrievers/document_compressors/embeddings_filter.js +13 -0
- package/dist/retrievers/document_compressors/index.cjs +26 -0
- package/dist/retrievers/document_compressors/index.d.ts +26 -0
- package/dist/retrievers/document_compressors/index.js +26 -0
- package/dist/retrievers/hyde.cjs +19 -0
- package/dist/retrievers/hyde.d.ts +19 -0
- package/dist/retrievers/hyde.js +19 -0
- package/dist/retrievers/metal.cjs +11 -0
- package/dist/retrievers/metal.d.ts +11 -0
- package/dist/retrievers/metal.js +11 -0
- package/dist/retrievers/multi_query.cjs +13 -1
- package/dist/retrievers/multi_query.d.ts +13 -0
- package/dist/retrievers/multi_query.js +13 -1
- package/dist/retrievers/multi_vector.cjs +13 -0
- package/dist/retrievers/multi_vector.d.ts +13 -0
- package/dist/retrievers/multi_vector.js +13 -0
- package/dist/retrievers/parent_document.cjs +21 -0
- package/dist/retrievers/parent_document.d.ts +21 -0
- package/dist/retrievers/parent_document.js +21 -0
- package/dist/retrievers/remote/chatgpt-plugin.cjs +10 -0
- package/dist/retrievers/remote/chatgpt-plugin.d.ts +10 -0
- package/dist/retrievers/remote/chatgpt-plugin.js +10 -0
- package/dist/retrievers/self_query/chroma.cjs +15 -0
- package/dist/retrievers/self_query/chroma.d.ts +15 -0
- package/dist/retrievers/self_query/chroma.js +15 -0
- package/dist/retrievers/self_query/functional.cjs +7 -0
- package/dist/retrievers/self_query/functional.d.ts +7 -0
- package/dist/retrievers/self_query/functional.js +7 -0
- package/dist/retrievers/self_query/index.cjs +13 -0
- package/dist/retrievers/self_query/index.d.ts +13 -0
- package/dist/retrievers/self_query/index.js +13 -0
- package/dist/retrievers/self_query/pinecone.cjs +14 -0
- package/dist/retrievers/self_query/pinecone.d.ts +14 -0
- package/dist/retrievers/self_query/pinecone.js +14 -0
- package/dist/retrievers/self_query/supabase.cjs +14 -0
- package/dist/retrievers/self_query/supabase.d.ts +14 -0
- package/dist/retrievers/self_query/supabase.js +14 -0
- package/dist/retrievers/self_query/weaviate.cjs +14 -0
- package/dist/retrievers/self_query/weaviate.d.ts +14 -0
- package/dist/retrievers/self_query/weaviate.js +14 -0
- package/dist/retrievers/time_weighted.cjs +16 -0
- package/dist/retrievers/time_weighted.d.ts +16 -0
- package/dist/retrievers/time_weighted.js +16 -0
- package/dist/retrievers/vespa.cjs +15 -0
- package/dist/retrievers/vespa.d.ts +15 -0
- package/dist/retrievers/vespa.js +15 -0
- package/dist/retrievers/zep.cjs +10 -0
- package/dist/retrievers/zep.d.ts +10 -0
- package/dist/retrievers/zep.js +10 -0
- package/dist/schema/index.d.ts +8 -6
- package/dist/schema/output_parser.cjs +15 -0
- package/dist/schema/output_parser.d.ts +15 -0
- package/dist/schema/output_parser.js +15 -0
- package/dist/schema/runnable/base.cjs +20 -0
- package/dist/schema/runnable/base.d.ts +20 -0
- package/dist/schema/runnable/base.js +20 -0
- package/dist/schema/runnable/branch.cjs +28 -0
- package/dist/schema/runnable/branch.d.ts +28 -0
- package/dist/schema/runnable/branch.js +28 -0
- package/dist/storage/file_system.cjs +21 -0
- package/dist/storage/file_system.d.ts +21 -0
- package/dist/storage/file_system.js +21 -0
- package/dist/storage/in_memory.cjs +15 -0
- package/dist/storage/in_memory.d.ts +15 -0
- package/dist/storage/in_memory.js +15 -0
- package/dist/storage/ioredis.cjs +24 -0
- package/dist/storage/ioredis.d.ts +24 -0
- package/dist/storage/ioredis.js +24 -0
- package/dist/storage/upstash_redis.cjs +19 -0
- package/dist/storage/upstash_redis.d.ts +19 -0
- package/dist/storage/upstash_redis.js +19 -0
- package/dist/storage/vercel_kv.cjs +16 -0
- package/dist/storage/vercel_kv.d.ts +16 -0
- package/dist/storage/vercel_kv.js +16 -0
- package/dist/stores/message/cassandra.cjs +25 -0
- package/dist/stores/message/cassandra.d.ts +25 -0
- package/dist/stores/message/cassandra.js +25 -0
- package/dist/stores/message/cloudflare_d1.cjs +18 -0
- package/dist/stores/message/cloudflare_d1.d.ts +18 -0
- package/dist/stores/message/cloudflare_d1.js +18 -0
- package/dist/stores/message/firestore.cjs +19 -0
- package/dist/stores/message/firestore.d.ts +19 -0
- package/dist/stores/message/firestore.js +19 -0
- package/dist/stores/message/ioredis.cjs +18 -0
- package/dist/stores/message/ioredis.d.ts +18 -0
- package/dist/stores/message/ioredis.js +18 -0
- package/dist/stores/message/momento.cjs +18 -0
- package/dist/stores/message/momento.d.ts +18 -0
- package/dist/stores/message/momento.js +18 -0
- package/dist/stores/message/mongodb.cjs +11 -0
- package/dist/stores/message/mongodb.d.ts +11 -0
- package/dist/stores/message/mongodb.js +11 -0
- package/dist/stores/message/planetscale.cjs +18 -0
- package/dist/stores/message/planetscale.d.ts +18 -0
- package/dist/stores/message/planetscale.js +18 -0
- package/dist/stores/message/redis.cjs +18 -0
- package/dist/stores/message/redis.d.ts +18 -0
- package/dist/stores/message/redis.js +18 -0
- package/dist/stores/message/xata.cjs +23 -0
- package/dist/stores/message/xata.d.ts +23 -0
- package/dist/stores/message/xata.js +23 -0
- package/dist/tools/calculator.cjs +6 -0
- package/dist/tools/calculator.d.ts +6 -0
- package/dist/tools/calculator.js +6 -0
- package/dist/tools/google_calendar/create.cjs +22 -0
- package/dist/tools/google_calendar/create.d.ts +22 -0
- package/dist/tools/google_calendar/create.js +22 -0
- package/dist/tools/google_calendar/view.cjs +20 -0
- package/dist/tools/google_calendar/view.d.ts +20 -0
- package/dist/tools/google_calendar/view.js +20 -0
- package/dist/tools/searchapi.cjs +27 -0
- package/dist/tools/searchapi.d.ts +27 -0
- package/dist/tools/searchapi.js +27 -0
- package/dist/tools/searxng_search.cjs +20 -1
- package/dist/tools/searxng_search.d.ts +20 -1
- package/dist/tools/searxng_search.js +20 -1
- package/dist/tools/webbrowser.cjs +8 -0
- package/dist/tools/webbrowser.d.ts +8 -0
- package/dist/tools/webbrowser.js +8 -0
- package/dist/tools/wikipedia_query_run.cjs +8 -0
- package/dist/tools/wikipedia_query_run.d.ts +8 -0
- package/dist/tools/wikipedia_query_run.js +8 -0
- package/dist/tools/wolframalpha.cjs +9 -0
- package/dist/tools/wolframalpha.d.ts +9 -0
- package/dist/tools/wolframalpha.js +9 -0
- package/dist/util/googlevertexai-connection.cjs +3 -9
- package/dist/util/googlevertexai-connection.js +3 -9
- package/dist/vectorstores/opensearch.cjs +2 -1
- package/dist/vectorstores/opensearch.js +2 -1
- package/dist/vectorstores/prisma.cjs +6 -0
- package/dist/vectorstores/prisma.d.ts +3 -0
- package/dist/vectorstores/prisma.js +6 -0
- package/package.json +1 -1
|
@@ -95,6 +95,24 @@ declare class GenerativeAgentMemoryChain extends BaseChain {
|
|
|
95
95
|
* formatting memories, getting memories until a token limit is reached,
|
|
96
96
|
* loading memory variables, saving the context of a model run to memory,
|
|
97
97
|
* and clearing memory contents.
|
|
98
|
+
* @example
|
|
99
|
+
* ```typescript
|
|
100
|
+
* const createNewMemoryRetriever = async () => {
|
|
101
|
+
* const vectorStore = new MemoryVectorStore(new OpenAIEmbeddings());
|
|
102
|
+
* const retriever = new TimeWeightedVectorStoreRetriever({
|
|
103
|
+
* vectorStore,
|
|
104
|
+
* otherScoreKeys: ["importance"],
|
|
105
|
+
* k: 15,
|
|
106
|
+
* });
|
|
107
|
+
* return retriever;
|
|
108
|
+
* };
|
|
109
|
+
* const tommiesMemory = new GenerativeAgentMemory(
|
|
110
|
+
* llm,
|
|
111
|
+
* await createNewMemoryRetriever(),
|
|
112
|
+
* { reflectionThreshold: 8 },
|
|
113
|
+
* );
|
|
114
|
+
* const summary = await tommiesMemory.getSummary();
|
|
115
|
+
* ```
|
|
98
116
|
*/
|
|
99
117
|
export declare class GenerativeAgentMemory extends BaseMemory {
|
|
100
118
|
llm: BaseLanguageModel;
|
|
@@ -243,6 +243,24 @@ class GenerativeAgentMemoryChain extends BaseChain {
|
|
|
243
243
|
* formatting memories, getting memories until a token limit is reached,
|
|
244
244
|
* loading memory variables, saving the context of a model run to memory,
|
|
245
245
|
* and clearing memory contents.
|
|
246
|
+
* @example
|
|
247
|
+
* ```typescript
|
|
248
|
+
* const createNewMemoryRetriever = async () => {
|
|
249
|
+
* const vectorStore = new MemoryVectorStore(new OpenAIEmbeddings());
|
|
250
|
+
* const retriever = new TimeWeightedVectorStoreRetriever({
|
|
251
|
+
* vectorStore,
|
|
252
|
+
* otherScoreKeys: ["importance"],
|
|
253
|
+
* k: 15,
|
|
254
|
+
* });
|
|
255
|
+
* return retriever;
|
|
256
|
+
* };
|
|
257
|
+
* const tommiesMemory = new GenerativeAgentMemory(
|
|
258
|
+
* llm,
|
|
259
|
+
* await createNewMemoryRetriever(),
|
|
260
|
+
* { reflectionThreshold: 8 },
|
|
261
|
+
* );
|
|
262
|
+
* const summary = await tommiesMemory.getSummary();
|
|
263
|
+
* ```
|
|
246
264
|
*/
|
|
247
265
|
export class GenerativeAgentMemory extends BaseMemory {
|
|
248
266
|
constructor(llm, memoryRetriever, config) {
|
package/dist/llms/base.cjs
CHANGED
|
@@ -2,8 +2,8 @@
|
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
3
|
exports.LLM = exports.BaseLLM = void 0;
|
|
4
4
|
const index_js_1 = require("../schema/index.cjs");
|
|
5
|
-
const index_js_2 = require("../base_language/index.cjs");
|
|
6
5
|
const manager_js_1 = require("../callbacks/manager.cjs");
|
|
6
|
+
const index_js_2 = require("../base_language/index.cjs");
|
|
7
7
|
const base_js_1 = require("../memory/base.cjs");
|
|
8
8
|
/**
|
|
9
9
|
* LLM Wrapper. Provides an {@link call} (an {@link generate}) function that takes in a prompt (or prompts) and returns a string.
|
|
@@ -54,6 +54,7 @@ class BaseLLM extends index_js_2.BaseLanguageModel {
|
|
|
54
54
|
const extra = {
|
|
55
55
|
options: callOptions,
|
|
56
56
|
invocation_params: this?.invocationParams(callOptions),
|
|
57
|
+
batch_size: 1,
|
|
57
58
|
};
|
|
58
59
|
const runManagers = await callbackManager_?.handleLLMStart(this.toJSON(), [prompt.toString()], undefined, undefined, extra, undefined, undefined, runnableConfig.runName);
|
|
59
60
|
let generation = new index_js_1.GenerationChunk({
|
|
@@ -128,6 +129,7 @@ class BaseLLM extends index_js_2.BaseLanguageModel {
|
|
|
128
129
|
const extra = {
|
|
129
130
|
options: parsedOptions,
|
|
130
131
|
invocation_params: this?.invocationParams(parsedOptions),
|
|
132
|
+
batch_size: prompts.length,
|
|
131
133
|
};
|
|
132
134
|
const runManagers = await callbackManager_?.handleLLMStart(this.toJSON(), prompts, undefined, undefined, extra, undefined, undefined, handledOptions?.runName);
|
|
133
135
|
let output;
|
package/dist/llms/base.d.ts
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { BaseMessage, BasePromptValue, GenerationChunk, LLMResult } from "../schema/index.js";
|
|
2
|
-
import { BaseLanguageModel, BaseLanguageModelCallOptions, BaseLanguageModelInput, BaseLanguageModelParams } from "../base_language/index.js";
|
|
3
2
|
import { BaseCallbackConfig, CallbackManagerForLLMRun, Callbacks } from "../callbacks/manager.js";
|
|
3
|
+
import { BaseLanguageModel, BaseLanguageModelCallOptions, BaseLanguageModelInput, BaseLanguageModelParams } from "../base_language/index.js";
|
|
4
4
|
import { RunnableConfig } from "../schema/runnable/config.js";
|
|
5
5
|
export type SerializedLLM = {
|
|
6
6
|
_model: string;
|
package/dist/llms/base.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { AIMessage, GenerationChunk, RUN_KEY, } from "../schema/index.js";
|
|
2
|
-
import { BaseLanguageModel, } from "../base_language/index.js";
|
|
3
2
|
import { CallbackManager, } from "../callbacks/manager.js";
|
|
3
|
+
import { BaseLanguageModel, } from "../base_language/index.js";
|
|
4
4
|
import { getBufferString } from "../memory/base.js";
|
|
5
5
|
/**
|
|
6
6
|
* LLM Wrapper. Provides an {@link call} (an {@link generate}) function that takes in a prompt (or prompts) and returns a string.
|
|
@@ -51,6 +51,7 @@ export class BaseLLM extends BaseLanguageModel {
|
|
|
51
51
|
const extra = {
|
|
52
52
|
options: callOptions,
|
|
53
53
|
invocation_params: this?.invocationParams(callOptions),
|
|
54
|
+
batch_size: 1,
|
|
54
55
|
};
|
|
55
56
|
const runManagers = await callbackManager_?.handleLLMStart(this.toJSON(), [prompt.toString()], undefined, undefined, extra, undefined, undefined, runnableConfig.runName);
|
|
56
57
|
let generation = new GenerationChunk({
|
|
@@ -125,6 +126,7 @@ export class BaseLLM extends BaseLanguageModel {
|
|
|
125
126
|
const extra = {
|
|
126
127
|
options: parsedOptions,
|
|
127
128
|
invocation_params: this?.invocationParams(parsedOptions),
|
|
129
|
+
batch_size: prompts.length,
|
|
128
130
|
};
|
|
129
131
|
const runManagers = await callbackManager_?.handleLLMStart(this.toJSON(), prompts, undefined, undefined, extra, undefined, undefined, handledOptions?.runName);
|
|
130
132
|
let output;
|
package/dist/llms/cohere.cjs
CHANGED
|
@@ -6,6 +6,19 @@ const base_js_1 = require("./base.cjs");
|
|
|
6
6
|
/**
|
|
7
7
|
* Class representing a Cohere Large Language Model (LLM). It interacts
|
|
8
8
|
* with the Cohere API to generate text completions.
|
|
9
|
+
* @example
|
|
10
|
+
* ```typescript
|
|
11
|
+
* const model = new Cohere({
|
|
12
|
+
* temperature: 0.7,
|
|
13
|
+
* maxTokens: 20,
|
|
14
|
+
* maxRetries: 5,
|
|
15
|
+
* });
|
|
16
|
+
*
|
|
17
|
+
* const res = await model.call(
|
|
18
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
19
|
+
* );
|
|
20
|
+
* console.log({ res });
|
|
21
|
+
* ```
|
|
9
22
|
*/
|
|
10
23
|
class Cohere extends base_js_1.LLM {
|
|
11
24
|
static lc_name() {
|
package/dist/llms/cohere.d.ts
CHANGED
|
@@ -16,6 +16,19 @@ export interface CohereInput extends BaseLLMParams {
|
|
|
16
16
|
/**
|
|
17
17
|
* Class representing a Cohere Large Language Model (LLM). It interacts
|
|
18
18
|
* with the Cohere API to generate text completions.
|
|
19
|
+
* @example
|
|
20
|
+
* ```typescript
|
|
21
|
+
* const model = new Cohere({
|
|
22
|
+
* temperature: 0.7,
|
|
23
|
+
* maxTokens: 20,
|
|
24
|
+
* maxRetries: 5,
|
|
25
|
+
* });
|
|
26
|
+
*
|
|
27
|
+
* const res = await model.call(
|
|
28
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
29
|
+
* );
|
|
30
|
+
* console.log({ res });
|
|
31
|
+
* ```
|
|
19
32
|
*/
|
|
20
33
|
export declare class Cohere extends LLM implements CohereInput {
|
|
21
34
|
static lc_name(): string;
|
package/dist/llms/cohere.js
CHANGED
|
@@ -3,6 +3,19 @@ import { LLM } from "./base.js";
|
|
|
3
3
|
/**
|
|
4
4
|
* Class representing a Cohere Large Language Model (LLM). It interacts
|
|
5
5
|
* with the Cohere API to generate text completions.
|
|
6
|
+
* @example
|
|
7
|
+
* ```typescript
|
|
8
|
+
* const model = new Cohere({
|
|
9
|
+
* temperature: 0.7,
|
|
10
|
+
* maxTokens: 20,
|
|
11
|
+
* maxRetries: 5,
|
|
12
|
+
* });
|
|
13
|
+
*
|
|
14
|
+
* const res = await model.call(
|
|
15
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
16
|
+
* );
|
|
17
|
+
* console.log({ res });
|
|
18
|
+
* ```
|
|
6
19
|
*/
|
|
7
20
|
export class Cohere extends LLM {
|
|
8
21
|
static lc_name() {
|
|
@@ -17,6 +17,18 @@ const googlevertexai_gauth_js_1 = require("../../util/googlevertexai-gauth.cjs")
|
|
|
17
17
|
* - The `GOOGLE_APPLICATION_CREDENTIALS` environment variable is set to the
|
|
18
18
|
* path of a credentials file for a service account permitted to the
|
|
19
19
|
* Google Cloud project using Vertex AI.
|
|
20
|
+
* @example
|
|
21
|
+
* ```typescript
|
|
22
|
+
* const model = new GoogleVertexAI({
|
|
23
|
+
* temperature: 0.7,
|
|
24
|
+
* });
|
|
25
|
+
* const stream = await model.stream(
|
|
26
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
27
|
+
* );
|
|
28
|
+
* for await (const chunk of stream) {
|
|
29
|
+
* console.log(chunk);
|
|
30
|
+
* }
|
|
31
|
+
* ```
|
|
20
32
|
*/
|
|
21
33
|
class GoogleVertexAI extends common_js_1.BaseGoogleVertexAI {
|
|
22
34
|
static lc_name() {
|
|
@@ -19,6 +19,18 @@ export interface GoogleVertexAITextInput extends GoogleVertexAIBaseLLMInput<Goog
|
|
|
19
19
|
* - The `GOOGLE_APPLICATION_CREDENTIALS` environment variable is set to the
|
|
20
20
|
* path of a credentials file for a service account permitted to the
|
|
21
21
|
* Google Cloud project using Vertex AI.
|
|
22
|
+
* @example
|
|
23
|
+
* ```typescript
|
|
24
|
+
* const model = new GoogleVertexAI({
|
|
25
|
+
* temperature: 0.7,
|
|
26
|
+
* });
|
|
27
|
+
* const stream = await model.stream(
|
|
28
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
29
|
+
* );
|
|
30
|
+
* for await (const chunk of stream) {
|
|
31
|
+
* console.log(chunk);
|
|
32
|
+
* }
|
|
33
|
+
* ```
|
|
22
34
|
*/
|
|
23
35
|
export declare class GoogleVertexAI extends BaseGoogleVertexAI<GoogleAuthOptions> {
|
|
24
36
|
static lc_name(): string;
|
|
@@ -14,6 +14,18 @@ import { GAuthClient } from "../../util/googlevertexai-gauth.js";
|
|
|
14
14
|
* - The `GOOGLE_APPLICATION_CREDENTIALS` environment variable is set to the
|
|
15
15
|
* path of a credentials file for a service account permitted to the
|
|
16
16
|
* Google Cloud project using Vertex AI.
|
|
17
|
+
* @example
|
|
18
|
+
* ```typescript
|
|
19
|
+
* const model = new GoogleVertexAI({
|
|
20
|
+
* temperature: 0.7,
|
|
21
|
+
* });
|
|
22
|
+
* const stream = await model.stream(
|
|
23
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
24
|
+
* );
|
|
25
|
+
* for await (const chunk of stream) {
|
|
26
|
+
* console.log(chunk);
|
|
27
|
+
* }
|
|
28
|
+
* ```
|
|
17
29
|
*/
|
|
18
30
|
export class GoogleVertexAI extends BaseGoogleVertexAI {
|
|
19
31
|
static lc_name() {
|
|
@@ -12,6 +12,18 @@ const common_js_1 = require("./common.cjs");
|
|
|
12
12
|
* functions where you do not have access to the file system. It supports passing
|
|
13
13
|
* service account credentials directly as a "GOOGLE_VERTEX_AI_WEB_CREDENTIALS"
|
|
14
14
|
* environment variable or directly as "authOptions.credentials".
|
|
15
|
+
* @example
|
|
16
|
+
* ```typescript
|
|
17
|
+
* const model = new GoogleVertexAI({
|
|
18
|
+
* temperature: 0.7,
|
|
19
|
+
* });
|
|
20
|
+
* const stream = await model.stream(
|
|
21
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
22
|
+
* );
|
|
23
|
+
* for await (const chunk of stream) {
|
|
24
|
+
* console.log(chunk);
|
|
25
|
+
* }
|
|
26
|
+
* ```
|
|
15
27
|
*/
|
|
16
28
|
class GoogleVertexAI extends common_js_1.BaseGoogleVertexAI {
|
|
17
29
|
static lc_name() {
|
|
@@ -14,6 +14,18 @@ export interface GoogleVertexAITextInput extends GoogleVertexAIBaseLLMInput<WebG
|
|
|
14
14
|
* functions where you do not have access to the file system. It supports passing
|
|
15
15
|
* service account credentials directly as a "GOOGLE_VERTEX_AI_WEB_CREDENTIALS"
|
|
16
16
|
* environment variable or directly as "authOptions.credentials".
|
|
17
|
+
* @example
|
|
18
|
+
* ```typescript
|
|
19
|
+
* const model = new GoogleVertexAI({
|
|
20
|
+
* temperature: 0.7,
|
|
21
|
+
* });
|
|
22
|
+
* const stream = await model.stream(
|
|
23
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
24
|
+
* );
|
|
25
|
+
* for await (const chunk of stream) {
|
|
26
|
+
* console.log(chunk);
|
|
27
|
+
* }
|
|
28
|
+
* ```
|
|
17
29
|
*/
|
|
18
30
|
export declare class GoogleVertexAI extends BaseGoogleVertexAI<WebGoogleAuthOptions> {
|
|
19
31
|
static lc_name(): string;
|
|
@@ -9,6 +9,18 @@ import { BaseGoogleVertexAI } from "./common.js";
|
|
|
9
9
|
* functions where you do not have access to the file system. It supports passing
|
|
10
10
|
* service account credentials directly as a "GOOGLE_VERTEX_AI_WEB_CREDENTIALS"
|
|
11
11
|
* environment variable or directly as "authOptions.credentials".
|
|
12
|
+
* @example
|
|
13
|
+
* ```typescript
|
|
14
|
+
* const model = new GoogleVertexAI({
|
|
15
|
+
* temperature: 0.7,
|
|
16
|
+
* });
|
|
17
|
+
* const stream = await model.stream(
|
|
18
|
+
* "What would be a good company name for a company that makes colorful socks?",
|
|
19
|
+
* );
|
|
20
|
+
* for await (const chunk of stream) {
|
|
21
|
+
* console.log(chunk);
|
|
22
|
+
* }
|
|
23
|
+
* ```
|
|
12
24
|
*/
|
|
13
25
|
export class GoogleVertexAI extends BaseGoogleVertexAI {
|
|
14
26
|
static lc_name() {
|
package/dist/llms/hf.cjs
CHANGED
|
@@ -6,6 +6,19 @@ const base_js_1 = require("./base.cjs");
|
|
|
6
6
|
/**
|
|
7
7
|
* Class implementing the Large Language Model (LLM) interface using the
|
|
8
8
|
* Hugging Face Inference API for text generation.
|
|
9
|
+
* @example
|
|
10
|
+
* ```typescript
|
|
11
|
+
* const model = new HuggingFaceInference({
|
|
12
|
+
* model: "gpt2",
|
|
13
|
+
* temperature: 0.7,
|
|
14
|
+
* maxTokens: 50,
|
|
15
|
+
* });
|
|
16
|
+
*
|
|
17
|
+
* const res = await model.call(
|
|
18
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
19
|
+
* );
|
|
20
|
+
* console.log({ res });
|
|
21
|
+
* ```
|
|
9
22
|
*/
|
|
10
23
|
class HuggingFaceInference extends base_js_1.LLM {
|
|
11
24
|
get lc_secrets() {
|
package/dist/llms/hf.d.ts
CHANGED
|
@@ -26,6 +26,19 @@ export interface HFInput {
|
|
|
26
26
|
/**
|
|
27
27
|
* Class implementing the Large Language Model (LLM) interface using the
|
|
28
28
|
* Hugging Face Inference API for text generation.
|
|
29
|
+
* @example
|
|
30
|
+
* ```typescript
|
|
31
|
+
* const model = new HuggingFaceInference({
|
|
32
|
+
* model: "gpt2",
|
|
33
|
+
* temperature: 0.7,
|
|
34
|
+
* maxTokens: 50,
|
|
35
|
+
* });
|
|
36
|
+
*
|
|
37
|
+
* const res = await model.call(
|
|
38
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
39
|
+
* );
|
|
40
|
+
* console.log({ res });
|
|
41
|
+
* ```
|
|
29
42
|
*/
|
|
30
43
|
export declare class HuggingFaceInference extends LLM implements HFInput {
|
|
31
44
|
get lc_secrets(): {
|
package/dist/llms/hf.js
CHANGED
|
@@ -3,6 +3,19 @@ import { LLM } from "./base.js";
|
|
|
3
3
|
/**
|
|
4
4
|
* Class implementing the Large Language Model (LLM) interface using the
|
|
5
5
|
* Hugging Face Inference API for text generation.
|
|
6
|
+
* @example
|
|
7
|
+
* ```typescript
|
|
8
|
+
* const model = new HuggingFaceInference({
|
|
9
|
+
* model: "gpt2",
|
|
10
|
+
* temperature: 0.7,
|
|
11
|
+
* maxTokens: 50,
|
|
12
|
+
* });
|
|
13
|
+
*
|
|
14
|
+
* const res = await model.call(
|
|
15
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
16
|
+
* );
|
|
17
|
+
* console.log({ res });
|
|
18
|
+
* ```
|
|
6
19
|
*/
|
|
7
20
|
export class HuggingFaceInference extends LLM {
|
|
8
21
|
get lc_secrets() {
|
package/dist/llms/ollama.cjs
CHANGED
|
@@ -7,6 +7,25 @@ const index_js_1 = require("../schema/index.cjs");
|
|
|
7
7
|
/**
|
|
8
8
|
* Class that represents the Ollama language model. It extends the base
|
|
9
9
|
* LLM class and implements the OllamaInput interface.
|
|
10
|
+
* @example
|
|
11
|
+
* ```typescript
|
|
12
|
+
* const ollama = new Ollama({
|
|
13
|
+
* baseUrl: "http://api.example.com",
|
|
14
|
+
* model: "llama2",
|
|
15
|
+
* });
|
|
16
|
+
*
|
|
17
|
+
* // Streaming translation from English to German
|
|
18
|
+
* const stream = await ollama.stream(
|
|
19
|
+
* `Translate "I love programming" into German.`
|
|
20
|
+
* );
|
|
21
|
+
*
|
|
22
|
+
* const chunks = [];
|
|
23
|
+
* for await (const chunk of stream) {
|
|
24
|
+
* chunks.push(chunk);
|
|
25
|
+
* }
|
|
26
|
+
*
|
|
27
|
+
* console.log(chunks.join(""));
|
|
28
|
+
* ```
|
|
10
29
|
*/
|
|
11
30
|
class Ollama extends base_js_1.LLM {
|
|
12
31
|
static lc_name() {
|
package/dist/llms/ollama.d.ts
CHANGED
|
@@ -6,6 +6,25 @@ import type { StringWithAutocomplete } from "../util/types.js";
|
|
|
6
6
|
/**
|
|
7
7
|
* Class that represents the Ollama language model. It extends the base
|
|
8
8
|
* LLM class and implements the OllamaInput interface.
|
|
9
|
+
* @example
|
|
10
|
+
* ```typescript
|
|
11
|
+
* const ollama = new Ollama({
|
|
12
|
+
* baseUrl: "http://api.example.com",
|
|
13
|
+
* model: "llama2",
|
|
14
|
+
* });
|
|
15
|
+
*
|
|
16
|
+
* // Streaming translation from English to German
|
|
17
|
+
* const stream = await ollama.stream(
|
|
18
|
+
* `Translate "I love programming" into German.`
|
|
19
|
+
* );
|
|
20
|
+
*
|
|
21
|
+
* const chunks = [];
|
|
22
|
+
* for await (const chunk of stream) {
|
|
23
|
+
* chunks.push(chunk);
|
|
24
|
+
* }
|
|
25
|
+
*
|
|
26
|
+
* console.log(chunks.join(""));
|
|
27
|
+
* ```
|
|
9
28
|
*/
|
|
10
29
|
export declare class Ollama extends LLM<OllamaCallOptions> implements OllamaInput {
|
|
11
30
|
static lc_name(): string;
|
package/dist/llms/ollama.js
CHANGED
|
@@ -4,6 +4,25 @@ import { GenerationChunk } from "../schema/index.js";
|
|
|
4
4
|
/**
|
|
5
5
|
* Class that represents the Ollama language model. It extends the base
|
|
6
6
|
* LLM class and implements the OllamaInput interface.
|
|
7
|
+
* @example
|
|
8
|
+
* ```typescript
|
|
9
|
+
* const ollama = new Ollama({
|
|
10
|
+
* baseUrl: "http://api.example.com",
|
|
11
|
+
* model: "llama2",
|
|
12
|
+
* });
|
|
13
|
+
*
|
|
14
|
+
* // Streaming translation from English to German
|
|
15
|
+
* const stream = await ollama.stream(
|
|
16
|
+
* `Translate "I love programming" into German.`
|
|
17
|
+
* );
|
|
18
|
+
*
|
|
19
|
+
* const chunks = [];
|
|
20
|
+
* for await (const chunk of stream) {
|
|
21
|
+
* chunks.push(chunk);
|
|
22
|
+
* }
|
|
23
|
+
*
|
|
24
|
+
* console.log(chunks.join(""));
|
|
25
|
+
* ```
|
|
7
26
|
*/
|
|
8
27
|
export class Ollama extends LLM {
|
|
9
28
|
static lc_name() {
|
|
@@ -29,6 +29,23 @@ const openai_js_1 = require("../util/openai.cjs");
|
|
|
29
29
|
* @augments BaseLLM
|
|
30
30
|
* @augments OpenAIInput
|
|
31
31
|
* @augments AzureOpenAIChatInput
|
|
32
|
+
* @example
|
|
33
|
+
* ```typescript
|
|
34
|
+
* const model = new OpenAIChat({
|
|
35
|
+
* prefixMessages: [
|
|
36
|
+
* {
|
|
37
|
+
* role: "system",
|
|
38
|
+
* content: "You are a helpful assistant that answers in pirate language",
|
|
39
|
+
* },
|
|
40
|
+
* ],
|
|
41
|
+
* maxTokens: 50,
|
|
42
|
+
* });
|
|
43
|
+
*
|
|
44
|
+
* const res = await model.call(
|
|
45
|
+
* "What would be a good company name for a company that makes colorful socks?"
|
|
46
|
+
* );
|
|
47
|
+
* console.log({ res });
|
|
48
|
+
* ```
|
|
32
49
|
*/
|
|
33
50
|
class OpenAIChat extends base_js_1.LLM {
|
|
34
51
|
static lc_name() {
|
|
@@ -33,6 +33,23 @@ export interface OpenAIChatCallOptions extends OpenAICallOptions {
|
|
|
33
33
|
* @augments BaseLLM
|
|
34
34
|
* @augments OpenAIInput
|
|
35
35
|
* @augments AzureOpenAIChatInput
|
|
36
|
+
* @example
|
|
37
|
+
* ```typescript
|
|
38
|
+
* const model = new OpenAIChat({
|
|
39
|
+
* prefixMessages: [
|
|
40
|
+
* {
|
|
41
|
+
* role: "system",
|
|
42
|
+
* content: "You are a helpful assistant that answers in pirate language",
|
|
43
|
+
* },
|
|
44
|
+
* ],
|
|
45
|
+
* maxTokens: 50,
|
|
46
|
+
* });
|
|
47
|
+
*
|
|
48
|
+
* const res = await model.call(
|
|
49
|
+
* "What would be a good company name for a company that makes colorful socks?"
|
|
50
|
+
* );
|
|
51
|
+
* console.log({ res });
|
|
52
|
+
* ```
|
|
36
53
|
*/
|
|
37
54
|
export declare class OpenAIChat extends LLM<OpenAIChatCallOptions> implements OpenAIChatInput, AzureOpenAIInput {
|
|
38
55
|
static lc_name(): string;
|
package/dist/llms/openai-chat.js
CHANGED
|
@@ -26,6 +26,23 @@ import { wrapOpenAIClientError } from "../util/openai.js";
|
|
|
26
26
|
* @augments BaseLLM
|
|
27
27
|
* @augments OpenAIInput
|
|
28
28
|
* @augments AzureOpenAIChatInput
|
|
29
|
+
* @example
|
|
30
|
+
* ```typescript
|
|
31
|
+
* const model = new OpenAIChat({
|
|
32
|
+
* prefixMessages: [
|
|
33
|
+
* {
|
|
34
|
+
* role: "system",
|
|
35
|
+
* content: "You are a helpful assistant that answers in pirate language",
|
|
36
|
+
* },
|
|
37
|
+
* ],
|
|
38
|
+
* maxTokens: 50,
|
|
39
|
+
* });
|
|
40
|
+
*
|
|
41
|
+
* const res = await model.call(
|
|
42
|
+
* "What would be a good company name for a company that makes colorful socks?"
|
|
43
|
+
* );
|
|
44
|
+
* console.log({ res });
|
|
45
|
+
* ```
|
|
29
46
|
*/
|
|
30
47
|
export class OpenAIChat extends LLM {
|
|
31
48
|
static lc_name() {
|
package/dist/llms/openai.cjs
CHANGED
|
@@ -28,6 +28,20 @@ const openai_js_1 = require("../util/openai.cjs");
|
|
|
28
28
|
* https://platform.openai.com/docs/api-reference/completions/create |
|
|
29
29
|
* `openai.createCompletion`} can be passed through {@link modelKwargs}, even
|
|
30
30
|
* if not explicitly available on this class.
|
|
31
|
+
* @example
|
|
32
|
+
* ```typescript
|
|
33
|
+
* const model = new OpenAI({
|
|
34
|
+
* modelName: "gpt-4",
|
|
35
|
+
* temperature: 0.7,
|
|
36
|
+
* maxTokens: 1000,
|
|
37
|
+
* maxRetries: 5,
|
|
38
|
+
* });
|
|
39
|
+
*
|
|
40
|
+
* const res = await model.call(
|
|
41
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
42
|
+
* );
|
|
43
|
+
* console.log({ res });
|
|
44
|
+
* ```
|
|
31
45
|
*/
|
|
32
46
|
class OpenAI extends base_js_1.BaseLLM {
|
|
33
47
|
static lc_name() {
|
package/dist/llms/openai.d.ts
CHANGED
|
@@ -21,6 +21,20 @@ export type { AzureOpenAIInput, OpenAICallOptions, OpenAIInput };
|
|
|
21
21
|
* https://platform.openai.com/docs/api-reference/completions/create |
|
|
22
22
|
* `openai.createCompletion`} can be passed through {@link modelKwargs}, even
|
|
23
23
|
* if not explicitly available on this class.
|
|
24
|
+
* @example
|
|
25
|
+
* ```typescript
|
|
26
|
+
* const model = new OpenAI({
|
|
27
|
+
* modelName: "gpt-4",
|
|
28
|
+
* temperature: 0.7,
|
|
29
|
+
* maxTokens: 1000,
|
|
30
|
+
* maxRetries: 5,
|
|
31
|
+
* });
|
|
32
|
+
*
|
|
33
|
+
* const res = await model.call(
|
|
34
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
35
|
+
* );
|
|
36
|
+
* console.log({ res });
|
|
37
|
+
* ```
|
|
24
38
|
*/
|
|
25
39
|
export declare class OpenAI<CallOptions extends OpenAICallOptions = OpenAICallOptions> extends BaseLLM<CallOptions> implements OpenAIInput, AzureOpenAIInput {
|
|
26
40
|
static lc_name(): string;
|
package/dist/llms/openai.js
CHANGED
|
@@ -25,6 +25,20 @@ import { wrapOpenAIClientError } from "../util/openai.js";
|
|
|
25
25
|
* https://platform.openai.com/docs/api-reference/completions/create |
|
|
26
26
|
* `openai.createCompletion`} can be passed through {@link modelKwargs}, even
|
|
27
27
|
* if not explicitly available on this class.
|
|
28
|
+
* @example
|
|
29
|
+
* ```typescript
|
|
30
|
+
* const model = new OpenAI({
|
|
31
|
+
* modelName: "gpt-4",
|
|
32
|
+
* temperature: 0.7,
|
|
33
|
+
* maxTokens: 1000,
|
|
34
|
+
* maxRetries: 5,
|
|
35
|
+
* });
|
|
36
|
+
*
|
|
37
|
+
* const res = await model.call(
|
|
38
|
+
* "Question: What would be a good company name for a company that makes colorful socks?\nAnswer:"
|
|
39
|
+
* );
|
|
40
|
+
* console.log({ res });
|
|
41
|
+
* ```
|
|
28
42
|
*/
|
|
29
43
|
export class OpenAI extends BaseLLM {
|
|
30
44
|
static lc_name() {
|
package/dist/llms/portkey.cjs
CHANGED
|
@@ -51,6 +51,30 @@ function getPortkeySession(options = {}) {
|
|
|
51
51
|
return session;
|
|
52
52
|
}
|
|
53
53
|
exports.getPortkeySession = getPortkeySession;
|
|
54
|
+
/**
|
|
55
|
+
* @example
|
|
56
|
+
* ```typescript
|
|
57
|
+
* const model = new Portkey({
|
|
58
|
+
* mode: "single",
|
|
59
|
+
* llms: [
|
|
60
|
+
* {
|
|
61
|
+
* provider: "openai",
|
|
62
|
+
* virtual_key: "open-ai-key-1234",
|
|
63
|
+
* model: "text-davinci-003",
|
|
64
|
+
* max_tokens: 2000,
|
|
65
|
+
* },
|
|
66
|
+
* ],
|
|
67
|
+
* });
|
|
68
|
+
*
|
|
69
|
+
* // Stream the output of the model and process it
|
|
70
|
+
* const res = await model.stream(
|
|
71
|
+
* "Question: Write a story about a king\nAnswer:"
|
|
72
|
+
* );
|
|
73
|
+
* for await (const i of res) {
|
|
74
|
+
* process.stdout.write(i);
|
|
75
|
+
* }
|
|
76
|
+
* ```
|
|
77
|
+
*/
|
|
54
78
|
class Portkey extends base_js_1.BaseLLM {
|
|
55
79
|
constructor(init) {
|
|
56
80
|
super(init ?? {});
|
package/dist/llms/portkey.d.ts
CHANGED
|
@@ -19,6 +19,30 @@ export declare class PortkeySession {
|
|
|
19
19
|
* @returns
|
|
20
20
|
*/
|
|
21
21
|
export declare function getPortkeySession(options?: PortkeyOptions): PortkeySession;
|
|
22
|
+
/**
|
|
23
|
+
* @example
|
|
24
|
+
* ```typescript
|
|
25
|
+
* const model = new Portkey({
|
|
26
|
+
* mode: "single",
|
|
27
|
+
* llms: [
|
|
28
|
+
* {
|
|
29
|
+
* provider: "openai",
|
|
30
|
+
* virtual_key: "open-ai-key-1234",
|
|
31
|
+
* model: "text-davinci-003",
|
|
32
|
+
* max_tokens: 2000,
|
|
33
|
+
* },
|
|
34
|
+
* ],
|
|
35
|
+
* });
|
|
36
|
+
*
|
|
37
|
+
* // Stream the output of the model and process it
|
|
38
|
+
* const res = await model.stream(
|
|
39
|
+
* "Question: Write a story about a king\nAnswer:"
|
|
40
|
+
* );
|
|
41
|
+
* for await (const i of res) {
|
|
42
|
+
* process.stdout.write(i);
|
|
43
|
+
* }
|
|
44
|
+
* ```
|
|
45
|
+
*/
|
|
22
46
|
export declare class Portkey extends BaseLLM {
|
|
23
47
|
apiKey?: string;
|
|
24
48
|
baseURL?: string;
|
package/dist/llms/portkey.js
CHANGED
|
@@ -43,6 +43,30 @@ export function getPortkeySession(options = {}) {
|
|
|
43
43
|
}
|
|
44
44
|
return session;
|
|
45
45
|
}
|
|
46
|
+
/**
|
|
47
|
+
* @example
|
|
48
|
+
* ```typescript
|
|
49
|
+
* const model = new Portkey({
|
|
50
|
+
* mode: "single",
|
|
51
|
+
* llms: [
|
|
52
|
+
* {
|
|
53
|
+
* provider: "openai",
|
|
54
|
+
* virtual_key: "open-ai-key-1234",
|
|
55
|
+
* model: "text-davinci-003",
|
|
56
|
+
* max_tokens: 2000,
|
|
57
|
+
* },
|
|
58
|
+
* ],
|
|
59
|
+
* });
|
|
60
|
+
*
|
|
61
|
+
* // Stream the output of the model and process it
|
|
62
|
+
* const res = await model.stream(
|
|
63
|
+
* "Question: Write a story about a king\nAnswer:"
|
|
64
|
+
* );
|
|
65
|
+
* for await (const i of res) {
|
|
66
|
+
* process.stdout.write(i);
|
|
67
|
+
* }
|
|
68
|
+
* ```
|
|
69
|
+
*/
|
|
46
70
|
export class Portkey extends BaseLLM {
|
|
47
71
|
constructor(init) {
|
|
48
72
|
super(init ?? {});
|