beeai-framework 0.1.19 → 0.1.21
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapters/a2a/agents/agent.cjs +4 -4
- package/dist/adapters/a2a/agents/agent.cjs.map +1 -1
- package/dist/adapters/a2a/agents/agent.d.cts +2 -2
- package/dist/adapters/a2a/agents/agent.d.ts +2 -2
- package/dist/adapters/a2a/agents/agent.js +4 -4
- package/dist/adapters/a2a/agents/agent.js.map +1 -1
- package/dist/adapters/a2a/agents/events.d.cts +1 -1
- package/dist/adapters/a2a/agents/events.d.ts +1 -1
- package/dist/adapters/a2a/agents/types.d.cts +2 -2
- package/dist/adapters/a2a/agents/types.d.ts +2 -2
- package/dist/adapters/a2a/agents/utils.cjs +2 -2
- package/dist/adapters/a2a/agents/utils.cjs.map +1 -1
- package/dist/adapters/a2a/agents/utils.js +2 -2
- package/dist/adapters/a2a/agents/utils.js.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.cjs +1 -3
- package/dist/adapters/a2a/serve/agent_executor.cjs.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.d.cts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.d.ts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.js +1 -3
- package/dist/adapters/a2a/serve/agent_executor.js.map +1 -1
- package/dist/adapters/a2a/serve/server.cjs +15 -0
- package/dist/adapters/a2a/serve/server.cjs.map +1 -1
- package/dist/adapters/a2a/serve/server.d.cts +4 -2
- package/dist/adapters/a2a/serve/server.d.ts +4 -2
- package/dist/adapters/a2a/serve/server.js +15 -0
- package/dist/adapters/a2a/serve/server.js.map +1 -1
- package/dist/adapters/acp/agents/agent.d.cts +2 -2
- package/dist/adapters/acp/agents/agent.d.ts +2 -2
- package/dist/adapters/acp/agents/events.d.cts +1 -1
- package/dist/adapters/acp/agents/events.d.ts +1 -1
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.cjs +4 -4
- package/dist/adapters/agentstack/agents/agent.cjs.map +1 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.cts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.ts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.js +4 -4
- package/dist/adapters/agentstack/agents/agent.js.map +1 -0
- package/dist/adapters/agentstack/agents/events.d.cts +23 -0
- package/dist/adapters/agentstack/agents/events.d.ts +23 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.cts +4 -4
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/chat.cjs +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/chat.d.cts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.d.ts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.js +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.js.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.d.cts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.js +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.js.map +1 -1
- package/dist/adapters/anthropic/backend/chat.cjs +3 -2
- package/dist/adapters/anthropic/backend/chat.cjs.map +1 -1
- package/dist/adapters/anthropic/backend/chat.d.cts +5 -6
- package/dist/adapters/anthropic/backend/chat.d.ts +5 -6
- package/dist/adapters/anthropic/backend/chat.js +3 -2
- package/dist/adapters/anthropic/backend/chat.js.map +1 -1
- package/dist/adapters/anthropic/backend/embedding.d.cts +2 -2
- package/dist/adapters/anthropic/backend/embedding.d.ts +2 -2
- package/dist/adapters/azure-openai/backend/chat.cjs +3 -2
- package/dist/adapters/azure-openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/chat.d.cts +5 -6
- package/dist/adapters/azure-openai/backend/chat.d.ts +5 -6
- package/dist/adapters/azure-openai/backend/chat.js +3 -2
- package/dist/adapters/azure-openai/backend/chat.js.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.cjs +2 -2
- package/dist/adapters/azure-openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.d.cts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.d.ts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.js +2 -2
- package/dist/adapters/azure-openai/backend/embedding.js.map +1 -1
- package/dist/adapters/dummy/backend/chat.d.cts +2 -2
- package/dist/adapters/dummy/backend/chat.d.ts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.cts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.ts +2 -2
- package/dist/adapters/google-vertex/backend/chat.cjs +3 -2
- package/dist/adapters/google-vertex/backend/chat.cjs.map +1 -1
- package/dist/adapters/google-vertex/backend/chat.d.cts +5 -6
- package/dist/adapters/google-vertex/backend/chat.d.ts +5 -6
- package/dist/adapters/google-vertex/backend/chat.js +3 -2
- package/dist/adapters/google-vertex/backend/chat.js.map +1 -1
- package/dist/adapters/google-vertex/backend/embedding.d.cts +2 -2
- package/dist/adapters/google-vertex/backend/embedding.d.ts +2 -2
- package/dist/adapters/groq/backend/chat.cjs +3 -2
- package/dist/adapters/groq/backend/chat.cjs.map +1 -1
- package/dist/adapters/groq/backend/chat.d.cts +5 -6
- package/dist/adapters/groq/backend/chat.d.ts +5 -6
- package/dist/adapters/groq/backend/chat.js +3 -2
- package/dist/adapters/groq/backend/chat.js.map +1 -1
- package/dist/adapters/groq/backend/embedding.d.cts +2 -2
- package/dist/adapters/groq/backend/embedding.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.cjs +6 -5
- package/dist/adapters/langchain/backend/chat.cjs.map +1 -1
- package/dist/adapters/langchain/backend/chat.d.cts +2 -2
- package/dist/adapters/langchain/backend/chat.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.js +6 -5
- package/dist/adapters/langchain/backend/chat.js.map +1 -1
- package/dist/adapters/langchain/backend/embedding.d.cts +3 -3
- package/dist/adapters/langchain/backend/embedding.d.ts +3 -3
- package/dist/adapters/langchain/tools.d.cts +1 -1
- package/dist/adapters/langchain/tools.d.ts +1 -1
- package/dist/adapters/mcp/serve/server.cjs +12 -4
- package/dist/adapters/mcp/serve/server.cjs.map +1 -1
- package/dist/adapters/mcp/serve/server.d.cts +1 -1
- package/dist/adapters/mcp/serve/server.d.ts +1 -1
- package/dist/adapters/mcp/serve/server.js +12 -4
- package/dist/adapters/mcp/serve/server.js.map +1 -1
- package/dist/adapters/ollama/backend/chat.cjs +1 -4
- package/dist/adapters/ollama/backend/chat.cjs.map +1 -1
- package/dist/adapters/ollama/backend/chat.d.cts +4 -4
- package/dist/adapters/ollama/backend/chat.d.ts +4 -4
- package/dist/adapters/ollama/backend/chat.js +1 -4
- package/dist/adapters/ollama/backend/chat.js.map +1 -1
- package/dist/adapters/ollama/backend/client.cjs +2 -2
- package/dist/adapters/ollama/backend/client.cjs.map +1 -1
- package/dist/adapters/ollama/backend/client.d.cts +1 -1
- package/dist/adapters/ollama/backend/client.d.ts +1 -1
- package/dist/adapters/ollama/backend/client.js +1 -1
- package/dist/adapters/ollama/backend/client.js.map +1 -1
- package/dist/adapters/ollama/backend/embedding.cjs.map +1 -1
- package/dist/adapters/ollama/backend/embedding.d.cts +3 -3
- package/dist/adapters/ollama/backend/embedding.d.ts +3 -3
- package/dist/adapters/ollama/backend/embedding.js.map +1 -1
- package/dist/adapters/openai/backend/chat.cjs +3 -2
- package/dist/adapters/openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/openai/backend/chat.d.cts +5 -6
- package/dist/adapters/openai/backend/chat.d.ts +5 -6
- package/dist/adapters/openai/backend/chat.js +3 -2
- package/dist/adapters/openai/backend/chat.js.map +1 -1
- package/dist/adapters/openai/backend/client.cjs +2 -11
- package/dist/adapters/openai/backend/client.cjs.map +1 -1
- package/dist/adapters/openai/backend/client.js +3 -12
- package/dist/adapters/openai/backend/client.js.map +1 -1
- package/dist/adapters/openai/backend/embedding.cjs +2 -2
- package/dist/adapters/openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/openai/backend/embedding.d.cts +4 -4
- package/dist/adapters/openai/backend/embedding.d.ts +4 -4
- package/dist/adapters/openai/backend/embedding.js +2 -2
- package/dist/adapters/openai/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/chat.cjs +26 -21
- package/dist/adapters/vercel/backend/chat.cjs.map +1 -1
- package/dist/adapters/vercel/backend/chat.d.cts +6 -4
- package/dist/adapters/vercel/backend/chat.d.ts +6 -4
- package/dist/adapters/vercel/backend/chat.js +27 -22
- package/dist/adapters/vercel/backend/chat.js.map +1 -1
- package/dist/adapters/vercel/backend/embedding.cjs.map +1 -1
- package/dist/adapters/vercel/backend/embedding.d.cts +3 -3
- package/dist/adapters/vercel/backend/embedding.d.ts +3 -3
- package/dist/adapters/vercel/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/utils.cjs +11 -0
- package/dist/adapters/vercel/backend/utils.cjs.map +1 -1
- package/dist/adapters/vercel/backend/utils.d.cts +19 -2
- package/dist/adapters/vercel/backend/utils.d.ts +19 -2
- package/dist/adapters/vercel/backend/utils.js +11 -1
- package/dist/adapters/vercel/backend/utils.js.map +1 -1
- package/dist/adapters/watsonx/backend/chat.cjs +10 -4
- package/dist/adapters/watsonx/backend/chat.cjs.map +1 -1
- package/dist/adapters/watsonx/backend/chat.d.cts +2 -2
- package/dist/adapters/watsonx/backend/chat.d.ts +2 -2
- package/dist/adapters/watsonx/backend/chat.js +10 -4
- package/dist/adapters/watsonx/backend/chat.js.map +1 -1
- package/dist/adapters/watsonx/backend/embedding.d.cts +2 -2
- package/dist/adapters/watsonx/backend/embedding.d.ts +2 -2
- package/dist/adapters/xai/backend/chat.cjs +3 -2
- package/dist/adapters/xai/backend/chat.cjs.map +1 -1
- package/dist/adapters/xai/backend/chat.d.cts +5 -6
- package/dist/adapters/xai/backend/chat.d.ts +5 -6
- package/dist/adapters/xai/backend/chat.js +3 -2
- package/dist/adapters/xai/backend/chat.js.map +1 -1
- package/dist/{agent-7NzfPC6d.d.ts → agent-CHebzsQB.d.ts} +4 -4
- package/dist/{agent-B6pOfx0K.d.cts → agent-DvvVGvLr.d.cts} +4 -4
- package/dist/agents/base.d.cts +2 -2
- package/dist/agents/base.d.ts +2 -2
- package/dist/agents/experimental/replan/agent.d.cts +3 -3
- package/dist/agents/experimental/replan/agent.d.ts +3 -3
- package/dist/agents/experimental/replan/prompts.d.cts +1 -1
- package/dist/agents/experimental/replan/prompts.d.ts +1 -1
- package/dist/agents/experimental/streamlit/agent.d.cts +3 -3
- package/dist/agents/experimental/streamlit/agent.d.ts +3 -3
- package/dist/agents/react/agent.d.cts +5 -4
- package/dist/agents/react/agent.d.ts +5 -4
- package/dist/agents/react/prompts.d.cts +1 -1
- package/dist/agents/react/prompts.d.ts +1 -1
- package/dist/agents/react/runners/base.cjs.map +1 -1
- package/dist/agents/react/runners/base.d.cts +5 -4
- package/dist/agents/react/runners/base.d.ts +5 -4
- package/dist/agents/react/runners/base.js.map +1 -1
- package/dist/agents/react/runners/deep-think/prompts.d.cts +1 -1
- package/dist/agents/react/runners/deep-think/prompts.d.ts +1 -1
- package/dist/agents/react/runners/deep-think/runner.d.cts +6 -5
- package/dist/agents/react/runners/deep-think/runner.d.ts +6 -5
- package/dist/agents/react/runners/default/runner.cjs.map +1 -1
- package/dist/agents/react/runners/default/runner.d.cts +6 -5
- package/dist/agents/react/runners/default/runner.d.ts +6 -5
- package/dist/agents/react/runners/default/runner.js.map +1 -1
- package/dist/agents/react/runners/granite/prompts.d.cts +1 -1
- package/dist/agents/react/runners/granite/prompts.d.ts +1 -1
- package/dist/agents/react/runners/granite/runner.cjs +4 -2
- package/dist/agents/react/runners/granite/runner.cjs.map +1 -1
- package/dist/agents/react/runners/granite/runner.d.cts +6 -5
- package/dist/agents/react/runners/granite/runner.d.ts +6 -5
- package/dist/agents/react/runners/granite/runner.js +4 -2
- package/dist/agents/react/runners/granite/runner.js.map +1 -1
- package/dist/agents/react/types.d.cts +3 -2
- package/dist/agents/react/types.d.ts +3 -2
- package/dist/agents/toolCalling/agent.cjs +9 -5
- package/dist/agents/toolCalling/agent.cjs.map +1 -1
- package/dist/agents/toolCalling/agent.d.cts +3 -3
- package/dist/agents/toolCalling/agent.d.ts +3 -3
- package/dist/agents/toolCalling/agent.js +9 -5
- package/dist/agents/toolCalling/agent.js.map +1 -1
- package/dist/agents/toolCalling/types.d.cts +1 -1
- package/dist/agents/toolCalling/types.d.ts +1 -1
- package/dist/agents/types.d.cts +2 -2
- package/dist/agents/types.d.ts +2 -2
- package/dist/backend/backend.d.cts +2 -2
- package/dist/backend/backend.d.ts +2 -2
- package/dist/backend/chat.cjs +1 -1
- package/dist/backend/chat.cjs.map +1 -1
- package/dist/backend/chat.d.cts +2 -2
- package/dist/backend/chat.d.ts +2 -2
- package/dist/backend/chat.js +1 -1
- package/dist/backend/chat.js.map +1 -1
- package/dist/backend/core.d.cts +2 -2
- package/dist/backend/core.d.ts +2 -2
- package/dist/backend/embedding.d.cts +2 -2
- package/dist/backend/embedding.d.ts +2 -2
- package/dist/backend/message.cjs +59 -3
- package/dist/backend/message.cjs.map +1 -1
- package/dist/backend/message.d.cts +3 -0
- package/dist/backend/message.d.ts +3 -0
- package/dist/backend/message.js +59 -3
- package/dist/backend/message.js.map +1 -1
- package/dist/backend/utils.d.cts +2 -2
- package/dist/backend/utils.d.ts +2 -2
- package/dist/{base-Bguef5Dg.d.ts → base-BQgDlcGa.d.ts} +1 -1
- package/dist/{base-C1hO5EJe.d.cts → base-CwcxwybB.d.cts} +1 -1
- package/dist/cache/fileCache.cjs.map +1 -1
- package/dist/cache/fileCache.js.map +1 -1
- package/dist/{chat-BBoOSvzm.d.ts → chat-C0s-o6ll.d.ts} +4 -2
- package/dist/{chat-BZ55YQab.d.cts → chat-CRb3vUVg.d.cts} +4 -2
- package/dist/context.d.cts +1 -1
- package/dist/context.d.ts +1 -1
- package/dist/emitter/emitter.d.cts +1 -1
- package/dist/emitter/emitter.d.ts +1 -1
- package/dist/emitter/types.d.cts +1 -1
- package/dist/emitter/types.d.ts +1 -1
- package/dist/{emitter-BqpLJQVb.d.ts → emitter-CZFbzlUi.d.ts} +3 -3
- package/dist/{emitter-pJzHC_AM.d.cts → emitter-DpqUYjXH.d.cts} +3 -3
- package/dist/index.d.cts +3 -3
- package/dist/index.d.ts +3 -3
- package/dist/internals/fetcher.d.cts +1 -1
- package/dist/internals/fetcher.d.ts +1 -1
- package/dist/internals/helpers/array.cjs +35 -0
- package/dist/internals/helpers/array.cjs.map +1 -1
- package/dist/internals/helpers/array.d.cts +6 -1
- package/dist/internals/helpers/array.d.ts +6 -1
- package/dist/internals/helpers/array.js +35 -1
- package/dist/internals/helpers/array.js.map +1 -1
- package/dist/internals/helpers/object.cjs +30 -5
- package/dist/internals/helpers/object.cjs.map +1 -1
- package/dist/internals/helpers/object.d.cts +3 -2
- package/dist/internals/helpers/object.d.ts +3 -2
- package/dist/internals/helpers/object.js +29 -5
- package/dist/internals/helpers/object.js.map +1 -1
- package/dist/logger/logger.cjs +8 -4
- package/dist/logger/logger.cjs.map +1 -1
- package/dist/logger/logger.d.cts +21 -15
- package/dist/logger/logger.d.ts +21 -15
- package/dist/logger/logger.js +6 -3
- package/dist/logger/logger.js.map +1 -1
- package/dist/logger/pretty.cjs +2 -2
- package/dist/logger/pretty.cjs.map +1 -1
- package/dist/logger/pretty.js +2 -2
- package/dist/logger/pretty.js.map +1 -1
- package/dist/memory/summarizeMemory.d.cts +2 -2
- package/dist/memory/summarizeMemory.d.ts +2 -2
- package/dist/parsers/linePrefix.cjs.map +1 -1
- package/dist/parsers/linePrefix.d.cts +4 -8
- package/dist/parsers/linePrefix.d.ts +4 -8
- package/dist/parsers/linePrefix.js.map +1 -1
- package/dist/serializer/utils.cjs.map +1 -1
- package/dist/serializer/utils.js.map +1 -1
- package/dist/tools/arxiv.cjs.map +1 -1
- package/dist/tools/arxiv.d.cts +1 -1
- package/dist/tools/arxiv.d.ts +1 -1
- package/dist/tools/arxiv.js.map +1 -1
- package/dist/tools/base.cjs.map +1 -1
- package/dist/tools/base.d.cts +1 -1
- package/dist/tools/base.d.ts +1 -1
- package/dist/tools/base.js.map +1 -1
- package/dist/tools/calculator.d.cts +1 -1
- package/dist/tools/calculator.d.ts +1 -1
- package/dist/tools/custom.d.cts +2 -2
- package/dist/tools/custom.d.ts +2 -2
- package/dist/tools/database/elasticsearch.cjs.map +1 -1
- package/dist/tools/database/elasticsearch.d.cts +1 -1
- package/dist/tools/database/elasticsearch.d.ts +1 -1
- package/dist/tools/database/elasticsearch.js.map +1 -1
- package/dist/tools/database/milvus.cjs +3 -3
- package/dist/tools/database/milvus.cjs.map +1 -1
- package/dist/tools/database/milvus.d.cts +1 -1
- package/dist/tools/database/milvus.d.ts +1 -1
- package/dist/tools/database/milvus.js +3 -3
- package/dist/tools/database/milvus.js.map +1 -1
- package/dist/tools/database/qdrant.cjs +2 -2
- package/dist/tools/database/qdrant.cjs.map +1 -1
- package/dist/tools/database/qdrant.d.cts +1 -1
- package/dist/tools/database/qdrant.d.ts +1 -1
- package/dist/tools/database/qdrant.js +2 -2
- package/dist/tools/database/qdrant.js.map +1 -1
- package/dist/tools/database/sql.cjs.map +1 -1
- package/dist/tools/database/sql.d.cts +1 -1
- package/dist/tools/database/sql.d.ts +1 -1
- package/dist/tools/database/sql.js.map +1 -1
- package/dist/tools/llm.d.cts +2 -2
- package/dist/tools/llm.d.ts +2 -2
- package/dist/tools/mcp.d.cts +1 -1
- package/dist/tools/mcp.d.ts +1 -1
- package/dist/tools/openapi.d.cts +1 -1
- package/dist/tools/openapi.d.ts +1 -1
- package/dist/tools/python/output.d.cts +1 -1
- package/dist/tools/python/output.d.ts +1 -1
- package/dist/tools/python/python.d.cts +2 -2
- package/dist/tools/python/python.d.ts +2 -2
- package/dist/tools/search/base.cjs.map +1 -1
- package/dist/tools/search/base.d.cts +1 -1
- package/dist/tools/search/base.d.ts +1 -1
- package/dist/tools/search/base.js.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.cjs.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.cts +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.ts +1 -1
- package/dist/tools/search/duckDuckGoSearch.js.map +1 -1
- package/dist/tools/search/googleSearch.cjs.map +1 -1
- package/dist/tools/search/googleSearch.d.cts +1 -1
- package/dist/tools/search/googleSearch.d.ts +1 -1
- package/dist/tools/search/googleSearch.js.map +1 -1
- package/dist/tools/search/searXNGSearch.d.cts +1 -1
- package/dist/tools/search/searXNGSearch.d.ts +1 -1
- package/dist/tools/search/wikipedia.cjs +25 -0
- package/dist/tools/search/wikipedia.cjs.map +1 -1
- package/dist/tools/search/wikipedia.d.cts +2 -3
- package/dist/tools/search/wikipedia.d.ts +2 -3
- package/dist/tools/search/wikipedia.js +23 -0
- package/dist/tools/search/wikipedia.js.map +1 -1
- package/dist/tools/similarity.d.cts +1 -1
- package/dist/tools/similarity.d.ts +1 -1
- package/dist/tools/weather/openMeteo.cjs.map +1 -1
- package/dist/tools/weather/openMeteo.d.cts +1 -1
- package/dist/tools/weather/openMeteo.d.ts +1 -1
- package/dist/tools/weather/openMeteo.js.map +1 -1
- package/dist/tools/web/webCrawler.cjs.map +1 -1
- package/dist/tools/web/webCrawler.d.cts +1 -1
- package/dist/tools/web/webCrawler.d.ts +1 -1
- package/dist/tools/web/webCrawler.js.map +1 -1
- package/dist/version.cjs +1 -1
- package/dist/version.js +1 -1
- package/dist/workflows/agent.d.cts +3 -3
- package/dist/workflows/agent.d.ts +3 -3
- package/dist/workflows/workflow.d.cts +1 -1
- package/dist/workflows/workflow.d.ts +1 -1
- package/package.json +68 -67
- package/dist/adapters/beeai_platform/agents/agent.cjs.map +0 -1
- package/dist/adapters/beeai_platform/agents/agent.js.map +0 -1
- package/dist/adapters/beeai_platform/agents/events.d.cts +0 -23
- package/dist/adapters/beeai_platform/agents/events.d.ts +0 -23
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js.map +0 -0
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/azure-openai/backend/embedding.ts"],"names":["AzureOpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/azure-openai/backend/embedding.ts"],"names":["AzureOpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","_settings","client","model","AzureOpenAIClient","ensure","instance","textEmbeddingModel"],"mappings":";;;;;;AAiBO,MAAMA,kCAAkCC,oBAAAA,CAAAA;EAjB/C;;;EAkBE,WAAA,CACEC,OAAAA,GAAuCC,OACrC,8BAAA,EACA,wBAAA,GAEFC,SAAAA,GAA+C,IAC/CC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,iBAAAA,CAAkBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,mBAAmBR,OAAAA,CAAAA;AAC3E,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport {\n AzureOpenAIClient,\n AzureOpenAIClientSettings,\n} from \"@/adapters/azure-openai/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { AzureOpenAIProvider as VercelAzureOpenAIProviderSettings } from \"@ai-sdk/azure\";\n\ntype AzureOpenAIParameters = Parameters<VercelAzureOpenAIProviderSettings[\"textEmbeddingModel\"]>;\nexport type AzureOpenAIEmbeddingModelId = NonNullable<AzureOpenAIParameters[0]>;\nexport type AzureOpenAIEmbeddingModelSettings = Record<string, any>;\n\nexport class AzureOpenAIEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: AzureOpenAIEmbeddingModelId = getEnv(\n \"AZURE_OPENAI_EMBEDDING_MODEL\",\n \"text-embedding-3-small\",\n ),\n _settings: AzureOpenAIEmbeddingModelSettings = {},\n client?: AzureOpenAIClient | AzureOpenAIClientSettings,\n ) {\n const model = AzureOpenAIClient.ensure(client).instance.textEmbeddingModel(modelId);\n super(model);\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { Logger } from '../../../logger/logger.cjs';
|
|
2
|
-
import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
2
|
+
import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
|
|
3
3
|
import { GetRunContext } from '../../../context.cjs';
|
|
4
|
-
import { E as Emitter } from '../../../emitter-
|
|
4
|
+
import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
|
|
5
5
|
import 'pino';
|
|
6
6
|
import '../../../errors.cjs';
|
|
7
7
|
import '../../../internals/types.cjs';
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { Logger } from '../../../logger/logger.js';
|
|
2
|
-
import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
2
|
+
import { m as ChatModel, C as ChatModelParameters, g as ChatModelEvents, d as ChatModelInput, n as ChatModelOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
|
|
3
3
|
import { GetRunContext } from '../../../context.js';
|
|
4
|
-
import { E as Emitter } from '../../../emitter-
|
|
4
|
+
import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
|
|
5
5
|
import 'pino';
|
|
6
6
|
import '../../../errors.js';
|
|
7
7
|
import '../../../internals/types.js';
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { GetRunContext } from '../../../context.cjs';
|
|
2
|
-
import { E as Emitter } from '../../../emitter-
|
|
2
|
+
import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
|
|
3
3
|
import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.cjs';
|
|
4
4
|
import '../../../internals/serializable.cjs';
|
|
5
5
|
import '../../../internals/types.cjs';
|
|
6
6
|
import '../../../internals/helpers/guards.cjs';
|
|
7
7
|
import '../../../internals/helpers/promise.cjs';
|
|
8
8
|
import '../../../errors.cjs';
|
|
9
|
-
import '../../../chat-
|
|
9
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
10
10
|
import '../../../backend/message.cjs';
|
|
11
11
|
import 'ai';
|
|
12
12
|
import 'promise-based-task';
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { GetRunContext } from '../../../context.js';
|
|
2
|
-
import { E as Emitter } from '../../../emitter-
|
|
2
|
+
import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
|
|
3
3
|
import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.js';
|
|
4
4
|
import '../../../internals/serializable.js';
|
|
5
5
|
import '../../../internals/types.js';
|
|
6
6
|
import '../../../internals/helpers/guards.js';
|
|
7
7
|
import '../../../internals/helpers/promise.js';
|
|
8
8
|
import '../../../errors.js';
|
|
9
|
-
import '../../../chat-
|
|
9
|
+
import '../../../chat-C0s-o6ll.js';
|
|
10
10
|
import '../../../backend/message.js';
|
|
11
11
|
import 'ai';
|
|
12
12
|
import 'promise-based-task';
|
|
@@ -10,9 +10,10 @@ class GoogleVertexChatModel extends chat_cjs.VercelChatModel {
|
|
|
10
10
|
static {
|
|
11
11
|
__name(this, "GoogleVertexChatModel");
|
|
12
12
|
}
|
|
13
|
-
constructor(modelId = env_cjs.getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"),
|
|
14
|
-
const model = client_cjs.GoogleVertexClient.ensure(client).instance.languageModel(modelId
|
|
13
|
+
constructor(modelId = env_cjs.getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), parameters = {}, client) {
|
|
14
|
+
const model = client_cjs.GoogleVertexClient.ensure(client).instance.languageModel(modelId);
|
|
15
15
|
super(model);
|
|
16
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
16
17
|
}
|
|
17
18
|
}
|
|
18
19
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GoogleVertexClient","ensure","instance","languageModel","Object","assign"],"mappings":";;;;;;;;AAiBO,MAAMA,8BAA8BC,wBAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,eAAO,0BAAA,EAA4B,gBAAA,GACtEC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,6BAAAA,CAAmBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AACvE,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n parameters: ChatModelParameters = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -1,17 +1,17 @@
|
|
|
1
1
|
import { GoogleVertexProvider } from '@ai-sdk/google-vertex';
|
|
2
2
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
3
3
|
import { GoogleVertexClientSettings, GoogleVertexClient } from './client.cjs';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
|
|
4
5
|
import '../../../logger/logger.cjs';
|
|
5
6
|
import 'pino';
|
|
6
7
|
import '../../../errors.cjs';
|
|
7
8
|
import '../../../internals/types.cjs';
|
|
8
9
|
import '../../../internals/helpers/guards.cjs';
|
|
9
10
|
import '../../../internals/serializable.cjs';
|
|
10
|
-
import '../../../chat-BZ55YQab.cjs';
|
|
11
|
-
import '../../../backend/message.cjs';
|
|
12
11
|
import 'ai';
|
|
12
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
13
|
+
import '../../../backend/message.cjs';
|
|
13
14
|
import '../../../context.cjs';
|
|
14
|
-
import '../../../emitter-pJzHC_AM.cjs';
|
|
15
15
|
import '../../../internals/helpers/promise.cjs';
|
|
16
16
|
import 'promise-based-task';
|
|
17
17
|
import '../../../cache/base.cjs';
|
|
@@ -31,9 +31,8 @@ import '../../../backend/client.cjs';
|
|
|
31
31
|
|
|
32
32
|
type GoogleVertexParameters = Parameters<GoogleVertexProvider["languageModel"]>;
|
|
33
33
|
type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;
|
|
34
|
-
type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;
|
|
35
34
|
declare class GoogleVertexChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: GoogleVertexChatModelId,
|
|
35
|
+
constructor(modelId?: GoogleVertexChatModelId, parameters?: ChatModelParameters, client?: GoogleVertexClientSettings | GoogleVertexClient);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { GoogleVertexChatModel, type GoogleVertexChatModelId
|
|
38
|
+
export { GoogleVertexChatModel, type GoogleVertexChatModelId };
|
|
@@ -1,17 +1,17 @@
|
|
|
1
1
|
import { GoogleVertexProvider } from '@ai-sdk/google-vertex';
|
|
2
2
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
3
3
|
import { GoogleVertexClientSettings, GoogleVertexClient } from './client.js';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
|
|
4
5
|
import '../../../logger/logger.js';
|
|
5
6
|
import 'pino';
|
|
6
7
|
import '../../../errors.js';
|
|
7
8
|
import '../../../internals/types.js';
|
|
8
9
|
import '../../../internals/helpers/guards.js';
|
|
9
10
|
import '../../../internals/serializable.js';
|
|
10
|
-
import '../../../chat-BBoOSvzm.js';
|
|
11
|
-
import '../../../backend/message.js';
|
|
12
11
|
import 'ai';
|
|
12
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
13
|
+
import '../../../backend/message.js';
|
|
13
14
|
import '../../../context.js';
|
|
14
|
-
import '../../../emitter-BqpLJQVb.js';
|
|
15
15
|
import '../../../internals/helpers/promise.js';
|
|
16
16
|
import 'promise-based-task';
|
|
17
17
|
import '../../../cache/base.js';
|
|
@@ -31,9 +31,8 @@ import '../../../backend/client.js';
|
|
|
31
31
|
|
|
32
32
|
type GoogleVertexParameters = Parameters<GoogleVertexProvider["languageModel"]>;
|
|
33
33
|
type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;
|
|
34
|
-
type GoogleVertexChatModelSettings = NonNullable<GoogleVertexParameters[1]>;
|
|
35
34
|
declare class GoogleVertexChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: GoogleVertexChatModelId,
|
|
35
|
+
constructor(modelId?: GoogleVertexChatModelId, parameters?: ChatModelParameters, client?: GoogleVertexClientSettings | GoogleVertexClient);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { GoogleVertexChatModel, type GoogleVertexChatModelId
|
|
38
|
+
export { GoogleVertexChatModel, type GoogleVertexChatModelId };
|
|
@@ -8,9 +8,10 @@ class GoogleVertexChatModel extends VercelChatModel {
|
|
|
8
8
|
static {
|
|
9
9
|
__name(this, "GoogleVertexChatModel");
|
|
10
10
|
}
|
|
11
|
-
constructor(modelId = getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"),
|
|
12
|
-
const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId
|
|
11
|
+
constructor(modelId = getEnv("GOOGLE_VERTEX_CHAT_MODEL", "gemini-1.5-pro"), parameters = {}, client) {
|
|
12
|
+
const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);
|
|
13
13
|
super(model);
|
|
14
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
14
15
|
}
|
|
15
16
|
}
|
|
16
17
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/google-vertex/backend/chat.ts"],"names":["GoogleVertexChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GoogleVertexClient","ensure","instance","languageModel","Object","assign"],"mappings":";;;;;;AAiBO,MAAMA,8BAA8BC,eAAAA,CAAAA;EAjB3C;;;EAkBE,WAAA,CACEC,OAAAA,GAAmCC,OAAO,0BAAA,EAA4B,gBAAA,GACtEC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,kBAAAA,CAAmBC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AACvE,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { GoogleVertexProvider } from \"@ai-sdk/google-vertex\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport {\n GoogleVertexClient,\n GoogleVertexClientSettings,\n} from \"@/adapters/google-vertex/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GoogleVertexParameters = Parameters<GoogleVertexProvider[\"languageModel\"]>;\nexport type GoogleVertexChatModelId = NonNullable<GoogleVertexParameters[0]>;\n\nexport class GoogleVertexChatModel extends VercelChatModel {\n constructor(\n modelId: GoogleVertexChatModelId = getEnv(\"GOOGLE_VERTEX_CHAT_MODEL\", \"gemini-1.5-pro\"),\n parameters: ChatModelParameters = {},\n client?: GoogleVertexClientSettings | GoogleVertexClient,\n ) {\n const model = GoogleVertexClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
|
|
|
7
7
|
import '../../../internals/helpers/guards.cjs';
|
|
8
8
|
import '../../../backend/embedding.cjs';
|
|
9
9
|
import '../../../context.cjs';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
11
11
|
import '../../../internals/helpers/promise.cjs';
|
|
12
12
|
import '../../../errors.cjs';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
14
14
|
import '../../../backend/message.cjs';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
|
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../backend/embedding.js';
|
|
9
9
|
import '../../../context.js';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
11
11
|
import '../../../internals/helpers/promise.js';
|
|
12
12
|
import '../../../errors.js';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-C0s-o6ll.js';
|
|
14
14
|
import '../../../backend/message.js';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -10,9 +10,10 @@ class GroqChatModel extends chat_cjs.VercelChatModel {
|
|
|
10
10
|
static {
|
|
11
11
|
__name(this, "GroqChatModel");
|
|
12
12
|
}
|
|
13
|
-
constructor(modelId = env_cjs.getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"),
|
|
14
|
-
const model = client_cjs.GroqClient.ensure(client).instance.languageModel(modelId
|
|
13
|
+
constructor(modelId = env_cjs.getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), parameters = {}, client) {
|
|
14
|
+
const model = client_cjs.GroqClient.ensure(client).instance.languageModel(modelId);
|
|
15
15
|
super(model);
|
|
16
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
16
17
|
}
|
|
17
18
|
static {
|
|
18
19
|
this.register();
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GroqClient","ensure","instance","languageModel","Object","assign","register"],"mappings":";;;;;;;;AAcO,MAAMA,sBAAsBC,wBAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,eAAO,iBAAA,EAAmB,cAAA,GACrDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,qBAAAA,CAAWC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AAC/D,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;EAEA;AACE,IAAA,IAAA,CAAKS,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n parameters: ChatModelParameters = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -1,17 +1,17 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
2
2
|
import { GroqClientSettings, GroqClient } from './client.cjs';
|
|
3
3
|
import { GroqProvider } from '@ai-sdk/groq';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
|
|
4
5
|
import '../../../logger/logger.cjs';
|
|
5
6
|
import 'pino';
|
|
6
7
|
import '../../../errors.cjs';
|
|
7
8
|
import '../../../internals/types.cjs';
|
|
8
9
|
import '../../../internals/helpers/guards.cjs';
|
|
9
10
|
import '../../../internals/serializable.cjs';
|
|
10
|
-
import '../../../chat-BZ55YQab.cjs';
|
|
11
|
-
import '../../../backend/message.cjs';
|
|
12
11
|
import 'ai';
|
|
12
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
13
|
+
import '../../../backend/message.cjs';
|
|
13
14
|
import '../../../context.cjs';
|
|
14
|
-
import '../../../emitter-pJzHC_AM.cjs';
|
|
15
15
|
import '../../../internals/helpers/promise.cjs';
|
|
16
16
|
import 'promise-based-task';
|
|
17
17
|
import '../../../cache/base.cjs';
|
|
@@ -31,9 +31,8 @@ import '../../../backend/client.cjs';
|
|
|
31
31
|
|
|
32
32
|
type GroqParameters = Parameters<GroqProvider["languageModel"]>;
|
|
33
33
|
type GroqChatModelId = NonNullable<GroqParameters[0]>;
|
|
34
|
-
type GroqChatModelSettings = NonNullable<GroqParameters[1]>;
|
|
35
34
|
declare class GroqChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: GroqChatModelId,
|
|
35
|
+
constructor(modelId?: GroqChatModelId, parameters?: ChatModelParameters, client?: GroqClientSettings | GroqClient);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { GroqChatModel, type GroqChatModelId
|
|
38
|
+
export { GroqChatModel, type GroqChatModelId };
|
|
@@ -1,17 +1,17 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
2
2
|
import { GroqClientSettings, GroqClient } from './client.js';
|
|
3
3
|
import { GroqProvider } from '@ai-sdk/groq';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
|
|
4
5
|
import '../../../logger/logger.js';
|
|
5
6
|
import 'pino';
|
|
6
7
|
import '../../../errors.js';
|
|
7
8
|
import '../../../internals/types.js';
|
|
8
9
|
import '../../../internals/helpers/guards.js';
|
|
9
10
|
import '../../../internals/serializable.js';
|
|
10
|
-
import '../../../chat-BBoOSvzm.js';
|
|
11
|
-
import '../../../backend/message.js';
|
|
12
11
|
import 'ai';
|
|
12
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
13
|
+
import '../../../backend/message.js';
|
|
13
14
|
import '../../../context.js';
|
|
14
|
-
import '../../../emitter-BqpLJQVb.js';
|
|
15
15
|
import '../../../internals/helpers/promise.js';
|
|
16
16
|
import 'promise-based-task';
|
|
17
17
|
import '../../../cache/base.js';
|
|
@@ -31,9 +31,8 @@ import '../../../backend/client.js';
|
|
|
31
31
|
|
|
32
32
|
type GroqParameters = Parameters<GroqProvider["languageModel"]>;
|
|
33
33
|
type GroqChatModelId = NonNullable<GroqParameters[0]>;
|
|
34
|
-
type GroqChatModelSettings = NonNullable<GroqParameters[1]>;
|
|
35
34
|
declare class GroqChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: GroqChatModelId,
|
|
35
|
+
constructor(modelId?: GroqChatModelId, parameters?: ChatModelParameters, client?: GroqClientSettings | GroqClient);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { GroqChatModel, type GroqChatModelId
|
|
38
|
+
export { GroqChatModel, type GroqChatModelId };
|
|
@@ -8,9 +8,10 @@ class GroqChatModel extends VercelChatModel {
|
|
|
8
8
|
static {
|
|
9
9
|
__name(this, "GroqChatModel");
|
|
10
10
|
}
|
|
11
|
-
constructor(modelId = getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"),
|
|
12
|
-
const model = GroqClient.ensure(client).instance.languageModel(modelId
|
|
11
|
+
constructor(modelId = getEnv("GROQ_CHAT_MODEL", "gemma2-9b-it"), parameters = {}, client) {
|
|
12
|
+
const model = GroqClient.ensure(client).instance.languageModel(modelId);
|
|
13
13
|
super(model);
|
|
14
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
14
15
|
}
|
|
15
16
|
static {
|
|
16
17
|
this.register();
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/groq/backend/chat.ts"],"names":["GroqChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","GroqClient","ensure","instance","languageModel","Object","assign","register"],"mappings":";;;;;;AAcO,MAAMA,sBAAsBC,eAAAA,CAAAA;EAdnC;;;EAeE,WAAA,CACEC,OAAAA,GAA2BC,OAAO,iBAAA,EAAmB,cAAA,GACrDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,UAAAA,CAAWC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,cAAcR,OAAAA,CAAAA;AAC/D,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;EAEA;AACE,IAAA,IAAA,CAAKS,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { GroqClient, GroqClientSettings } from \"@/adapters/groq/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { GroqProvider } from \"@ai-sdk/groq\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype GroqParameters = Parameters<GroqProvider[\"languageModel\"]>;\nexport type GroqChatModelId = NonNullable<GroqParameters[0]>;\n\nexport class GroqChatModel extends VercelChatModel {\n constructor(\n modelId: GroqChatModelId = getEnv(\"GROQ_CHAT_MODEL\", \"gemma2-9b-it\"),\n parameters: ChatModelParameters = {},\n client?: GroqClientSettings | GroqClient,\n ) {\n const model = GroqClient.ensure(client).instance.languageModel(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -3,13 +3,13 @@ import { GroqClientSettings, GroqClient } from './client.cjs';
|
|
|
3
3
|
import { GroqProvider } from '@ai-sdk/groq';
|
|
4
4
|
import '../../../backend/embedding.cjs';
|
|
5
5
|
import '../../../context.cjs';
|
|
6
|
-
import '../../../emitter-
|
|
6
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
7
7
|
import '../../../internals/types.cjs';
|
|
8
8
|
import '../../../internals/helpers/guards.cjs';
|
|
9
9
|
import '../../../internals/serializable.cjs';
|
|
10
10
|
import '../../../internals/helpers/promise.cjs';
|
|
11
11
|
import '../../../errors.cjs';
|
|
12
|
-
import '../../../chat-
|
|
12
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
13
13
|
import '../../../backend/message.cjs';
|
|
14
14
|
import 'ai';
|
|
15
15
|
import 'promise-based-task';
|
|
@@ -3,13 +3,13 @@ import { GroqClientSettings, GroqClient } from './client.js';
|
|
|
3
3
|
import { GroqProvider } from '@ai-sdk/groq';
|
|
4
4
|
import '../../../backend/embedding.js';
|
|
5
5
|
import '../../../context.js';
|
|
6
|
-
import '../../../emitter-
|
|
6
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
7
7
|
import '../../../internals/types.js';
|
|
8
8
|
import '../../../internals/helpers/guards.js';
|
|
9
9
|
import '../../../internals/serializable.js';
|
|
10
10
|
import '../../../internals/helpers/promise.js';
|
|
11
11
|
import '../../../errors.js';
|
|
12
|
-
import '../../../chat-
|
|
12
|
+
import '../../../chat-C0s-o6ll.js';
|
|
13
13
|
import '../../../backend/message.js';
|
|
14
14
|
import 'ai';
|
|
15
15
|
import 'promise-based-task';
|
|
@@ -82,13 +82,14 @@ class LangChainChatModel extends chat_cjs.ChatModel {
|
|
|
82
82
|
}
|
|
83
83
|
})));
|
|
84
84
|
}
|
|
85
|
-
const
|
|
85
|
+
const stop = output.response_metadata.stop_sequence || "stop";
|
|
86
|
+
return new chat_cjs.ChatModelOutput(messages, {
|
|
86
87
|
totalTokens: output.usage_metadata?.total_tokens ?? 0,
|
|
87
88
|
promptTokens: output.usage_metadata?.input_tokens ?? 0,
|
|
88
|
-
completionTokens: output.usage_metadata?.output_tokens ?? 0
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
89
|
+
completionTokens: output.usage_metadata?.output_tokens ?? 0,
|
|
90
|
+
reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,
|
|
91
|
+
cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read
|
|
92
|
+
}, stop);
|
|
92
93
|
}
|
|
93
94
|
async _createStructure(input, run) {
|
|
94
95
|
const { messages, options } = this.prepareInput(input, run);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","response_metadata","stop_sequence","ChatModelOutput","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","reasoningTokens","output_token_details","reasoning","cachedPromptTokens","input_token_details","cache_read","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;;;AAyBO,MAAMA,2BAA2BC,kBAAAA,CAAAA;EAzBxC;;;;;AA0BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,mBAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,4BAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,4BAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,qBAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAME,IAAAA,GAA8BK,MAAAA,CAAOQ,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAC9E,IAAA,OAAO,IAAIC,yBACT3B,QAAAA,EACA;MACE4B,WAAAA,EAAaX,MAAAA,CAAOY,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcd,MAAAA,CAAOY,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBhB,MAAAA,CAAOY,gBAAgBK,aAAAA,IAAiB,CAAA;MAC1DC,eAAAA,EAAiBlB,MAAAA,CAAOY,gBAAgBO,oBAAAA,EAAsBC,SAAAA;MAC9DC,kBAAAA,EAAoBrB,MAAAA,CAAOY,gBAAgBU,mBAAAA,EAAqBC;AAClE,KAAA,EACA5B,IAAAA,CAAAA;AAEJ;EAEA,MAAgB6B,gBAAAA,CACdjD,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAEiD,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAK9D,KAAAA,CAChC+D,oBAAAA,CAA0BpD,KAAAA,CAAMqD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACCjD,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAEgD,MAAAA,EAAQN,MAAAA;MAAa1B,MAAAA,EAAQ,IAAA,CAAKf,cAAcwC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBtE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAsE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n return new ChatModelOutput(\n messages,\n {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,\n cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read,\n },\n stop,\n );\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { Logger } from '../../../logger/logger.cjs';
|
|
2
|
-
import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
2
|
+
import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
|
|
3
3
|
import { RunContext } from '../../../context.cjs';
|
|
4
4
|
import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
|
|
5
5
|
import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
|
|
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.cjs';
|
|
|
10
10
|
import '../../../internals/serializable.cjs';
|
|
11
11
|
import '../../../backend/message.cjs';
|
|
12
12
|
import 'ai';
|
|
13
|
-
import '../../../emitter-
|
|
13
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
14
14
|
import 'promise-based-task';
|
|
15
15
|
import '../../../cache/base.cjs';
|
|
16
16
|
import '../../../backend/constants.cjs';
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { Logger } from '../../../logger/logger.js';
|
|
2
|
-
import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
2
|
+
import { m as ChatModel, C as ChatModelParameters, h as ChatModelEmitter, d as ChatModelInput, n as ChatModelOutput, a as ChatModelObjectInput, b as ChatModelObjectOutput, i as ChatModelCache, l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
|
|
3
3
|
import { RunContext } from '../../../context.js';
|
|
4
4
|
import { BaseChatModel, BaseChatModelCallOptions } from '@langchain/core/language_models/chat_models';
|
|
5
5
|
import { BaseMessageLike, AIMessageChunk } from '@langchain/core/messages';
|
|
@@ -10,7 +10,7 @@ import '../../../internals/helpers/guards.js';
|
|
|
10
10
|
import '../../../internals/serializable.js';
|
|
11
11
|
import '../../../backend/message.js';
|
|
12
12
|
import 'ai';
|
|
13
|
-
import '../../../emitter-
|
|
13
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
14
14
|
import 'promise-based-task';
|
|
15
15
|
import '../../../cache/base.js';
|
|
16
16
|
import '../../../backend/constants.js';
|
|
@@ -80,13 +80,14 @@ class LangChainChatModel extends ChatModel {
|
|
|
80
80
|
}
|
|
81
81
|
})));
|
|
82
82
|
}
|
|
83
|
-
const
|
|
83
|
+
const stop = output.response_metadata.stop_sequence || "stop";
|
|
84
|
+
return new ChatModelOutput(messages, {
|
|
84
85
|
totalTokens: output.usage_metadata?.total_tokens ?? 0,
|
|
85
86
|
promptTokens: output.usage_metadata?.input_tokens ?? 0,
|
|
86
|
-
completionTokens: output.usage_metadata?.output_tokens ?? 0
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
87
|
+
completionTokens: output.usage_metadata?.output_tokens ?? 0,
|
|
88
|
+
reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,
|
|
89
|
+
cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read
|
|
90
|
+
}, stop);
|
|
90
91
|
}
|
|
91
92
|
async _createStructure(input, run) {
|
|
92
93
|
const { messages, options } = this.prepareInput(input, run);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/langchain/backend/chat.ts"],"names":["LangChainChatModel","ChatModel","emitter","lcLLM","parameters","Emitter","root","child","namespace","creator","modelId","_modelType","providerId","_create","input","run","preparedInput","prepareInput","response","bindTools","tools","invoke","messages","options","prepareOutput","_createStream","stream","chunk","map","msg","role","content","type","runId","stop","stopSequences","signal","tool_choice","toolChoice","output","push","AssistantMessage","message","text","image_url","toString","ValueError","response_metadata","stop_sequence","ChatModelOutput","totalTokens","usage_metadata","total_tokens","promptTokens","input_tokens","completionTokens","output_tokens","reasoningTokens","output_token_details","reasoning","cachedPromptTokens","input_token_details","cache_read","_createStructure","raw","parsed","withStructuredOutput","schema","method","strict","includeRaw","object","createSnapshot","loadSnapshot","snapshot","Object","assign"],"mappings":";;;;;;;AAyBO,MAAMA,2BAA2BC,SAAAA,CAAAA;EAzBxC;;;;;AA0BkBC,EAAAA,OAAAA;EAEhB,WAAA,CACqBC,KAAAA,EACHC,UAAAA,GAAkC,EAAC,EACnD;AACA,IAAA,KAAA,EAAK,EAAA,IAAA,CAHcD,KAAAA,GAAAA,KAAAA,EAAAA,KACHC,UAAAA,GAAAA,UAAAA;AAGhB,IAAA,IAAA,CAAKF,OAAAA,GAAUG,OAAAA,CAAQC,IAAAA,CAAKC,KAAAA,CAAM;MAChCC,SAAAA,EAAW;AAAC,QAAA,SAAA;AAAW,QAAA,WAAA;AAAa,QAAA;;MACpCC,OAAAA,EAAS;KACX,CAAA;AACF;AAEA,EAAA,IAAIC,OAAAA,GAAkB;AACpB,IAAA,OAAO,IAAA,CAAKP,MAAMQ,UAAAA,EAAU;AAC9B;AAEA,EAAA,IAAIC,UAAAA,GAAa;AACf,IAAA,OAAO,WAAA;AACT;EAEA,MAAgBC,OAAAA,CAAQC,OAAuBC,GAAAA,EAAiD;AAC9F,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAC/C,IAAA,MAAMG,QAAAA,GAAW,IAAA,CAAKf,KAAAA,CAAMgB,SAAAA,GACxB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BC,OAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMkB,MAAAA,CAAOL,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,OAAO,IAAA,CAAKC,cAAcN,QAAAA,CAAAA;AAC5B;EAEA,OAAiBO,aAAAA,CACfX,OACAC,GAAAA,EACiC;AACjC,IAAA,MAAMC,aAAAA,GAAgB,IAAA,CAAKC,YAAAA,CAAaH,KAAAA,EAAOC,GAAAA,CAAAA;AAE/C,IAAA,MAAMW,MAAAA,GAAS,IAAA,CAAKvB,KAAAA,CAAMgB,SAAAA,GACtB,MAAM,IAAA,CAAKhB,KAAAA,CACRgB,SAAAA,CAAUL,KAAAA,CAAMM,KAAAA,IAAS,EAAE,CAAA,CAC3BM,OAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA,GACvD,MAAM,IAAA,CAAKpB,KAAAA,CAAMuB,MAAAA,CAAOV,aAAAA,CAAcM,QAAAA,EAAUN,aAAAA,CAAcO,OAAO,CAAA;AAEzE,IAAA,WAAA,MAAiBL,YAAYQ,MAAAA,EAAQ;AACnC,MAAA,MAAMC,KAAAA,GAAQ,IAAA,CAAKH,aAAAA,CAAcN,QAAAA,CAAAA;AACjC,MAAA,MAAMS,KAAAA;AACR;AACF;AAEUV,EAAAA,YAAAA,CAAaH,OAAuBC,GAAAA,EAAuB;AACnE,IAAA,MAAMO,QAAAA,GAA8BR,KAAAA,CAAMQ,QAAAA,CAASM,GAAAA,CAAI,CAACC,GAAAA,MAAS;AAC/DC,MAAAA,IAAAA,EAAMD,GAAAA,CAAIC,IAAAA;AACVC,MAAAA,OAAAA,EAASF,GAAAA,CAAIE,OAAAA;AACbC,MAAAA,IAAAA,EAAMH,GAAAA,CAAIC;KAEZ,CAAA,CAAA;AAEA,IAAA,MAAMP,OAAAA,GAAoC;AACxCU,MAAAA,KAAAA,EAAOlB,GAAAA,CAAIkB,KAAAA;AACXC,MAAAA,IAAAA,EAAMpB,KAAAA,CAAMqB,aAAAA;AACZC,MAAAA,MAAAA,EAAQrB,GAAAA,CAAIqB,MAAAA;AACZC,MAAAA,WAAAA,EAAavB,KAAAA,CAAMwB;AACrB,KAAA;AAEA,IAAA,OAAO;AAAEhB,MAAAA,QAAAA;AAAUC,MAAAA;AAAQ,KAAA;AAC7B;AAEUC,EAAAA,aAAAA,CAAce,MAAAA,EAAwB;AAC9C,IAAA,MAAMjB,WAAsB,EAAA;AAC5B,IAAA,IAAI,OAAOiB,MAAAA,CAAOR,OAAAA,KAAY,QAAA,EAAU;AACtCT,MAAAA,QAAAA,CAASkB,IAAAA,CAAK,IAAIC,gBAAAA,CAAiBF,MAAAA,CAAOR,OAAO,CAAA,CAAA;KACnD,MAAO;AACLT,MAAAA,QAAAA,CAASkB,KACP,IAAIC,gBAAAA,CACFF,OAAOR,OAAAA,CAAQH,GAAAA,CAAI,CAACc,OAAAA,KAAAA;AAClB,QAAA,IAAIA,OAAAA,CAAQV,SAAS,MAAA,EAAQ;AAC3B,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;AAAQW,YAAAA,IAAAA,EAAMD,OAAAA,CAAQC;AAAK,WAAA;SAC5C,MAAA,IAAWD,OAAAA,CAAQV,SAAS,WAAA,EAAa;AACvC,UAAA,OAAO;YAAEA,IAAAA,EAAM,MAAA;YAAQW,IAAAA,EAAMD,OAAAA,CAAQE,UAAUC,QAAAA;AAAW,WAAA;SAC5D,MAAO;AACL,UAAA,MAAM,IAAIC,UAAAA,CAAW,CAAA,sBAAA,EAAyBJ,OAAAA,CAAQV,IAAI,CAAA,CAAA,CAAG,CAAA;AAC/D;AACF,OAAA,CAAA,CAAA,CAAA;AAGN;AAEA,IAAA,MAAME,IAAAA,GAA8BK,MAAAA,CAAOQ,iBAAAA,CAAkBC,aAAAA,IAAiB,MAAA;AAC9E,IAAA,OAAO,IAAIC,gBACT3B,QAAAA,EACA;MACE4B,WAAAA,EAAaX,MAAAA,CAAOY,gBAAgBC,YAAAA,IAAgB,CAAA;MACpDC,YAAAA,EAAcd,MAAAA,CAAOY,gBAAgBG,YAAAA,IAAgB,CAAA;MACrDC,gBAAAA,EAAkBhB,MAAAA,CAAOY,gBAAgBK,aAAAA,IAAiB,CAAA;MAC1DC,eAAAA,EAAiBlB,MAAAA,CAAOY,gBAAgBO,oBAAAA,EAAsBC,SAAAA;MAC9DC,kBAAAA,EAAoBrB,MAAAA,CAAOY,gBAAgBU,mBAAAA,EAAqBC;AAClE,KAAA,EACA5B,IAAAA,CAAAA;AAEJ;EAEA,MAAgB6B,gBAAAA,CACdjD,OACAC,GAAAA,EACmC;AACnC,IAAA,MAAM,EAAEO,QAAAA,EAAUC,OAAAA,KAAY,IAAA,CAAKN,YAAAA,CAAaH,OAAOC,GAAAA,CAAAA;AACvD,IAAA,MAAM,EAAEiD,KAAKC,MAAAA,EAAM,GAAK,MAAM,IAAA,CAAK9D,KAAAA,CAChC+D,oBAAAA,CAA0BpD,KAAAA,CAAMqD,MAAAA,EAAQ;MACvCC,MAAAA,EAAQ,YAAA;MACRC,MAAAA,EAAQ,KAAA;MACRC,UAAAA,EAAY;KACd,CAAA,CACCjD,MAAAA,CAAOC,QAAAA,EAAUC,OAAAA,CAAAA;AAEpB,IAAA,OAAO;MAAEgD,MAAAA,EAAQN,MAAAA;MAAa1B,MAAAA,EAAQ,IAAA,CAAKf,cAAcwC,GAAAA;AAAuB,KAAA;AAClF;EAEAQ,cAAAA,GAAiB;AACf,IAAA,OAAO;AAAE,MAAA,GAAG,MAAMA,cAAAA,EAAAA;AAAkBtE,MAAAA,OAAAA,EAAS,IAAA,CAAKA,OAAAA;AAASC,MAAAA,KAAAA,EAAO,IAAA,CAAKA;AAAM,KAAA;AAC/E;AAEAsE,EAAAA,YAAAA,CAAaC,QAAAA,EAAwD;AACnEC,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,QAAAA,CAAAA;AACtB;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport {\n ChatModel,\n ChatModelEmitter,\n ChatModelFinishReason,\n ChatModelInput,\n ChatModelObjectInput,\n ChatModelObjectOutput,\n ChatModelOutput,\n ChatModelParameters,\n} from \"@/backend/chat.js\";\nimport { RunContext } from \"@/context.js\";\nimport { Emitter } from \"@/emitter/emitter.js\";\nimport {\n BaseChatModel,\n BaseChatModelCallOptions,\n} from \"@langchain/core/language_models/chat_models\";\nimport { AIMessageChunk, BaseMessageLike } from \"@langchain/core/messages\";\nimport { AssistantMessage, Message } from \"@/backend/message.js\";\nimport { ValueError } from \"@/errors.js\";\n\nexport class LangChainChatModel extends ChatModel {\n public readonly emitter: ChatModelEmitter;\n\n constructor(\n protected readonly lcLLM: BaseChatModel,\n public readonly parameters: ChatModelParameters = {},\n ) {\n super();\n this.emitter = Emitter.root.child({\n namespace: [\"backend\", \"langchain\", \"chat\"],\n creator: this,\n });\n }\n\n get modelId(): string {\n return this.lcLLM._modelType();\n }\n\n get providerId() {\n return \"langchain\";\n }\n\n protected async _create(input: ChatModelInput, run: RunContext<this>): Promise<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n const response = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .invoke(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.invoke(preparedInput.messages, preparedInput.options);\n\n return this.prepareOutput(response);\n }\n\n protected async *_createStream(\n input: ChatModelInput,\n run: RunContext<this>,\n ): AsyncGenerator<ChatModelOutput> {\n const preparedInput = this.prepareInput(input, run);\n\n const stream = this.lcLLM.bindTools\n ? await this.lcLLM\n .bindTools(input.tools ?? [])\n .stream(preparedInput.messages, preparedInput.options)\n : await this.lcLLM.stream(preparedInput.messages, preparedInput.options);\n\n for await (const response of stream) {\n const chunk = this.prepareOutput(response);\n yield chunk;\n }\n }\n\n protected prepareInput(input: ChatModelInput, run: RunContext<this>) {\n const messages: BaseMessageLike[] = input.messages.map((msg) => ({\n role: msg.role,\n content: msg.content,\n type: msg.role,\n // TODO\n }));\n\n const options: BaseChatModelCallOptions = {\n runId: run.runId,\n stop: input.stopSequences,\n signal: run.signal,\n tool_choice: input.toolChoice,\n };\n\n return { messages, options };\n }\n\n protected prepareOutput(output: AIMessageChunk) {\n const messages: Message[] = [];\n if (typeof output.content === \"string\") {\n messages.push(new AssistantMessage(output.content));\n } else {\n messages.push(\n new AssistantMessage(\n output.content.map((message) => {\n if (message.type === \"text\") {\n return { type: \"text\", text: message.text };\n } else if (message.type === \"image_url\") {\n return { type: \"text\", text: message.image_url.toString() };\n } else {\n throw new ValueError(`Unknown message type \"${message.type}\"`);\n }\n }),\n ),\n );\n }\n\n const stop: ChatModelFinishReason = output.response_metadata.stop_sequence || \"stop\";\n return new ChatModelOutput(\n messages,\n {\n totalTokens: output.usage_metadata?.total_tokens ?? 0,\n promptTokens: output.usage_metadata?.input_tokens ?? 0,\n completionTokens: output.usage_metadata?.output_tokens ?? 0,\n reasoningTokens: output.usage_metadata?.output_token_details?.reasoning,\n cachedPromptTokens: output.usage_metadata?.input_token_details?.cache_read,\n },\n stop,\n );\n }\n\n protected async _createStructure<T>(\n input: ChatModelObjectInput<T>,\n run: RunContext<this>,\n ): Promise<ChatModelObjectOutput<T>> {\n const { messages, options } = this.prepareInput(input, run);\n const { raw, parsed } = await this.lcLLM\n .withStructuredOutput<any>(input.schema, {\n method: \"jsonSchema\",\n strict: false,\n includeRaw: true,\n })\n .invoke(messages, options);\n\n return { object: parsed as T, output: this.prepareOutput(raw as AIMessageChunk) };\n }\n\n createSnapshot() {\n return { ...super.createSnapshot(), emitter: this.emitter, lcLLM: this.lcLLM };\n }\n\n loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void {\n Object.assign(this, snapshot);\n }\n}\n"]}
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.cjs';
|
|
2
2
|
import { RunContext } from '../../../context.cjs';
|
|
3
|
-
import { E as Emitter } from '../../../emitter-
|
|
3
|
+
import { E as Emitter } from '../../../emitter-DpqUYjXH.cjs';
|
|
4
4
|
import { Embeddings } from '@langchain/core/embeddings';
|
|
5
5
|
import '../../../internals/serializable.cjs';
|
|
6
6
|
import '../../../internals/types.cjs';
|
|
7
7
|
import '../../../internals/helpers/guards.cjs';
|
|
8
8
|
import '../../../errors.cjs';
|
|
9
|
-
import '../../../chat-
|
|
9
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
10
10
|
import '../../../backend/message.cjs';
|
|
11
11
|
import 'ai';
|
|
12
12
|
import 'promise-based-task';
|
|
@@ -35,7 +35,7 @@ declare class LangChainEmbeddingModel extends EmbeddingModel {
|
|
|
35
35
|
get providerId(): string;
|
|
36
36
|
protected _create(input: EmbeddingModelInput, run: RunContext<this>): Promise<EmbeddingModelOutput>;
|
|
37
37
|
createSnapshot(): {
|
|
38
|
-
lcEmbedding: Embeddings
|
|
38
|
+
lcEmbedding: Embeddings<number[]>;
|
|
39
39
|
emitter: Emitter<EmbeddingModelEvents>;
|
|
40
40
|
};
|
|
41
41
|
loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void;
|