beeai-framework 0.1.19 → 0.1.21
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapters/a2a/agents/agent.cjs +4 -4
- package/dist/adapters/a2a/agents/agent.cjs.map +1 -1
- package/dist/adapters/a2a/agents/agent.d.cts +2 -2
- package/dist/adapters/a2a/agents/agent.d.ts +2 -2
- package/dist/adapters/a2a/agents/agent.js +4 -4
- package/dist/adapters/a2a/agents/agent.js.map +1 -1
- package/dist/adapters/a2a/agents/events.d.cts +1 -1
- package/dist/adapters/a2a/agents/events.d.ts +1 -1
- package/dist/adapters/a2a/agents/types.d.cts +2 -2
- package/dist/adapters/a2a/agents/types.d.ts +2 -2
- package/dist/adapters/a2a/agents/utils.cjs +2 -2
- package/dist/adapters/a2a/agents/utils.cjs.map +1 -1
- package/dist/adapters/a2a/agents/utils.js +2 -2
- package/dist/adapters/a2a/agents/utils.js.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.cjs +1 -3
- package/dist/adapters/a2a/serve/agent_executor.cjs.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.d.cts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.d.ts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.js +1 -3
- package/dist/adapters/a2a/serve/agent_executor.js.map +1 -1
- package/dist/adapters/a2a/serve/server.cjs +15 -0
- package/dist/adapters/a2a/serve/server.cjs.map +1 -1
- package/dist/adapters/a2a/serve/server.d.cts +4 -2
- package/dist/adapters/a2a/serve/server.d.ts +4 -2
- package/dist/adapters/a2a/serve/server.js +15 -0
- package/dist/adapters/a2a/serve/server.js.map +1 -1
- package/dist/adapters/acp/agents/agent.d.cts +2 -2
- package/dist/adapters/acp/agents/agent.d.ts +2 -2
- package/dist/adapters/acp/agents/events.d.cts +1 -1
- package/dist/adapters/acp/agents/events.d.ts +1 -1
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.cjs +4 -4
- package/dist/adapters/agentstack/agents/agent.cjs.map +1 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.cts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.ts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.js +4 -4
- package/dist/adapters/agentstack/agents/agent.js.map +1 -0
- package/dist/adapters/agentstack/agents/events.d.cts +23 -0
- package/dist/adapters/agentstack/agents/events.d.ts +23 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.cts +4 -4
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/chat.cjs +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/chat.d.cts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.d.ts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.js +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.js.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.d.cts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.js +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.js.map +1 -1
- package/dist/adapters/anthropic/backend/chat.cjs +3 -2
- package/dist/adapters/anthropic/backend/chat.cjs.map +1 -1
- package/dist/adapters/anthropic/backend/chat.d.cts +5 -6
- package/dist/adapters/anthropic/backend/chat.d.ts +5 -6
- package/dist/adapters/anthropic/backend/chat.js +3 -2
- package/dist/adapters/anthropic/backend/chat.js.map +1 -1
- package/dist/adapters/anthropic/backend/embedding.d.cts +2 -2
- package/dist/adapters/anthropic/backend/embedding.d.ts +2 -2
- package/dist/adapters/azure-openai/backend/chat.cjs +3 -2
- package/dist/adapters/azure-openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/chat.d.cts +5 -6
- package/dist/adapters/azure-openai/backend/chat.d.ts +5 -6
- package/dist/adapters/azure-openai/backend/chat.js +3 -2
- package/dist/adapters/azure-openai/backend/chat.js.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.cjs +2 -2
- package/dist/adapters/azure-openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.d.cts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.d.ts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.js +2 -2
- package/dist/adapters/azure-openai/backend/embedding.js.map +1 -1
- package/dist/adapters/dummy/backend/chat.d.cts +2 -2
- package/dist/adapters/dummy/backend/chat.d.ts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.cts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.ts +2 -2
- package/dist/adapters/google-vertex/backend/chat.cjs +3 -2
- package/dist/adapters/google-vertex/backend/chat.cjs.map +1 -1
- package/dist/adapters/google-vertex/backend/chat.d.cts +5 -6
- package/dist/adapters/google-vertex/backend/chat.d.ts +5 -6
- package/dist/adapters/google-vertex/backend/chat.js +3 -2
- package/dist/adapters/google-vertex/backend/chat.js.map +1 -1
- package/dist/adapters/google-vertex/backend/embedding.d.cts +2 -2
- package/dist/adapters/google-vertex/backend/embedding.d.ts +2 -2
- package/dist/adapters/groq/backend/chat.cjs +3 -2
- package/dist/adapters/groq/backend/chat.cjs.map +1 -1
- package/dist/adapters/groq/backend/chat.d.cts +5 -6
- package/dist/adapters/groq/backend/chat.d.ts +5 -6
- package/dist/adapters/groq/backend/chat.js +3 -2
- package/dist/adapters/groq/backend/chat.js.map +1 -1
- package/dist/adapters/groq/backend/embedding.d.cts +2 -2
- package/dist/adapters/groq/backend/embedding.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.cjs +6 -5
- package/dist/adapters/langchain/backend/chat.cjs.map +1 -1
- package/dist/adapters/langchain/backend/chat.d.cts +2 -2
- package/dist/adapters/langchain/backend/chat.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.js +6 -5
- package/dist/adapters/langchain/backend/chat.js.map +1 -1
- package/dist/adapters/langchain/backend/embedding.d.cts +3 -3
- package/dist/adapters/langchain/backend/embedding.d.ts +3 -3
- package/dist/adapters/langchain/tools.d.cts +1 -1
- package/dist/adapters/langchain/tools.d.ts +1 -1
- package/dist/adapters/mcp/serve/server.cjs +12 -4
- package/dist/adapters/mcp/serve/server.cjs.map +1 -1
- package/dist/adapters/mcp/serve/server.d.cts +1 -1
- package/dist/adapters/mcp/serve/server.d.ts +1 -1
- package/dist/adapters/mcp/serve/server.js +12 -4
- package/dist/adapters/mcp/serve/server.js.map +1 -1
- package/dist/adapters/ollama/backend/chat.cjs +1 -4
- package/dist/adapters/ollama/backend/chat.cjs.map +1 -1
- package/dist/adapters/ollama/backend/chat.d.cts +4 -4
- package/dist/adapters/ollama/backend/chat.d.ts +4 -4
- package/dist/adapters/ollama/backend/chat.js +1 -4
- package/dist/adapters/ollama/backend/chat.js.map +1 -1
- package/dist/adapters/ollama/backend/client.cjs +2 -2
- package/dist/adapters/ollama/backend/client.cjs.map +1 -1
- package/dist/adapters/ollama/backend/client.d.cts +1 -1
- package/dist/adapters/ollama/backend/client.d.ts +1 -1
- package/dist/adapters/ollama/backend/client.js +1 -1
- package/dist/adapters/ollama/backend/client.js.map +1 -1
- package/dist/adapters/ollama/backend/embedding.cjs.map +1 -1
- package/dist/adapters/ollama/backend/embedding.d.cts +3 -3
- package/dist/adapters/ollama/backend/embedding.d.ts +3 -3
- package/dist/adapters/ollama/backend/embedding.js.map +1 -1
- package/dist/adapters/openai/backend/chat.cjs +3 -2
- package/dist/adapters/openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/openai/backend/chat.d.cts +5 -6
- package/dist/adapters/openai/backend/chat.d.ts +5 -6
- package/dist/adapters/openai/backend/chat.js +3 -2
- package/dist/adapters/openai/backend/chat.js.map +1 -1
- package/dist/adapters/openai/backend/client.cjs +2 -11
- package/dist/adapters/openai/backend/client.cjs.map +1 -1
- package/dist/adapters/openai/backend/client.js +3 -12
- package/dist/adapters/openai/backend/client.js.map +1 -1
- package/dist/adapters/openai/backend/embedding.cjs +2 -2
- package/dist/adapters/openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/openai/backend/embedding.d.cts +4 -4
- package/dist/adapters/openai/backend/embedding.d.ts +4 -4
- package/dist/adapters/openai/backend/embedding.js +2 -2
- package/dist/adapters/openai/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/chat.cjs +26 -21
- package/dist/adapters/vercel/backend/chat.cjs.map +1 -1
- package/dist/adapters/vercel/backend/chat.d.cts +6 -4
- package/dist/adapters/vercel/backend/chat.d.ts +6 -4
- package/dist/adapters/vercel/backend/chat.js +27 -22
- package/dist/adapters/vercel/backend/chat.js.map +1 -1
- package/dist/adapters/vercel/backend/embedding.cjs.map +1 -1
- package/dist/adapters/vercel/backend/embedding.d.cts +3 -3
- package/dist/adapters/vercel/backend/embedding.d.ts +3 -3
- package/dist/adapters/vercel/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/utils.cjs +11 -0
- package/dist/adapters/vercel/backend/utils.cjs.map +1 -1
- package/dist/adapters/vercel/backend/utils.d.cts +19 -2
- package/dist/adapters/vercel/backend/utils.d.ts +19 -2
- package/dist/adapters/vercel/backend/utils.js +11 -1
- package/dist/adapters/vercel/backend/utils.js.map +1 -1
- package/dist/adapters/watsonx/backend/chat.cjs +10 -4
- package/dist/adapters/watsonx/backend/chat.cjs.map +1 -1
- package/dist/adapters/watsonx/backend/chat.d.cts +2 -2
- package/dist/adapters/watsonx/backend/chat.d.ts +2 -2
- package/dist/adapters/watsonx/backend/chat.js +10 -4
- package/dist/adapters/watsonx/backend/chat.js.map +1 -1
- package/dist/adapters/watsonx/backend/embedding.d.cts +2 -2
- package/dist/adapters/watsonx/backend/embedding.d.ts +2 -2
- package/dist/adapters/xai/backend/chat.cjs +3 -2
- package/dist/adapters/xai/backend/chat.cjs.map +1 -1
- package/dist/adapters/xai/backend/chat.d.cts +5 -6
- package/dist/adapters/xai/backend/chat.d.ts +5 -6
- package/dist/adapters/xai/backend/chat.js +3 -2
- package/dist/adapters/xai/backend/chat.js.map +1 -1
- package/dist/{agent-7NzfPC6d.d.ts → agent-CHebzsQB.d.ts} +4 -4
- package/dist/{agent-B6pOfx0K.d.cts → agent-DvvVGvLr.d.cts} +4 -4
- package/dist/agents/base.d.cts +2 -2
- package/dist/agents/base.d.ts +2 -2
- package/dist/agents/experimental/replan/agent.d.cts +3 -3
- package/dist/agents/experimental/replan/agent.d.ts +3 -3
- package/dist/agents/experimental/replan/prompts.d.cts +1 -1
- package/dist/agents/experimental/replan/prompts.d.ts +1 -1
- package/dist/agents/experimental/streamlit/agent.d.cts +3 -3
- package/dist/agents/experimental/streamlit/agent.d.ts +3 -3
- package/dist/agents/react/agent.d.cts +5 -4
- package/dist/agents/react/agent.d.ts +5 -4
- package/dist/agents/react/prompts.d.cts +1 -1
- package/dist/agents/react/prompts.d.ts +1 -1
- package/dist/agents/react/runners/base.cjs.map +1 -1
- package/dist/agents/react/runners/base.d.cts +5 -4
- package/dist/agents/react/runners/base.d.ts +5 -4
- package/dist/agents/react/runners/base.js.map +1 -1
- package/dist/agents/react/runners/deep-think/prompts.d.cts +1 -1
- package/dist/agents/react/runners/deep-think/prompts.d.ts +1 -1
- package/dist/agents/react/runners/deep-think/runner.d.cts +6 -5
- package/dist/agents/react/runners/deep-think/runner.d.ts +6 -5
- package/dist/agents/react/runners/default/runner.cjs.map +1 -1
- package/dist/agents/react/runners/default/runner.d.cts +6 -5
- package/dist/agents/react/runners/default/runner.d.ts +6 -5
- package/dist/agents/react/runners/default/runner.js.map +1 -1
- package/dist/agents/react/runners/granite/prompts.d.cts +1 -1
- package/dist/agents/react/runners/granite/prompts.d.ts +1 -1
- package/dist/agents/react/runners/granite/runner.cjs +4 -2
- package/dist/agents/react/runners/granite/runner.cjs.map +1 -1
- package/dist/agents/react/runners/granite/runner.d.cts +6 -5
- package/dist/agents/react/runners/granite/runner.d.ts +6 -5
- package/dist/agents/react/runners/granite/runner.js +4 -2
- package/dist/agents/react/runners/granite/runner.js.map +1 -1
- package/dist/agents/react/types.d.cts +3 -2
- package/dist/agents/react/types.d.ts +3 -2
- package/dist/agents/toolCalling/agent.cjs +9 -5
- package/dist/agents/toolCalling/agent.cjs.map +1 -1
- package/dist/agents/toolCalling/agent.d.cts +3 -3
- package/dist/agents/toolCalling/agent.d.ts +3 -3
- package/dist/agents/toolCalling/agent.js +9 -5
- package/dist/agents/toolCalling/agent.js.map +1 -1
- package/dist/agents/toolCalling/types.d.cts +1 -1
- package/dist/agents/toolCalling/types.d.ts +1 -1
- package/dist/agents/types.d.cts +2 -2
- package/dist/agents/types.d.ts +2 -2
- package/dist/backend/backend.d.cts +2 -2
- package/dist/backend/backend.d.ts +2 -2
- package/dist/backend/chat.cjs +1 -1
- package/dist/backend/chat.cjs.map +1 -1
- package/dist/backend/chat.d.cts +2 -2
- package/dist/backend/chat.d.ts +2 -2
- package/dist/backend/chat.js +1 -1
- package/dist/backend/chat.js.map +1 -1
- package/dist/backend/core.d.cts +2 -2
- package/dist/backend/core.d.ts +2 -2
- package/dist/backend/embedding.d.cts +2 -2
- package/dist/backend/embedding.d.ts +2 -2
- package/dist/backend/message.cjs +59 -3
- package/dist/backend/message.cjs.map +1 -1
- package/dist/backend/message.d.cts +3 -0
- package/dist/backend/message.d.ts +3 -0
- package/dist/backend/message.js +59 -3
- package/dist/backend/message.js.map +1 -1
- package/dist/backend/utils.d.cts +2 -2
- package/dist/backend/utils.d.ts +2 -2
- package/dist/{base-Bguef5Dg.d.ts → base-BQgDlcGa.d.ts} +1 -1
- package/dist/{base-C1hO5EJe.d.cts → base-CwcxwybB.d.cts} +1 -1
- package/dist/cache/fileCache.cjs.map +1 -1
- package/dist/cache/fileCache.js.map +1 -1
- package/dist/{chat-BBoOSvzm.d.ts → chat-C0s-o6ll.d.ts} +4 -2
- package/dist/{chat-BZ55YQab.d.cts → chat-CRb3vUVg.d.cts} +4 -2
- package/dist/context.d.cts +1 -1
- package/dist/context.d.ts +1 -1
- package/dist/emitter/emitter.d.cts +1 -1
- package/dist/emitter/emitter.d.ts +1 -1
- package/dist/emitter/types.d.cts +1 -1
- package/dist/emitter/types.d.ts +1 -1
- package/dist/{emitter-BqpLJQVb.d.ts → emitter-CZFbzlUi.d.ts} +3 -3
- package/dist/{emitter-pJzHC_AM.d.cts → emitter-DpqUYjXH.d.cts} +3 -3
- package/dist/index.d.cts +3 -3
- package/dist/index.d.ts +3 -3
- package/dist/internals/fetcher.d.cts +1 -1
- package/dist/internals/fetcher.d.ts +1 -1
- package/dist/internals/helpers/array.cjs +35 -0
- package/dist/internals/helpers/array.cjs.map +1 -1
- package/dist/internals/helpers/array.d.cts +6 -1
- package/dist/internals/helpers/array.d.ts +6 -1
- package/dist/internals/helpers/array.js +35 -1
- package/dist/internals/helpers/array.js.map +1 -1
- package/dist/internals/helpers/object.cjs +30 -5
- package/dist/internals/helpers/object.cjs.map +1 -1
- package/dist/internals/helpers/object.d.cts +3 -2
- package/dist/internals/helpers/object.d.ts +3 -2
- package/dist/internals/helpers/object.js +29 -5
- package/dist/internals/helpers/object.js.map +1 -1
- package/dist/logger/logger.cjs +8 -4
- package/dist/logger/logger.cjs.map +1 -1
- package/dist/logger/logger.d.cts +21 -15
- package/dist/logger/logger.d.ts +21 -15
- package/dist/logger/logger.js +6 -3
- package/dist/logger/logger.js.map +1 -1
- package/dist/logger/pretty.cjs +2 -2
- package/dist/logger/pretty.cjs.map +1 -1
- package/dist/logger/pretty.js +2 -2
- package/dist/logger/pretty.js.map +1 -1
- package/dist/memory/summarizeMemory.d.cts +2 -2
- package/dist/memory/summarizeMemory.d.ts +2 -2
- package/dist/parsers/linePrefix.cjs.map +1 -1
- package/dist/parsers/linePrefix.d.cts +4 -8
- package/dist/parsers/linePrefix.d.ts +4 -8
- package/dist/parsers/linePrefix.js.map +1 -1
- package/dist/serializer/utils.cjs.map +1 -1
- package/dist/serializer/utils.js.map +1 -1
- package/dist/tools/arxiv.cjs.map +1 -1
- package/dist/tools/arxiv.d.cts +1 -1
- package/dist/tools/arxiv.d.ts +1 -1
- package/dist/tools/arxiv.js.map +1 -1
- package/dist/tools/base.cjs.map +1 -1
- package/dist/tools/base.d.cts +1 -1
- package/dist/tools/base.d.ts +1 -1
- package/dist/tools/base.js.map +1 -1
- package/dist/tools/calculator.d.cts +1 -1
- package/dist/tools/calculator.d.ts +1 -1
- package/dist/tools/custom.d.cts +2 -2
- package/dist/tools/custom.d.ts +2 -2
- package/dist/tools/database/elasticsearch.cjs.map +1 -1
- package/dist/tools/database/elasticsearch.d.cts +1 -1
- package/dist/tools/database/elasticsearch.d.ts +1 -1
- package/dist/tools/database/elasticsearch.js.map +1 -1
- package/dist/tools/database/milvus.cjs +3 -3
- package/dist/tools/database/milvus.cjs.map +1 -1
- package/dist/tools/database/milvus.d.cts +1 -1
- package/dist/tools/database/milvus.d.ts +1 -1
- package/dist/tools/database/milvus.js +3 -3
- package/dist/tools/database/milvus.js.map +1 -1
- package/dist/tools/database/qdrant.cjs +2 -2
- package/dist/tools/database/qdrant.cjs.map +1 -1
- package/dist/tools/database/qdrant.d.cts +1 -1
- package/dist/tools/database/qdrant.d.ts +1 -1
- package/dist/tools/database/qdrant.js +2 -2
- package/dist/tools/database/qdrant.js.map +1 -1
- package/dist/tools/database/sql.cjs.map +1 -1
- package/dist/tools/database/sql.d.cts +1 -1
- package/dist/tools/database/sql.d.ts +1 -1
- package/dist/tools/database/sql.js.map +1 -1
- package/dist/tools/llm.d.cts +2 -2
- package/dist/tools/llm.d.ts +2 -2
- package/dist/tools/mcp.d.cts +1 -1
- package/dist/tools/mcp.d.ts +1 -1
- package/dist/tools/openapi.d.cts +1 -1
- package/dist/tools/openapi.d.ts +1 -1
- package/dist/tools/python/output.d.cts +1 -1
- package/dist/tools/python/output.d.ts +1 -1
- package/dist/tools/python/python.d.cts +2 -2
- package/dist/tools/python/python.d.ts +2 -2
- package/dist/tools/search/base.cjs.map +1 -1
- package/dist/tools/search/base.d.cts +1 -1
- package/dist/tools/search/base.d.ts +1 -1
- package/dist/tools/search/base.js.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.cjs.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.cts +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.ts +1 -1
- package/dist/tools/search/duckDuckGoSearch.js.map +1 -1
- package/dist/tools/search/googleSearch.cjs.map +1 -1
- package/dist/tools/search/googleSearch.d.cts +1 -1
- package/dist/tools/search/googleSearch.d.ts +1 -1
- package/dist/tools/search/googleSearch.js.map +1 -1
- package/dist/tools/search/searXNGSearch.d.cts +1 -1
- package/dist/tools/search/searXNGSearch.d.ts +1 -1
- package/dist/tools/search/wikipedia.cjs +25 -0
- package/dist/tools/search/wikipedia.cjs.map +1 -1
- package/dist/tools/search/wikipedia.d.cts +2 -3
- package/dist/tools/search/wikipedia.d.ts +2 -3
- package/dist/tools/search/wikipedia.js +23 -0
- package/dist/tools/search/wikipedia.js.map +1 -1
- package/dist/tools/similarity.d.cts +1 -1
- package/dist/tools/similarity.d.ts +1 -1
- package/dist/tools/weather/openMeteo.cjs.map +1 -1
- package/dist/tools/weather/openMeteo.d.cts +1 -1
- package/dist/tools/weather/openMeteo.d.ts +1 -1
- package/dist/tools/weather/openMeteo.js.map +1 -1
- package/dist/tools/web/webCrawler.cjs.map +1 -1
- package/dist/tools/web/webCrawler.d.cts +1 -1
- package/dist/tools/web/webCrawler.d.ts +1 -1
- package/dist/tools/web/webCrawler.js.map +1 -1
- package/dist/version.cjs +1 -1
- package/dist/version.js +1 -1
- package/dist/workflows/agent.d.cts +3 -3
- package/dist/workflows/agent.d.ts +3 -3
- package/dist/workflows/workflow.d.cts +1 -1
- package/dist/workflows/workflow.d.ts +1 -1
- package/package.json +68 -67
- package/dist/adapters/beeai_platform/agents/agent.cjs.map +0 -1
- package/dist/adapters/beeai_platform/agents/agent.js.map +0 -1
- package/dist/adapters/beeai_platform/agents/events.d.cts +0 -23
- package/dist/adapters/beeai_platform/agents/events.d.ts +0 -23
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js.map +0 -0
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.js';
|
|
2
2
|
import { RunContext } from '../../../context.js';
|
|
3
|
-
import { E as Emitter } from '../../../emitter-
|
|
3
|
+
import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
|
|
4
4
|
import { Embeddings } from '@langchain/core/embeddings';
|
|
5
5
|
import '../../../internals/serializable.js';
|
|
6
6
|
import '../../../internals/types.js';
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../errors.js';
|
|
9
|
-
import '../../../chat-
|
|
9
|
+
import '../../../chat-C0s-o6ll.js';
|
|
10
10
|
import '../../../backend/message.js';
|
|
11
11
|
import 'ai';
|
|
12
12
|
import 'promise-based-task';
|
|
@@ -35,7 +35,7 @@ declare class LangChainEmbeddingModel extends EmbeddingModel {
|
|
|
35
35
|
get providerId(): string;
|
|
36
36
|
protected _create(input: EmbeddingModelInput, run: RunContext<this>): Promise<EmbeddingModelOutput>;
|
|
37
37
|
createSnapshot(): {
|
|
38
|
-
lcEmbedding: Embeddings
|
|
38
|
+
lcEmbedding: Embeddings<number[]>;
|
|
39
39
|
emitter: Emitter<EmbeddingModelEvents>;
|
|
40
40
|
};
|
|
41
41
|
loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void;
|
|
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
|
|
|
4
4
|
import { GetRunContext } from '../../context.cjs';
|
|
5
5
|
import { RunnableConfig } from '@langchain/core/runnables';
|
|
6
6
|
import * as LCTools from '@langchain/core/tools';
|
|
7
|
-
import { E as Emitter } from '../../emitter-
|
|
7
|
+
import { E as Emitter } from '../../emitter-DpqUYjXH.cjs';
|
|
8
8
|
import '../../internals/serializable.cjs';
|
|
9
9
|
import '../../internals/types.cjs';
|
|
10
10
|
import '../../internals/helpers/guards.cjs';
|
|
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
|
|
|
4
4
|
import { GetRunContext } from '../../context.js';
|
|
5
5
|
import { RunnableConfig } from '@langchain/core/runnables';
|
|
6
6
|
import * as LCTools from '@langchain/core/tools';
|
|
7
|
-
import { E as Emitter } from '../../emitter-
|
|
7
|
+
import { E as Emitter } from '../../emitter-CZFbzlUi.js';
|
|
8
8
|
import '../../internals/serializable.js';
|
|
9
9
|
import '../../internals/types.js';
|
|
10
10
|
import '../../internals/helpers/guards.js';
|
|
@@ -65,10 +65,18 @@ class MCPServer extends server_cjs.Server {
|
|
|
65
65
|
throw new errors_cjs.ValueError("Input type is not supported by this server.");
|
|
66
66
|
}
|
|
67
67
|
}
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
68
|
+
switch (this.config.transport) {
|
|
69
|
+
case "stdio":
|
|
70
|
+
await this.server.connect(new stdio_js.StdioServerTransport());
|
|
71
|
+
break;
|
|
72
|
+
case "sse":
|
|
73
|
+
case "streamable-http":
|
|
74
|
+
http_server_cjs.runServer(this.server, this.config.hostname, this.config.port);
|
|
75
|
+
break;
|
|
76
|
+
default: {
|
|
77
|
+
this.config.transport;
|
|
78
|
+
throw new Error("Unsupported transport type.");
|
|
79
|
+
}
|
|
72
80
|
}
|
|
73
81
|
}
|
|
74
82
|
getFactory(member) {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/mcp/serve/server.ts"],"names":["MCPServerConfig","transport","hostname","port","name","version","settings","partial","Object","assign","MCPServer","Server","server","config","McpServer","serve","member","members","factory","getFactory","entry","type","tool","description","paramsSchema","callback","prompt","argsSchema","resource","uri","template","ValueError","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/mcp/serve/server.ts"],"names":["MCPServerConfig","transport","hostname","port","name","version","settings","partial","Object","assign","MCPServer","Server","server","config","McpServer","serve","member","members","factory","getFactory","entry","type","tool","description","paramsSchema","callback","prompt","argsSchema","resource","uri","template","ValueError","connect","StdioServerTransport","runServer","Error","factories","has","Tool","get","toolFactory","schema","inputSchema","ZodType","shape","args","result","run","content","text","getTextContent","registerFactory"],"mappings":";;;;;;;;;;;;AA8DO,MAAMA,eAAAA,CAAAA;EA9Db;;;EA+DEC,SAAAA,GAAiD,OAAA;EACjDC,QAAAA,GAAW,WAAA;EACXC,IAAAA,GAAO,GAAA;EACPC,IAAAA,GAAO,YAAA;EACPC,OAAAA,GAAU,OAAA;AACVC,EAAAA,QAAAA;AAEA,EAAA,WAAA,CAAYC,OAAAA,EAAoC;AAC9C,IAAA,IAAIA,OAAAA,EAAS;AACXC,MAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,OAAAA,CAAAA;AACtB;AACF;AACF;AAEO,MAAMG,kBAAkBC,iBAAAA,CAAAA;EA7E/B;;;AA8EYC,EAAAA,MAAAA;AAEV,EAAA,WAAA,CAAYC,MAAAA,EAA0B;AACpC,IAAA,KAAA,CAAMA,MAAAA,IAAU,IAAIb,eAAAA,EAAAA,CAAAA;AACpB,IAAA,IAAA,CAAKY,MAAAA,GAAS,IAAIE,gBAAAA,CAAU;AAC1BV,MAAAA,IAAAA,EAAM,KAAKS,MAAAA,CAAOT,IAAAA;AAClBC,MAAAA,OAAAA,EAAS,KAAKQ,MAAAA,CAAOR,OAAAA;AACrB,MAAA,GAAG,KAAKQ,MAAAA,CAAOP;KACjB,CAAA;AACF;AAEA,EAAA,MAAMS,KAAAA,GAAQ;AACZ,IAAA,KAAA,MAAWC,MAAAA,IAAU,KAAKC,OAAAA,EAAS;AACjC,MAAA,MAAMC,OAAAA,GAAU,IAAA,CAAKC,UAAAA,CAAWH,MAAAA,CAAAA;AAChC,MAAA,MAAMI,KAAAA,GAAQ,MAAMF,OAAAA,CAAQF,MAAAA,CAAAA;AAE5B,MAAA,QAAQI,MAAMC,IAAAA;QACZ,KAAK,MAAA;AACH,UAAA,IAAA,CAAKT,MAAAA,CAAOU,KAAKF,KAAAA,CAAMhB,IAAAA,EAAMgB,MAAMG,WAAAA,EAAaH,KAAAA,CAAMI,YAAAA,EAAcJ,KAAAA,CAAMK,QAAQ,CAAA;AAClF,UAAA;QACF,KAAK,QAAA;AACH,UAAA,IAAI,gBAAgBL,KAAAA,EAAO;AACzB,YAAA,IAAA,CAAKR,MAAAA,CAAOc,OAAON,KAAAA,CAAMhB,IAAAA,EAAMgB,MAAMG,WAAAA,EAAaH,KAAAA,CAAMO,UAAAA,EAAYP,KAAAA,CAAMK,QAAQ,CAAA;WACpF,MAAO;AACL,YAAA,IAAA,CAAKb,OAAOc,MAAAA,CAAON,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMG,WAAAA,EAAaH,MAAMK,QAAQ,CAAA;AAClE;AACA,UAAA;QACF,KAAK,UAAA;AACH,UAAA,IAAI,SAASL,KAAAA,EAAO;AAClB,YAAA,IAAA,CAAKR,OAAOgB,QAAAA,CAASR,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMS,GAAAA,EAAKT,MAAMK,QAAQ,CAAA;WAC5D,MAAO;AACL,YAAA,IAAA,CAAKb,OAAOgB,QAAAA,CAASR,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMU,QAAAA,EAAUV,MAAMK,QAAQ,CAAA;AACjE;AACA,UAAA;AACF,QAAA;AACE,UAAA,MAAM,IAAIM,sBAAW,6CAAA,CAAA;AACzB;AACF;AAEA,IAAA,QAAQ,IAAA,CAAKlB,OAAOZ,SAAAA;MAClB,KAAK,OAAA;AACH,QAAA,MAAM,IAAA,CAAKW,MAAAA,CAAOoB,OAAAA,CAAQ,IAAIC,+BAAAA,CAAAA;AAC9B,QAAA;MACF,KAAK,KAAA;MACL,KAAK,iBAAA;AACHC,QAAAA,yBAAAA,CAAU,KAAKtB,MAAAA,EAAQ,IAAA,CAAKC,OAAOX,QAAAA,EAAU,IAAA,CAAKW,OAAOV,IAAI,CAAA;AAC7D,QAAA;MACF,SAAS;AAEP,QAAgC,KAAKU,MAAAA,CAAOZ;AAC5C,QAAA,MAAM,IAAIkC,MAAM,6BAAA,CAAA;AAClB;AACF;AACF;AAEAhB,EAAAA,UAAAA,CAAWH,MAAAA,EAAa;AACtB,IAAA,MAAMoB,SAAAA,GAAa,KAAK,WAAA,CAA8BA,SAAAA;AACtD,IAAA,OAAO,CAACA,SAAAA,CAAUC,GAAAA,CAAIrB,OAAO,WAAW,CAAA,IAAKA,kBAAkBsB,aAAAA,IAAQF,SAAAA,CAAUC,GAAAA,CAAIC,aAAAA,IACjFF,SAAAA,CAAUG,GAAAA,CAAID,aAAAA,CAAAA,GACd,KAAA,CAAMnB,WAAWH,MAAAA,CAAAA;AACvB;AACF;AAEA,eAAewB,YAAYlB,IAAAA,EAAa;AACtC,EAAA,MAAMmB,MAAAA,GAAS,MAAMnB,IAAAA,CAAKoB,WAAAA,EAAW;AACrC,EAAA,IAAI,EAAED,kBAAkBE,WAAAA,CAAAA,EAAU;AAChC,IAAA,MAAM,IAAIZ,sBAAW,4CAAA,CAAA;AACvB;AACA,EAAA,MAAMP,eAAeiB,MAAAA,CAAOG,KAAAA;AAC5B,EAAA,OAAO;IACLvB,IAAAA,EAAM,MAAA;AACNjB,IAAAA,IAAAA,EAAMkB,IAAAA,CAAKlB,IAAAA;AACXmB,IAAAA,WAAAA,EAAaD,IAAAA,CAAKC,WAAAA;AAClBC,IAAAA,YAAAA;AACAC,IAAAA,QAAAA,mCAAoBoB,IAAAA,KAAAA;AAClB,MAAA,MAAMC,MAAAA,GAAS,MAAMxB,IAAAA,CAAKyB,GAAAA,CAAG,GAAIF,IAAAA,CAAAA;AACjC,MAAA,OAAO;QACLG,OAAAA,EAAS;AACP,UAAA;YACE3B,IAAAA,EAAM,MAAA;AACN4B,YAAAA,IAAAA,EAAMH,OAAOI,cAAAA;AACf;;AAEJ,OAAA;KACF,EAVU,UAAA;AAWZ,GAAA;AACF;AAvBeV,MAAAA,CAAAA,WAAAA,EAAAA,aAAAA,CAAAA;AAyBf9B,SAAAA,CAAUyC,eAAAA,CAAgBb,eAAME,WAAAA,CAAAA","file":"server.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ValueError } from \"@/errors.js\";\nimport { Server } from \"@/serve/server.js\";\nimport { ServerOptions } from \"@modelcontextprotocol/sdk/server/index.js\";\nimport {\n McpServer,\n PromptCallback,\n ReadResourceCallback,\n ReadResourceTemplateCallback,\n ResourceTemplate,\n ToolCallback,\n} from \"@modelcontextprotocol/sdk/server/mcp.js\";\nimport { StdioServerTransport } from \"@modelcontextprotocol/sdk/server/stdio.js\";\nimport { AnyTool, Tool } from \"@/tools/base.js\";\nimport { runServer } from \"./http_server.js\";\nimport { ZodRawShape, ZodType } from \"zod\";\nimport { ToolAnnotations } from \"@modelcontextprotocol/sdk/types.js\";\n\ntype MCPServerPrompt =\n | {\n type: \"prompt\";\n name: string;\n description: string;\n callback: PromptCallback;\n }\n | {\n type: \"prompt\";\n name: string;\n description: string;\n argsSchema: ZodRawShape;\n callback: PromptCallback<ZodRawShape>;\n };\n\ntype MCPServerResource =\n | {\n type: \"resource\";\n name: string;\n uri: string;\n callback: ReadResourceCallback;\n }\n | {\n type: \"resource\";\n name: string;\n template: ResourceTemplate;\n callback: ReadResourceTemplateCallback;\n };\n\ninterface MCPServerTool {\n type: \"tool\";\n name: string;\n description: string;\n paramsSchema: ZodRawShape | ToolAnnotations;\n callback: ToolCallback<ZodRawShape>;\n}\n\ntype MCPServerEntry = MCPServerPrompt | MCPServerResource | MCPServerTool;\n\n// Configuration for the MCPServer.\nexport class MCPServerConfig {\n transport: \"stdio\" | \"sse\" | \"streamable-http\" = \"stdio\";\n hostname = \"127.0.0.1\";\n port = 8000;\n name = \"MCP Server\";\n version = \"1.0.0\";\n settings?: ServerOptions;\n\n constructor(partial?: Partial<MCPServerConfig>) {\n if (partial) {\n Object.assign(this, partial);\n }\n }\n}\n\nexport class MCPServer extends Server<any, MCPServerEntry, MCPServerConfig, never> {\n protected server: McpServer;\n\n constructor(config?: MCPServerConfig) {\n super(config || new MCPServerConfig());\n this.server = new McpServer({\n name: this.config.name,\n version: this.config.version,\n ...this.config.settings,\n });\n }\n\n async serve() {\n for (const member of this.members) {\n const factory = this.getFactory(member);\n const entry = await factory(member);\n\n switch (entry.type) {\n case \"tool\":\n this.server.tool(entry.name, entry.description, entry.paramsSchema, entry.callback);\n break;\n case \"prompt\":\n if (\"argsSchema\" in entry) {\n this.server.prompt(entry.name, entry.description, entry.argsSchema, entry.callback);\n } else {\n this.server.prompt(entry.name, entry.description, entry.callback);\n }\n break;\n case \"resource\":\n if (\"uri\" in entry) {\n this.server.resource(entry.name, entry.uri, entry.callback);\n } else {\n this.server.resource(entry.name, entry.template, entry.callback);\n }\n break;\n default:\n throw new ValueError(\"Input type is not supported by this server.\");\n }\n }\n\n switch (this.config.transport) {\n case \"stdio\":\n await this.server.connect(new StdioServerTransport());\n break;\n case \"sse\":\n case \"streamable-http\":\n runServer(this.server, this.config.hostname, this.config.port);\n break;\n default: {\n // The following line ensures exhaustiveness checking at compile time.\n const _exhaustiveCheck: never = this.config.transport;\n throw new Error(\"Unsupported transport type.\");\n }\n }\n }\n\n getFactory(member: any) {\n const factories = (this.constructor as typeof Server).factories;\n return !factories.has(member.constructor) && member instanceof Tool && factories.has(Tool)\n ? factories.get(Tool)!\n : super.getFactory(member);\n }\n}\n\nasync function toolFactory(tool: AnyTool): Promise<MCPServerEntry> {\n const schema = await tool.inputSchema();\n if (!(schema instanceof ZodType)) {\n throw new ValueError(\"JsonSchema is not supported for MCP tools.\");\n }\n const paramsSchema = schema.shape;\n return {\n type: \"tool\",\n name: tool.name,\n description: tool.description,\n paramsSchema: paramsSchema,\n callback: async (...args: Parameters<typeof tool.run>) => {\n const result = await tool.run(...args);\n return {\n content: [\n {\n type: \"text\",\n text: result.getTextContent(),\n },\n ],\n };\n },\n };\n}\n\nMCPServer.registerFactory(Tool, toolFactory);\n"]}
|
|
@@ -41,7 +41,7 @@ interface MCPServerTool {
|
|
|
41
41
|
}
|
|
42
42
|
type MCPServerEntry = MCPServerPrompt | MCPServerResource | MCPServerTool;
|
|
43
43
|
declare class MCPServerConfig {
|
|
44
|
-
transport: "stdio" | "sse";
|
|
44
|
+
transport: "stdio" | "sse" | "streamable-http";
|
|
45
45
|
hostname: string;
|
|
46
46
|
port: number;
|
|
47
47
|
name: string;
|
|
@@ -41,7 +41,7 @@ interface MCPServerTool {
|
|
|
41
41
|
}
|
|
42
42
|
type MCPServerEntry = MCPServerPrompt | MCPServerResource | MCPServerTool;
|
|
43
43
|
declare class MCPServerConfig {
|
|
44
|
-
transport: "stdio" | "sse";
|
|
44
|
+
transport: "stdio" | "sse" | "streamable-http";
|
|
45
45
|
hostname: string;
|
|
46
46
|
port: number;
|
|
47
47
|
name: string;
|
|
@@ -63,10 +63,18 @@ class MCPServer extends Server {
|
|
|
63
63
|
throw new ValueError("Input type is not supported by this server.");
|
|
64
64
|
}
|
|
65
65
|
}
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
66
|
+
switch (this.config.transport) {
|
|
67
|
+
case "stdio":
|
|
68
|
+
await this.server.connect(new StdioServerTransport());
|
|
69
|
+
break;
|
|
70
|
+
case "sse":
|
|
71
|
+
case "streamable-http":
|
|
72
|
+
runServer(this.server, this.config.hostname, this.config.port);
|
|
73
|
+
break;
|
|
74
|
+
default: {
|
|
75
|
+
this.config.transport;
|
|
76
|
+
throw new Error("Unsupported transport type.");
|
|
77
|
+
}
|
|
70
78
|
}
|
|
71
79
|
}
|
|
72
80
|
getFactory(member) {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/mcp/serve/server.ts"],"names":["MCPServerConfig","transport","hostname","port","name","version","settings","partial","Object","assign","MCPServer","Server","server","config","McpServer","serve","member","members","factory","getFactory","entry","type","tool","description","paramsSchema","callback","prompt","argsSchema","resource","uri","template","ValueError","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/mcp/serve/server.ts"],"names":["MCPServerConfig","transport","hostname","port","name","version","settings","partial","Object","assign","MCPServer","Server","server","config","McpServer","serve","member","members","factory","getFactory","entry","type","tool","description","paramsSchema","callback","prompt","argsSchema","resource","uri","template","ValueError","connect","StdioServerTransport","runServer","Error","factories","has","Tool","get","toolFactory","schema","inputSchema","ZodType","shape","args","result","run","content","text","getTextContent","registerFactory"],"mappings":";;;;;;;;;;AA8DO,MAAMA,eAAAA,CAAAA;EA9Db;;;EA+DEC,SAAAA,GAAiD,OAAA;EACjDC,QAAAA,GAAW,WAAA;EACXC,IAAAA,GAAO,GAAA;EACPC,IAAAA,GAAO,YAAA;EACPC,OAAAA,GAAU,OAAA;AACVC,EAAAA,QAAAA;AAEA,EAAA,WAAA,CAAYC,OAAAA,EAAoC;AAC9C,IAAA,IAAIA,OAAAA,EAAS;AACXC,MAAAA,MAAAA,CAAOC,MAAAA,CAAO,MAAMF,OAAAA,CAAAA;AACtB;AACF;AACF;AAEO,MAAMG,kBAAkBC,MAAAA,CAAAA;EA7E/B;;;AA8EYC,EAAAA,MAAAA;AAEV,EAAA,WAAA,CAAYC,MAAAA,EAA0B;AACpC,IAAA,KAAA,CAAMA,MAAAA,IAAU,IAAIb,eAAAA,EAAAA,CAAAA;AACpB,IAAA,IAAA,CAAKY,MAAAA,GAAS,IAAIE,SAAAA,CAAU;AAC1BV,MAAAA,IAAAA,EAAM,KAAKS,MAAAA,CAAOT,IAAAA;AAClBC,MAAAA,OAAAA,EAAS,KAAKQ,MAAAA,CAAOR,OAAAA;AACrB,MAAA,GAAG,KAAKQ,MAAAA,CAAOP;KACjB,CAAA;AACF;AAEA,EAAA,MAAMS,KAAAA,GAAQ;AACZ,IAAA,KAAA,MAAWC,MAAAA,IAAU,KAAKC,OAAAA,EAAS;AACjC,MAAA,MAAMC,OAAAA,GAAU,IAAA,CAAKC,UAAAA,CAAWH,MAAAA,CAAAA;AAChC,MAAA,MAAMI,KAAAA,GAAQ,MAAMF,OAAAA,CAAQF,MAAAA,CAAAA;AAE5B,MAAA,QAAQI,MAAMC,IAAAA;QACZ,KAAK,MAAA;AACH,UAAA,IAAA,CAAKT,MAAAA,CAAOU,KAAKF,KAAAA,CAAMhB,IAAAA,EAAMgB,MAAMG,WAAAA,EAAaH,KAAAA,CAAMI,YAAAA,EAAcJ,KAAAA,CAAMK,QAAQ,CAAA;AAClF,UAAA;QACF,KAAK,QAAA;AACH,UAAA,IAAI,gBAAgBL,KAAAA,EAAO;AACzB,YAAA,IAAA,CAAKR,MAAAA,CAAOc,OAAON,KAAAA,CAAMhB,IAAAA,EAAMgB,MAAMG,WAAAA,EAAaH,KAAAA,CAAMO,UAAAA,EAAYP,KAAAA,CAAMK,QAAQ,CAAA;WACpF,MAAO;AACL,YAAA,IAAA,CAAKb,OAAOc,MAAAA,CAAON,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMG,WAAAA,EAAaH,MAAMK,QAAQ,CAAA;AAClE;AACA,UAAA;QACF,KAAK,UAAA;AACH,UAAA,IAAI,SAASL,KAAAA,EAAO;AAClB,YAAA,IAAA,CAAKR,OAAOgB,QAAAA,CAASR,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMS,GAAAA,EAAKT,MAAMK,QAAQ,CAAA;WAC5D,MAAO;AACL,YAAA,IAAA,CAAKb,OAAOgB,QAAAA,CAASR,KAAAA,CAAMhB,MAAMgB,KAAAA,CAAMU,QAAAA,EAAUV,MAAMK,QAAQ,CAAA;AACjE;AACA,UAAA;AACF,QAAA;AACE,UAAA,MAAM,IAAIM,WAAW,6CAAA,CAAA;AACzB;AACF;AAEA,IAAA,QAAQ,IAAA,CAAKlB,OAAOZ,SAAAA;MAClB,KAAK,OAAA;AACH,QAAA,MAAM,IAAA,CAAKW,MAAAA,CAAOoB,OAAAA,CAAQ,IAAIC,sBAAAA,CAAAA;AAC9B,QAAA;MACF,KAAK,KAAA;MACL,KAAK,iBAAA;AACHC,QAAAA,SAAAA,CAAU,KAAKtB,MAAAA,EAAQ,IAAA,CAAKC,OAAOX,QAAAA,EAAU,IAAA,CAAKW,OAAOV,IAAI,CAAA;AAC7D,QAAA;MACF,SAAS;AAEP,QAAgC,KAAKU,MAAAA,CAAOZ;AAC5C,QAAA,MAAM,IAAIkC,MAAM,6BAAA,CAAA;AAClB;AACF;AACF;AAEAhB,EAAAA,UAAAA,CAAWH,MAAAA,EAAa;AACtB,IAAA,MAAMoB,SAAAA,GAAa,KAAK,WAAA,CAA8BA,SAAAA;AACtD,IAAA,OAAO,CAACA,SAAAA,CAAUC,GAAAA,CAAIrB,OAAO,WAAW,CAAA,IAAKA,kBAAkBsB,IAAAA,IAAQF,SAAAA,CAAUC,GAAAA,CAAIC,IAAAA,IACjFF,SAAAA,CAAUG,GAAAA,CAAID,IAAAA,CAAAA,GACd,KAAA,CAAMnB,WAAWH,MAAAA,CAAAA;AACvB;AACF;AAEA,eAAewB,YAAYlB,IAAAA,EAAa;AACtC,EAAA,MAAMmB,MAAAA,GAAS,MAAMnB,IAAAA,CAAKoB,WAAAA,EAAW;AACrC,EAAA,IAAI,EAAED,kBAAkBE,OAAAA,CAAAA,EAAU;AAChC,IAAA,MAAM,IAAIZ,WAAW,4CAAA,CAAA;AACvB;AACA,EAAA,MAAMP,eAAeiB,MAAAA,CAAOG,KAAAA;AAC5B,EAAA,OAAO;IACLvB,IAAAA,EAAM,MAAA;AACNjB,IAAAA,IAAAA,EAAMkB,IAAAA,CAAKlB,IAAAA;AACXmB,IAAAA,WAAAA,EAAaD,IAAAA,CAAKC,WAAAA;AAClBC,IAAAA,YAAAA;AACAC,IAAAA,QAAAA,mCAAoBoB,IAAAA,KAAAA;AAClB,MAAA,MAAMC,MAAAA,GAAS,MAAMxB,IAAAA,CAAKyB,GAAAA,CAAG,GAAIF,IAAAA,CAAAA;AACjC,MAAA,OAAO;QACLG,OAAAA,EAAS;AACP,UAAA;YACE3B,IAAAA,EAAM,MAAA;AACN4B,YAAAA,IAAAA,EAAMH,OAAOI,cAAAA;AACf;;AAEJ,OAAA;KACF,EAVU,UAAA;AAWZ,GAAA;AACF;AAvBeV,MAAAA,CAAAA,WAAAA,EAAAA,aAAAA,CAAAA;AAyBf9B,SAAAA,CAAUyC,eAAAA,CAAgBb,MAAME,WAAAA,CAAAA","file":"server.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { ValueError } from \"@/errors.js\";\nimport { Server } from \"@/serve/server.js\";\nimport { ServerOptions } from \"@modelcontextprotocol/sdk/server/index.js\";\nimport {\n McpServer,\n PromptCallback,\n ReadResourceCallback,\n ReadResourceTemplateCallback,\n ResourceTemplate,\n ToolCallback,\n} from \"@modelcontextprotocol/sdk/server/mcp.js\";\nimport { StdioServerTransport } from \"@modelcontextprotocol/sdk/server/stdio.js\";\nimport { AnyTool, Tool } from \"@/tools/base.js\";\nimport { runServer } from \"./http_server.js\";\nimport { ZodRawShape, ZodType } from \"zod\";\nimport { ToolAnnotations } from \"@modelcontextprotocol/sdk/types.js\";\n\ntype MCPServerPrompt =\n | {\n type: \"prompt\";\n name: string;\n description: string;\n callback: PromptCallback;\n }\n | {\n type: \"prompt\";\n name: string;\n description: string;\n argsSchema: ZodRawShape;\n callback: PromptCallback<ZodRawShape>;\n };\n\ntype MCPServerResource =\n | {\n type: \"resource\";\n name: string;\n uri: string;\n callback: ReadResourceCallback;\n }\n | {\n type: \"resource\";\n name: string;\n template: ResourceTemplate;\n callback: ReadResourceTemplateCallback;\n };\n\ninterface MCPServerTool {\n type: \"tool\";\n name: string;\n description: string;\n paramsSchema: ZodRawShape | ToolAnnotations;\n callback: ToolCallback<ZodRawShape>;\n}\n\ntype MCPServerEntry = MCPServerPrompt | MCPServerResource | MCPServerTool;\n\n// Configuration for the MCPServer.\nexport class MCPServerConfig {\n transport: \"stdio\" | \"sse\" | \"streamable-http\" = \"stdio\";\n hostname = \"127.0.0.1\";\n port = 8000;\n name = \"MCP Server\";\n version = \"1.0.0\";\n settings?: ServerOptions;\n\n constructor(partial?: Partial<MCPServerConfig>) {\n if (partial) {\n Object.assign(this, partial);\n }\n }\n}\n\nexport class MCPServer extends Server<any, MCPServerEntry, MCPServerConfig, never> {\n protected server: McpServer;\n\n constructor(config?: MCPServerConfig) {\n super(config || new MCPServerConfig());\n this.server = new McpServer({\n name: this.config.name,\n version: this.config.version,\n ...this.config.settings,\n });\n }\n\n async serve() {\n for (const member of this.members) {\n const factory = this.getFactory(member);\n const entry = await factory(member);\n\n switch (entry.type) {\n case \"tool\":\n this.server.tool(entry.name, entry.description, entry.paramsSchema, entry.callback);\n break;\n case \"prompt\":\n if (\"argsSchema\" in entry) {\n this.server.prompt(entry.name, entry.description, entry.argsSchema, entry.callback);\n } else {\n this.server.prompt(entry.name, entry.description, entry.callback);\n }\n break;\n case \"resource\":\n if (\"uri\" in entry) {\n this.server.resource(entry.name, entry.uri, entry.callback);\n } else {\n this.server.resource(entry.name, entry.template, entry.callback);\n }\n break;\n default:\n throw new ValueError(\"Input type is not supported by this server.\");\n }\n }\n\n switch (this.config.transport) {\n case \"stdio\":\n await this.server.connect(new StdioServerTransport());\n break;\n case \"sse\":\n case \"streamable-http\":\n runServer(this.server, this.config.hostname, this.config.port);\n break;\n default: {\n // The following line ensures exhaustiveness checking at compile time.\n const _exhaustiveCheck: never = this.config.transport;\n throw new Error(\"Unsupported transport type.\");\n }\n }\n }\n\n getFactory(member: any) {\n const factories = (this.constructor as typeof Server).factories;\n return !factories.has(member.constructor) && member instanceof Tool && factories.has(Tool)\n ? factories.get(Tool)!\n : super.getFactory(member);\n }\n}\n\nasync function toolFactory(tool: AnyTool): Promise<MCPServerEntry> {\n const schema = await tool.inputSchema();\n if (!(schema instanceof ZodType)) {\n throw new ValueError(\"JsonSchema is not supported for MCP tools.\");\n }\n const paramsSchema = schema.shape;\n return {\n type: \"tool\",\n name: tool.name,\n description: tool.description,\n paramsSchema: paramsSchema,\n callback: async (...args: Parameters<typeof tool.run>) => {\n const result = await tool.run(...args);\n return {\n content: [\n {\n type: \"text\",\n text: result.getTextContent(),\n },\n ],\n };\n },\n };\n}\n\nMCPServer.registerFactory(Tool, toolFactory);\n"]}
|
|
@@ -16,10 +16,7 @@ class OllamaChatModel extends chat_cjs.VercelChatModel {
|
|
|
16
16
|
"auto"
|
|
17
17
|
];
|
|
18
18
|
constructor(modelId = env_cjs.getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
|
|
19
|
-
const model = client_cjs.OllamaClient.ensure(client).instance.chat(modelId,
|
|
20
|
-
...settings,
|
|
21
|
-
structuredOutputs: true
|
|
22
|
-
});
|
|
19
|
+
const model = client_cjs.OllamaClient.ensure(client).instance.chat(modelId, settings);
|
|
23
20
|
super(model);
|
|
24
21
|
}
|
|
25
22
|
static {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;;;AAeO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
2
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
2
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { OllamaClient, OllamaClientSettings } from './client.cjs';
|
|
4
|
-
import { l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
4
|
+
import { l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
|
|
5
5
|
import '../../../logger/logger.cjs';
|
|
6
6
|
import 'pino';
|
|
7
7
|
import '../../../errors.cjs';
|
|
@@ -9,7 +9,7 @@ import '../../../internals/types.cjs';
|
|
|
9
9
|
import '../../../internals/helpers/guards.cjs';
|
|
10
10
|
import '../../../internals/serializable.cjs';
|
|
11
11
|
import 'ai';
|
|
12
|
-
import '../../../emitter-
|
|
12
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
13
13
|
import '../../../backend/message.cjs';
|
|
14
14
|
import '../../../context.cjs';
|
|
15
15
|
import '../../../internals/helpers/promise.cjs';
|
|
@@ -29,7 +29,7 @@ import '../../../backend/client.cjs';
|
|
|
29
29
|
* SPDX-License-Identifier: Apache-2.0
|
|
30
30
|
*/
|
|
31
31
|
|
|
32
|
-
type OllamaParameters = Parameters<OllamaProvider["
|
|
32
|
+
type OllamaParameters = Parameters<OllamaProvider["chat"]>;
|
|
33
33
|
type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
|
|
34
34
|
type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
|
|
35
35
|
declare class OllamaChatModel extends VercelChatModel {
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
2
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
2
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { OllamaClient, OllamaClientSettings } from './client.js';
|
|
4
|
-
import { l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
4
|
+
import { l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
|
|
5
5
|
import '../../../logger/logger.js';
|
|
6
6
|
import 'pino';
|
|
7
7
|
import '../../../errors.js';
|
|
@@ -9,7 +9,7 @@ import '../../../internals/types.js';
|
|
|
9
9
|
import '../../../internals/helpers/guards.js';
|
|
10
10
|
import '../../../internals/serializable.js';
|
|
11
11
|
import 'ai';
|
|
12
|
-
import '../../../emitter-
|
|
12
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
13
13
|
import '../../../backend/message.js';
|
|
14
14
|
import '../../../context.js';
|
|
15
15
|
import '../../../internals/helpers/promise.js';
|
|
@@ -29,7 +29,7 @@ import '../../../backend/client.js';
|
|
|
29
29
|
* SPDX-License-Identifier: Apache-2.0
|
|
30
30
|
*/
|
|
31
31
|
|
|
32
|
-
type OllamaParameters = Parameters<OllamaProvider["
|
|
32
|
+
type OllamaParameters = Parameters<OllamaProvider["chat"]>;
|
|
33
33
|
type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
|
|
34
34
|
type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
|
|
35
35
|
declare class OllamaChatModel extends VercelChatModel {
|
|
@@ -14,10 +14,7 @@ class OllamaChatModel extends VercelChatModel {
|
|
|
14
14
|
"auto"
|
|
15
15
|
];
|
|
16
16
|
constructor(modelId = getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
|
|
17
|
-
const model = OllamaClient.ensure(client).instance.chat(modelId,
|
|
18
|
-
...settings,
|
|
19
|
-
structuredOutputs: true
|
|
20
|
-
});
|
|
17
|
+
const model = OllamaClient.ensure(client).instance.chat(modelId, settings);
|
|
21
18
|
super(model);
|
|
22
19
|
}
|
|
23
20
|
static {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;AAeO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
'use strict';
|
|
2
2
|
|
|
3
3
|
var env_cjs = require('../../../internals/env.cjs');
|
|
4
|
-
var
|
|
4
|
+
var ollamaAiProviderV2 = require('ollama-ai-provider-v2');
|
|
5
5
|
var client_cjs = require('../../../backend/client.cjs');
|
|
6
6
|
var utils_cjs = require('../../vercel/backend/utils.cjs');
|
|
7
7
|
|
|
@@ -14,7 +14,7 @@ class OllamaClient extends client_cjs.BackendClient {
|
|
|
14
14
|
create() {
|
|
15
15
|
const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};
|
|
16
16
|
const apiKey = _apiKey || env_cjs.getEnv("OLLAMA_API_KEY");
|
|
17
|
-
return
|
|
17
|
+
return ollamaAiProviderV2.createOllama({
|
|
18
18
|
...settings,
|
|
19
19
|
baseURL: baseURL || env_cjs.getEnv("OLLAMA_BASE_URL"),
|
|
20
20
|
fetch: utils_cjs.vercelFetcher(this.settings?.fetch),
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;;;AAYO,MAAMA,qBAAqBC,wBAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,cAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;;;AAYO,MAAMA,qBAAqBC,wBAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,cAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,+BAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,eAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,uBAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,8BAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider-v2\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { BackendClient } from '../../../backend/client.cjs';
|
|
3
3
|
import '../../../internals/serializable.cjs';
|
|
4
4
|
import '../../../internals/types.cjs';
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { BackendClient } from '../../../backend/client.js';
|
|
3
3
|
import '../../../internals/serializable.js';
|
|
4
4
|
import '../../../internals/types.js';
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { getEnv } from '../../../internals/env.js';
|
|
2
|
-
import { createOllama } from 'ollama-ai-provider';
|
|
2
|
+
import { createOllama } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { BackendClient } from '../../../backend/client.js';
|
|
4
4
|
import { parseHeadersFromEnv, vercelFetcher } from '../../vercel/backend/utils.js';
|
|
5
5
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;AAYO,MAAMA,qBAAqBC,aAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,MAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,YAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,OAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,aAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,oBAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;AAYO,MAAMA,qBAAqBC,aAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,MAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,YAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,OAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,aAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,oBAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider-v2\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;;;AAcO,MAAMA,6BAA6BC,kCAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,eAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;;;AAcO,MAAMA,6BAA6BC,kCAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,eAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { OllamaClient, OllamaClientSettings } from './client.cjs';
|
|
3
3
|
import { VercelEmbeddingModel } from '../../vercel/backend/embedding.cjs';
|
|
4
4
|
import '../../../backend/client.cjs';
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
|
|
|
7
7
|
import '../../../internals/helpers/guards.cjs';
|
|
8
8
|
import '../../../backend/embedding.cjs';
|
|
9
9
|
import '../../../context.cjs';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
11
11
|
import '../../../internals/helpers/promise.cjs';
|
|
12
12
|
import '../../../errors.cjs';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
14
14
|
import '../../../backend/message.cjs';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { OllamaClient, OllamaClientSettings } from './client.js';
|
|
3
3
|
import { VercelEmbeddingModel } from '../../vercel/backend/embedding.js';
|
|
4
4
|
import '../../../backend/client.js';
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
|
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../backend/embedding.js';
|
|
9
9
|
import '../../../context.js';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
11
11
|
import '../../../internals/helpers/promise.js';
|
|
12
12
|
import '../../../errors.js';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-C0s-o6ll.js';
|
|
14
14
|
import '../../../backend/message.js';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;AAcO,MAAMA,6BAA6BC,oBAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,OAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;AAcO,MAAMA,6BAA6BC,oBAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,OAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
@@ -10,9 +10,10 @@ class OpenAIChatModel extends chat_cjs.VercelChatModel {
|
|
|
10
10
|
static {
|
|
11
11
|
__name(this, "OpenAIChatModel");
|
|
12
12
|
}
|
|
13
|
-
constructor(modelId = env_cjs.getEnv("OPENAI_CHAT_MODEL", "gpt-4o"),
|
|
14
|
-
const model = client_cjs.OpenAIClient.ensure(client).instance.chat(modelId
|
|
13
|
+
constructor(modelId = env_cjs.getEnv("OPENAI_CHAT_MODEL", "gpt-4o"), parameters = {}, client) {
|
|
14
|
+
const model = client_cjs.OpenAIClient.ensure(client).instance.chat(modelId);
|
|
15
15
|
super(model);
|
|
16
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
16
17
|
}
|
|
17
18
|
}
|
|
18
19
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","OpenAIClient","ensure","instance","chat","Object","assign"],"mappings":";;;;;;;;AAcO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAdrC;;;EAeE,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,QAAA,GACzDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,uBAAAA,CAAaC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,KAAKR,OAAAA,CAAAA;AACxD,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OpenAIProvider } from \"@ai-sdk/openai\";\nimport { OpenAIClient, OpenAIClientSettings } from \"@/adapters/openai/backend/client.js\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype OpenAIParameters = Parameters<OpenAIProvider[\"chat\"]>;\nexport type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;\n\nexport class OpenAIChatModel extends VercelChatModel {\n constructor(\n modelId: OpenAIChatModelId = getEnv(\"OPENAI_CHAT_MODEL\", \"gpt-4o\"),\n parameters: ChatModelParameters = {},\n client?: OpenAIClient | OpenAIClientSettings,\n ) {\n const model = OpenAIClient.ensure(client).instance.chat(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import { OpenAIProvider } from '@ai-sdk/openai';
|
|
2
2
|
import { OpenAIClient, OpenAIClientSettings } from './client.cjs';
|
|
3
3
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
|
|
4
5
|
import '../../../backend/client.cjs';
|
|
5
6
|
import '../../../internals/serializable.cjs';
|
|
6
7
|
import '../../../internals/types.cjs';
|
|
@@ -8,11 +9,10 @@ import '../../../internals/helpers/guards.cjs';
|
|
|
8
9
|
import '../../../logger/logger.cjs';
|
|
9
10
|
import 'pino';
|
|
10
11
|
import '../../../errors.cjs';
|
|
11
|
-
import '../../../chat-BZ55YQab.cjs';
|
|
12
|
-
import '../../../backend/message.cjs';
|
|
13
12
|
import 'ai';
|
|
13
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
14
|
+
import '../../../backend/message.cjs';
|
|
14
15
|
import '../../../context.cjs';
|
|
15
|
-
import '../../../emitter-pJzHC_AM.cjs';
|
|
16
16
|
import '../../../internals/helpers/promise.cjs';
|
|
17
17
|
import 'promise-based-task';
|
|
18
18
|
import '../../../cache/base.cjs';
|
|
@@ -31,9 +31,8 @@ import '../../../template.cjs';
|
|
|
31
31
|
|
|
32
32
|
type OpenAIParameters = Parameters<OpenAIProvider["chat"]>;
|
|
33
33
|
type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;
|
|
34
|
-
type OpenAIChatModelSettings = NonNullable<OpenAIParameters[1]>;
|
|
35
34
|
declare class OpenAIChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: OpenAIChatModelId,
|
|
35
|
+
constructor(modelId?: OpenAIChatModelId, parameters?: ChatModelParameters, client?: OpenAIClient | OpenAIClientSettings);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { OpenAIChatModel, type OpenAIChatModelId
|
|
38
|
+
export { OpenAIChatModel, type OpenAIChatModelId };
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import { OpenAIProvider } from '@ai-sdk/openai';
|
|
2
2
|
import { OpenAIClient, OpenAIClientSettings } from './client.js';
|
|
3
3
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
|
|
4
5
|
import '../../../backend/client.js';
|
|
5
6
|
import '../../../internals/serializable.js';
|
|
6
7
|
import '../../../internals/types.js';
|
|
@@ -8,11 +9,10 @@ import '../../../internals/helpers/guards.js';
|
|
|
8
9
|
import '../../../logger/logger.js';
|
|
9
10
|
import 'pino';
|
|
10
11
|
import '../../../errors.js';
|
|
11
|
-
import '../../../chat-BBoOSvzm.js';
|
|
12
|
-
import '../../../backend/message.js';
|
|
13
12
|
import 'ai';
|
|
13
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
14
|
+
import '../../../backend/message.js';
|
|
14
15
|
import '../../../context.js';
|
|
15
|
-
import '../../../emitter-BqpLJQVb.js';
|
|
16
16
|
import '../../../internals/helpers/promise.js';
|
|
17
17
|
import 'promise-based-task';
|
|
18
18
|
import '../../../cache/base.js';
|
|
@@ -31,9 +31,8 @@ import '../../../template.js';
|
|
|
31
31
|
|
|
32
32
|
type OpenAIParameters = Parameters<OpenAIProvider["chat"]>;
|
|
33
33
|
type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;
|
|
34
|
-
type OpenAIChatModelSettings = NonNullable<OpenAIParameters[1]>;
|
|
35
34
|
declare class OpenAIChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: OpenAIChatModelId,
|
|
35
|
+
constructor(modelId?: OpenAIChatModelId, parameters?: ChatModelParameters, client?: OpenAIClient | OpenAIClientSettings);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { OpenAIChatModel, type OpenAIChatModelId
|
|
38
|
+
export { OpenAIChatModel, type OpenAIChatModelId };
|
|
@@ -8,9 +8,10 @@ class OpenAIChatModel extends VercelChatModel {
|
|
|
8
8
|
static {
|
|
9
9
|
__name(this, "OpenAIChatModel");
|
|
10
10
|
}
|
|
11
|
-
constructor(modelId = getEnv("OPENAI_CHAT_MODEL", "gpt-4o"),
|
|
12
|
-
const model = OpenAIClient.ensure(client).instance.chat(modelId
|
|
11
|
+
constructor(modelId = getEnv("OPENAI_CHAT_MODEL", "gpt-4o"), parameters = {}, client) {
|
|
12
|
+
const model = OpenAIClient.ensure(client).instance.chat(modelId);
|
|
13
13
|
super(model);
|
|
14
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
14
15
|
}
|
|
15
16
|
}
|
|
16
17
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","OpenAIClient","ensure","instance","chat","Object","assign"],"mappings":";;;;;;AAcO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAdrC;;;EAeE,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,QAAA,GACzDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,YAAAA,CAAaC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,KAAKR,OAAAA,CAAAA;AACxD,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OpenAIProvider } from \"@ai-sdk/openai\";\nimport { OpenAIClient, OpenAIClientSettings } from \"@/adapters/openai/backend/client.js\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype OpenAIParameters = Parameters<OpenAIProvider[\"chat\"]>;\nexport type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;\n\nexport class OpenAIChatModel extends VercelChatModel {\n constructor(\n modelId: OpenAIChatModelId = getEnv(\"OPENAI_CHAT_MODEL\", \"gpt-4o\"),\n parameters: ChatModelParameters = {},\n client?: OpenAIClient | OpenAIClientSettings,\n ) {\n const model = OpenAIClient.ensure(client).instance.chat(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -12,21 +12,12 @@ class OpenAIClient extends client_cjs.BackendClient {
|
|
|
12
12
|
__name(this, "OpenAIClient");
|
|
13
13
|
}
|
|
14
14
|
create() {
|
|
15
|
-
const extraHeaders = utils_cjs.parseHeadersFromEnv("OPENAI_API_HEADERS");
|
|
16
|
-
const baseURL = this.settings?.baseURL || env_cjs.getEnv("OPENAI_API_ENDPOINT");
|
|
17
|
-
let compatibility = this.settings?.compatibility || env_cjs.getEnv("OPENAI_COMPATIBILITY_MODE");
|
|
18
|
-
if (baseURL && !compatibility) {
|
|
19
|
-
compatibility = "compatible";
|
|
20
|
-
} else if (!baseURL && !compatibility) {
|
|
21
|
-
compatibility = "strict";
|
|
22
|
-
}
|
|
23
15
|
return openai.createOpenAI({
|
|
24
16
|
...this.settings,
|
|
25
|
-
compatibility,
|
|
26
17
|
apiKey: this.settings?.apiKey || env_cjs.getEnv("OPENAI_API_KEY"),
|
|
27
|
-
baseURL,
|
|
18
|
+
baseURL: this.settings?.baseURL || env_cjs.getEnv("OPENAI_API_ENDPOINT"),
|
|
28
19
|
headers: {
|
|
29
|
-
...
|
|
20
|
+
...utils_cjs.parseHeadersFromEnv("OPENAI_API_HEADERS"),
|
|
30
21
|
...this.settings?.headers
|
|
31
22
|
},
|
|
32
23
|
fetch: utils_cjs.vercelFetcher(this.settings?.fetch)
|