beeai-framework 0.1.20 → 0.1.21
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapters/a2a/agents/agent.cjs +4 -4
- package/dist/adapters/a2a/agents/agent.cjs.map +1 -1
- package/dist/adapters/a2a/agents/agent.d.cts +2 -2
- package/dist/adapters/a2a/agents/agent.d.ts +2 -2
- package/dist/adapters/a2a/agents/agent.js +4 -4
- package/dist/adapters/a2a/agents/agent.js.map +1 -1
- package/dist/adapters/a2a/agents/events.d.cts +1 -1
- package/dist/adapters/a2a/agents/events.d.ts +1 -1
- package/dist/adapters/a2a/agents/types.d.cts +2 -2
- package/dist/adapters/a2a/agents/types.d.ts +2 -2
- package/dist/adapters/a2a/agents/utils.cjs +2 -2
- package/dist/adapters/a2a/agents/utils.cjs.map +1 -1
- package/dist/adapters/a2a/agents/utils.js +2 -2
- package/dist/adapters/a2a/agents/utils.js.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.cjs +1 -3
- package/dist/adapters/a2a/serve/agent_executor.cjs.map +1 -1
- package/dist/adapters/a2a/serve/agent_executor.d.cts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.d.ts +4 -3
- package/dist/adapters/a2a/serve/agent_executor.js +1 -3
- package/dist/adapters/a2a/serve/agent_executor.js.map +1 -1
- package/dist/adapters/a2a/serve/server.cjs +15 -0
- package/dist/adapters/a2a/serve/server.cjs.map +1 -1
- package/dist/adapters/a2a/serve/server.d.cts +4 -2
- package/dist/adapters/a2a/serve/server.d.ts +4 -2
- package/dist/adapters/a2a/serve/server.js +15 -0
- package/dist/adapters/a2a/serve/server.js.map +1 -1
- package/dist/adapters/acp/agents/agent.d.cts +2 -2
- package/dist/adapters/acp/agents/agent.d.ts +2 -2
- package/dist/adapters/acp/agents/events.d.cts +1 -1
- package/dist/adapters/acp/agents/events.d.ts +1 -1
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.cjs +4 -4
- package/dist/adapters/agentstack/agents/agent.cjs.map +1 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.cts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.d.ts +12 -12
- package/dist/adapters/{beeai_platform → agentstack}/agents/agent.js +4 -4
- package/dist/adapters/agentstack/agents/agent.js.map +1 -0
- package/dist/adapters/agentstack/agents/events.d.cts +23 -0
- package/dist/adapters/agentstack/agents/events.d.ts +23 -0
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.cts +4 -4
- package/dist/adapters/{beeai_platform → agentstack}/agents/types.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/chat.cjs +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/chat.d.cts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.d.ts +5 -6
- package/dist/adapters/amazon-bedrock/backend/chat.js +3 -2
- package/dist/adapters/amazon-bedrock/backend/chat.js.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.cjs.map +1 -1
- package/dist/adapters/amazon-bedrock/backend/embedding.d.cts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.d.ts +4 -4
- package/dist/adapters/amazon-bedrock/backend/embedding.js +2 -2
- package/dist/adapters/amazon-bedrock/backend/embedding.js.map +1 -1
- package/dist/adapters/anthropic/backend/chat.cjs +3 -2
- package/dist/adapters/anthropic/backend/chat.cjs.map +1 -1
- package/dist/adapters/anthropic/backend/chat.d.cts +5 -6
- package/dist/adapters/anthropic/backend/chat.d.ts +5 -6
- package/dist/adapters/anthropic/backend/chat.js +3 -2
- package/dist/adapters/anthropic/backend/chat.js.map +1 -1
- package/dist/adapters/anthropic/backend/embedding.d.cts +2 -2
- package/dist/adapters/anthropic/backend/embedding.d.ts +2 -2
- package/dist/adapters/azure-openai/backend/chat.cjs +3 -2
- package/dist/adapters/azure-openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/chat.d.cts +5 -6
- package/dist/adapters/azure-openai/backend/chat.d.ts +5 -6
- package/dist/adapters/azure-openai/backend/chat.js +3 -2
- package/dist/adapters/azure-openai/backend/chat.js.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.cjs +2 -2
- package/dist/adapters/azure-openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/azure-openai/backend/embedding.d.cts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.d.ts +3 -3
- package/dist/adapters/azure-openai/backend/embedding.js +2 -2
- package/dist/adapters/azure-openai/backend/embedding.js.map +1 -1
- package/dist/adapters/dummy/backend/chat.d.cts +2 -2
- package/dist/adapters/dummy/backend/chat.d.ts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.cts +2 -2
- package/dist/adapters/dummy/backend/embedding.d.ts +2 -2
- package/dist/adapters/google-vertex/backend/chat.cjs +3 -2
- package/dist/adapters/google-vertex/backend/chat.cjs.map +1 -1
- package/dist/adapters/google-vertex/backend/chat.d.cts +5 -6
- package/dist/adapters/google-vertex/backend/chat.d.ts +5 -6
- package/dist/adapters/google-vertex/backend/chat.js +3 -2
- package/dist/adapters/google-vertex/backend/chat.js.map +1 -1
- package/dist/adapters/google-vertex/backend/embedding.d.cts +2 -2
- package/dist/adapters/google-vertex/backend/embedding.d.ts +2 -2
- package/dist/adapters/groq/backend/chat.cjs +3 -2
- package/dist/adapters/groq/backend/chat.cjs.map +1 -1
- package/dist/adapters/groq/backend/chat.d.cts +5 -6
- package/dist/adapters/groq/backend/chat.d.ts +5 -6
- package/dist/adapters/groq/backend/chat.js +3 -2
- package/dist/adapters/groq/backend/chat.js.map +1 -1
- package/dist/adapters/groq/backend/embedding.d.cts +2 -2
- package/dist/adapters/groq/backend/embedding.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.cjs +6 -5
- package/dist/adapters/langchain/backend/chat.cjs.map +1 -1
- package/dist/adapters/langchain/backend/chat.d.cts +2 -2
- package/dist/adapters/langchain/backend/chat.d.ts +2 -2
- package/dist/adapters/langchain/backend/chat.js +6 -5
- package/dist/adapters/langchain/backend/chat.js.map +1 -1
- package/dist/adapters/langchain/backend/embedding.d.cts +3 -3
- package/dist/adapters/langchain/backend/embedding.d.ts +3 -3
- package/dist/adapters/langchain/tools.d.cts +1 -1
- package/dist/adapters/langchain/tools.d.ts +1 -1
- package/dist/adapters/ollama/backend/chat.cjs +1 -4
- package/dist/adapters/ollama/backend/chat.cjs.map +1 -1
- package/dist/adapters/ollama/backend/chat.d.cts +4 -4
- package/dist/adapters/ollama/backend/chat.d.ts +4 -4
- package/dist/adapters/ollama/backend/chat.js +1 -4
- package/dist/adapters/ollama/backend/chat.js.map +1 -1
- package/dist/adapters/ollama/backend/client.cjs +2 -2
- package/dist/adapters/ollama/backend/client.cjs.map +1 -1
- package/dist/adapters/ollama/backend/client.d.cts +1 -1
- package/dist/adapters/ollama/backend/client.d.ts +1 -1
- package/dist/adapters/ollama/backend/client.js +1 -1
- package/dist/adapters/ollama/backend/client.js.map +1 -1
- package/dist/adapters/ollama/backend/embedding.cjs.map +1 -1
- package/dist/adapters/ollama/backend/embedding.d.cts +3 -3
- package/dist/adapters/ollama/backend/embedding.d.ts +3 -3
- package/dist/adapters/ollama/backend/embedding.js.map +1 -1
- package/dist/adapters/openai/backend/chat.cjs +3 -2
- package/dist/adapters/openai/backend/chat.cjs.map +1 -1
- package/dist/adapters/openai/backend/chat.d.cts +5 -6
- package/dist/adapters/openai/backend/chat.d.ts +5 -6
- package/dist/adapters/openai/backend/chat.js +3 -2
- package/dist/adapters/openai/backend/chat.js.map +1 -1
- package/dist/adapters/openai/backend/client.cjs +2 -11
- package/dist/adapters/openai/backend/client.cjs.map +1 -1
- package/dist/adapters/openai/backend/client.js +3 -12
- package/dist/adapters/openai/backend/client.js.map +1 -1
- package/dist/adapters/openai/backend/embedding.cjs +2 -2
- package/dist/adapters/openai/backend/embedding.cjs.map +1 -1
- package/dist/adapters/openai/backend/embedding.d.cts +4 -4
- package/dist/adapters/openai/backend/embedding.d.ts +4 -4
- package/dist/adapters/openai/backend/embedding.js +2 -2
- package/dist/adapters/openai/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/chat.cjs +26 -21
- package/dist/adapters/vercel/backend/chat.cjs.map +1 -1
- package/dist/adapters/vercel/backend/chat.d.cts +6 -4
- package/dist/adapters/vercel/backend/chat.d.ts +6 -4
- package/dist/adapters/vercel/backend/chat.js +27 -22
- package/dist/adapters/vercel/backend/chat.js.map +1 -1
- package/dist/adapters/vercel/backend/embedding.cjs.map +1 -1
- package/dist/adapters/vercel/backend/embedding.d.cts +3 -3
- package/dist/adapters/vercel/backend/embedding.d.ts +3 -3
- package/dist/adapters/vercel/backend/embedding.js.map +1 -1
- package/dist/adapters/vercel/backend/utils.cjs +11 -0
- package/dist/adapters/vercel/backend/utils.cjs.map +1 -1
- package/dist/adapters/vercel/backend/utils.d.cts +19 -2
- package/dist/adapters/vercel/backend/utils.d.ts +19 -2
- package/dist/adapters/vercel/backend/utils.js +11 -1
- package/dist/adapters/vercel/backend/utils.js.map +1 -1
- package/dist/adapters/watsonx/backend/chat.cjs +10 -4
- package/dist/adapters/watsonx/backend/chat.cjs.map +1 -1
- package/dist/adapters/watsonx/backend/chat.d.cts +2 -2
- package/dist/adapters/watsonx/backend/chat.d.ts +2 -2
- package/dist/adapters/watsonx/backend/chat.js +10 -4
- package/dist/adapters/watsonx/backend/chat.js.map +1 -1
- package/dist/adapters/watsonx/backend/embedding.d.cts +2 -2
- package/dist/adapters/watsonx/backend/embedding.d.ts +2 -2
- package/dist/adapters/xai/backend/chat.cjs +3 -2
- package/dist/adapters/xai/backend/chat.cjs.map +1 -1
- package/dist/adapters/xai/backend/chat.d.cts +5 -6
- package/dist/adapters/xai/backend/chat.d.ts +5 -6
- package/dist/adapters/xai/backend/chat.js +3 -2
- package/dist/adapters/xai/backend/chat.js.map +1 -1
- package/dist/{agent-CtrCeN7o.d.ts → agent-CHebzsQB.d.ts} +3 -3
- package/dist/{agent-I0Saf4A1.d.cts → agent-DvvVGvLr.d.cts} +3 -3
- package/dist/agents/base.d.cts +2 -2
- package/dist/agents/base.d.ts +2 -2
- package/dist/agents/experimental/replan/agent.d.cts +3 -3
- package/dist/agents/experimental/replan/agent.d.ts +3 -3
- package/dist/agents/experimental/replan/prompts.d.cts +1 -1
- package/dist/agents/experimental/replan/prompts.d.ts +1 -1
- package/dist/agents/experimental/streamlit/agent.d.cts +3 -3
- package/dist/agents/experimental/streamlit/agent.d.ts +3 -3
- package/dist/agents/react/agent.d.cts +5 -4
- package/dist/agents/react/agent.d.ts +5 -4
- package/dist/agents/react/runners/base.cjs.map +1 -1
- package/dist/agents/react/runners/base.d.cts +5 -4
- package/dist/agents/react/runners/base.d.ts +5 -4
- package/dist/agents/react/runners/base.js.map +1 -1
- package/dist/agents/react/runners/deep-think/runner.d.cts +5 -4
- package/dist/agents/react/runners/deep-think/runner.d.ts +5 -4
- package/dist/agents/react/runners/default/runner.cjs.map +1 -1
- package/dist/agents/react/runners/default/runner.d.cts +5 -4
- package/dist/agents/react/runners/default/runner.d.ts +5 -4
- package/dist/agents/react/runners/default/runner.js.map +1 -1
- package/dist/agents/react/runners/granite/runner.cjs +4 -2
- package/dist/agents/react/runners/granite/runner.cjs.map +1 -1
- package/dist/agents/react/runners/granite/runner.d.cts +5 -4
- package/dist/agents/react/runners/granite/runner.d.ts +5 -4
- package/dist/agents/react/runners/granite/runner.js +4 -2
- package/dist/agents/react/runners/granite/runner.js.map +1 -1
- package/dist/agents/react/types.d.cts +3 -2
- package/dist/agents/react/types.d.ts +3 -2
- package/dist/agents/toolCalling/agent.cjs +9 -5
- package/dist/agents/toolCalling/agent.cjs.map +1 -1
- package/dist/agents/toolCalling/agent.d.cts +3 -3
- package/dist/agents/toolCalling/agent.d.ts +3 -3
- package/dist/agents/toolCalling/agent.js +9 -5
- package/dist/agents/toolCalling/agent.js.map +1 -1
- package/dist/agents/toolCalling/types.d.cts +1 -1
- package/dist/agents/toolCalling/types.d.ts +1 -1
- package/dist/agents/types.d.cts +2 -2
- package/dist/agents/types.d.ts +2 -2
- package/dist/backend/backend.d.cts +2 -2
- package/dist/backend/backend.d.ts +2 -2
- package/dist/backend/chat.cjs +1 -1
- package/dist/backend/chat.cjs.map +1 -1
- package/dist/backend/chat.d.cts +2 -2
- package/dist/backend/chat.d.ts +2 -2
- package/dist/backend/chat.js +1 -1
- package/dist/backend/chat.js.map +1 -1
- package/dist/backend/core.d.cts +2 -2
- package/dist/backend/core.d.ts +2 -2
- package/dist/backend/embedding.d.cts +2 -2
- package/dist/backend/embedding.d.ts +2 -2
- package/dist/backend/message.cjs +59 -3
- package/dist/backend/message.cjs.map +1 -1
- package/dist/backend/message.d.cts +3 -0
- package/dist/backend/message.d.ts +3 -0
- package/dist/backend/message.js +59 -3
- package/dist/backend/message.js.map +1 -1
- package/dist/backend/utils.d.cts +2 -2
- package/dist/backend/utils.d.ts +2 -2
- package/dist/{base-BsY2baWy.d.ts → base-BQgDlcGa.d.ts} +1 -1
- package/dist/{base-BP7v-NSL.d.cts → base-CwcxwybB.d.cts} +1 -1
- package/dist/cache/fileCache.cjs.map +1 -1
- package/dist/cache/fileCache.js.map +1 -1
- package/dist/{chat-W9j8T6UU.d.ts → chat-C0s-o6ll.d.ts} +4 -2
- package/dist/{chat-CR4yfOR9.d.cts → chat-CRb3vUVg.d.cts} +4 -2
- package/dist/context.d.cts +1 -1
- package/dist/context.d.ts +1 -1
- package/dist/emitter/emitter.d.cts +1 -1
- package/dist/emitter/emitter.d.ts +1 -1
- package/dist/emitter/types.d.cts +1 -1
- package/dist/emitter/types.d.ts +1 -1
- package/dist/{emitter-C9nwGLt6.d.ts → emitter-CZFbzlUi.d.ts} +3 -3
- package/dist/{emitter-DWEsd-ud.d.cts → emitter-DpqUYjXH.d.cts} +3 -3
- package/dist/index.d.cts +3 -3
- package/dist/index.d.ts +3 -3
- package/dist/internals/fetcher.d.cts +1 -1
- package/dist/internals/fetcher.d.ts +1 -1
- package/dist/internals/helpers/array.cjs +35 -0
- package/dist/internals/helpers/array.cjs.map +1 -1
- package/dist/internals/helpers/array.d.cts +6 -1
- package/dist/internals/helpers/array.d.ts +6 -1
- package/dist/internals/helpers/array.js +35 -1
- package/dist/internals/helpers/array.js.map +1 -1
- package/dist/internals/helpers/object.cjs +30 -5
- package/dist/internals/helpers/object.cjs.map +1 -1
- package/dist/internals/helpers/object.d.cts +3 -2
- package/dist/internals/helpers/object.d.ts +3 -2
- package/dist/internals/helpers/object.js +29 -5
- package/dist/internals/helpers/object.js.map +1 -1
- package/dist/logger/logger.cjs +8 -4
- package/dist/logger/logger.cjs.map +1 -1
- package/dist/logger/logger.d.cts +21 -15
- package/dist/logger/logger.d.ts +21 -15
- package/dist/logger/logger.js +6 -3
- package/dist/logger/logger.js.map +1 -1
- package/dist/logger/pretty.cjs +2 -2
- package/dist/logger/pretty.cjs.map +1 -1
- package/dist/logger/pretty.js +2 -2
- package/dist/logger/pretty.js.map +1 -1
- package/dist/memory/slidingMemory.d.cts +1 -1
- package/dist/memory/slidingMemory.d.ts +1 -1
- package/dist/memory/summarizeMemory.d.cts +2 -2
- package/dist/memory/summarizeMemory.d.ts +2 -2
- package/dist/parsers/linePrefix.cjs.map +1 -1
- package/dist/parsers/linePrefix.d.cts +4 -8
- package/dist/parsers/linePrefix.d.ts +4 -8
- package/dist/parsers/linePrefix.js.map +1 -1
- package/dist/serializer/utils.cjs.map +1 -1
- package/dist/serializer/utils.js.map +1 -1
- package/dist/tools/arxiv.cjs.map +1 -1
- package/dist/tools/arxiv.d.cts +1 -1
- package/dist/tools/arxiv.d.ts +1 -1
- package/dist/tools/arxiv.js.map +1 -1
- package/dist/tools/base.cjs.map +1 -1
- package/dist/tools/base.d.cts +1 -1
- package/dist/tools/base.d.ts +1 -1
- package/dist/tools/base.js.map +1 -1
- package/dist/tools/calculator.d.cts +1 -1
- package/dist/tools/calculator.d.ts +1 -1
- package/dist/tools/custom.d.cts +2 -2
- package/dist/tools/custom.d.ts +2 -2
- package/dist/tools/database/elasticsearch.cjs.map +1 -1
- package/dist/tools/database/elasticsearch.d.cts +1 -1
- package/dist/tools/database/elasticsearch.d.ts +1 -1
- package/dist/tools/database/elasticsearch.js.map +1 -1
- package/dist/tools/database/milvus.cjs +3 -3
- package/dist/tools/database/milvus.cjs.map +1 -1
- package/dist/tools/database/milvus.d.cts +3 -3
- package/dist/tools/database/milvus.d.ts +3 -3
- package/dist/tools/database/milvus.js +3 -3
- package/dist/tools/database/milvus.js.map +1 -1
- package/dist/tools/database/qdrant.cjs +2 -2
- package/dist/tools/database/qdrant.cjs.map +1 -1
- package/dist/tools/database/qdrant.d.cts +1 -1
- package/dist/tools/database/qdrant.d.ts +1 -1
- package/dist/tools/database/qdrant.js +2 -2
- package/dist/tools/database/qdrant.js.map +1 -1
- package/dist/tools/database/sql.cjs.map +1 -1
- package/dist/tools/database/sql.d.cts +1 -1
- package/dist/tools/database/sql.d.ts +1 -1
- package/dist/tools/database/sql.js.map +1 -1
- package/dist/tools/llm.d.cts +2 -2
- package/dist/tools/llm.d.ts +2 -2
- package/dist/tools/mcp.d.cts +1 -1
- package/dist/tools/mcp.d.ts +1 -1
- package/dist/tools/openapi.d.cts +1 -1
- package/dist/tools/openapi.d.ts +1 -1
- package/dist/tools/python/output.d.cts +1 -1
- package/dist/tools/python/output.d.ts +1 -1
- package/dist/tools/python/python.d.cts +2 -2
- package/dist/tools/python/python.d.ts +2 -2
- package/dist/tools/search/base.cjs.map +1 -1
- package/dist/tools/search/base.d.cts +1 -1
- package/dist/tools/search/base.d.ts +1 -1
- package/dist/tools/search/base.js.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.cjs.map +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.cts +1 -1
- package/dist/tools/search/duckDuckGoSearch.d.ts +1 -1
- package/dist/tools/search/duckDuckGoSearch.js.map +1 -1
- package/dist/tools/search/googleSearch.cjs.map +1 -1
- package/dist/tools/search/googleSearch.d.cts +1 -1
- package/dist/tools/search/googleSearch.d.ts +1 -1
- package/dist/tools/search/googleSearch.js.map +1 -1
- package/dist/tools/search/searXNGSearch.d.cts +1 -1
- package/dist/tools/search/searXNGSearch.d.ts +1 -1
- package/dist/tools/search/wikipedia.cjs +25 -0
- package/dist/tools/search/wikipedia.cjs.map +1 -1
- package/dist/tools/search/wikipedia.d.cts +2 -3
- package/dist/tools/search/wikipedia.d.ts +2 -3
- package/dist/tools/search/wikipedia.js +23 -0
- package/dist/tools/search/wikipedia.js.map +1 -1
- package/dist/tools/similarity.d.cts +1 -1
- package/dist/tools/similarity.d.ts +1 -1
- package/dist/tools/weather/openMeteo.cjs.map +1 -1
- package/dist/tools/weather/openMeteo.d.cts +1 -1
- package/dist/tools/weather/openMeteo.d.ts +1 -1
- package/dist/tools/weather/openMeteo.js.map +1 -1
- package/dist/tools/web/webCrawler.cjs.map +1 -1
- package/dist/tools/web/webCrawler.d.cts +1 -1
- package/dist/tools/web/webCrawler.d.ts +1 -1
- package/dist/tools/web/webCrawler.js.map +1 -1
- package/dist/version.cjs +1 -1
- package/dist/version.js +1 -1
- package/dist/workflows/agent.d.cts +3 -3
- package/dist/workflows/agent.d.ts +3 -3
- package/dist/workflows/workflow.d.cts +1 -1
- package/dist/workflows/workflow.d.ts +1 -1
- package/package.json +68 -67
- package/dist/adapters/beeai_platform/agents/agent.cjs.map +0 -1
- package/dist/adapters/beeai_platform/agents/agent.js.map +0 -1
- package/dist/adapters/beeai_platform/agents/events.d.cts +0 -23
- package/dist/adapters/beeai_platform/agents/events.d.ts +0 -23
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/events.js.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.cjs.map +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js +0 -0
- /package/dist/adapters/{beeai_platform → agentstack}/agents/types.js.map +0 -0
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { EmbeddingModel, EmbeddingModelEvents, EmbeddingModelInput, EmbeddingModelOutput } from '../../../backend/embedding.js';
|
|
2
2
|
import { RunContext } from '../../../context.js';
|
|
3
|
-
import { E as Emitter } from '../../../emitter-
|
|
3
|
+
import { E as Emitter } from '../../../emitter-CZFbzlUi.js';
|
|
4
4
|
import { Embeddings } from '@langchain/core/embeddings';
|
|
5
5
|
import '../../../internals/serializable.js';
|
|
6
6
|
import '../../../internals/types.js';
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../errors.js';
|
|
9
|
-
import '../../../chat-
|
|
9
|
+
import '../../../chat-C0s-o6ll.js';
|
|
10
10
|
import '../../../backend/message.js';
|
|
11
11
|
import 'ai';
|
|
12
12
|
import 'promise-based-task';
|
|
@@ -35,7 +35,7 @@ declare class LangChainEmbeddingModel extends EmbeddingModel {
|
|
|
35
35
|
get providerId(): string;
|
|
36
36
|
protected _create(input: EmbeddingModelInput, run: RunContext<this>): Promise<EmbeddingModelOutput>;
|
|
37
37
|
createSnapshot(): {
|
|
38
|
-
lcEmbedding: Embeddings
|
|
38
|
+
lcEmbedding: Embeddings<number[]>;
|
|
39
39
|
emitter: Emitter<EmbeddingModelEvents>;
|
|
40
40
|
};
|
|
41
41
|
loadSnapshot(snapshot: ReturnType<typeof this.createSnapshot>): void;
|
|
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
|
|
|
4
4
|
import { GetRunContext } from '../../context.cjs';
|
|
5
5
|
import { RunnableConfig } from '@langchain/core/runnables';
|
|
6
6
|
import * as LCTools from '@langchain/core/tools';
|
|
7
|
-
import { E as Emitter } from '../../emitter-
|
|
7
|
+
import { E as Emitter } from '../../emitter-DpqUYjXH.cjs';
|
|
8
8
|
import '../../internals/serializable.cjs';
|
|
9
9
|
import '../../internals/types.cjs';
|
|
10
10
|
import '../../internals/helpers/guards.cjs';
|
|
@@ -4,7 +4,7 @@ import { BaseToolRunOptions, BaseToolOptions, JSONToolOutput, Tool, ToolEmitter,
|
|
|
4
4
|
import { GetRunContext } from '../../context.js';
|
|
5
5
|
import { RunnableConfig } from '@langchain/core/runnables';
|
|
6
6
|
import * as LCTools from '@langchain/core/tools';
|
|
7
|
-
import { E as Emitter } from '../../emitter-
|
|
7
|
+
import { E as Emitter } from '../../emitter-CZFbzlUi.js';
|
|
8
8
|
import '../../internals/serializable.js';
|
|
9
9
|
import '../../internals/types.js';
|
|
10
10
|
import '../../internals/helpers/guards.js';
|
|
@@ -16,10 +16,7 @@ class OllamaChatModel extends chat_cjs.VercelChatModel {
|
|
|
16
16
|
"auto"
|
|
17
17
|
];
|
|
18
18
|
constructor(modelId = env_cjs.getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
|
|
19
|
-
const model = client_cjs.OllamaClient.ensure(client).instance.chat(modelId,
|
|
20
|
-
...settings,
|
|
21
|
-
structuredOutputs: true
|
|
22
|
-
});
|
|
19
|
+
const model = client_cjs.OllamaClient.ensure(client).instance.chat(modelId, settings);
|
|
23
20
|
super(model);
|
|
24
21
|
}
|
|
25
22
|
static {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;;;AAeO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
2
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
2
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { OllamaClient, OllamaClientSettings } from './client.cjs';
|
|
4
|
-
import { l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
4
|
+
import { l as ChatModelToolChoiceSupport } from '../../../chat-CRb3vUVg.cjs';
|
|
5
5
|
import '../../../logger/logger.cjs';
|
|
6
6
|
import 'pino';
|
|
7
7
|
import '../../../errors.cjs';
|
|
@@ -9,7 +9,7 @@ import '../../../internals/types.cjs';
|
|
|
9
9
|
import '../../../internals/helpers/guards.cjs';
|
|
10
10
|
import '../../../internals/serializable.cjs';
|
|
11
11
|
import 'ai';
|
|
12
|
-
import '../../../emitter-
|
|
12
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
13
13
|
import '../../../backend/message.cjs';
|
|
14
14
|
import '../../../context.cjs';
|
|
15
15
|
import '../../../internals/helpers/promise.cjs';
|
|
@@ -29,7 +29,7 @@ import '../../../backend/client.cjs';
|
|
|
29
29
|
* SPDX-License-Identifier: Apache-2.0
|
|
30
30
|
*/
|
|
31
31
|
|
|
32
|
-
type OllamaParameters = Parameters<OllamaProvider["
|
|
32
|
+
type OllamaParameters = Parameters<OllamaProvider["chat"]>;
|
|
33
33
|
type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
|
|
34
34
|
type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
|
|
35
35
|
declare class OllamaChatModel extends VercelChatModel {
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
2
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
2
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { OllamaClient, OllamaClientSettings } from './client.js';
|
|
4
|
-
import { l as ChatModelToolChoiceSupport } from '../../../chat-
|
|
4
|
+
import { l as ChatModelToolChoiceSupport } from '../../../chat-C0s-o6ll.js';
|
|
5
5
|
import '../../../logger/logger.js';
|
|
6
6
|
import 'pino';
|
|
7
7
|
import '../../../errors.js';
|
|
@@ -9,7 +9,7 @@ import '../../../internals/types.js';
|
|
|
9
9
|
import '../../../internals/helpers/guards.js';
|
|
10
10
|
import '../../../internals/serializable.js';
|
|
11
11
|
import 'ai';
|
|
12
|
-
import '../../../emitter-
|
|
12
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
13
13
|
import '../../../backend/message.js';
|
|
14
14
|
import '../../../context.js';
|
|
15
15
|
import '../../../internals/helpers/promise.js';
|
|
@@ -29,7 +29,7 @@ import '../../../backend/client.js';
|
|
|
29
29
|
* SPDX-License-Identifier: Apache-2.0
|
|
30
30
|
*/
|
|
31
31
|
|
|
32
|
-
type OllamaParameters = Parameters<OllamaProvider["
|
|
32
|
+
type OllamaParameters = Parameters<OllamaProvider["chat"]>;
|
|
33
33
|
type OllamaChatModelId = NonNullable<OllamaParameters[0]>;
|
|
34
34
|
type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;
|
|
35
35
|
declare class OllamaChatModel extends VercelChatModel {
|
|
@@ -14,10 +14,7 @@ class OllamaChatModel extends VercelChatModel {
|
|
|
14
14
|
"auto"
|
|
15
15
|
];
|
|
16
16
|
constructor(modelId = getEnv("OLLAMA_CHAT_MODEL", "llama3.1:8b"), settings = {}, client) {
|
|
17
|
-
const model = OllamaClient.ensure(client).instance.chat(modelId,
|
|
18
|
-
...settings,
|
|
19
|
-
structuredOutputs: true
|
|
20
|
-
});
|
|
17
|
+
const model = OllamaClient.ensure(client).instance.chat(modelId, settings);
|
|
21
18
|
super(model);
|
|
22
19
|
}
|
|
23
20
|
static {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/chat.ts"],"names":["OllamaChatModel","VercelChatModel","supportsToolStreaming","toolChoiceSupport","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","chat","register"],"mappings":";;;;;;AAeO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAfrC;;;EAgBWC,qBAAAA,GAAwB,KAAA;EACjBC,iBAAAA,GAAkD;AAAC,IAAA,MAAA;AAAQ,IAAA;;EAE3E,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,aAAA,GACzDC,QAAAA,GAAoC,IACpCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,IAAAA,CAAKR,SAASE,QAAAA,CAAAA;AACjE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;EAEA;AACE,IAAA,IAAA,CAAKK,QAAAA,EAAQ;AACf;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelToolChoiceSupport } from \"@/backend/chat.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"chat\"]>;\nexport type OllamaChatModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaChatModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaChatModel extends VercelChatModel {\n readonly supportsToolStreaming = false;\n public readonly toolChoiceSupport: ChatModelToolChoiceSupport[] = [\"none\", \"auto\"];\n\n constructor(\n modelId: OllamaChatModelId = getEnv(\"OLLAMA_CHAT_MODEL\", \"llama3.1:8b\"),\n settings: OllamaChatModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.chat(modelId, settings);\n super(model);\n }\n\n static {\n this.register();\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
'use strict';
|
|
2
2
|
|
|
3
3
|
var env_cjs = require('../../../internals/env.cjs');
|
|
4
|
-
var
|
|
4
|
+
var ollamaAiProviderV2 = require('ollama-ai-provider-v2');
|
|
5
5
|
var client_cjs = require('../../../backend/client.cjs');
|
|
6
6
|
var utils_cjs = require('../../vercel/backend/utils.cjs');
|
|
7
7
|
|
|
@@ -14,7 +14,7 @@ class OllamaClient extends client_cjs.BackendClient {
|
|
|
14
14
|
create() {
|
|
15
15
|
const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};
|
|
16
16
|
const apiKey = _apiKey || env_cjs.getEnv("OLLAMA_API_KEY");
|
|
17
|
-
return
|
|
17
|
+
return ollamaAiProviderV2.createOllama({
|
|
18
18
|
...settings,
|
|
19
19
|
baseURL: baseURL || env_cjs.getEnv("OLLAMA_BASE_URL"),
|
|
20
20
|
fetch: utils_cjs.vercelFetcher(this.settings?.fetch),
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;;;AAYO,MAAMA,qBAAqBC,wBAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,cAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;;;AAYO,MAAMA,qBAAqBC,wBAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,cAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,+BAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,eAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,uBAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,8BAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider-v2\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { BackendClient } from '../../../backend/client.cjs';
|
|
3
3
|
import '../../../internals/serializable.cjs';
|
|
4
4
|
import '../../../internals/types.cjs';
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProviderSettings, OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { BackendClient } from '../../../backend/client.js';
|
|
3
3
|
import '../../../internals/serializable.js';
|
|
4
4
|
import '../../../internals/types.js';
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { getEnv } from '../../../internals/env.js';
|
|
2
|
-
import { createOllama } from 'ollama-ai-provider';
|
|
2
|
+
import { createOllama } from 'ollama-ai-provider-v2';
|
|
3
3
|
import { BackendClient } from '../../../backend/client.js';
|
|
4
4
|
import { parseHeadersFromEnv, vercelFetcher } from '../../vercel/backend/utils.js';
|
|
5
5
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;AAYO,MAAMA,qBAAqBC,aAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,MAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,YAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,OAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,aAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,oBAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/client.ts"],"names":["OllamaClient","BackendClient","create","apiKey","_apiKey","baseURL","headers","settings","getEnv","createOllama","fetch","vercelFetcher","parseHeadersFromEnv","Authorization"],"mappings":";;;;;;;AAYO,MAAMA,qBAAqBC,aAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,MAAM,EAAEC,MAAAA,EAAQC,OAAAA,EAASC,OAAAA,EAASC,OAAAA,EAAS,GAAGC,QAAAA,EAAAA,GAAa,IAAA,CAAKA,QAAAA,IAAY,EAAC;AAC7E,IAAA,MAAMJ,MAAAA,GAASC,OAAAA,IAAWI,MAAAA,CAAO,gBAAA,CAAA;AAEjC,IAAA,OAAOC,YAAAA,CAAa;MAClB,GAAGF,QAAAA;MACHF,OAAAA,EAASA,OAAAA,IAAWG,OAAO,iBAAA,CAAA;MAC3BE,KAAAA,EAAOC,aAAAA,CAAc,IAAA,CAAKJ,QAAAA,EAAUG,KAAAA,CAAAA;MACpCJ,OAAAA,EAAS;AACP,QAAA,GAAGM,oBAAoB,oBAAA,CAAA;QACvB,GAAGN,OAAAA;AACH,QAAA,GAAIH,MAAAA,IAAU;AAAEU,UAAAA,aAAAA,EAAe,UAAUV,MAAAA,CAAAA;AAAS;AACpD;KACF,CAAA;AACF;AACF","file":"client.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { getEnv } from \"@/internals/env.js\";\nimport { createOllama, OllamaProvider, OllamaProviderSettings } from \"ollama-ai-provider-v2\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OllamaClientSettings = OllamaProviderSettings & { apiKey?: string };\n\nexport class OllamaClient extends BackendClient<OllamaClientSettings, OllamaProvider> {\n protected create(): OllamaProvider {\n const { apiKey: _apiKey, baseURL, headers, ...settings } = this.settings ?? {};\n const apiKey = _apiKey || getEnv(\"OLLAMA_API_KEY\");\n\n return createOllama({\n ...settings,\n baseURL: baseURL || getEnv(\"OLLAMA_BASE_URL\"),\n fetch: vercelFetcher(this.settings?.fetch),\n headers: {\n ...parseHeadersFromEnv(\"OLLAMA_API_HEADERS\"),\n ...headers,\n ...(apiKey && { Authorization: `Bearer ${apiKey}` }),\n },\n });\n }\n}\n"]}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;;;AAcO,MAAMA,6BAA6BC,kCAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,eAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;;;AAcO,MAAMA,6BAA6BC,kCAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,eAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,wBAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { OllamaClient, OllamaClientSettings } from './client.cjs';
|
|
3
3
|
import { VercelEmbeddingModel } from '../../vercel/backend/embedding.cjs';
|
|
4
4
|
import '../../../backend/client.cjs';
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
|
|
|
7
7
|
import '../../../internals/helpers/guards.cjs';
|
|
8
8
|
import '../../../backend/embedding.cjs';
|
|
9
9
|
import '../../../context.cjs';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
11
11
|
import '../../../internals/helpers/promise.cjs';
|
|
12
12
|
import '../../../errors.cjs';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
14
14
|
import '../../../backend/message.cjs';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { OllamaProvider } from 'ollama-ai-provider';
|
|
1
|
+
import { OllamaProvider } from 'ollama-ai-provider-v2';
|
|
2
2
|
import { OllamaClient, OllamaClientSettings } from './client.js';
|
|
3
3
|
import { VercelEmbeddingModel } from '../../vercel/backend/embedding.js';
|
|
4
4
|
import '../../../backend/client.js';
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
|
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../backend/embedding.js';
|
|
9
9
|
import '../../../context.js';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
11
11
|
import '../../../internals/helpers/promise.js';
|
|
12
12
|
import '../../../errors.js';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-C0s-o6ll.js';
|
|
14
14
|
import '../../../backend/message.js';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;AAcO,MAAMA,6BAA6BC,oBAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,OAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/ollama/backend/embedding.ts"],"names":["OllamaEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","settings","client","model","OllamaClient","ensure","instance","embedding"],"mappings":";;;;;;AAcO,MAAMA,6BAA6BC,oBAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,OAAO,wBAAA,EAA0B,kBAAA,GACnEC,QAAAA,GAAyC,IACzCC,MAAAA,EACA;AACA,IAAA,MAAMC,KAAAA,GAAQC,aAAaC,MAAAA,CAAOH,MAAAA,EAAQI,QAAAA,CAASC,SAAAA,CAAUR,SAASE,QAAAA,CAAAA;AACtE,IAAA,KAAA,CAAME,KAAAA,CAAAA;AACR;AACF","file":"embedding.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OllamaProvider } from \"ollama-ai-provider-v2\";\nimport { OllamaClient, OllamaClientSettings } from \"@/adapters/ollama/backend/client.js\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OllamaParameters = Parameters<OllamaProvider[\"textEmbeddingModel\"]>;\nexport type OllamaEmbeddingModelId = NonNullable<OllamaParameters[0]>;\nexport type OllamaEmbeddingModelSettings = NonNullable<OllamaParameters[1]>;\n\nexport class OllamaEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OllamaEmbeddingModelId = getEnv(\"OLLAMA_EMBEDDING_MODEL\", \"nomic-embed-text\"),\n settings: OllamaEmbeddingModelSettings = {},\n client?: OllamaClient | OllamaClientSettings,\n ) {\n const model = OllamaClient.ensure(client).instance.embedding(modelId, settings);\n super(model);\n }\n}\n"]}
|
|
@@ -10,9 +10,10 @@ class OpenAIChatModel extends chat_cjs.VercelChatModel {
|
|
|
10
10
|
static {
|
|
11
11
|
__name(this, "OpenAIChatModel");
|
|
12
12
|
}
|
|
13
|
-
constructor(modelId = env_cjs.getEnv("OPENAI_CHAT_MODEL", "gpt-4o"),
|
|
14
|
-
const model = client_cjs.OpenAIClient.ensure(client).instance.chat(modelId
|
|
13
|
+
constructor(modelId = env_cjs.getEnv("OPENAI_CHAT_MODEL", "gpt-4o"), parameters = {}, client) {
|
|
14
|
+
const model = client_cjs.OpenAIClient.ensure(client).instance.chat(modelId);
|
|
15
15
|
super(model);
|
|
16
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
16
17
|
}
|
|
17
18
|
}
|
|
18
19
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","OpenAIClient","ensure","instance","chat","Object","assign"],"mappings":";;;;;;;;AAcO,MAAMA,wBAAwBC,wBAAAA,CAAAA;EAdrC;;;EAeE,WAAA,CACEC,OAAAA,GAA6BC,eAAO,mBAAA,EAAqB,QAAA,GACzDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,uBAAAA,CAAaC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,KAAKR,OAAAA,CAAAA;AACxD,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OpenAIProvider } from \"@ai-sdk/openai\";\nimport { OpenAIClient, OpenAIClientSettings } from \"@/adapters/openai/backend/client.js\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype OpenAIParameters = Parameters<OpenAIProvider[\"chat\"]>;\nexport type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;\n\nexport class OpenAIChatModel extends VercelChatModel {\n constructor(\n modelId: OpenAIChatModelId = getEnv(\"OPENAI_CHAT_MODEL\", \"gpt-4o\"),\n parameters: ChatModelParameters = {},\n client?: OpenAIClient | OpenAIClientSettings,\n ) {\n const model = OpenAIClient.ensure(client).instance.chat(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import { OpenAIProvider } from '@ai-sdk/openai';
|
|
2
2
|
import { OpenAIClient, OpenAIClientSettings } from './client.cjs';
|
|
3
3
|
import { VercelChatModel } from '../../vercel/backend/chat.cjs';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-CRb3vUVg.cjs';
|
|
4
5
|
import '../../../backend/client.cjs';
|
|
5
6
|
import '../../../internals/serializable.cjs';
|
|
6
7
|
import '../../../internals/types.cjs';
|
|
@@ -8,11 +9,10 @@ import '../../../internals/helpers/guards.cjs';
|
|
|
8
9
|
import '../../../logger/logger.cjs';
|
|
9
10
|
import 'pino';
|
|
10
11
|
import '../../../errors.cjs';
|
|
11
|
-
import '../../../chat-CR4yfOR9.cjs';
|
|
12
|
-
import '../../../backend/message.cjs';
|
|
13
12
|
import 'ai';
|
|
13
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
14
|
+
import '../../../backend/message.cjs';
|
|
14
15
|
import '../../../context.cjs';
|
|
15
|
-
import '../../../emitter-DWEsd-ud.cjs';
|
|
16
16
|
import '../../../internals/helpers/promise.cjs';
|
|
17
17
|
import 'promise-based-task';
|
|
18
18
|
import '../../../cache/base.cjs';
|
|
@@ -31,9 +31,8 @@ import '../../../template.cjs';
|
|
|
31
31
|
|
|
32
32
|
type OpenAIParameters = Parameters<OpenAIProvider["chat"]>;
|
|
33
33
|
type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;
|
|
34
|
-
type OpenAIChatModelSettings = NonNullable<OpenAIParameters[1]>;
|
|
35
34
|
declare class OpenAIChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: OpenAIChatModelId,
|
|
35
|
+
constructor(modelId?: OpenAIChatModelId, parameters?: ChatModelParameters, client?: OpenAIClient | OpenAIClientSettings);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { OpenAIChatModel, type OpenAIChatModelId
|
|
38
|
+
export { OpenAIChatModel, type OpenAIChatModelId };
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import { OpenAIProvider } from '@ai-sdk/openai';
|
|
2
2
|
import { OpenAIClient, OpenAIClientSettings } from './client.js';
|
|
3
3
|
import { VercelChatModel } from '../../vercel/backend/chat.js';
|
|
4
|
+
import { C as ChatModelParameters } from '../../../chat-C0s-o6ll.js';
|
|
4
5
|
import '../../../backend/client.js';
|
|
5
6
|
import '../../../internals/serializable.js';
|
|
6
7
|
import '../../../internals/types.js';
|
|
@@ -8,11 +9,10 @@ import '../../../internals/helpers/guards.js';
|
|
|
8
9
|
import '../../../logger/logger.js';
|
|
9
10
|
import 'pino';
|
|
10
11
|
import '../../../errors.js';
|
|
11
|
-
import '../../../chat-W9j8T6UU.js';
|
|
12
|
-
import '../../../backend/message.js';
|
|
13
12
|
import 'ai';
|
|
13
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
14
|
+
import '../../../backend/message.js';
|
|
14
15
|
import '../../../context.js';
|
|
15
|
-
import '../../../emitter-C9nwGLt6.js';
|
|
16
16
|
import '../../../internals/helpers/promise.js';
|
|
17
17
|
import 'promise-based-task';
|
|
18
18
|
import '../../../cache/base.js';
|
|
@@ -31,9 +31,8 @@ import '../../../template.js';
|
|
|
31
31
|
|
|
32
32
|
type OpenAIParameters = Parameters<OpenAIProvider["chat"]>;
|
|
33
33
|
type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;
|
|
34
|
-
type OpenAIChatModelSettings = NonNullable<OpenAIParameters[1]>;
|
|
35
34
|
declare class OpenAIChatModel extends VercelChatModel {
|
|
36
|
-
constructor(modelId?: OpenAIChatModelId,
|
|
35
|
+
constructor(modelId?: OpenAIChatModelId, parameters?: ChatModelParameters, client?: OpenAIClient | OpenAIClientSettings);
|
|
37
36
|
}
|
|
38
37
|
|
|
39
|
-
export { OpenAIChatModel, type OpenAIChatModelId
|
|
38
|
+
export { OpenAIChatModel, type OpenAIChatModelId };
|
|
@@ -8,9 +8,10 @@ class OpenAIChatModel extends VercelChatModel {
|
|
|
8
8
|
static {
|
|
9
9
|
__name(this, "OpenAIChatModel");
|
|
10
10
|
}
|
|
11
|
-
constructor(modelId = getEnv("OPENAI_CHAT_MODEL", "gpt-4o"),
|
|
12
|
-
const model = OpenAIClient.ensure(client).instance.chat(modelId
|
|
11
|
+
constructor(modelId = getEnv("OPENAI_CHAT_MODEL", "gpt-4o"), parameters = {}, client) {
|
|
12
|
+
const model = OpenAIClient.ensure(client).instance.chat(modelId);
|
|
13
13
|
super(model);
|
|
14
|
+
Object.assign(this.parameters, parameters ?? {});
|
|
14
15
|
}
|
|
15
16
|
}
|
|
16
17
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/chat.ts"],"names":["OpenAIChatModel","VercelChatModel","modelId","getEnv","parameters","client","model","OpenAIClient","ensure","instance","chat","Object","assign"],"mappings":";;;;;;AAcO,MAAMA,wBAAwBC,eAAAA,CAAAA;EAdrC;;;EAeE,WAAA,CACEC,OAAAA,GAA6BC,OAAO,mBAAA,EAAqB,QAAA,GACzDC,UAAAA,GAAkC,IAClCC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,YAAAA,CAAaC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,KAAKR,OAAAA,CAAAA;AACxD,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACNK,IAAAA,MAAAA,CAAOC,MAAAA,CAAO,IAAA,CAAKR,UAAAA,EAAYA,UAAAA,IAAc,EAAC,CAAA;AAChD;AACF","file":"chat.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OpenAIProvider } from \"@ai-sdk/openai\";\nimport { OpenAIClient, OpenAIClientSettings } from \"@/adapters/openai/backend/client.js\";\nimport { VercelChatModel } from \"@/adapters/vercel/backend/chat.js\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { ChatModelParameters } from \"@/backend/chat.js\";\n\ntype OpenAIParameters = Parameters<OpenAIProvider[\"chat\"]>;\nexport type OpenAIChatModelId = NonNullable<OpenAIParameters[0]>;\n\nexport class OpenAIChatModel extends VercelChatModel {\n constructor(\n modelId: OpenAIChatModelId = getEnv(\"OPENAI_CHAT_MODEL\", \"gpt-4o\"),\n parameters: ChatModelParameters = {},\n client?: OpenAIClient | OpenAIClientSettings,\n ) {\n const model = OpenAIClient.ensure(client).instance.chat(modelId);\n super(model);\n Object.assign(this.parameters, parameters ?? {});\n }\n}\n"]}
|
|
@@ -12,21 +12,12 @@ class OpenAIClient extends client_cjs.BackendClient {
|
|
|
12
12
|
__name(this, "OpenAIClient");
|
|
13
13
|
}
|
|
14
14
|
create() {
|
|
15
|
-
const extraHeaders = utils_cjs.parseHeadersFromEnv("OPENAI_API_HEADERS");
|
|
16
|
-
const baseURL = this.settings?.baseURL || env_cjs.getEnv("OPENAI_API_ENDPOINT");
|
|
17
|
-
let compatibility = this.settings?.compatibility || env_cjs.getEnv("OPENAI_COMPATIBILITY_MODE");
|
|
18
|
-
if (baseURL && !compatibility) {
|
|
19
|
-
compatibility = "compatible";
|
|
20
|
-
} else if (!baseURL && !compatibility) {
|
|
21
|
-
compatibility = "strict";
|
|
22
|
-
}
|
|
23
15
|
return openai.createOpenAI({
|
|
24
16
|
...this.settings,
|
|
25
|
-
compatibility,
|
|
26
17
|
apiKey: this.settings?.apiKey || env_cjs.getEnv("OPENAI_API_KEY"),
|
|
27
|
-
baseURL,
|
|
18
|
+
baseURL: this.settings?.baseURL || env_cjs.getEnv("OPENAI_API_ENDPOINT"),
|
|
28
19
|
headers: {
|
|
29
|
-
...
|
|
20
|
+
...utils_cjs.parseHeadersFromEnv("OPENAI_API_HEADERS"),
|
|
30
21
|
...this.settings?.headers
|
|
31
22
|
},
|
|
32
23
|
fetch: utils_cjs.vercelFetcher(this.settings?.fetch)
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/client.ts"],"names":["OpenAIClient","BackendClient","create","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/client.ts"],"names":["OpenAIClient","BackendClient","create","createOpenAI","settings","apiKey","getEnv","baseURL","headers","parseHeadersFromEnv","fetch","vercelFetcher"],"mappings":";;;;;;;;;AAYO,MAAMA,qBAAqBC,wBAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,OAAOC,mBAAAA,CAAa;AAClB,MAAA,GAAG,IAAA,CAAKC,QAAAA;AACRC,MAAAA,MAAAA,EAAQ,IAAA,CAAKD,QAAAA,EAAUC,MAAAA,IAAUC,cAAAA,CAAO,gBAAA,CAAA;AACxCC,MAAAA,OAAAA,EAAS,IAAA,CAAKH,QAAAA,EAAUG,OAAAA,IAAWD,cAAAA,CAAO,qBAAA,CAAA;MAC1CE,OAAAA,EAAS;AACP,QAAA,GAAGC,8BAAoB,oBAAA,CAAA;AACvB,QAAA,GAAG,KAAKL,QAAAA,EAAUI;AACpB,OAAA;MACAE,KAAAA,EAAOC,uBAAAA,CAAc,IAAA,CAAKP,QAAAA,EAAUM,KAAAA;KACtC,CAAA;AACF;AACF","file":"client.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { createOpenAI, OpenAIProvider, OpenAIProviderSettings } from \"@ai-sdk/openai\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OpenAIClientSettings = OpenAIProviderSettings;\n\nexport class OpenAIClient extends BackendClient<OpenAIClientSettings, OpenAIProvider> {\n protected create(): OpenAIProvider {\n return createOpenAI({\n ...this.settings,\n apiKey: this.settings?.apiKey || getEnv(\"OPENAI_API_KEY\"),\n baseURL: this.settings?.baseURL || getEnv(\"OPENAI_API_ENDPOINT\"),\n headers: {\n ...parseHeadersFromEnv(\"OPENAI_API_HEADERS\"),\n ...this.settings?.headers,\n },\n fetch: vercelFetcher(this.settings?.fetch),\n });\n }\n}\n"]}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { createOpenAI } from '@ai-sdk/openai';
|
|
2
2
|
import { getEnv } from '../../../internals/env.js';
|
|
3
3
|
import { BackendClient } from '../../../backend/client.js';
|
|
4
|
-
import {
|
|
4
|
+
import { vercelFetcher, parseHeadersFromEnv } from '../../vercel/backend/utils.js';
|
|
5
5
|
|
|
6
6
|
var __defProp = Object.defineProperty;
|
|
7
7
|
var __name = (target, value) => __defProp(target, "name", { value, configurable: true });
|
|
@@ -10,21 +10,12 @@ class OpenAIClient extends BackendClient {
|
|
|
10
10
|
__name(this, "OpenAIClient");
|
|
11
11
|
}
|
|
12
12
|
create() {
|
|
13
|
-
const extraHeaders = parseHeadersFromEnv("OPENAI_API_HEADERS");
|
|
14
|
-
const baseURL = this.settings?.baseURL || getEnv("OPENAI_API_ENDPOINT");
|
|
15
|
-
let compatibility = this.settings?.compatibility || getEnv("OPENAI_COMPATIBILITY_MODE");
|
|
16
|
-
if (baseURL && !compatibility) {
|
|
17
|
-
compatibility = "compatible";
|
|
18
|
-
} else if (!baseURL && !compatibility) {
|
|
19
|
-
compatibility = "strict";
|
|
20
|
-
}
|
|
21
13
|
return createOpenAI({
|
|
22
14
|
...this.settings,
|
|
23
|
-
compatibility,
|
|
24
15
|
apiKey: this.settings?.apiKey || getEnv("OPENAI_API_KEY"),
|
|
25
|
-
baseURL,
|
|
16
|
+
baseURL: this.settings?.baseURL || getEnv("OPENAI_API_ENDPOINT"),
|
|
26
17
|
headers: {
|
|
27
|
-
...
|
|
18
|
+
...parseHeadersFromEnv("OPENAI_API_HEADERS"),
|
|
28
19
|
...this.settings?.headers
|
|
29
20
|
},
|
|
30
21
|
fetch: vercelFetcher(this.settings?.fetch)
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/client.ts"],"names":["OpenAIClient","BackendClient","create","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/client.ts"],"names":["OpenAIClient","BackendClient","create","createOpenAI","settings","apiKey","getEnv","baseURL","headers","parseHeadersFromEnv","fetch","vercelFetcher"],"mappings":";;;;;;;AAYO,MAAMA,qBAAqBC,aAAAA,CAAAA;EAZlC;;;EAaYC,MAAAA,GAAyB;AACjC,IAAA,OAAOC,YAAAA,CAAa;AAClB,MAAA,GAAG,IAAA,CAAKC,QAAAA;AACRC,MAAAA,MAAAA,EAAQ,IAAA,CAAKD,QAAAA,EAAUC,MAAAA,IAAUC,MAAAA,CAAO,gBAAA,CAAA;AACxCC,MAAAA,OAAAA,EAAS,IAAA,CAAKH,QAAAA,EAAUG,OAAAA,IAAWD,MAAAA,CAAO,qBAAA,CAAA;MAC1CE,OAAAA,EAAS;AACP,QAAA,GAAGC,oBAAoB,oBAAA,CAAA;AACvB,QAAA,GAAG,KAAKL,QAAAA,EAAUI;AACpB,OAAA;MACAE,KAAAA,EAAOC,aAAAA,CAAc,IAAA,CAAKP,QAAAA,EAAUM,KAAAA;KACtC,CAAA;AACF;AACF","file":"client.js","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { createOpenAI, OpenAIProvider, OpenAIProviderSettings } from \"@ai-sdk/openai\";\nimport { getEnv } from \"@/internals/env.js\";\nimport { BackendClient } from \"@/backend/client.js\";\nimport { parseHeadersFromEnv, vercelFetcher } from \"@/adapters/vercel/backend/utils.js\";\n\nexport type OpenAIClientSettings = OpenAIProviderSettings;\n\nexport class OpenAIClient extends BackendClient<OpenAIClientSettings, OpenAIProvider> {\n protected create(): OpenAIProvider {\n return createOpenAI({\n ...this.settings,\n apiKey: this.settings?.apiKey || getEnv(\"OPENAI_API_KEY\"),\n baseURL: this.settings?.baseURL || getEnv(\"OPENAI_API_ENDPOINT\"),\n headers: {\n ...parseHeadersFromEnv(\"OPENAI_API_HEADERS\"),\n ...this.settings?.headers,\n },\n fetch: vercelFetcher(this.settings?.fetch),\n });\n }\n}\n"]}
|
|
@@ -10,8 +10,8 @@ class OpenAIEmbeddingModel extends embedding_cjs.VercelEmbeddingModel {
|
|
|
10
10
|
static {
|
|
11
11
|
__name(this, "OpenAIEmbeddingModel");
|
|
12
12
|
}
|
|
13
|
-
constructor(modelId = env_cjs.getEnv("OPENAI_EMBEDDING_MODEL", "text-embedding-3-small"),
|
|
14
|
-
const model = client_cjs.OpenAIClient.ensure(client).instance.embedding(modelId
|
|
13
|
+
constructor(modelId = env_cjs.getEnv("OPENAI_EMBEDDING_MODEL", "text-embedding-3-small"), _settings = {}, client) {
|
|
14
|
+
const model = client_cjs.OpenAIClient.ensure(client).instance.embedding(modelId);
|
|
15
15
|
super(model);
|
|
16
16
|
}
|
|
17
17
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../../../src/adapters/openai/backend/embedding.ts"],"names":["OpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","
|
|
1
|
+
{"version":3,"sources":["../../../../src/adapters/openai/backend/embedding.ts"],"names":["OpenAIEmbeddingModel","VercelEmbeddingModel","modelId","getEnv","_settings","client","model","OpenAIClient","ensure","instance","embedding"],"mappings":";;;;;;;;AAcO,MAAMA,6BAA6BC,kCAAAA,CAAAA;EAd1C;;;EAeE,WAAA,CACEC,OAAAA,GAAkCC,eAAO,wBAAA,EAA0B,wBAAA,GACnEC,SAAAA,GAA0C,IAC1CC,MAAAA,EACA;AACA,IAAA,MAAMC,QAAQC,uBAAAA,CAAaC,MAAAA,CAAOH,MAAAA,CAAAA,CAAQI,QAAAA,CAASC,UAAUR,OAAAA,CAAAA;AAC7D,IAAA,KAAA,CAAMI,KAAAA,CAAAA;AACR;AACF","file":"embedding.cjs","sourcesContent":["/**\n * Copyright 2025 © BeeAI a Series of LF Projects, LLC\n * SPDX-License-Identifier: Apache-2.0\n */\n\nimport { OpenAIClient } from \"@/adapters/openai/backend/client.js\";\nimport { OpenAIProvider, OpenAIProviderSettings } from \"@ai-sdk/openai\";\nimport { VercelEmbeddingModel } from \"@/adapters/vercel/backend/embedding.js\";\nimport { getEnv } from \"@/internals/env.js\";\n\ntype OpenAIParameters = Parameters<OpenAIProvider[\"embedding\"]>;\nexport type OpenAIEmbeddingModelId = NonNullable<OpenAIParameters[0]>;\nexport type OpenAIEmbeddingModelSettings = any;\n\nexport class OpenAIEmbeddingModel extends VercelEmbeddingModel {\n constructor(\n modelId: OpenAIEmbeddingModelId = getEnv(\"OPENAI_EMBEDDING_MODEL\", \"text-embedding-3-small\"),\n _settings: OpenAIEmbeddingModelSettings = {},\n client?: OpenAIProviderSettings | OpenAIClient,\n ) {\n const model = OpenAIClient.ensure(client).instance.embedding(modelId);\n super(model);\n }\n}\n"]}
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.cjs';
|
|
|
7
7
|
import '../../../internals/helpers/guards.cjs';
|
|
8
8
|
import '../../../backend/embedding.cjs';
|
|
9
9
|
import '../../../context.cjs';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-DpqUYjXH.cjs';
|
|
11
11
|
import '../../../internals/helpers/promise.cjs';
|
|
12
12
|
import '../../../errors.cjs';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-CRb3vUVg.cjs';
|
|
14
14
|
import '../../../backend/message.cjs';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -32,9 +32,9 @@ import 'pino';
|
|
|
32
32
|
|
|
33
33
|
type OpenAIParameters = Parameters<OpenAIProvider["embedding"]>;
|
|
34
34
|
type OpenAIEmbeddingModelId = NonNullable<OpenAIParameters[0]>;
|
|
35
|
-
type OpenAIEmbeddingModelSettings =
|
|
35
|
+
type OpenAIEmbeddingModelSettings = any;
|
|
36
36
|
declare class OpenAIEmbeddingModel extends VercelEmbeddingModel {
|
|
37
|
-
constructor(modelId?: OpenAIEmbeddingModelId,
|
|
37
|
+
constructor(modelId?: OpenAIEmbeddingModelId, _settings?: OpenAIEmbeddingModelSettings, client?: OpenAIProviderSettings | OpenAIClient);
|
|
38
38
|
}
|
|
39
39
|
|
|
40
40
|
export { OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingModelSettings };
|
|
@@ -7,10 +7,10 @@ import '../../../internals/types.js';
|
|
|
7
7
|
import '../../../internals/helpers/guards.js';
|
|
8
8
|
import '../../../backend/embedding.js';
|
|
9
9
|
import '../../../context.js';
|
|
10
|
-
import '../../../emitter-
|
|
10
|
+
import '../../../emitter-CZFbzlUi.js';
|
|
11
11
|
import '../../../internals/helpers/promise.js';
|
|
12
12
|
import '../../../errors.js';
|
|
13
|
-
import '../../../chat-
|
|
13
|
+
import '../../../chat-C0s-o6ll.js';
|
|
14
14
|
import '../../../backend/message.js';
|
|
15
15
|
import 'ai';
|
|
16
16
|
import 'promise-based-task';
|
|
@@ -32,9 +32,9 @@ import 'pino';
|
|
|
32
32
|
|
|
33
33
|
type OpenAIParameters = Parameters<OpenAIProvider["embedding"]>;
|
|
34
34
|
type OpenAIEmbeddingModelId = NonNullable<OpenAIParameters[0]>;
|
|
35
|
-
type OpenAIEmbeddingModelSettings =
|
|
35
|
+
type OpenAIEmbeddingModelSettings = any;
|
|
36
36
|
declare class OpenAIEmbeddingModel extends VercelEmbeddingModel {
|
|
37
|
-
constructor(modelId?: OpenAIEmbeddingModelId,
|
|
37
|
+
constructor(modelId?: OpenAIEmbeddingModelId, _settings?: OpenAIEmbeddingModelSettings, client?: OpenAIProviderSettings | OpenAIClient);
|
|
38
38
|
}
|
|
39
39
|
|
|
40
40
|
export { OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingModelSettings };
|
|
@@ -8,8 +8,8 @@ class OpenAIEmbeddingModel extends VercelEmbeddingModel {
|
|
|
8
8
|
static {
|
|
9
9
|
__name(this, "OpenAIEmbeddingModel");
|
|
10
10
|
}
|
|
11
|
-
constructor(modelId = getEnv("OPENAI_EMBEDDING_MODEL", "text-embedding-3-small"),
|
|
12
|
-
const model = OpenAIClient.ensure(client).instance.embedding(modelId
|
|
11
|
+
constructor(modelId = getEnv("OPENAI_EMBEDDING_MODEL", "text-embedding-3-small"), _settings = {}, client) {
|
|
12
|
+
const model = OpenAIClient.ensure(client).instance.embedding(modelId);
|
|
13
13
|
super(model);
|
|
14
14
|
}
|
|
15
15
|
}
|