ag2 0.9.1a1__py3-none-any.whl → 0.9.1.post0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of ag2 might be problematic. Click here for more details.
- {ag2-0.9.1a1.dist-info → ag2-0.9.1.post0.dist-info}/METADATA +264 -73
- ag2-0.9.1.post0.dist-info/RECORD +392 -0
- {ag2-0.9.1a1.dist-info → ag2-0.9.1.post0.dist-info}/WHEEL +1 -2
- autogen/__init__.py +89 -0
- autogen/_website/__init__.py +3 -0
- autogen/_website/generate_api_references.py +427 -0
- autogen/_website/generate_mkdocs.py +1174 -0
- autogen/_website/notebook_processor.py +476 -0
- autogen/_website/process_notebooks.py +656 -0
- autogen/_website/utils.py +412 -0
- autogen/agentchat/__init__.py +44 -0
- autogen/agentchat/agent.py +182 -0
- autogen/agentchat/assistant_agent.py +85 -0
- autogen/agentchat/chat.py +309 -0
- autogen/agentchat/contrib/__init__.py +5 -0
- autogen/agentchat/contrib/agent_eval/README.md +7 -0
- autogen/agentchat/contrib/agent_eval/agent_eval.py +108 -0
- autogen/agentchat/contrib/agent_eval/criterion.py +43 -0
- autogen/agentchat/contrib/agent_eval/critic_agent.py +44 -0
- autogen/agentchat/contrib/agent_eval/quantifier_agent.py +39 -0
- autogen/agentchat/contrib/agent_eval/subcritic_agent.py +45 -0
- autogen/agentchat/contrib/agent_eval/task.py +42 -0
- autogen/agentchat/contrib/agent_optimizer.py +429 -0
- autogen/agentchat/contrib/capabilities/__init__.py +5 -0
- autogen/agentchat/contrib/capabilities/agent_capability.py +20 -0
- autogen/agentchat/contrib/capabilities/generate_images.py +301 -0
- autogen/agentchat/contrib/capabilities/teachability.py +393 -0
- autogen/agentchat/contrib/capabilities/text_compressors.py +66 -0
- autogen/agentchat/contrib/capabilities/tools_capability.py +22 -0
- autogen/agentchat/contrib/capabilities/transform_messages.py +93 -0
- autogen/agentchat/contrib/capabilities/transforms.py +566 -0
- autogen/agentchat/contrib/capabilities/transforms_util.py +122 -0
- autogen/agentchat/contrib/capabilities/vision_capability.py +214 -0
- autogen/agentchat/contrib/captainagent/__init__.py +9 -0
- autogen/agentchat/contrib/captainagent/agent_builder.py +790 -0
- autogen/agentchat/contrib/captainagent/captainagent.py +512 -0
- autogen/agentchat/contrib/captainagent/tool_retriever.py +335 -0
- autogen/agentchat/contrib/captainagent/tools/README.md +44 -0
- autogen/agentchat/contrib/captainagent/tools/__init__.py +5 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/calculate_correlation.py +40 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/calculate_skewness_and_kurtosis.py +28 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/detect_outlier_iqr.py +28 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/detect_outlier_zscore.py +28 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/explore_csv.py +21 -0
- autogen/agentchat/contrib/captainagent/tools/data_analysis/shapiro_wilk_test.py +30 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/arxiv_download.py +27 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/arxiv_search.py +53 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/extract_pdf_image.py +53 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/extract_pdf_text.py +38 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/get_wikipedia_text.py +21 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/get_youtube_caption.py +34 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/image_qa.py +60 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/optical_character_recognition.py +61 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/perform_web_search.py +47 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/scrape_wikipedia_tables.py +33 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/transcribe_audio_file.py +21 -0
- autogen/agentchat/contrib/captainagent/tools/information_retrieval/youtube_download.py +35 -0
- autogen/agentchat/contrib/captainagent/tools/math/calculate_circle_area_from_diameter.py +21 -0
- autogen/agentchat/contrib/captainagent/tools/math/calculate_day_of_the_week.py +18 -0
- autogen/agentchat/contrib/captainagent/tools/math/calculate_fraction_sum.py +28 -0
- autogen/agentchat/contrib/captainagent/tools/math/calculate_matrix_power.py +31 -0
- autogen/agentchat/contrib/captainagent/tools/math/calculate_reflected_point.py +16 -0
- autogen/agentchat/contrib/captainagent/tools/math/complex_numbers_product.py +25 -0
- autogen/agentchat/contrib/captainagent/tools/math/compute_currency_conversion.py +23 -0
- autogen/agentchat/contrib/captainagent/tools/math/count_distinct_permutations.py +27 -0
- autogen/agentchat/contrib/captainagent/tools/math/evaluate_expression.py +28 -0
- autogen/agentchat/contrib/captainagent/tools/math/find_continuity_point.py +34 -0
- autogen/agentchat/contrib/captainagent/tools/math/fraction_to_mixed_numbers.py +39 -0
- autogen/agentchat/contrib/captainagent/tools/math/modular_inverse_sum.py +23 -0
- autogen/agentchat/contrib/captainagent/tools/math/simplify_mixed_numbers.py +36 -0
- autogen/agentchat/contrib/captainagent/tools/math/sum_of_digit_factorials.py +15 -0
- autogen/agentchat/contrib/captainagent/tools/math/sum_of_primes_below.py +15 -0
- autogen/agentchat/contrib/captainagent/tools/requirements.txt +10 -0
- autogen/agentchat/contrib/captainagent/tools/tool_description.tsv +34 -0
- autogen/agentchat/contrib/gpt_assistant_agent.py +526 -0
- autogen/agentchat/contrib/graph_rag/__init__.py +9 -0
- autogen/agentchat/contrib/graph_rag/document.py +29 -0
- autogen/agentchat/contrib/graph_rag/falkor_graph_query_engine.py +170 -0
- autogen/agentchat/contrib/graph_rag/falkor_graph_rag_capability.py +103 -0
- autogen/agentchat/contrib/graph_rag/graph_query_engine.py +53 -0
- autogen/agentchat/contrib/graph_rag/graph_rag_capability.py +63 -0
- autogen/agentchat/contrib/graph_rag/neo4j_graph_query_engine.py +268 -0
- autogen/agentchat/contrib/graph_rag/neo4j_graph_rag_capability.py +83 -0
- autogen/agentchat/contrib/graph_rag/neo4j_native_graph_query_engine.py +210 -0
- autogen/agentchat/contrib/graph_rag/neo4j_native_graph_rag_capability.py +93 -0
- autogen/agentchat/contrib/img_utils.py +397 -0
- autogen/agentchat/contrib/llamaindex_conversable_agent.py +117 -0
- autogen/agentchat/contrib/llava_agent.py +187 -0
- autogen/agentchat/contrib/math_user_proxy_agent.py +464 -0
- autogen/agentchat/contrib/multimodal_conversable_agent.py +125 -0
- autogen/agentchat/contrib/qdrant_retrieve_user_proxy_agent.py +324 -0
- autogen/agentchat/contrib/rag/__init__.py +10 -0
- autogen/agentchat/contrib/rag/chromadb_query_engine.py +272 -0
- autogen/agentchat/contrib/rag/llamaindex_query_engine.py +198 -0
- autogen/agentchat/contrib/rag/mongodb_query_engine.py +329 -0
- autogen/agentchat/contrib/rag/query_engine.py +74 -0
- autogen/agentchat/contrib/retrieve_assistant_agent.py +56 -0
- autogen/agentchat/contrib/retrieve_user_proxy_agent.py +703 -0
- autogen/agentchat/contrib/society_of_mind_agent.py +199 -0
- autogen/agentchat/contrib/swarm_agent.py +1425 -0
- autogen/agentchat/contrib/text_analyzer_agent.py +79 -0
- autogen/agentchat/contrib/vectordb/__init__.py +5 -0
- autogen/agentchat/contrib/vectordb/base.py +232 -0
- autogen/agentchat/contrib/vectordb/chromadb.py +315 -0
- autogen/agentchat/contrib/vectordb/couchbase.py +407 -0
- autogen/agentchat/contrib/vectordb/mongodb.py +550 -0
- autogen/agentchat/contrib/vectordb/pgvectordb.py +928 -0
- autogen/agentchat/contrib/vectordb/qdrant.py +320 -0
- autogen/agentchat/contrib/vectordb/utils.py +126 -0
- autogen/agentchat/contrib/web_surfer.py +303 -0
- autogen/agentchat/conversable_agent.py +4020 -0
- autogen/agentchat/group/__init__.py +64 -0
- autogen/agentchat/group/available_condition.py +91 -0
- autogen/agentchat/group/context_condition.py +77 -0
- autogen/agentchat/group/context_expression.py +238 -0
- autogen/agentchat/group/context_str.py +41 -0
- autogen/agentchat/group/context_variables.py +192 -0
- autogen/agentchat/group/group_tool_executor.py +202 -0
- autogen/agentchat/group/group_utils.py +591 -0
- autogen/agentchat/group/handoffs.py +244 -0
- autogen/agentchat/group/llm_condition.py +93 -0
- autogen/agentchat/group/multi_agent_chat.py +237 -0
- autogen/agentchat/group/on_condition.py +58 -0
- autogen/agentchat/group/on_context_condition.py +54 -0
- autogen/agentchat/group/patterns/__init__.py +18 -0
- autogen/agentchat/group/patterns/auto.py +159 -0
- autogen/agentchat/group/patterns/manual.py +176 -0
- autogen/agentchat/group/patterns/pattern.py +288 -0
- autogen/agentchat/group/patterns/random.py +106 -0
- autogen/agentchat/group/patterns/round_robin.py +117 -0
- autogen/agentchat/group/reply_result.py +26 -0
- autogen/agentchat/group/speaker_selection_result.py +41 -0
- autogen/agentchat/group/targets/__init__.py +4 -0
- autogen/agentchat/group/targets/group_chat_target.py +132 -0
- autogen/agentchat/group/targets/group_manager_target.py +151 -0
- autogen/agentchat/group/targets/transition_target.py +413 -0
- autogen/agentchat/group/targets/transition_utils.py +6 -0
- autogen/agentchat/groupchat.py +1694 -0
- autogen/agentchat/realtime/__init__.py +3 -0
- autogen/agentchat/realtime/experimental/__init__.py +20 -0
- autogen/agentchat/realtime/experimental/audio_adapters/__init__.py +8 -0
- autogen/agentchat/realtime/experimental/audio_adapters/twilio_audio_adapter.py +148 -0
- autogen/agentchat/realtime/experimental/audio_adapters/websocket_audio_adapter.py +139 -0
- autogen/agentchat/realtime/experimental/audio_observer.py +42 -0
- autogen/agentchat/realtime/experimental/clients/__init__.py +15 -0
- autogen/agentchat/realtime/experimental/clients/gemini/__init__.py +7 -0
- autogen/agentchat/realtime/experimental/clients/gemini/client.py +274 -0
- autogen/agentchat/realtime/experimental/clients/oai/__init__.py +8 -0
- autogen/agentchat/realtime/experimental/clients/oai/base_client.py +220 -0
- autogen/agentchat/realtime/experimental/clients/oai/rtc_client.py +243 -0
- autogen/agentchat/realtime/experimental/clients/oai/utils.py +48 -0
- autogen/agentchat/realtime/experimental/clients/realtime_client.py +190 -0
- autogen/agentchat/realtime/experimental/function_observer.py +85 -0
- autogen/agentchat/realtime/experimental/realtime_agent.py +158 -0
- autogen/agentchat/realtime/experimental/realtime_events.py +42 -0
- autogen/agentchat/realtime/experimental/realtime_observer.py +100 -0
- autogen/agentchat/realtime/experimental/realtime_swarm.py +475 -0
- autogen/agentchat/realtime/experimental/websockets.py +21 -0
- autogen/agentchat/realtime_agent/__init__.py +21 -0
- autogen/agentchat/user_proxy_agent.py +111 -0
- autogen/agentchat/utils.py +206 -0
- autogen/agents/__init__.py +3 -0
- autogen/agents/contrib/__init__.py +10 -0
- autogen/agents/contrib/time/__init__.py +8 -0
- autogen/agents/contrib/time/time_reply_agent.py +73 -0
- autogen/agents/contrib/time/time_tool_agent.py +51 -0
- autogen/agents/experimental/__init__.py +27 -0
- autogen/agents/experimental/deep_research/__init__.py +7 -0
- autogen/agents/experimental/deep_research/deep_research.py +52 -0
- autogen/agents/experimental/discord/__init__.py +7 -0
- autogen/agents/experimental/discord/discord.py +66 -0
- autogen/agents/experimental/document_agent/__init__.py +19 -0
- autogen/agents/experimental/document_agent/chroma_query_engine.py +316 -0
- autogen/agents/experimental/document_agent/docling_doc_ingest_agent.py +118 -0
- autogen/agents/experimental/document_agent/document_agent.py +461 -0
- autogen/agents/experimental/document_agent/document_conditions.py +50 -0
- autogen/agents/experimental/document_agent/document_utils.py +380 -0
- autogen/agents/experimental/document_agent/inmemory_query_engine.py +220 -0
- autogen/agents/experimental/document_agent/parser_utils.py +130 -0
- autogen/agents/experimental/document_agent/url_utils.py +426 -0
- autogen/agents/experimental/reasoning/__init__.py +7 -0
- autogen/agents/experimental/reasoning/reasoning_agent.py +1178 -0
- autogen/agents/experimental/slack/__init__.py +7 -0
- autogen/agents/experimental/slack/slack.py +73 -0
- autogen/agents/experimental/telegram/__init__.py +7 -0
- autogen/agents/experimental/telegram/telegram.py +77 -0
- autogen/agents/experimental/websurfer/__init__.py +7 -0
- autogen/agents/experimental/websurfer/websurfer.py +62 -0
- autogen/agents/experimental/wikipedia/__init__.py +7 -0
- autogen/agents/experimental/wikipedia/wikipedia.py +90 -0
- autogen/browser_utils.py +309 -0
- autogen/cache/__init__.py +10 -0
- autogen/cache/abstract_cache_base.py +75 -0
- autogen/cache/cache.py +203 -0
- autogen/cache/cache_factory.py +88 -0
- autogen/cache/cosmos_db_cache.py +144 -0
- autogen/cache/disk_cache.py +102 -0
- autogen/cache/in_memory_cache.py +58 -0
- autogen/cache/redis_cache.py +123 -0
- autogen/code_utils.py +596 -0
- autogen/coding/__init__.py +22 -0
- autogen/coding/base.py +119 -0
- autogen/coding/docker_commandline_code_executor.py +268 -0
- autogen/coding/factory.py +47 -0
- autogen/coding/func_with_reqs.py +202 -0
- autogen/coding/jupyter/__init__.py +23 -0
- autogen/coding/jupyter/base.py +36 -0
- autogen/coding/jupyter/docker_jupyter_server.py +167 -0
- autogen/coding/jupyter/embedded_ipython_code_executor.py +182 -0
- autogen/coding/jupyter/import_utils.py +82 -0
- autogen/coding/jupyter/jupyter_client.py +231 -0
- autogen/coding/jupyter/jupyter_code_executor.py +160 -0
- autogen/coding/jupyter/local_jupyter_server.py +172 -0
- autogen/coding/local_commandline_code_executor.py +405 -0
- autogen/coding/markdown_code_extractor.py +45 -0
- autogen/coding/utils.py +56 -0
- autogen/doc_utils.py +34 -0
- autogen/events/__init__.py +7 -0
- autogen/events/agent_events.py +1010 -0
- autogen/events/base_event.py +99 -0
- autogen/events/client_events.py +167 -0
- autogen/events/helpers.py +36 -0
- autogen/events/print_event.py +46 -0
- autogen/exception_utils.py +73 -0
- autogen/extensions/__init__.py +5 -0
- autogen/fast_depends/__init__.py +16 -0
- autogen/fast_depends/_compat.py +80 -0
- autogen/fast_depends/core/__init__.py +14 -0
- autogen/fast_depends/core/build.py +225 -0
- autogen/fast_depends/core/model.py +576 -0
- autogen/fast_depends/dependencies/__init__.py +15 -0
- autogen/fast_depends/dependencies/model.py +29 -0
- autogen/fast_depends/dependencies/provider.py +39 -0
- autogen/fast_depends/library/__init__.py +10 -0
- autogen/fast_depends/library/model.py +46 -0
- autogen/fast_depends/py.typed +6 -0
- autogen/fast_depends/schema.py +66 -0
- autogen/fast_depends/use.py +280 -0
- autogen/fast_depends/utils.py +187 -0
- autogen/formatting_utils.py +83 -0
- autogen/function_utils.py +13 -0
- autogen/graph_utils.py +178 -0
- autogen/import_utils.py +526 -0
- autogen/interop/__init__.py +22 -0
- autogen/interop/crewai/__init__.py +7 -0
- autogen/interop/crewai/crewai.py +88 -0
- autogen/interop/interoperability.py +71 -0
- autogen/interop/interoperable.py +46 -0
- autogen/interop/langchain/__init__.py +8 -0
- autogen/interop/langchain/langchain_chat_model_factory.py +155 -0
- autogen/interop/langchain/langchain_tool.py +82 -0
- autogen/interop/litellm/__init__.py +7 -0
- autogen/interop/litellm/litellm_config_factory.py +113 -0
- autogen/interop/pydantic_ai/__init__.py +7 -0
- autogen/interop/pydantic_ai/pydantic_ai.py +168 -0
- autogen/interop/registry.py +69 -0
- autogen/io/__init__.py +15 -0
- autogen/io/base.py +151 -0
- autogen/io/console.py +56 -0
- autogen/io/processors/__init__.py +12 -0
- autogen/io/processors/base.py +21 -0
- autogen/io/processors/console_event_processor.py +56 -0
- autogen/io/run_response.py +293 -0
- autogen/io/thread_io_stream.py +63 -0
- autogen/io/websockets.py +213 -0
- autogen/json_utils.py +43 -0
- autogen/llm_config.py +379 -0
- autogen/logger/__init__.py +11 -0
- autogen/logger/base_logger.py +128 -0
- autogen/logger/file_logger.py +261 -0
- autogen/logger/logger_factory.py +42 -0
- autogen/logger/logger_utils.py +57 -0
- autogen/logger/sqlite_logger.py +523 -0
- autogen/math_utils.py +339 -0
- autogen/mcp/__init__.py +7 -0
- autogen/mcp/mcp_client.py +208 -0
- autogen/messages/__init__.py +7 -0
- autogen/messages/agent_messages.py +948 -0
- autogen/messages/base_message.py +107 -0
- autogen/messages/client_messages.py +171 -0
- autogen/messages/print_message.py +49 -0
- autogen/oai/__init__.py +53 -0
- autogen/oai/anthropic.py +714 -0
- autogen/oai/bedrock.py +628 -0
- autogen/oai/cerebras.py +299 -0
- autogen/oai/client.py +1435 -0
- autogen/oai/client_utils.py +169 -0
- autogen/oai/cohere.py +479 -0
- autogen/oai/gemini.py +990 -0
- autogen/oai/gemini_types.py +129 -0
- autogen/oai/groq.py +305 -0
- autogen/oai/mistral.py +303 -0
- autogen/oai/oai_models/__init__.py +11 -0
- autogen/oai/oai_models/_models.py +16 -0
- autogen/oai/oai_models/chat_completion.py +87 -0
- autogen/oai/oai_models/chat_completion_audio.py +32 -0
- autogen/oai/oai_models/chat_completion_message.py +86 -0
- autogen/oai/oai_models/chat_completion_message_tool_call.py +37 -0
- autogen/oai/oai_models/chat_completion_token_logprob.py +63 -0
- autogen/oai/oai_models/completion_usage.py +60 -0
- autogen/oai/ollama.py +643 -0
- autogen/oai/openai_utils.py +881 -0
- autogen/oai/together.py +370 -0
- autogen/retrieve_utils.py +491 -0
- autogen/runtime_logging.py +160 -0
- autogen/token_count_utils.py +267 -0
- autogen/tools/__init__.py +20 -0
- autogen/tools/contrib/__init__.py +9 -0
- autogen/tools/contrib/time/__init__.py +7 -0
- autogen/tools/contrib/time/time.py +41 -0
- autogen/tools/dependency_injection.py +254 -0
- autogen/tools/experimental/__init__.py +43 -0
- autogen/tools/experimental/browser_use/__init__.py +7 -0
- autogen/tools/experimental/browser_use/browser_use.py +161 -0
- autogen/tools/experimental/crawl4ai/__init__.py +7 -0
- autogen/tools/experimental/crawl4ai/crawl4ai.py +153 -0
- autogen/tools/experimental/deep_research/__init__.py +7 -0
- autogen/tools/experimental/deep_research/deep_research.py +328 -0
- autogen/tools/experimental/duckduckgo/__init__.py +7 -0
- autogen/tools/experimental/duckduckgo/duckduckgo_search.py +109 -0
- autogen/tools/experimental/google/__init__.py +14 -0
- autogen/tools/experimental/google/authentication/__init__.py +11 -0
- autogen/tools/experimental/google/authentication/credentials_hosted_provider.py +43 -0
- autogen/tools/experimental/google/authentication/credentials_local_provider.py +91 -0
- autogen/tools/experimental/google/authentication/credentials_provider.py +35 -0
- autogen/tools/experimental/google/drive/__init__.py +9 -0
- autogen/tools/experimental/google/drive/drive_functions.py +124 -0
- autogen/tools/experimental/google/drive/toolkit.py +88 -0
- autogen/tools/experimental/google/model.py +17 -0
- autogen/tools/experimental/google/toolkit_protocol.py +19 -0
- autogen/tools/experimental/google_search/__init__.py +8 -0
- autogen/tools/experimental/google_search/google_search.py +93 -0
- autogen/tools/experimental/google_search/youtube_search.py +181 -0
- autogen/tools/experimental/messageplatform/__init__.py +17 -0
- autogen/tools/experimental/messageplatform/discord/__init__.py +7 -0
- autogen/tools/experimental/messageplatform/discord/discord.py +288 -0
- autogen/tools/experimental/messageplatform/slack/__init__.py +7 -0
- autogen/tools/experimental/messageplatform/slack/slack.py +391 -0
- autogen/tools/experimental/messageplatform/telegram/__init__.py +7 -0
- autogen/tools/experimental/messageplatform/telegram/telegram.py +275 -0
- autogen/tools/experimental/perplexity/__init__.py +7 -0
- autogen/tools/experimental/perplexity/perplexity_search.py +260 -0
- autogen/tools/experimental/tavily/__init__.py +7 -0
- autogen/tools/experimental/tavily/tavily_search.py +183 -0
- autogen/tools/experimental/web_search_preview/__init__.py +7 -0
- autogen/tools/experimental/web_search_preview/web_search_preview.py +114 -0
- autogen/tools/experimental/wikipedia/__init__.py +7 -0
- autogen/tools/experimental/wikipedia/wikipedia.py +287 -0
- autogen/tools/function_utils.py +411 -0
- autogen/tools/tool.py +187 -0
- autogen/tools/toolkit.py +86 -0
- autogen/types.py +29 -0
- autogen/version.py +7 -0
- ag2-0.9.1a1.dist-info/RECORD +0 -6
- ag2-0.9.1a1.dist-info/top_level.txt +0 -1
- {ag2-0.9.1a1.dist-info → ag2-0.9.1.post0.dist-info/licenses}/LICENSE +0 -0
- {ag2-0.9.1a1.dist-info → ag2-0.9.1.post0.dist-info/licenses}/NOTICE.md +0 -0
|
@@ -0,0 +1,169 @@
|
|
|
1
|
+
# Copyright (c) 2023 - 2025, AG2ai, Inc., AG2ai open-source projects maintainers and core contributors
|
|
2
|
+
#
|
|
3
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
4
|
+
#
|
|
5
|
+
# Portions derived from https://github.com/microsoft/autogen are under the MIT License.
|
|
6
|
+
# SPDX-License-Identifier: MIT
|
|
7
|
+
"""Utilities for client classes"""
|
|
8
|
+
|
|
9
|
+
import logging
|
|
10
|
+
import warnings
|
|
11
|
+
from typing import Any, Optional, Protocol, runtime_checkable
|
|
12
|
+
|
|
13
|
+
|
|
14
|
+
@runtime_checkable
|
|
15
|
+
class FormatterProtocol(Protocol):
|
|
16
|
+
"""Structured Output classes with a format method"""
|
|
17
|
+
|
|
18
|
+
def format(self) -> str: ...
|
|
19
|
+
|
|
20
|
+
|
|
21
|
+
def validate_parameter(
|
|
22
|
+
params: dict[str, Any],
|
|
23
|
+
param_name: str,
|
|
24
|
+
allowed_types: tuple[Any, ...],
|
|
25
|
+
allow_None: bool, # noqa: N803
|
|
26
|
+
default_value: Any,
|
|
27
|
+
numerical_bound: Optional[tuple[Optional[float], Optional[float]]],
|
|
28
|
+
allowed_values: Optional[list[Any]],
|
|
29
|
+
) -> Any:
|
|
30
|
+
"""Validates a given config parameter, checking its type, values, and setting defaults
|
|
31
|
+
Parameters:
|
|
32
|
+
params (Dict[str, Any]): Dictionary containing parameters to validate.
|
|
33
|
+
param_name (str): The name of the parameter to validate.
|
|
34
|
+
allowed_types (Tuple): Tuple of acceptable types for the parameter.
|
|
35
|
+
allow_None (bool): Whether the parameter can be `None`.
|
|
36
|
+
default_value (Any): The default value to use if the parameter is invalid or missing.
|
|
37
|
+
numerical_bound (Optional[Tuple[Optional[float], Optional[float]]]):
|
|
38
|
+
A tuple specifying the lower and upper bounds for numerical parameters.
|
|
39
|
+
Each bound can be `None` if not applicable.
|
|
40
|
+
allowed_values (Optional[List[Any]]): A list of acceptable values for the parameter.
|
|
41
|
+
Can be `None` if no specific values are required.
|
|
42
|
+
|
|
43
|
+
Returns:
|
|
44
|
+
Any: The validated parameter value or the default value if validation fails.
|
|
45
|
+
|
|
46
|
+
Raises:
|
|
47
|
+
TypeError: If `allowed_values` is provided but is not a list.
|
|
48
|
+
|
|
49
|
+
Example Usage:
|
|
50
|
+
```python
|
|
51
|
+
# Validating a numerical parameter within specific bounds
|
|
52
|
+
params = {"temperature": 0.5, "safety_model": "Meta-Llama/Llama-Guard-7b"}
|
|
53
|
+
temperature = validate_parameter(params, "temperature", (int, float), True, 0.7, (0, 1), None)
|
|
54
|
+
# Result: 0.5
|
|
55
|
+
|
|
56
|
+
# Validating a parameter that can be one of a list of allowed values
|
|
57
|
+
model = validate_parameter(
|
|
58
|
+
params, "safety_model", str, True, None, None, ["Meta-Llama/Llama-Guard-7b", "Meta-Llama/Llama-Guard-13b"]
|
|
59
|
+
)
|
|
60
|
+
# If "safety_model" is missing or invalid in params, defaults to "default"
|
|
61
|
+
```
|
|
62
|
+
"""
|
|
63
|
+
if allowed_values is not None and not isinstance(allowed_values, list):
|
|
64
|
+
raise TypeError(f"allowed_values should be a list or None, got {type(allowed_values).__name__}")
|
|
65
|
+
|
|
66
|
+
param_value = params.get(param_name, default_value)
|
|
67
|
+
warning = ""
|
|
68
|
+
|
|
69
|
+
if param_value is None and allow_None:
|
|
70
|
+
pass
|
|
71
|
+
elif param_value is None:
|
|
72
|
+
if not allow_None:
|
|
73
|
+
warning = "cannot be None"
|
|
74
|
+
elif not isinstance(param_value, allowed_types):
|
|
75
|
+
# Check types and list possible types if invalid
|
|
76
|
+
if isinstance(allowed_types, tuple):
|
|
77
|
+
formatted_types = "(" + ", ".join(f"{t.__name__}" for t in allowed_types) + ")"
|
|
78
|
+
else:
|
|
79
|
+
formatted_types = f"{allowed_types.__name__}"
|
|
80
|
+
warning = f"must be of type {formatted_types}{' or None' if allow_None else ''}"
|
|
81
|
+
elif numerical_bound:
|
|
82
|
+
# Check the value fits in possible bounds
|
|
83
|
+
lower_bound, upper_bound = numerical_bound
|
|
84
|
+
if (lower_bound is not None and param_value < lower_bound) or (
|
|
85
|
+
upper_bound is not None and param_value > upper_bound
|
|
86
|
+
):
|
|
87
|
+
warning = "has numerical bounds"
|
|
88
|
+
if lower_bound is not None:
|
|
89
|
+
warning += f", >= {lower_bound!s}"
|
|
90
|
+
if upper_bound is not None:
|
|
91
|
+
if lower_bound is not None:
|
|
92
|
+
warning += " and"
|
|
93
|
+
warning += f" <= {upper_bound!s}"
|
|
94
|
+
if allow_None:
|
|
95
|
+
warning += ", or can be None"
|
|
96
|
+
|
|
97
|
+
elif allowed_values: # noqa: SIM102
|
|
98
|
+
# Check if the value matches any allowed values
|
|
99
|
+
if not (allow_None and param_value is None) and param_value not in allowed_values:
|
|
100
|
+
warning = f"must be one of these values [{allowed_values}]{', or can be None' if allow_None else ''}"
|
|
101
|
+
|
|
102
|
+
# If we failed any checks, warn and set to default value
|
|
103
|
+
if warning:
|
|
104
|
+
warnings.warn(
|
|
105
|
+
f"Config error - {param_name} {warning}, defaulting to {default_value}.",
|
|
106
|
+
UserWarning,
|
|
107
|
+
)
|
|
108
|
+
param_value = default_value
|
|
109
|
+
|
|
110
|
+
return param_value
|
|
111
|
+
|
|
112
|
+
|
|
113
|
+
def should_hide_tools(messages: list[dict[str, Any]], tools: list[dict[str, Any]], hide_tools_param: str) -> bool:
|
|
114
|
+
"""Determines if tools should be hidden. This function is used to hide tools when they have been run, minimising the chance of the LLM choosing them when they shouldn't.
|
|
115
|
+
Parameters:
|
|
116
|
+
messages (List[Dict[str, Any]]): List of messages
|
|
117
|
+
tools (List[Dict[str, Any]]): List of tools
|
|
118
|
+
hide_tools_param (str): "hide_tools" parameter value. Can be "if_all_run" (hide tools if all tools have been run), "if_any_run" (hide tools if any of the tools have been run), "never" (never hide tools). Default is "never".
|
|
119
|
+
|
|
120
|
+
Returns:
|
|
121
|
+
bool: Indicates whether the tools should be excluded from the response create request
|
|
122
|
+
|
|
123
|
+
Example Usage:
|
|
124
|
+
```python
|
|
125
|
+
# Validating a numerical parameter within specific bounds
|
|
126
|
+
messages = params.get("messages", [])
|
|
127
|
+
tools = params.get("tools", None)
|
|
128
|
+
hide_tools = should_hide_tools(messages, tools, params["hide_tools"])
|
|
129
|
+
"""
|
|
130
|
+
if hide_tools_param == "never" or tools is None or len(tools) == 0:
|
|
131
|
+
return False
|
|
132
|
+
elif hide_tools_param == "if_any_run":
|
|
133
|
+
# Return True if any tool_call_id exists, indicating a tool call has been executed. False otherwise.
|
|
134
|
+
return any(["tool_call_id" in dictionary for dictionary in messages])
|
|
135
|
+
elif hide_tools_param == "if_all_run":
|
|
136
|
+
# Return True if all tools have been executed at least once. False otherwise.
|
|
137
|
+
|
|
138
|
+
# Get the list of tool names
|
|
139
|
+
check_tool_names = [item["function"]["name"] for item in tools]
|
|
140
|
+
|
|
141
|
+
# Prepare a list of tool call ids and related function names
|
|
142
|
+
tool_call_ids = {}
|
|
143
|
+
|
|
144
|
+
# Loop through the messages and check if the tools have been run, removing them as we go
|
|
145
|
+
for message in messages:
|
|
146
|
+
if "tool_calls" in message:
|
|
147
|
+
# Register the tool ids and the function names (there could be multiple tool calls)
|
|
148
|
+
for tool_call in message["tool_calls"]:
|
|
149
|
+
tool_call_ids[tool_call["id"]] = tool_call["function"]["name"]
|
|
150
|
+
elif "tool_call_id" in message:
|
|
151
|
+
# Tool called, get the name of the function based on the id
|
|
152
|
+
tool_name_called = tool_call_ids[message["tool_call_id"]]
|
|
153
|
+
|
|
154
|
+
# If we had not yet called the tool, check and remove it to indicate we have
|
|
155
|
+
if tool_name_called in check_tool_names:
|
|
156
|
+
check_tool_names.remove(tool_name_called)
|
|
157
|
+
|
|
158
|
+
# Return True if all tools have been called at least once (accounted for)
|
|
159
|
+
return len(check_tool_names) == 0
|
|
160
|
+
else:
|
|
161
|
+
raise TypeError(
|
|
162
|
+
f"hide_tools_param is not a valid value ['if_all_run','if_any_run','never'], got '{hide_tools_param}'"
|
|
163
|
+
)
|
|
164
|
+
|
|
165
|
+
|
|
166
|
+
# Logging format (originally from FLAML)
|
|
167
|
+
logging_formatter = logging.Formatter(
|
|
168
|
+
"[%(name)s: %(asctime)s] {%(lineno)d} %(levelname)s - %(message)s", "%m-%d %H:%M:%S"
|
|
169
|
+
)
|
autogen/oai/cohere.py
ADDED
|
@@ -0,0 +1,479 @@
|
|
|
1
|
+
# Copyright (c) 2023 - 2025, AG2ai, Inc., AG2ai open-source projects maintainers and core contributors
|
|
2
|
+
#
|
|
3
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
4
|
+
#
|
|
5
|
+
# Portions derived from https://github.com/microsoft/autogen are under the MIT License.
|
|
6
|
+
# SPDX-License-Identifier: MIT
|
|
7
|
+
"""Create an OpenAI-compatible client using Cohere's API.
|
|
8
|
+
|
|
9
|
+
Example:
|
|
10
|
+
```python
|
|
11
|
+
llm_config={
|
|
12
|
+
"config_list": [{
|
|
13
|
+
"api_type": "cohere",
|
|
14
|
+
"model": "command-r-plus",
|
|
15
|
+
"api_key": os.environ.get("COHERE_API_KEY")
|
|
16
|
+
"client_name": "autogen-cohere", # Optional parameter
|
|
17
|
+
}
|
|
18
|
+
]}
|
|
19
|
+
|
|
20
|
+
agent = autogen.AssistantAgent("my_agent", llm_config=llm_config)
|
|
21
|
+
```
|
|
22
|
+
|
|
23
|
+
Install Cohere's python library using: pip install --upgrade cohere
|
|
24
|
+
|
|
25
|
+
Resources:
|
|
26
|
+
- https://docs.cohere.com/reference/chat
|
|
27
|
+
"""
|
|
28
|
+
|
|
29
|
+
from __future__ import annotations
|
|
30
|
+
|
|
31
|
+
import json
|
|
32
|
+
import logging
|
|
33
|
+
import os
|
|
34
|
+
import sys
|
|
35
|
+
import time
|
|
36
|
+
import warnings
|
|
37
|
+
from typing import Any, Literal, Optional, Type
|
|
38
|
+
|
|
39
|
+
from pydantic import BaseModel, Field
|
|
40
|
+
|
|
41
|
+
from autogen.oai.client_utils import FormatterProtocol, logging_formatter, validate_parameter
|
|
42
|
+
|
|
43
|
+
from ..import_utils import optional_import_block, require_optional_import
|
|
44
|
+
from ..llm_config import LLMConfigEntry, register_llm_config
|
|
45
|
+
from .oai_models import ChatCompletion, ChatCompletionMessage, ChatCompletionMessageToolCall, Choice, CompletionUsage
|
|
46
|
+
|
|
47
|
+
with optional_import_block():
|
|
48
|
+
from cohere import ClientV2 as CohereV2
|
|
49
|
+
from cohere.types import ToolResult
|
|
50
|
+
|
|
51
|
+
logger = logging.getLogger(__name__)
|
|
52
|
+
if not logger.handlers:
|
|
53
|
+
# Add the console handler.
|
|
54
|
+
_ch = logging.StreamHandler(stream=sys.stdout)
|
|
55
|
+
_ch.setFormatter(logging_formatter)
|
|
56
|
+
logger.addHandler(_ch)
|
|
57
|
+
|
|
58
|
+
|
|
59
|
+
COHERE_PRICING_1K = {
|
|
60
|
+
"command-r-plus": (0.003, 0.015),
|
|
61
|
+
"command-r": (0.0005, 0.0015),
|
|
62
|
+
"command-nightly": (0.00025, 0.00125),
|
|
63
|
+
"command": (0.015, 0.075),
|
|
64
|
+
"command-light": (0.008, 0.024),
|
|
65
|
+
"command-light-nightly": (0.008, 0.024),
|
|
66
|
+
}
|
|
67
|
+
|
|
68
|
+
|
|
69
|
+
@register_llm_config
|
|
70
|
+
class CohereLLMConfigEntry(LLMConfigEntry):
|
|
71
|
+
api_type: Literal["cohere"] = "cohere"
|
|
72
|
+
temperature: float = Field(default=0.3, ge=0)
|
|
73
|
+
max_tokens: Optional[int] = Field(default=None, ge=0)
|
|
74
|
+
k: int = Field(default=0, ge=0, le=500)
|
|
75
|
+
p: float = Field(default=0.75, ge=0.01, le=0.99)
|
|
76
|
+
seed: Optional[int] = None
|
|
77
|
+
frequency_penalty: float = Field(default=0, ge=0, le=1)
|
|
78
|
+
presence_penalty: float = Field(default=0, ge=0, le=1)
|
|
79
|
+
client_name: Optional[str] = None
|
|
80
|
+
strict_tools: bool = False
|
|
81
|
+
stream: bool = False
|
|
82
|
+
tool_choice: Optional[Literal["NONE", "REQUIRED"]] = None
|
|
83
|
+
|
|
84
|
+
def create_client(self):
|
|
85
|
+
raise NotImplementedError("CohereLLMConfigEntry.create_client is not implemented.")
|
|
86
|
+
|
|
87
|
+
|
|
88
|
+
class CohereClient:
|
|
89
|
+
"""Client for Cohere's API."""
|
|
90
|
+
|
|
91
|
+
def __init__(self, **kwargs):
|
|
92
|
+
"""Requires api_key or environment variable to be set
|
|
93
|
+
|
|
94
|
+
Args:
|
|
95
|
+
**kwargs: The keyword arguments to pass to the Cohere API.
|
|
96
|
+
"""
|
|
97
|
+
# Ensure we have the api_key upon instantiation
|
|
98
|
+
self.api_key = kwargs.get("api_key")
|
|
99
|
+
if not self.api_key:
|
|
100
|
+
self.api_key = os.getenv("COHERE_API_KEY")
|
|
101
|
+
|
|
102
|
+
assert self.api_key, (
|
|
103
|
+
"Please include the api_key in your config list entry for Cohere or set the COHERE_API_KEY env variable."
|
|
104
|
+
)
|
|
105
|
+
|
|
106
|
+
# Store the response format, if provided (for structured outputs)
|
|
107
|
+
self._response_format: Optional[Type[BaseModel]] = None
|
|
108
|
+
|
|
109
|
+
def message_retrieval(self, response) -> list:
|
|
110
|
+
"""Retrieve and return a list of strings or a list of Choice.Message from the response.
|
|
111
|
+
|
|
112
|
+
NOTE: if a list of Choice.Message is returned, it currently needs to contain the fields of OpenAI's ChatCompletion Message object,
|
|
113
|
+
since that is expected for function or tool calling in the rest of the codebase at the moment, unless a custom agent is being used.
|
|
114
|
+
"""
|
|
115
|
+
return [choice.message for choice in response.choices]
|
|
116
|
+
|
|
117
|
+
def cost(self, response) -> float:
|
|
118
|
+
return response.cost
|
|
119
|
+
|
|
120
|
+
@staticmethod
|
|
121
|
+
def get_usage(response) -> dict:
|
|
122
|
+
"""Return usage summary of the response using RESPONSE_USAGE_KEYS."""
|
|
123
|
+
# ... # pragma: no cover
|
|
124
|
+
return {
|
|
125
|
+
"prompt_tokens": response.usage.prompt_tokens,
|
|
126
|
+
"completion_tokens": response.usage.completion_tokens,
|
|
127
|
+
"total_tokens": response.usage.total_tokens,
|
|
128
|
+
"cost": response.cost,
|
|
129
|
+
"model": response.model,
|
|
130
|
+
}
|
|
131
|
+
|
|
132
|
+
def parse_params(self, params: dict[str, Any]) -> dict[str, Any]:
|
|
133
|
+
"""Loads the parameters for Cohere API from the passed in parameters and returns a validated set. Checks types, ranges, and sets defaults"""
|
|
134
|
+
cohere_params = {}
|
|
135
|
+
|
|
136
|
+
# Check that we have what we need to use Cohere's API
|
|
137
|
+
# We won't enforce the available models as they are likely to change
|
|
138
|
+
cohere_params["model"] = params.get("model")
|
|
139
|
+
assert cohere_params["model"], (
|
|
140
|
+
"Please specify the 'model' in your config list entry to nominate the Cohere model to use."
|
|
141
|
+
)
|
|
142
|
+
|
|
143
|
+
# Handle structured output response format from Pydantic model
|
|
144
|
+
if "response_format" in params and params["response_format"] is not None:
|
|
145
|
+
self._response_format = params.get("response_format")
|
|
146
|
+
|
|
147
|
+
response_format = params["response_format"]
|
|
148
|
+
|
|
149
|
+
# Check if it's a Pydantic model
|
|
150
|
+
if hasattr(response_format, "model_json_schema"):
|
|
151
|
+
# Get the JSON schema from the Pydantic model
|
|
152
|
+
schema = response_format.model_json_schema()
|
|
153
|
+
|
|
154
|
+
def resolve_ref(ref: str, defs: dict) -> dict:
|
|
155
|
+
"""Resolve a $ref to its actual schema definition"""
|
|
156
|
+
# Extract the definition name from "#/$defs/Name"
|
|
157
|
+
def_name = ref.split("/")[-1]
|
|
158
|
+
return defs[def_name]
|
|
159
|
+
|
|
160
|
+
def ensure_type_fields(obj: dict, defs: dict) -> dict:
|
|
161
|
+
"""Recursively ensure all objects in the schema have a type and properties field"""
|
|
162
|
+
if isinstance(obj, dict):
|
|
163
|
+
# If it has a $ref, replace it with the actual definition
|
|
164
|
+
if "$ref" in obj:
|
|
165
|
+
ref_def = resolve_ref(obj["$ref"], defs)
|
|
166
|
+
# Merge the reference definition with any existing fields
|
|
167
|
+
obj = {**ref_def, **obj}
|
|
168
|
+
# Remove the $ref as we've replaced it
|
|
169
|
+
del obj["$ref"]
|
|
170
|
+
|
|
171
|
+
# Process each value recursively
|
|
172
|
+
return {
|
|
173
|
+
k: ensure_type_fields(v, defs) if isinstance(v, (dict, list)) else v for k, v in obj.items()
|
|
174
|
+
}
|
|
175
|
+
elif isinstance(obj, list):
|
|
176
|
+
return [ensure_type_fields(item, defs) for item in obj]
|
|
177
|
+
return obj
|
|
178
|
+
|
|
179
|
+
# Make a copy of $defs before processing
|
|
180
|
+
defs = schema.get("$defs", {})
|
|
181
|
+
|
|
182
|
+
# Process the schema
|
|
183
|
+
processed_schema = ensure_type_fields(schema, defs)
|
|
184
|
+
|
|
185
|
+
cohere_params["response_format"] = {"type": "json_object", "json_schema": processed_schema}
|
|
186
|
+
else:
|
|
187
|
+
raise ValueError("response_format must be a Pydantic BaseModel")
|
|
188
|
+
|
|
189
|
+
# Handle strict tools parameter for structured outputs with tools
|
|
190
|
+
if "tools" in params:
|
|
191
|
+
cohere_params["strict_tools"] = validate_parameter(params, "strict_tools", bool, False, False, None, None)
|
|
192
|
+
|
|
193
|
+
# Validate allowed Cohere parameters
|
|
194
|
+
# https://docs.cohere.com/reference/chat
|
|
195
|
+
if "temperature" in params:
|
|
196
|
+
cohere_params["temperature"] = validate_parameter(
|
|
197
|
+
params, "temperature", (int, float), False, 0.3, (0, None), None
|
|
198
|
+
)
|
|
199
|
+
|
|
200
|
+
if "max_tokens" in params:
|
|
201
|
+
cohere_params["max_tokens"] = validate_parameter(params, "max_tokens", int, True, None, (0, None), None)
|
|
202
|
+
|
|
203
|
+
if "k" in params:
|
|
204
|
+
cohere_params["k"] = validate_parameter(params, "k", int, False, 0, (0, 500), None)
|
|
205
|
+
|
|
206
|
+
if "p" in params:
|
|
207
|
+
cohere_params["p"] = validate_parameter(params, "p", (int, float), False, 0.75, (0.01, 0.99), None)
|
|
208
|
+
|
|
209
|
+
if "seed" in params:
|
|
210
|
+
cohere_params["seed"] = validate_parameter(params, "seed", int, True, None, None, None)
|
|
211
|
+
|
|
212
|
+
if "frequency_penalty" in params:
|
|
213
|
+
cohere_params["frequency_penalty"] = validate_parameter(
|
|
214
|
+
params, "frequency_penalty", (int, float), True, 0, (0, 1), None
|
|
215
|
+
)
|
|
216
|
+
|
|
217
|
+
if "presence_penalty" in params:
|
|
218
|
+
cohere_params["presence_penalty"] = validate_parameter(
|
|
219
|
+
params, "presence_penalty", (int, float), True, 0, (0, 1), None
|
|
220
|
+
)
|
|
221
|
+
|
|
222
|
+
if "tool_choice" in params:
|
|
223
|
+
cohere_params["tool_choice"] = validate_parameter(
|
|
224
|
+
params, "tool_choice", str, True, None, None, ["NONE", "REQUIRED"]
|
|
225
|
+
)
|
|
226
|
+
|
|
227
|
+
return cohere_params
|
|
228
|
+
|
|
229
|
+
@require_optional_import("cohere", "cohere")
|
|
230
|
+
def create(self, params: dict) -> ChatCompletion:
|
|
231
|
+
messages = params.get("messages", [])
|
|
232
|
+
client_name = params.get("client_name") or "AG2"
|
|
233
|
+
cohere_tool_names = set()
|
|
234
|
+
tool_calls_modified_ids = set()
|
|
235
|
+
|
|
236
|
+
# Parse parameters to the Cohere API's parameters
|
|
237
|
+
cohere_params = self.parse_params(params)
|
|
238
|
+
|
|
239
|
+
cohere_params["messages"] = messages
|
|
240
|
+
|
|
241
|
+
if "tools" in params:
|
|
242
|
+
cohere_tool_names = set([tool["function"]["name"] for tool in params["tools"]])
|
|
243
|
+
cohere_params["tools"] = params["tools"]
|
|
244
|
+
|
|
245
|
+
# Strip out name
|
|
246
|
+
for message in cohere_params["messages"]:
|
|
247
|
+
message_name = message.pop("name", "")
|
|
248
|
+
# Extract and prepend name to content or tool_plan if available
|
|
249
|
+
message["content"] = (
|
|
250
|
+
f"{message_name}: {(message.get('content') or message.get('tool_plan'))}"
|
|
251
|
+
if message_name
|
|
252
|
+
else (message.get("content") or message.get("tool_plan"))
|
|
253
|
+
)
|
|
254
|
+
|
|
255
|
+
# Handle tool calls
|
|
256
|
+
if message.get("tool_calls") is not None and len(message["tool_calls"]) > 0:
|
|
257
|
+
message["tool_plan"] = message.get("tool_plan", message["content"])
|
|
258
|
+
del message["content"] # Remove content as tool_plan is prioritized
|
|
259
|
+
|
|
260
|
+
# If tool call name is missing or not recognized, modify role and content
|
|
261
|
+
for tool_call in message["tool_calls"] or []:
|
|
262
|
+
if (not tool_call.get("function", {}).get("name")) or tool_call.get("function", {}).get(
|
|
263
|
+
"name"
|
|
264
|
+
) not in cohere_tool_names:
|
|
265
|
+
message["role"] = "assistant"
|
|
266
|
+
message["content"] = f"{message.pop('tool_plan', '')}{str(message['tool_calls'])}"
|
|
267
|
+
tool_calls_modified_ids = tool_calls_modified_ids.union(
|
|
268
|
+
set([tool_call.get("id") for tool_call in message["tool_calls"]])
|
|
269
|
+
)
|
|
270
|
+
del message["tool_calls"]
|
|
271
|
+
break
|
|
272
|
+
|
|
273
|
+
# Adjust role if message comes from a tool with a modified ID
|
|
274
|
+
if message.get("role") == "tool":
|
|
275
|
+
tool_id = message.get("tool_call_id")
|
|
276
|
+
if tool_id in tool_calls_modified_ids:
|
|
277
|
+
message["role"] = "user"
|
|
278
|
+
del message["tool_call_id"] # Remove the tool call ID
|
|
279
|
+
|
|
280
|
+
# We use chat model by default
|
|
281
|
+
client = CohereV2(api_key=self.api_key, client_name=client_name)
|
|
282
|
+
|
|
283
|
+
# Token counts will be returned
|
|
284
|
+
prompt_tokens = 0
|
|
285
|
+
completion_tokens = 0
|
|
286
|
+
total_tokens = 0
|
|
287
|
+
|
|
288
|
+
# Stream if in parameters
|
|
289
|
+
streaming = params.get("stream")
|
|
290
|
+
cohere_finish = "stop"
|
|
291
|
+
tool_calls = None
|
|
292
|
+
ans = None
|
|
293
|
+
if streaming:
|
|
294
|
+
response = client.chat_stream(**cohere_params)
|
|
295
|
+
# Streaming...
|
|
296
|
+
ans = ""
|
|
297
|
+
plan = ""
|
|
298
|
+
prompt_tokens = 0
|
|
299
|
+
completion_tokens = 0
|
|
300
|
+
for chunk in response:
|
|
301
|
+
if chunk.type == "content-delta":
|
|
302
|
+
ans = ans + chunk.delta.message.content.text
|
|
303
|
+
elif chunk.type == "tool-plan-delta":
|
|
304
|
+
plan = plan + chunk.delta.message.tool_plan
|
|
305
|
+
elif chunk.type == "tool-call-start":
|
|
306
|
+
cohere_finish = "tool_calls"
|
|
307
|
+
|
|
308
|
+
# Initialize a new tool call
|
|
309
|
+
tool_call = chunk.delta.message.tool_calls
|
|
310
|
+
current_tool = {
|
|
311
|
+
"id": tool_call.id,
|
|
312
|
+
"type": "function",
|
|
313
|
+
"function": {"name": tool_call.function.name, "arguments": ""},
|
|
314
|
+
}
|
|
315
|
+
elif chunk.type == "tool-call-delta":
|
|
316
|
+
# Progressively build the arguments as they stream in
|
|
317
|
+
if current_tool is not None:
|
|
318
|
+
current_tool["function"]["arguments"] += chunk.delta.message.tool_calls.function.arguments
|
|
319
|
+
elif chunk.type == "tool-call-end":
|
|
320
|
+
# Append the finished tool call to the list
|
|
321
|
+
if current_tool is not None:
|
|
322
|
+
if tool_calls is None:
|
|
323
|
+
tool_calls = []
|
|
324
|
+
tool_calls.append(ChatCompletionMessageToolCall(**current_tool))
|
|
325
|
+
current_tool = None
|
|
326
|
+
elif chunk.type == "message-start":
|
|
327
|
+
response_id = chunk.id
|
|
328
|
+
elif chunk.type == "message-end":
|
|
329
|
+
prompt_tokens = (
|
|
330
|
+
chunk.delta.usage.billed_units.input_tokens
|
|
331
|
+
) # Note total (billed+non-billed) available with ...usage.tokens...
|
|
332
|
+
completion_tokens = chunk.delta.usage.billed_units.output_tokens
|
|
333
|
+
|
|
334
|
+
total_tokens = prompt_tokens + completion_tokens
|
|
335
|
+
else:
|
|
336
|
+
response = client.chat(**cohere_params)
|
|
337
|
+
|
|
338
|
+
if response.message.tool_calls is not None:
|
|
339
|
+
ans = response.message.tool_plan
|
|
340
|
+
cohere_finish = "tool_calls"
|
|
341
|
+
tool_calls = []
|
|
342
|
+
for tool_call in response.message.tool_calls:
|
|
343
|
+
# if parameters are null, clear them out (Cohere can return a string "null" if no parameter values)
|
|
344
|
+
|
|
345
|
+
tool_calls.append(
|
|
346
|
+
ChatCompletionMessageToolCall(
|
|
347
|
+
id=tool_call.id,
|
|
348
|
+
function={
|
|
349
|
+
"name": tool_call.function.name,
|
|
350
|
+
"arguments": (
|
|
351
|
+
"" if tool_call.function.arguments is None else tool_call.function.arguments
|
|
352
|
+
),
|
|
353
|
+
},
|
|
354
|
+
type="function",
|
|
355
|
+
)
|
|
356
|
+
)
|
|
357
|
+
else:
|
|
358
|
+
ans: str = response.message.content[0].text
|
|
359
|
+
|
|
360
|
+
# Not using billed_units, but that may be better for cost purposes
|
|
361
|
+
prompt_tokens = (
|
|
362
|
+
response.usage.billed_units.input_tokens
|
|
363
|
+
) # Note total (billed+non-billed) available with ...usage.tokens...
|
|
364
|
+
completion_tokens = response.usage.billed_units.output_tokens
|
|
365
|
+
total_tokens = prompt_tokens + completion_tokens
|
|
366
|
+
|
|
367
|
+
response_id = response.id
|
|
368
|
+
|
|
369
|
+
# Clean up structured output if needed
|
|
370
|
+
if self._response_format:
|
|
371
|
+
# ans = clean_return_response_format(ans)
|
|
372
|
+
try:
|
|
373
|
+
parsed_response = self._convert_json_response(ans)
|
|
374
|
+
ans = _format_json_response(parsed_response, ans)
|
|
375
|
+
except ValueError as e:
|
|
376
|
+
ans = str(e)
|
|
377
|
+
|
|
378
|
+
# 3. convert output
|
|
379
|
+
message = ChatCompletionMessage(
|
|
380
|
+
role="assistant",
|
|
381
|
+
content=ans,
|
|
382
|
+
function_call=None,
|
|
383
|
+
tool_calls=tool_calls,
|
|
384
|
+
)
|
|
385
|
+
choices = [Choice(finish_reason=cohere_finish, index=0, message=message)]
|
|
386
|
+
|
|
387
|
+
response_oai = ChatCompletion(
|
|
388
|
+
id=response_id,
|
|
389
|
+
model=cohere_params["model"],
|
|
390
|
+
created=int(time.time()),
|
|
391
|
+
object="chat.completion",
|
|
392
|
+
choices=choices,
|
|
393
|
+
usage=CompletionUsage(
|
|
394
|
+
prompt_tokens=prompt_tokens,
|
|
395
|
+
completion_tokens=completion_tokens,
|
|
396
|
+
total_tokens=total_tokens,
|
|
397
|
+
),
|
|
398
|
+
cost=calculate_cohere_cost(prompt_tokens, completion_tokens, cohere_params["model"]),
|
|
399
|
+
)
|
|
400
|
+
|
|
401
|
+
return response_oai
|
|
402
|
+
|
|
403
|
+
def _convert_json_response(self, response: str) -> Any:
|
|
404
|
+
"""Extract and validate JSON response from the output for structured outputs.
|
|
405
|
+
Args:
|
|
406
|
+
response (str): The response from the API.
|
|
407
|
+
Returns:
|
|
408
|
+
Any: The parsed JSON response.
|
|
409
|
+
"""
|
|
410
|
+
if not self._response_format:
|
|
411
|
+
return response
|
|
412
|
+
|
|
413
|
+
try:
|
|
414
|
+
# Parse JSON and validate against the Pydantic model
|
|
415
|
+
json_data = json.loads(response)
|
|
416
|
+
return self._response_format.model_validate(json_data)
|
|
417
|
+
except Exception as e:
|
|
418
|
+
raise ValueError(
|
|
419
|
+
f"Failed to parse response as valid JSON matching the schema for Structured Output: {str(e)}"
|
|
420
|
+
)
|
|
421
|
+
|
|
422
|
+
|
|
423
|
+
def _format_json_response(response: Any, original_answer: str) -> str:
|
|
424
|
+
"""Formats the JSON response for structured outputs using the format method if it exists."""
|
|
425
|
+
return (
|
|
426
|
+
response.format() if isinstance(response, FormatterProtocol) else clean_return_response_format(original_answer)
|
|
427
|
+
)
|
|
428
|
+
|
|
429
|
+
|
|
430
|
+
def extract_to_cohere_tool_results(tool_call_id: str, content_output: str, all_tool_calls) -> list[dict[str, Any]]:
|
|
431
|
+
temp_tool_results = []
|
|
432
|
+
|
|
433
|
+
for tool_call in all_tool_calls:
|
|
434
|
+
if tool_call["id"] == tool_call_id:
|
|
435
|
+
call = {
|
|
436
|
+
"name": tool_call["function"]["name"],
|
|
437
|
+
"parameters": json.loads(
|
|
438
|
+
tool_call["function"]["arguments"] if tool_call["function"]["arguments"] != "" else "{}"
|
|
439
|
+
),
|
|
440
|
+
}
|
|
441
|
+
output = [{"value": content_output}]
|
|
442
|
+
temp_tool_results.append(ToolResult(call=call, outputs=output))
|
|
443
|
+
return temp_tool_results
|
|
444
|
+
|
|
445
|
+
|
|
446
|
+
def calculate_cohere_cost(input_tokens: int, output_tokens: int, model: str) -> float:
|
|
447
|
+
"""Calculate the cost of the completion using the Cohere pricing."""
|
|
448
|
+
total = 0.0
|
|
449
|
+
|
|
450
|
+
if model in COHERE_PRICING_1K:
|
|
451
|
+
input_cost_per_k, output_cost_per_k = COHERE_PRICING_1K[model]
|
|
452
|
+
input_cost = (input_tokens / 1000) * input_cost_per_k
|
|
453
|
+
output_cost = (output_tokens / 1000) * output_cost_per_k
|
|
454
|
+
total = input_cost + output_cost
|
|
455
|
+
else:
|
|
456
|
+
warnings.warn(f"Cost calculation not available for {model} model", UserWarning)
|
|
457
|
+
|
|
458
|
+
return total
|
|
459
|
+
|
|
460
|
+
|
|
461
|
+
def clean_return_response_format(response_str: str) -> str:
|
|
462
|
+
"""Clean up the response string by parsing through json library."""
|
|
463
|
+
# Parse the string to a JSON object to handle escapes
|
|
464
|
+
data = json.loads(response_str)
|
|
465
|
+
|
|
466
|
+
# Convert back to JSON string with minimal formatting
|
|
467
|
+
return json.dumps(data)
|
|
468
|
+
|
|
469
|
+
|
|
470
|
+
class CohereError(Exception):
|
|
471
|
+
"""Base class for other Cohere exceptions"""
|
|
472
|
+
|
|
473
|
+
pass
|
|
474
|
+
|
|
475
|
+
|
|
476
|
+
class CohereRateLimitError(CohereError):
|
|
477
|
+
"""Raised when rate limit is exceeded"""
|
|
478
|
+
|
|
479
|
+
pass
|