ag2 0.5.3b1__tar.gz → 0.6.0b2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of ag2 might be problematic. Click here for more details.

@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ag2
3
- Version: 0.5.3b1
3
+ Version: 0.6.0b2
4
4
  Summary: Alias package for pyautogen
5
5
  Home-page: https://github.com/ag2ai/ag2
6
6
  Author: Chi Wang & Qingyun Wu
@@ -10,7 +10,7 @@ Platform: UNKNOWN
10
10
  Classifier: Programming Language :: Python :: 3
11
11
  Classifier: License :: OSI Approved :: Apache Software License
12
12
  Classifier: Operating System :: OS Independent
13
- Requires-Python: >=3.8,<3.14
13
+ Requires-Python: >=3.9,<3.14
14
14
  Description-Content-Type: text/markdown
15
15
  Provides-Extra: anthropic
16
16
  Provides-Extra: autobuild
@@ -24,6 +24,10 @@ Provides-Extra: gemini
24
24
  Provides-Extra: graph
25
25
  Provides-Extra: graph-rag-falkor-db
26
26
  Provides-Extra: groq
27
+ Provides-Extra: interop
28
+ Provides-Extra: interop-crewai
29
+ Provides-Extra: interop-langchain
30
+ Provides-Extra: interop-pydantic-ai
27
31
  Provides-Extra: jupyter-executor
28
32
  Provides-Extra: lmm
29
33
  Provides-Extra: long-context
@@ -39,6 +43,7 @@ Provides-Extra: retrievechat-qdrant
39
43
  Provides-Extra: teachable
40
44
  Provides-Extra: test
41
45
  Provides-Extra: together
46
+ Provides-Extra: twilio
42
47
  Provides-Extra: types
43
48
  Provides-Extra: websockets
44
49
  Provides-Extra: websurfer
@@ -50,7 +55,7 @@ License-File: NOTICE.md
50
55
  ![Pypi Downloads](https://img.shields.io/pypi/dm/pyautogen?label=PyPI%20downloads)
51
56
  [![PyPI version](https://badge.fury.io/py/autogen.svg)](https://badge.fury.io/py/autogen)
52
57
  [![Build](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml/badge.svg)](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml)
53
- ![Python Version](https://img.shields.io/badge/3.8%20%7C%203.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
58
+ ![Python Version](https://img.shields.io/badge/3.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
54
59
  [![Discord](https://img.shields.io/discord/1153072414184452236?logo=discord&style=flat)](https://discord.gg/pAbnFJrkgZ)
55
60
  [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/cloudposse.svg?style=social&label=Follow%20%40ag2ai)](https://x.com/ag2ai)
56
61
 
@@ -94,7 +99,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
94
99
 
95
100
  :tada: Apr 17, 2024: Andrew Ng cited AutoGen in [The Batch newsletter](https://www.deeplearning.ai/the-batch/issue-245/) and [What's next for AI agentic workflows](https://youtu.be/sal78ACtGTc?si=JduUzN_1kDnMq0vF) at Sequoia Capital's AI Ascent (Mar 26).
96
101
 
97
- :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://ag2ai.github.io/ag2/blog/2024/03/03/AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
102
+ :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://docs.ag2.ai/blog/2024-03-03-AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
98
103
 
99
104
  <!-- :tada: Mar 1, 2024: the first AutoGen multi-agent experiment on the challenging [GAIA](https://huggingface.co/spaces/gaia-benchmark/leaderboard) benchmark achieved the No. 1 accuracy in all the three levels. -->
100
105
 
@@ -104,7 +109,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
104
109
 
105
110
  <!-- :fire: Nov 24: pyautogen [v0.2](https://github.com/ag2ai/ag2/releases/tag/v0.2.0) is released with many updates and new features compared to v0.1.1. It switches to using openai-python v1. Please read the [migration guide](https://ag2ai.github.io/ag2/docs/Installation#python). -->
106
111
 
107
- <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://ag2ai.github.io/ag2/blog/2023/11/13/OAI-assistants) for details and examples. -->
112
+ <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://docs.ag2.ai/blog/2023-11-13-OAI-assistants) for details and examples. -->
108
113
 
109
114
  :tada: Nov 8, 2023: AutoGen is selected into [Open100: Top 100 Open Source achievements](https://www.benchcouncil.org/evaluation/opencs/annual.html) 35 days after spinoff from [FLAML](https://github.com/microsoft/FLAML).
110
115
 
@@ -172,7 +177,7 @@ Find detailed instructions for users [here](https://ag2ai.github.io/ag2/docs/ins
172
177
 
173
178
  ### Option 2. Install AG2 Locally
174
179
 
175
- AG2 requires **Python version >= 3.8, < 3.14**. It can be installed from pip:
180
+ AG2 requires **Python version >= 3.9, < 3.14**. It can be installed from pip:
176
181
 
177
182
  ```bash
178
183
  pip install ag2
@@ -281,7 +286,7 @@ You can find detailed documentation about AG2 [here](https://ag2ai.github.io/ag2
281
286
 
282
287
  In addition, you can find:
283
288
 
284
- - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://ag2ai.github.io/ag2/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
289
+ - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://docs.ag2.ai/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
285
290
 
286
291
  - [Discord](https://discord.gg/pAbnFJrkgZ)
287
292
 
@@ -3,7 +3,7 @@
3
3
  ![Pypi Downloads](https://img.shields.io/pypi/dm/pyautogen?label=PyPI%20downloads)
4
4
  [![PyPI version](https://badge.fury.io/py/autogen.svg)](https://badge.fury.io/py/autogen)
5
5
  [![Build](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml/badge.svg)](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml)
6
- ![Python Version](https://img.shields.io/badge/3.8%20%7C%203.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
6
+ ![Python Version](https://img.shields.io/badge/3.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
7
7
  [![Discord](https://img.shields.io/discord/1153072414184452236?logo=discord&style=flat)](https://discord.gg/pAbnFJrkgZ)
8
8
  [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/cloudposse.svg?style=social&label=Follow%20%40ag2ai)](https://x.com/ag2ai)
9
9
 
@@ -47,7 +47,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
47
47
 
48
48
  :tada: Apr 17, 2024: Andrew Ng cited AutoGen in [The Batch newsletter](https://www.deeplearning.ai/the-batch/issue-245/) and [What's next for AI agentic workflows](https://youtu.be/sal78ACtGTc?si=JduUzN_1kDnMq0vF) at Sequoia Capital's AI Ascent (Mar 26).
49
49
 
50
- :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://ag2ai.github.io/ag2/blog/2024/03/03/AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
50
+ :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://docs.ag2.ai/blog/2024-03-03-AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
51
51
 
52
52
  <!-- :tada: Mar 1, 2024: the first AutoGen multi-agent experiment on the challenging [GAIA](https://huggingface.co/spaces/gaia-benchmark/leaderboard) benchmark achieved the No. 1 accuracy in all the three levels. -->
53
53
 
@@ -57,7 +57,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
57
57
 
58
58
  <!-- :fire: Nov 24: pyautogen [v0.2](https://github.com/ag2ai/ag2/releases/tag/v0.2.0) is released with many updates and new features compared to v0.1.1. It switches to using openai-python v1. Please read the [migration guide](https://ag2ai.github.io/ag2/docs/Installation#python). -->
59
59
 
60
- <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://ag2ai.github.io/ag2/blog/2023/11/13/OAI-assistants) for details and examples. -->
60
+ <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://docs.ag2.ai/blog/2023-11-13-OAI-assistants) for details and examples. -->
61
61
 
62
62
  :tada: Nov 8, 2023: AutoGen is selected into [Open100: Top 100 Open Source achievements](https://www.benchcouncil.org/evaluation/opencs/annual.html) 35 days after spinoff from [FLAML](https://github.com/microsoft/FLAML).
63
63
 
@@ -125,7 +125,7 @@ Find detailed instructions for users [here](https://ag2ai.github.io/ag2/docs/ins
125
125
 
126
126
  ### Option 2. Install AG2 Locally
127
127
 
128
- AG2 requires **Python version >= 3.8, < 3.14**. It can be installed from pip:
128
+ AG2 requires **Python version >= 3.9, < 3.14**. It can be installed from pip:
129
129
 
130
130
  ```bash
131
131
  pip install ag2
@@ -234,7 +234,7 @@ You can find detailed documentation about AG2 [here](https://ag2ai.github.io/ag2
234
234
 
235
235
  In addition, you can find:
236
236
 
237
- - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://ag2ai.github.io/ag2/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
237
+ - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://docs.ag2.ai/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
238
238
 
239
239
  - [Discord](https://discord.gg/pAbnFJrkgZ)
240
240
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ag2
3
- Version: 0.5.3b1
3
+ Version: 0.6.0b2
4
4
  Summary: Alias package for pyautogen
5
5
  Home-page: https://github.com/ag2ai/ag2
6
6
  Author: Chi Wang & Qingyun Wu
@@ -10,7 +10,7 @@ Platform: UNKNOWN
10
10
  Classifier: Programming Language :: Python :: 3
11
11
  Classifier: License :: OSI Approved :: Apache Software License
12
12
  Classifier: Operating System :: OS Independent
13
- Requires-Python: >=3.8,<3.14
13
+ Requires-Python: >=3.9,<3.14
14
14
  Description-Content-Type: text/markdown
15
15
  Provides-Extra: anthropic
16
16
  Provides-Extra: autobuild
@@ -24,6 +24,10 @@ Provides-Extra: gemini
24
24
  Provides-Extra: graph
25
25
  Provides-Extra: graph-rag-falkor-db
26
26
  Provides-Extra: groq
27
+ Provides-Extra: interop
28
+ Provides-Extra: interop-crewai
29
+ Provides-Extra: interop-langchain
30
+ Provides-Extra: interop-pydantic-ai
27
31
  Provides-Extra: jupyter-executor
28
32
  Provides-Extra: lmm
29
33
  Provides-Extra: long-context
@@ -39,6 +43,7 @@ Provides-Extra: retrievechat-qdrant
39
43
  Provides-Extra: teachable
40
44
  Provides-Extra: test
41
45
  Provides-Extra: together
46
+ Provides-Extra: twilio
42
47
  Provides-Extra: types
43
48
  Provides-Extra: websockets
44
49
  Provides-Extra: websurfer
@@ -50,7 +55,7 @@ License-File: NOTICE.md
50
55
  ![Pypi Downloads](https://img.shields.io/pypi/dm/pyautogen?label=PyPI%20downloads)
51
56
  [![PyPI version](https://badge.fury.io/py/autogen.svg)](https://badge.fury.io/py/autogen)
52
57
  [![Build](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml/badge.svg)](https://github.com/ag2ai/ag2/actions/workflows/python-package.yml)
53
- ![Python Version](https://img.shields.io/badge/3.8%20%7C%203.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
58
+ ![Python Version](https://img.shields.io/badge/3.9%20%7C%203.10%20%7C%203.11%20%7C%203.12-blue)
54
59
  [![Discord](https://img.shields.io/discord/1153072414184452236?logo=discord&style=flat)](https://discord.gg/pAbnFJrkgZ)
55
60
  [![Twitter](https://img.shields.io/twitter/url/https/twitter.com/cloudposse.svg?style=social&label=Follow%20%40ag2ai)](https://x.com/ag2ai)
56
61
 
@@ -94,7 +99,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
94
99
 
95
100
  :tada: Apr 17, 2024: Andrew Ng cited AutoGen in [The Batch newsletter](https://www.deeplearning.ai/the-batch/issue-245/) and [What's next for AI agentic workflows](https://youtu.be/sal78ACtGTc?si=JduUzN_1kDnMq0vF) at Sequoia Capital's AI Ascent (Mar 26).
96
101
 
97
- :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://ag2ai.github.io/ag2/blog/2024/03/03/AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
102
+ :tada: Mar 3, 2024: What's new in AutoGen? 📰[Blog](https://docs.ag2.ai/blog/2024-03-03-AutoGen-Update); 📺[Youtube](https://www.youtube.com/watch?v=j_mtwQiaLGU).
98
103
 
99
104
  <!-- :tada: Mar 1, 2024: the first AutoGen multi-agent experiment on the challenging [GAIA](https://huggingface.co/spaces/gaia-benchmark/leaderboard) benchmark achieved the No. 1 accuracy in all the three levels. -->
100
105
 
@@ -104,7 +109,7 @@ We adopt the Apache 2.0 license from v0.3. This enhances our commitment to open-
104
109
 
105
110
  <!-- :fire: Nov 24: pyautogen [v0.2](https://github.com/ag2ai/ag2/releases/tag/v0.2.0) is released with many updates and new features compared to v0.1.1. It switches to using openai-python v1. Please read the [migration guide](https://ag2ai.github.io/ag2/docs/Installation#python). -->
106
111
 
107
- <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://ag2ai.github.io/ag2/blog/2023/11/13/OAI-assistants) for details and examples. -->
112
+ <!-- :fire: Nov 11: OpenAI's Assistants are available in AutoGen and interoperatable with other AutoGen agents! Checkout our [blogpost](https://docs.ag2.ai/blog/2023-11-13-OAI-assistants) for details and examples. -->
108
113
 
109
114
  :tada: Nov 8, 2023: AutoGen is selected into [Open100: Top 100 Open Source achievements](https://www.benchcouncil.org/evaluation/opencs/annual.html) 35 days after spinoff from [FLAML](https://github.com/microsoft/FLAML).
110
115
 
@@ -172,7 +177,7 @@ Find detailed instructions for users [here](https://ag2ai.github.io/ag2/docs/ins
172
177
 
173
178
  ### Option 2. Install AG2 Locally
174
179
 
175
- AG2 requires **Python version >= 3.8, < 3.14**. It can be installed from pip:
180
+ AG2 requires **Python version >= 3.9, < 3.14**. It can be installed from pip:
176
181
 
177
182
  ```bash
178
183
  pip install ag2
@@ -281,7 +286,7 @@ You can find detailed documentation about AG2 [here](https://ag2ai.github.io/ag2
281
286
 
282
287
  In addition, you can find:
283
288
 
284
- - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://ag2ai.github.io/ag2/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
289
+ - [Research](https://ag2ai.github.io/ag2/docs/Research), [blogposts](https://docs.ag2.ai/blog) around AG2, and [Transparency FAQs](https://github.com/ag2ai/ag2/blob/main/TRANSPARENCY_FAQS.md)
285
290
 
286
291
  - [Discord](https://discord.gg/pAbnFJrkgZ)
287
292
 
@@ -0,0 +1,106 @@
1
+ pyautogen==0.6.0b2
2
+
3
+ [anthropic]
4
+ pyautogen[anthropic]==0.6.0b2
5
+
6
+ [autobuild]
7
+ pyautogen[autobuild]==0.6.0b2
8
+
9
+ [bedrock]
10
+ pyautogen[bedrock]==0.6.0b2
11
+
12
+ [blendsearch]
13
+ pyautogen[blendsearch]==0.6.0b2
14
+
15
+ [captainagent]
16
+ pyautogen[captainagent]==0.6.0b2
17
+
18
+ [cerebras]
19
+ pyautogen[cerebras]==0.6.0b2
20
+
21
+ [cohere]
22
+ pyautogen[cohere]==0.6.0b2
23
+
24
+ [cosmosdb]
25
+ pyautogen[cosmosdb]==0.6.0b2
26
+
27
+ [gemini]
28
+ pyautogen[gemini]==0.6.0b2
29
+
30
+ [graph]
31
+ pyautogen[graph]==0.6.0b2
32
+
33
+ [graph-rag-falkor-db]
34
+ pyautogen[graph-rag-falkor-db]==0.6.0b2
35
+
36
+ [groq]
37
+ pyautogen[groq]==0.6.0b2
38
+
39
+ [interop]
40
+ pyautogen[interop]==0.6.0b2
41
+
42
+ [interop-crewai]
43
+ pyautogen[interop-crewai]==0.6.0b2
44
+
45
+ [interop-langchain]
46
+ pyautogen[interop-langchain]==0.6.0b2
47
+
48
+ [interop-pydantic-ai]
49
+ pyautogen[interop-pydantic-ai]==0.6.0b2
50
+
51
+ [jupyter-executor]
52
+ pyautogen[jupyter-executor]==0.6.0b2
53
+
54
+ [lmm]
55
+ pyautogen[lmm]==0.6.0b2
56
+
57
+ [long-context]
58
+ pyautogen[long-context]==0.6.0b2
59
+
60
+ [mathchat]
61
+ pyautogen[mathchat]==0.6.0b2
62
+
63
+ [mistral]
64
+ pyautogen[mistral]==0.6.0b2
65
+
66
+ [neo4j]
67
+ pyautogen[neo4j]==0.6.0b2
68
+
69
+ [ollama]
70
+ pyautogen[ollama]==0.6.0b2
71
+
72
+ [redis]
73
+ pyautogen[redis]==0.6.0b2
74
+
75
+ [retrievechat]
76
+ pyautogen[retrievechat]==0.6.0b2
77
+
78
+ [retrievechat-mongodb]
79
+ pyautogen[retrievechat-mongodb]==0.6.0b2
80
+
81
+ [retrievechat-pgvector]
82
+ pyautogen[retrievechat-pgvector]==0.6.0b2
83
+
84
+ [retrievechat-qdrant]
85
+ pyautogen[retrievechat-qdrant]==0.6.0b2
86
+
87
+ [teachable]
88
+ pyautogen[teachable]==0.6.0b2
89
+
90
+ [test]
91
+ pyautogen[test]==0.6.0b2
92
+
93
+ [together]
94
+ pyautogen[together]==0.6.0b2
95
+
96
+ [twilio]
97
+ pyautogen[twilio]==0.6.0b2
98
+
99
+ [types]
100
+ pyautogen[types]==0.6.0b2
101
+
102
+ [websockets]
103
+ pyautogen[websockets]==0.6.0b2
104
+
105
+ [websurfer]
106
+ pyautogen[websurfer]==0.6.0b2
@@ -60,9 +60,15 @@ files = [
60
60
  "autogen/_pydantic.py",
61
61
  "autogen/function_utils.py",
62
62
  "autogen/io",
63
+ "autogen/tools",
64
+ "autogen/interop",
65
+ "autogen/agentchat/realtime_agent",
63
66
  "test/test_pydantic.py",
64
67
  "test/test_function_utils.py",
65
68
  "test/io",
69
+ "test/tools",
70
+ "test/interop",
71
+ "test/agentchat/realtime_agent",
66
72
  ]
67
73
  exclude = [
68
74
  "autogen/math_utils\\.py",
@@ -73,7 +79,6 @@ exclude = [
73
79
  ]
74
80
 
75
81
  strict = true
76
- python_version = "3.8"
77
82
  ignore_missing_imports = true
78
83
  install_types = true
79
84
  non_interactive = true
@@ -88,7 +93,8 @@ no_implicit_optional = true
88
93
  check_untyped_defs = true
89
94
  warn_return_any = true
90
95
  show_error_codes = true
91
- warn_unused_ignores = true
96
+
97
+ warn_unused_ignores = false
92
98
 
93
99
  disallow_incomplete_defs = true
94
100
  disallow_untyped_decorators = true
@@ -25,7 +25,7 @@ __version__ = version["__version__"]
25
25
  current_os = platform.system()
26
26
 
27
27
  install_requires = [
28
- "openai>=1.3",
28
+ "openai>=1.57",
29
29
  "diskcache",
30
30
  "termcolor",
31
31
  "flaml",
@@ -38,6 +38,19 @@ install_requires = [
38
38
  "pydantic>=1.10,<3,!=2.6.0", # could be both V1 and V2
39
39
  "docker",
40
40
  "packaging",
41
+ "websockets>=14,<15",
42
+ "asyncer>=0.0.8",
43
+ ]
44
+
45
+ test = [
46
+ "ipykernel",
47
+ "nbconvert",
48
+ "nbformat",
49
+ "pre-commit",
50
+ "pytest-cov>=5",
51
+ "pytest-asyncio",
52
+ "pytest>=8,<9",
53
+ "pandas",
41
54
  ]
42
55
 
43
56
  jupyter_executor = [
@@ -69,6 +82,16 @@ neo4j = [
69
82
  "llama-index-core==0.12.5",
70
83
  ]
71
84
 
85
+ # used for agentchat_realtime_swarm notebook and realtime agent twilio demo
86
+ twilio = ["fastapi>=0.115.0,<1", "uvicorn>=0.30.6,<1", "twilio>=9.3.2"]
87
+
88
+ interop_crewai = ["crewai[tools]>=0.86,<1; python_version>='3.10' and python_version<'3.13'"]
89
+ interop_langchain = ["langchain-community>=0.3.12,<1"]
90
+ interop_pydantic_ai = ["pydantic-ai==0.0.13"]
91
+ interop = interop_crewai + interop_langchain + interop_pydantic_ai
92
+
93
+ types = ["mypy==1.9.0"] + test + jupyter_executor + interop + ["fastapi>=0.115.0,<1"]
94
+
72
95
  if current_os in ["Windows", "Darwin"]:
73
96
  retrieve_chat_pgvector.extend(["psycopg[binary]>=3.1.18"])
74
97
  elif current_os == "Linux":
@@ -84,16 +107,7 @@ autobuild = ["chromadb", "sentence-transformers", "huggingface-hub", "pysqlite3-
84
107
  # PLEASE add them in the setup_ag2.py and setup_autogen.py files
85
108
 
86
109
  extra_require = {
87
- "test": [
88
- "ipykernel",
89
- "nbconvert",
90
- "nbformat",
91
- "pre-commit",
92
- "pytest-cov>=5",
93
- "pytest-asyncio",
94
- "pytest>=6.1.1,<8",
95
- "pandas",
96
- ],
110
+ "test": test,
97
111
  "blendsearch": ["flaml[blendsearch]"],
98
112
  "mathchat": ["sympy", "pydantic==1.10.9", "wolframalpha"],
99
113
  "retrievechat": retrieve_chat,
@@ -111,9 +125,9 @@ extra_require = {
111
125
  "websurfer": ["beautifulsoup4", "markdownify", "pdfminer.six", "pathvalidate"],
112
126
  "redis": ["redis"],
113
127
  "cosmosdb": ["azure-cosmos>=4.2.0"],
114
- "websockets": ["websockets>=12.0,<13"],
128
+ "websockets": ["websockets>=14.0,<15"],
115
129
  "jupyter-executor": jupyter_executor,
116
- "types": ["mypy==1.9.0", "pytest>=6.1.1,<8"] + jupyter_executor,
130
+ "types": types,
117
131
  "long-context": ["llmlingua<0.3"],
118
132
  "anthropic": ["anthropic>=0.23.1"],
119
133
  "cerebras": ["cerebras_cloud_sdk>=1.0.0"],
@@ -122,6 +136,11 @@ extra_require = {
122
136
  "cohere": ["cohere>=5.5.8"],
123
137
  "ollama": ["ollama>=0.3.3", "fix_busted_json>=0.0.18"],
124
138
  "bedrock": ["boto3>=1.34.149"],
139
+ "twilio": twilio,
140
+ "interop-crewai": interop_crewai,
141
+ "interop-langchain": interop_langchain,
142
+ "interop-pydantic-ai": interop_pydantic_ai,
143
+ "interop": interop,
125
144
  "neo4j": neo4j,
126
145
  }
127
146
 
@@ -156,5 +175,5 @@ setuptools.setup(
156
175
  "Operating System :: OS Independent",
157
176
  ],
158
177
  license="Apache Software License 2.0",
159
- python_requires=">=3.8,<3.14",
178
+ python_requires=">=3.9,<3.14",
160
179
  )
@@ -54,6 +54,11 @@ setuptools.setup(
54
54
  "cohere": ["pyautogen[cohere]==" + __version__],
55
55
  "ollama": ["pyautogen[ollama]==" + __version__],
56
56
  "bedrock": ["pyautogen[bedrock]==" + __version__],
57
+ "twilio": ["pyautogen[twilio]==" + __version__],
58
+ "interop-crewai": ["pyautogen[interop-crewai]==" + __version__],
59
+ "interop-langchain": ["pyautogen[interop-langchain]==" + __version__],
60
+ "interop-pydantic-ai": ["pyautogen[interop-pydantic-ai]==" + __version__],
61
+ "interop": ["pyautogen[interop]==" + __version__],
57
62
  "neo4j": ["pyautogen[neo4j]==" + __version__],
58
63
  },
59
64
  url="https://github.com/ag2ai/ag2",
@@ -65,5 +70,5 @@ setuptools.setup(
65
70
  "Operating System :: OS Independent",
66
71
  ],
67
72
  license="Apache Software License 2.0",
68
- python_requires=">=3.8,<3.14",
73
+ python_requires=">=3.9,<3.14",
69
74
  )
@@ -16,8 +16,8 @@ import pytest
16
16
  import requests
17
17
  from agentchat.test_assistant_agent import KEY_LOC # noqa: E402
18
18
 
19
- BLOG_POST_URL = "https://ag2ai.github.io/ag2/blog/2023/04/21/LLM-tuning-math"
20
- BLOG_POST_TITLE = "Does Model and Inference Parameter Matter in LLM Applications? - A Case Study for MATH | AG2"
19
+ BLOG_POST_URL = "https://docs.ag2.ai/blog/2023-04-21-LLM-tuning-math"
20
+ BLOG_POST_TITLE = "Does Model and Inference Parameter Matter in LLM Applications? - A Case Study for MATH - AG2"
21
21
  BLOG_POST_STRING = "Large language models (LLMs) are powerful tools that can generate natural language texts for various applications, such as chatbots, summarization, translation, and more. GPT-4 is currently the state of the art LLM in the world. Is model selection irrelevant? What about inference parameters?"
22
22
 
23
23
  WIKIPEDIA_URL = "https://en.wikipedia.org/wiki/Microsoft"
@@ -156,7 +156,7 @@ def test_simple_text_browser():
156
156
 
157
157
 
158
158
  @pytest.mark.skipif(
159
- skip_bing,
159
+ skip_all or skip_bing,
160
160
  reason="do not run bing tests if key is missing",
161
161
  )
162
162
  def test_bing_search():
@@ -54,7 +54,7 @@ else:
54
54
  # OAI_CONFIG_LIST,
55
55
  # file_location=KEY_LOC,
56
56
  # filter_dict={
57
- # "model": ["gpt-4", "gpt4", "gpt-4-32k", "gpt-4-32k-0314"],
57
+ # "model": ["gpt-4o", "gpt4", "gpt-4-32k", "gpt-4-32k-0314"],
58
58
  # },
59
59
  # )
60
60
  # # config_list = autogen.config_list_from_json(
@@ -7,11 +7,10 @@
7
7
  import asyncio
8
8
  import inspect
9
9
  import unittest.mock
10
- from typing import Any, Dict, List, Literal, Optional, Tuple
10
+ from typing import Annotated, Any, Dict, List, Literal, Optional, Tuple
11
11
 
12
12
  import pytest
13
13
  from pydantic import BaseModel, Field
14
- from typing_extensions import Annotated
15
14
 
16
15
  from autogen._pydantic import PYDANTIC_V1, model_dump
17
16
  from autogen.function_utils import (
@@ -40,7 +39,7 @@ def g( # type: ignore[empty-body]
40
39
  b: int = 2,
41
40
  c: Annotated[float, "Parameter c"] = 0.1,
42
41
  *,
43
- d: Dict[str, Tuple[Optional[int], List[float]]],
42
+ d: dict[str, tuple[Optional[int], list[float]]],
44
43
  ) -> str:
45
44
  pass
46
45
 
@@ -50,7 +49,7 @@ async def a_g( # type: ignore[empty-body]
50
49
  b: int = 2,
51
50
  c: Annotated[float, "Parameter c"] = 0.1,
52
51
  *,
53
- d: Dict[str, Tuple[Optional[int], List[float]]],
52
+ d: dict[str, tuple[Optional[int], list[float]]],
54
53
  ) -> str:
55
54
  pass
56
55
 
@@ -89,7 +88,7 @@ def test_get_parameter_json_schema() -> None:
89
88
  b: float
90
89
  c: str
91
90
 
92
- expected: Dict[str, Any] = {
91
+ expected: dict[str, Any] = {
93
92
  "description": "b",
94
93
  "properties": {"b": {"title": "B", "type": "number"}, "c": {"title": "C", "type": "string"}},
95
94
  "required": ["b", "c"],
@@ -367,7 +366,7 @@ def test_load_basemodels_if_needed_sync() -> None:
367
366
  def f(
368
367
  base: Annotated[Currency, "Base currency"],
369
368
  quote_currency: Annotated[CurrencySymbol, "Quote currency"] = "EUR",
370
- ) -> Tuple[Currency, CurrencySymbol]:
369
+ ) -> tuple[Currency, CurrencySymbol]:
371
370
  return base, quote_currency
372
371
 
373
372
  assert not inspect.iscoroutinefunction(f)
@@ -385,7 +384,7 @@ async def test_load_basemodels_if_needed_async() -> None:
385
384
  async def f(
386
385
  base: Annotated[Currency, "Base currency"],
387
386
  quote_currency: Annotated[CurrencySymbol, "Quote currency"] = "EUR",
388
- ) -> Tuple[Currency, CurrencySymbol]:
387
+ ) -> tuple[Currency, CurrencySymbol]:
389
388
  return base, quote_currency
390
389
 
391
390
  assert inspect.iscoroutinefunction(f)
@@ -35,7 +35,7 @@ SAMPLE_CHAT_REQUEST = json.loads(
35
35
  "role": "user"
36
36
  }
37
37
  ],
38
- "model": "gpt-4"
38
+ "model": "gpt-4o"
39
39
  }
40
40
  """
41
41
  )
@@ -58,7 +58,7 @@ SAMPLE_CHAT_RESPONSE = json.loads(
58
58
  }
59
59
  ],
60
60
  "created": 1705993480,
61
- "model": "gpt-4",
61
+ "model": "gpt-4o",
62
62
  "object": "chat.completion",
63
63
  "system_fingerprint": "fp_6d044fb900",
64
64
  "usage": {
@@ -159,7 +159,7 @@ def test_log_new_agent(db_connection):
159
159
 
160
160
  cur = db_connection.cursor()
161
161
  agent_name = "some_assistant"
162
- config_list = [{"model": "gpt-4", "api_key": "some_key"}]
162
+ config_list = [{"model": "gpt-4o", "api_key": "some_key"}]
163
163
 
164
164
  agent = AssistantAgent(agent_name, llm_config={"config_list": config_list})
165
165
  init_args = {"foo": "bar", "baz": {"other_key": "other_val"}, "a": None}
@@ -184,7 +184,7 @@ def test_log_oai_wrapper(db_connection):
184
184
 
185
185
  cur = db_connection.cursor()
186
186
 
187
- llm_config = {"config_list": [{"model": "gpt-4", "api_key": "some_key", "base_url": "some url"}]}
187
+ llm_config = {"config_list": [{"model": "gpt-4o", "api_key": "some_key", "base_url": "some url"}]}
188
188
  init_args = {"llm_config": llm_config, "base_config": {}}
189
189
  wrapper = OpenAIWrapper(**llm_config)
190
190
 
@@ -210,8 +210,8 @@ def test_log_oai_client(db_connection):
210
210
 
211
211
  openai_config = {
212
212
  "api_key": "some_key",
213
- "api_version": "2024-02-01",
214
- "azure_deployment": "gpt-4",
213
+ "api_version": "2024-08-06",
214
+ "azure_deployment": "gpt-4o",
215
215
  "azure_endpoint": "https://foobar.openai.azure.com/",
216
216
  }
217
217
  client = AzureOpenAI(**openai_config)
@@ -264,7 +264,7 @@ def test_to_dict():
264
264
  self.extra_key = "remove this key"
265
265
  self.path = Path("/to/something")
266
266
 
267
- class Bar(object):
267
+ class Bar:
268
268
  def init(self):
269
269
  pass
270
270
 
@@ -4,10 +4,9 @@
4
4
  #
5
5
  # Portions derived from https://github.com/microsoft/autogen are under the MIT License.
6
6
  # SPDX-License-Identifier: MIT
7
- from typing import Dict, List, Optional, Tuple, Union
7
+ from typing import Annotated, Dict, List, Optional, Tuple, Union
8
8
 
9
9
  from pydantic import BaseModel, Field
10
- from typing_extensions import Annotated
11
10
 
12
11
  from autogen._pydantic import model_dump, model_dump_json, type2schema
13
12
 
@@ -19,14 +18,14 @@ def test_type2schema() -> None:
19
18
  assert type2schema(bool) == {"type": "boolean"}
20
19
  assert type2schema(None) == {"type": "null"}
21
20
  assert type2schema(Optional[int]) == {"anyOf": [{"type": "integer"}, {"type": "null"}]}
22
- assert type2schema(List[int]) == {"items": {"type": "integer"}, "type": "array"}
23
- assert type2schema(Tuple[int, float, str]) == {
21
+ assert type2schema(list[int]) == {"items": {"type": "integer"}, "type": "array"}
22
+ assert type2schema(tuple[int, float, str]) == {
24
23
  "maxItems": 3,
25
24
  "minItems": 3,
26
25
  "prefixItems": [{"type": "integer"}, {"type": "number"}, {"type": "string"}],
27
26
  "type": "array",
28
27
  }
29
- assert type2schema(Dict[str, int]) == {"additionalProperties": {"type": "integer"}, "type": "object"}
28
+ assert type2schema(dict[str, int]) == {"additionalProperties": {"type": "integer"}, "type": "object"}
30
29
  assert type2schema(Annotated[str, "some text"]) == {"type": "string"}
31
30
  assert type2schema(Union[int, float]) == {"anyOf": [{"type": "integer"}, {"type": "number"}]}
32
31
 
@@ -226,7 +226,7 @@ class TestRetrieveUtils:
226
226
  dir_path="./website/docs",
227
227
  client=client,
228
228
  collection_name="autogen-docs",
229
- custom_text_types=["txt", "md", "rtf", "rst"],
229
+ custom_text_types=["txt", "md", "rtf", "rst", "mdx"],
230
230
  get_or_create=True,
231
231
  )
232
232
  results = query_vector_db(
@@ -141,6 +141,8 @@ def test_model_aliases():
141
141
  assert get_max_token_limit("gpt-35-turbo") == get_max_token_limit("gpt-3.5-turbo")
142
142
  assert get_max_token_limit("gpt4") == get_max_token_limit("gpt-4")
143
143
  assert get_max_token_limit("gpt4-32k") == get_max_token_limit("gpt-4-32k")
144
+ assert get_max_token_limit("gpt4o") == get_max_token_limit("gpt-4o")
145
+ assert get_max_token_limit("gpt4o-mini") == get_max_token_limit("gpt-4o-mini")
144
146
 
145
147
 
146
148
  if __name__ == "__main__":
@@ -1,91 +0,0 @@
1
- pyautogen==0.5.3b1
2
-
3
- [anthropic]
4
- pyautogen[anthropic]==0.5.3b1
5
-
6
- [autobuild]
7
- pyautogen[autobuild]==0.5.3b1
8
-
9
- [bedrock]
10
- pyautogen[bedrock]==0.5.3b1
11
-
12
- [blendsearch]
13
- pyautogen[blendsearch]==0.5.3b1
14
-
15
- [captainagent]
16
- pyautogen[captainagent]==0.5.3b1
17
-
18
- [cerebras]
19
- pyautogen[cerebras]==0.5.3b1
20
-
21
- [cohere]
22
- pyautogen[cohere]==0.5.3b1
23
-
24
- [cosmosdb]
25
- pyautogen[cosmosdb]==0.5.3b1
26
-
27
- [gemini]
28
- pyautogen[gemini]==0.5.3b1
29
-
30
- [graph]
31
- pyautogen[graph]==0.5.3b1
32
-
33
- [graph-rag-falkor-db]
34
- pyautogen[graph-rag-falkor-db]==0.5.3b1
35
-
36
- [groq]
37
- pyautogen[groq]==0.5.3b1
38
-
39
- [jupyter-executor]
40
- pyautogen[jupyter-executor]==0.5.3b1
41
-
42
- [lmm]
43
- pyautogen[lmm]==0.5.3b1
44
-
45
- [long-context]
46
- pyautogen[long-context]==0.5.3b1
47
-
48
- [mathchat]
49
- pyautogen[mathchat]==0.5.3b1
50
-
51
- [mistral]
52
- pyautogen[mistral]==0.5.3b1
53
-
54
- [neo4j]
55
- pyautogen[neo4j]==0.5.3b1
56
-
57
- [ollama]
58
- pyautogen[ollama]==0.5.3b1
59
-
60
- [redis]
61
- pyautogen[redis]==0.5.3b1
62
-
63
- [retrievechat]
64
- pyautogen[retrievechat]==0.5.3b1
65
-
66
- [retrievechat-mongodb]
67
- pyautogen[retrievechat-mongodb]==0.5.3b1
68
-
69
- [retrievechat-pgvector]
70
- pyautogen[retrievechat-pgvector]==0.5.3b1
71
-
72
- [retrievechat-qdrant]
73
- pyautogen[retrievechat-qdrant]==0.5.3b1
74
-
75
- [teachable]
76
- pyautogen[teachable]==0.5.3b1
77
-
78
- [test]
79
- pyautogen[test]==0.5.3b1
80
-
81
- [together]
82
- pyautogen[together]==0.5.3b1
83
-
84
- [types]
85
- pyautogen[types]==0.5.3b1
86
-
87
- [websockets]
88
- pyautogen[websockets]==0.5.3b1
89
-
90
- [websurfer]
91
- pyautogen[websurfer]==0.5.3b1
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes