ag2 0.8.1__tar.gz → 0.8.2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of ag2 might be problematic. Click here for more details.

@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ag2
3
- Version: 0.8.1
3
+ Version: 0.8.2
4
4
  Summary: Alias package for pyautogen
5
5
  Home-page: https://github.com/ag2ai/ag2
6
6
  Author: Chi Wang & Qingyun Wu
@@ -25,6 +25,7 @@ Provides-Extra: graph-rag-falkor-db
25
25
  Provides-Extra: rag
26
26
  Provides-Extra: crawl4ai
27
27
  Provides-Extra: browser-use
28
+ Provides-Extra: google-search
28
29
  Provides-Extra: neo4j
29
30
  Provides-Extra: twilio
30
31
  Provides-Extra: interop-crewai
@@ -160,14 +161,14 @@ You can use the sample file `OAI_CONFIG_LIST_sample` as a template.
160
161
  Create a script or a Jupyter Notebook and run your first agent.
161
162
 
162
163
  ```python
163
- from autogen import AssistantAgent, UserProxyAgent, config_list_from_json
164
+ from autogen import AssistantAgent, UserProxyAgent, LLMConfig
165
+
166
+ llm_config = LLMConfig.from_json(path="OAI_CONFIG_LIST")
164
167
 
165
- llm_config = {
166
- "config_list": config_list_from_json(env_or_file="OAI_CONFIG_LIST")
167
- }
168
168
 
169
- assistant = AssistantAgent("assistant", llm_config=llm_config)
170
- user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
169
+ with llm_config:
170
+ assistant = AssistantAgent("assistant", llm_config=llm_config)
171
+ user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
171
172
  user_proxy.initiate_chat(assistant, message="Plot a chart of NVDA and TESLA stock price change YTD.")
172
173
  # This initiates an automated chat between the two agents to solve the task
173
174
  ```
@@ -93,14 +93,14 @@ You can use the sample file `OAI_CONFIG_LIST_sample` as a template.
93
93
  Create a script or a Jupyter Notebook and run your first agent.
94
94
 
95
95
  ```python
96
- from autogen import AssistantAgent, UserProxyAgent, config_list_from_json
96
+ from autogen import AssistantAgent, UserProxyAgent, LLMConfig
97
+
98
+ llm_config = LLMConfig.from_json(path="OAI_CONFIG_LIST")
97
99
 
98
- llm_config = {
99
- "config_list": config_list_from_json(env_or_file="OAI_CONFIG_LIST")
100
- }
101
100
 
102
- assistant = AssistantAgent("assistant", llm_config=llm_config)
103
- user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
101
+ with llm_config:
102
+ assistant = AssistantAgent("assistant", llm_config=llm_config)
103
+ user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
104
104
  user_proxy.initiate_chat(assistant, message="Plot a chart of NVDA and TESLA stock price change YTD.")
105
105
  # This initiates an automated chat between the two agents to solve the task
106
106
  ```
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ag2
3
- Version: 0.8.1
3
+ Version: 0.8.2
4
4
  Summary: Alias package for pyautogen
5
5
  Home-page: https://github.com/ag2ai/ag2
6
6
  Author: Chi Wang & Qingyun Wu
@@ -25,6 +25,7 @@ Provides-Extra: graph-rag-falkor-db
25
25
  Provides-Extra: rag
26
26
  Provides-Extra: crawl4ai
27
27
  Provides-Extra: browser-use
28
+ Provides-Extra: google-search
28
29
  Provides-Extra: neo4j
29
30
  Provides-Extra: twilio
30
31
  Provides-Extra: interop-crewai
@@ -160,14 +161,14 @@ You can use the sample file `OAI_CONFIG_LIST_sample` as a template.
160
161
  Create a script or a Jupyter Notebook and run your first agent.
161
162
 
162
163
  ```python
163
- from autogen import AssistantAgent, UserProxyAgent, config_list_from_json
164
+ from autogen import AssistantAgent, UserProxyAgent, LLMConfig
165
+
166
+ llm_config = LLMConfig.from_json(path="OAI_CONFIG_LIST")
164
167
 
165
- llm_config = {
166
- "config_list": config_list_from_json(env_or_file="OAI_CONFIG_LIST")
167
- }
168
168
 
169
- assistant = AssistantAgent("assistant", llm_config=llm_config)
170
- user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
169
+ with llm_config:
170
+ assistant = AssistantAgent("assistant", llm_config=llm_config)
171
+ user_proxy = UserProxyAgent("user_proxy", code_execution_config={"work_dir": "coding", "use_docker": False})
171
172
  user_proxy.initiate_chat(assistant, message="Plot a chart of NVDA and TESLA stock price change YTD.")
172
173
  # This initiates an automated chat between the two agents to solve the task
173
174
  ```
@@ -14,6 +14,8 @@ test/test_conftest.py
14
14
  test/test_graph_utils.py
15
15
  test/test_import.py
16
16
  test/test_import_utils.py
17
+ test/test_json_utils.py
18
+ test/test_llm_config.py
17
19
  test/test_logging.py
18
20
  test/test_notebook.py
19
21
  test/test_retrieve_utils.py
@@ -0,0 +1,154 @@
1
+ pyautogen==0.8.2
2
+
3
+ [anthropic]
4
+ pyautogen[anthropic]==0.8.2
5
+
6
+ [autobuild]
7
+ pyautogen[autobuild]==0.8.2
8
+
9
+ [bedrock]
10
+ pyautogen[bedrock]==0.8.2
11
+
12
+ [blendsearch]
13
+ pyautogen[blendsearch]==0.8.2
14
+
15
+ [browser-use]
16
+ pyautogen[browser-use]==0.8.2
17
+
18
+ [captainagent]
19
+ pyautogen[captainagent]==0.8.2
20
+
21
+ [cerebras]
22
+ pyautogen[cerebras]==0.8.2
23
+
24
+ [cohere]
25
+ pyautogen[cohere]==0.8.2
26
+
27
+ [commsagent-discord]
28
+ pyautogen[commsagent-discord]==0.8.2
29
+
30
+ [commsagent-slack]
31
+ pyautogen[commsagent-slack]==0.8.2
32
+
33
+ [commsagent-telegram]
34
+ pyautogen[commsagent-telegram]==0.8.2
35
+
36
+ [cosmosdb]
37
+ pyautogen[cosmosdb]==0.8.2
38
+
39
+ [crawl4ai]
40
+ pyautogen[crawl4ai]==0.8.2
41
+
42
+ [deepseek]
43
+ pyautogen[deepseek]==0.8.2
44
+
45
+ [dev]
46
+ pyautogen[dev]==0.8.2
47
+
48
+ [docs]
49
+ pyautogen[docs]==0.8.2
50
+
51
+ [flaml]
52
+ pyautogen[flaml]==0.8.2
53
+
54
+ [gemini]
55
+ pyautogen[gemini]==0.8.2
56
+
57
+ [gemini-realtime]
58
+ pyautogen[gemini-realtime]==0.8.2
59
+
60
+ [google-search]
61
+ pyautogen[google-search]==0.8.2
62
+
63
+ [graph]
64
+ pyautogen[graph]==0.8.2
65
+
66
+ [graph-rag-falkor-db]
67
+ pyautogen[graph-rag-falkor-db]==0.8.2
68
+
69
+ [groq]
70
+ pyautogen[groq]==0.8.2
71
+
72
+ [interop]
73
+ pyautogen[interop]==0.8.2
74
+
75
+ [interop-crewai]
76
+ pyautogen[interop-crewai]==0.8.2
77
+
78
+ [interop-langchain]
79
+ pyautogen[interop-langchain]==0.8.2
80
+
81
+ [interop-pydantic-ai]
82
+ pyautogen[interop-pydantic-ai]==0.8.2
83
+
84
+ [jupyter-executor]
85
+ pyautogen[jupyter-executor]==0.8.2
86
+
87
+ [lint]
88
+ pyautogen[lint]==0.8.2
89
+
90
+ [lmm]
91
+ pyautogen[lmm]==0.8.2
92
+
93
+ [long-context]
94
+ pyautogen[long-context]==0.8.2
95
+
96
+ [mathchat]
97
+ pyautogen[mathchat]==0.8.2
98
+
99
+ [mistral]
100
+ pyautogen[mistral]==0.8.2
101
+
102
+ [neo4j]
103
+ pyautogen[neo4j]==0.8.2
104
+
105
+ [ollama]
106
+ pyautogen[ollama]==0.8.2
107
+
108
+ [openai]
109
+ pyautogen[openai]==0.8.2
110
+
111
+ [openai-realtime]
112
+ pyautogen[openai-realtime]==0.8.2
113
+
114
+ [rag]
115
+ pyautogen[rag]==0.8.2
116
+
117
+ [redis]
118
+ pyautogen[redis]==0.8.2
119
+
120
+ [retrievechat]
121
+ pyautogen[retrievechat]==0.8.2
122
+
123
+ [retrievechat-couchbase]
124
+ pyautogen[retrievechat-couchbase]==0.8.2
125
+
126
+ [retrievechat-mongodb]
127
+ pyautogen[retrievechat-mongodb]==0.8.2
128
+
129
+ [retrievechat-pgvector]
130
+ pyautogen[retrievechat-pgvector]==0.8.2
131
+
132
+ [retrievechat-qdrant]
133
+ pyautogen[retrievechat-qdrant]==0.8.2
134
+
135
+ [teachable]
136
+ pyautogen[teachable]==0.8.2
137
+
138
+ [test]
139
+ pyautogen[test]==0.8.2
140
+
141
+ [together]
142
+ pyautogen[together]==0.8.2
143
+
144
+ [twilio]
145
+ pyautogen[twilio]==0.8.2
146
+
147
+ [types]
148
+ pyautogen[types]==0.8.2
149
+
150
+ [websockets]
151
+ pyautogen[websockets]==0.8.2
152
+
153
+ [websurfer]
154
+ pyautogen[websurfer]==0.8.2
@@ -60,8 +60,9 @@ dependencies = [
60
60
  "docker",
61
61
  "packaging",
62
62
  "asyncer==0.0.8",
63
- "fast-depends>=2.4.12,<3",
63
+ # "fast-depends>=2.4.12,<3", # integrated into the package
64
64
  "httpx>=0.28.1,<1",
65
+ "anyio>=3.0.0,<5.0.0" # needed by the internal fast-depends
65
66
  ]
66
67
 
67
68
  [project.optional-dependencies]
@@ -76,7 +77,7 @@ flaml = [
76
77
  # public distributions
77
78
 
78
79
  openai = [
79
- "openai>=1.58",
80
+ "openai>=1.66.2",
80
81
  ]
81
82
 
82
83
  openai-realtime = [
@@ -156,11 +157,14 @@ browser-use = [
156
157
  "browser-use==0.1.37",
157
158
  ]
158
159
 
160
+ google-search = [
161
+ "google-api-python-client>=2.163.0,<3.0",
162
+ ]
163
+
159
164
  neo4j = [
160
165
  "docx2txt==0.8",
161
- "llama-index==0.12.22",
166
+ "llama-index>=0.12,<1",
162
167
  "llama-index-graph-stores-neo4j==0.4.6",
163
- "llama-index-core==0.12.22",
164
168
  "llama-index-readers-web==0.3.7",
165
169
  ]
166
170
 
@@ -181,8 +185,7 @@ interop =[
181
185
  "pyautogen[interop-crewai, interop-langchain, interop-pydantic-ai]",
182
186
  ]
183
187
 
184
- # pysqlite3-binary used so it doesn't need to compile pysqlite3
185
- autobuild = ["chromadb", "sentence-transformers", "huggingface-hub", "pysqlite3-binary"]
188
+ autobuild = ["chromadb", "sentence-transformers", "huggingface-hub"]
186
189
 
187
190
  blendsearch = ["flaml[blendsearch]"]
188
191
  mathchat = ["sympy", "wolframalpha"]
@@ -197,7 +200,6 @@ gemini = [
197
200
  "google-auth",
198
201
  "pillow",
199
202
  "jsonschema",
200
- "jsonref>=1,<2",
201
203
  ]
202
204
 
203
205
  gemini-realtime = [
@@ -216,7 +218,7 @@ cerebras = ["cerebras_cloud_sdk>=1.0.0"]
216
218
  mistral = ["mistralai>=1.0.1"]
217
219
  groq = ["groq>=0.9.0"]
218
220
  cohere = ["cohere>=5.13.5"]
219
- ollama = ["ollama>=0.4.5", "fix_busted_json>=0.0.18"]
221
+ ollama = ["ollama>=0.4.7", "fix_busted_json>=0.0.18"]
220
222
  bedrock = ["boto3>=1.34.149"]
221
223
  deepseek = ["pyautogen[openai]"]
222
224
 
@@ -234,16 +236,18 @@ test = [
234
236
  "pytest-cov==6.0.0",
235
237
  "pytest-asyncio==0.25.3",
236
238
  "pytest==8.3.5",
237
- "mock==5.1.0",
239
+ "mock==5.2.0",
238
240
  "pandas==2.2.3",
239
241
  "fastapi==0.115.11",
242
+ "dirty-equals==0.9.0",
240
243
  ]
241
244
 
242
245
  docs = [
243
- "mkdocs-material==9.6.7",
244
- "mkdocstrings[python]==0.28.2",
246
+ "mkdocs-material==9.6.9",
247
+ "mkdocstrings[python]==0.29.0",
245
248
  "mkdocs-literate-nav==0.6.1",
246
249
  "mdx-include==1.4.2",
250
+ # currently problematic and cannot be upgraded
247
251
  "mkdocs-git-revision-date-localized-plugin==1.3.0",
248
252
  "mike==2.1.3",
249
253
  "typer==0.15.2",
@@ -254,7 +258,7 @@ docs = [
254
258
  "pillow", # required for mkdocs-glightbo
255
259
  "cairosvg", # required for mkdocs-glightbo
256
260
  "pdoc3==0.11.5",
257
- "jinja2==3.1.5",
261
+ "jinja2==3.1.6",
258
262
  "pyyaml==6.0.2",
259
263
  "termcolor==2.5.0",
260
264
  "nbclient==0.10.2",
@@ -276,7 +280,7 @@ dev = [
276
280
  "pyautogen[lint,test,types,docs]",
277
281
  "pre-commit==4.1.0",
278
282
  "detect-secrets==1.5.0",
279
- "uv==0.6.4",
283
+ "uv==0.6.6",
280
284
  ]
281
285
 
282
286
 
@@ -341,6 +345,7 @@ markers = [
341
345
  "interop",
342
346
  "browser_use",
343
347
  "crawl4ai",
348
+ "google_search",
344
349
  "websockets",
345
350
  "commsagent_discord",
346
351
  "commsagent_slack",
@@ -430,6 +435,7 @@ convention = "google"
430
435
  [tool.mypy]
431
436
  files = [
432
437
  "autogen/agentchat/agent.py",
438
+ # "autogen/agentchat/cenversable_agent.py",
433
439
  "autogen/agentchat/contrib/rag",
434
440
  "autogen/agentchat/contrib/graph_rag",
435
441
  "autogen/agentchat/contrib/swarm_agent.py",
@@ -438,9 +444,12 @@ files = [
438
444
  "autogen/agents",
439
445
  "autogen/coding",
440
446
  "autogen/exception_utils.py",
447
+ "autogen/fast_depends",
441
448
  "autogen/import_utils.py",
442
449
  "autogen/interop",
443
450
  "autogen/io",
451
+ "autogen/json_utils.py",
452
+ "autogen/llm_config.py",
444
453
  "autogen/logger",
445
454
  "autogen/messages",
446
455
  "autogen/oai/oai_models",
@@ -454,8 +463,11 @@ files = [
454
463
  "test/agentchat/realtime_agent",
455
464
  "test/agents",
456
465
  "test/conftest.py",
466
+ # "test/fast_depends",
457
467
  "test/interop",
458
468
  "test/io",
469
+ "test/test_json_utils.py",
470
+ "test/test_llm_config.py",
459
471
  "test/messages",
460
472
  "test/test_import_utils.py",
461
473
  "test/test_import.py",
@@ -41,6 +41,7 @@ setuptools.setup(
41
41
  "rag": ["pyautogen[rag]==" + __version__],
42
42
  "crawl4ai": ["pyautogen[crawl4ai]==" + __version__],
43
43
  "browser-use": ["pyautogen[browser-use]==" + __version__],
44
+ "google-search": ["pyautogen[google-search]==" + __version__],
44
45
  "neo4j": ["pyautogen[neo4j]==" + __version__],
45
46
  "twilio": ["pyautogen[twilio]==" + __version__],
46
47
  "interop-crewai": ["pyautogen[interop-crewai]==" + __version__],