universal-mcp-agents 0.1.23rc12__tar.gz → 0.1.24rc1__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of universal-mcp-agents might be problematic. Click here for more details.
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/PKG-INFO +1 -1
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/pyproject.toml +1 -1
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/prompts.py +24 -12
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/tools.py +1 -1
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/utils.py +1 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/llm.py +7 -3
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/applications/llm/app.py +4 -4
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/uv.lock +1 -1
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/evals.yml +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/lint.yml +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/release-please.yml +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/tests.yml +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.gitignore +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.pre-commit-config.yaml +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/GEMINI.md +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/PROMPTS.md +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/README.md +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/bump_and_release.sh +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/dataset.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/exact.jsonl +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/tasks.jsonl +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/test.jsonl +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/evaluators.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/prompts.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/run.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/utils.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/tests/test_agents.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/tests/test_sandbox.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/base.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/__main__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/agent.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/context.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/graph.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/prompts.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/state.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/bigtool/tools.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/builder/__main__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/builder/builder.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/builder/helper.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/builder/prompts.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/builder/state.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/cli.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/__main__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/agent.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/config.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/langgraph_agent.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/llm_tool.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/sandbox.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/codeact0/state.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/hil.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/react.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/sandbox.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/shared/__main__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/shared/prompts.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/shared/tool_node.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/simple.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/utils.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/applications/filesystem/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/applications/filesystem/app.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/applications/llm/__init__.py +0 -0
- {universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/applications/ui/app.py +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: universal-mcp-agents
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.24rc1
|
|
4
4
|
Summary: Add your description here
|
|
5
5
|
Project-URL: Homepage, https://github.com/universal-mcp/applications
|
|
6
6
|
Project-URL: Repository, https://github.com/universal-mcp/applications
|
|
@@ -12,21 +12,33 @@ Your job is to answer the user's question or perform the task they ask for.
|
|
|
12
12
|
- For writing, text/document generation (like HTML/markdown document generation) or language processing tasks DO NOT answer directly. Instead you MUST use `execute_ipython_cell` tool with the llm functions provided to you for tasks like summarizing, text generation, classification, data extraction from text or unstructured data, etc. Avoid hardcoded approaches to classification, data extraction, or creative writing.
|
|
13
13
|
- You also have access to two tools for finding and loading more python functions- `search_functions` and `load_functions`, which you must use for finding functions for using different external applications or additional functionality.
|
|
14
14
|
- Prioritize connected applications over unconnected ones from the output of `search_functions`. However, if the user specifically asks for an application, you MUST use that irrespective of connection status.
|
|
15
|
-
- When multiple apps are connected, or none of the apps are connected, YOU MUST ask the user to choose the application(s). The search results
|
|
16
|
-
- The code you write will be executed in a sandbox environment, and you can use the output of previous executions in your code. variables, functions, imports are retained.
|
|
17
|
-
- Read and understand the output of the previous code snippet and use it to answer the user's request. Note that the code output is NOT visible to the user, so after the task is complete, you have to give the output to the user in a markdown format. Similarly, you should only use print/smart_print for your own analysis, the user does not get the output.
|
|
15
|
+
- When multiple relevant apps are connected, or none of the apps are connected, YOU MUST ask the user to choose the application(s). The search results may also inform you when such a case occurs, and you must stop and ask the user if multiple apps are relevant.
|
|
18
16
|
- If needed, feel free to ask for more information from the user (without using the `execute_ipython_cell` tool) to clarify the task.
|
|
19
17
|
|
|
20
|
-
**Code
|
|
21
|
-
- The code you write will be executed in a sandbox environment, and you can use the output of previous executions in your code. Variables, functions, imports are retained.
|
|
22
|
-
- Read and understand the output of the previous code snippet and use it to answer the user's request. Note that the code output is NOT visible to the user, so after the task is complete, you have to give the output to the user in a markdown format. Similarly, you should only use print/smart_print for your own analysis, the user does not get the output.
|
|
23
|
-
- If needed, feel free to ask for more information from the user (without using the `execute_ipython_cell` tool) to clarify the task.
|
|
24
|
-
- DO NOT use the code execution to communicate/show anything to the user. The user is not able to see the output of the code cells, it is only for your processing and actions.
|
|
25
|
-
- Whenever you need to generate a large body of text, such as a document, an HTML file, or a report, use llm functions and save the output file using save functions. Do not generate text yourself and do not print the entire text in order to save your memory.
|
|
26
|
-
|
|
27
|
-
**Coding Best Practices:**
|
|
18
|
+
**Code Design**:
|
|
28
19
|
- Structure your code into multiple small, well-defined functions within a single execution snippet. This ensures modularity and makes it easier to debug or update specific logic without rewriting or re-executing large portions of code. You can only rewrite the function/portion that you need to edit since the others are retained in context.
|
|
29
|
-
-
|
|
20
|
+
- Every snippet you execute using `execute_ipython_cell` MUST follow this structure:
|
|
21
|
+
- Break down logic into 3-5 small helper functions (30 lines each max)
|
|
22
|
+
- Each helper function should do ONE thing
|
|
23
|
+
- Example:
|
|
24
|
+
def _helper_function_1(...):
|
|
25
|
+
...
|
|
26
|
+
|
|
27
|
+
def _helper_function_2(...):
|
|
28
|
+
...
|
|
29
|
+
result1 = _helper_function_1(...)
|
|
30
|
+
smart_print(result1[:1]) #As an example, to check if it has been processed correctly
|
|
31
|
+
result2 = _helper_function_2(...)
|
|
32
|
+
smart_print(result2[:1])
|
|
33
|
+
final_result = ...
|
|
34
|
+
smart_print(final_result)
|
|
35
|
+
- Thus, while debugging, if you face an error in result2, you do not need to rewrite _helper_function_1.
|
|
36
|
+
|
|
37
|
+
|
|
38
|
+
**Code Writing Rules:**
|
|
39
|
+
- The code you write will be executed in a sandbox environment, and you can use the output of previous executions in your code. Variables, defined functions, imports, loaded functions are retained.
|
|
40
|
+
- DO NOT use the code execution to communicate/show anything to the user. The user is not able to see the output of the code cells, it is only for your processing and actions. Similarly, you should only use print/smart_print for your own analysis, the user does not get the output.
|
|
41
|
+
- Whenever you need to generate a large body of text, such as a document, an HTML file, or a report, use llm functions and save the output file using save/upload functions. Do not generate text yourself and do not print the entire text in order to save your memory.
|
|
30
42
|
- External functions which return a dict or list[dict] are ambiguous. Therefore, you MUST explore the structure of the returned data using `smart_print()` statements before using it, printing keys and values. `smart_print` truncates long strings from data, preventing huge output logs.
|
|
31
43
|
- When an operation involves running a fixed set of steps on a list of items, run one run correctly and then use a for loop to run the steps on each item in the list.
|
|
32
44
|
- You can only import libraries that come pre-installed with Python. However, do consider searching for external functions first, using the search and load tools to access them in the code.
|
|
@@ -503,7 +503,7 @@ async def get_valid_tools(tool_ids: list[str], registry: AgentrRegistry) -> tupl
|
|
|
503
503
|
start = text.find(":") + 1
|
|
504
504
|
end = text.find(". R", start)
|
|
505
505
|
url = text[start:end].strip()
|
|
506
|
-
markdown_link = f"
|
|
506
|
+
markdown_link = f"[Connect to {app.capitalize()}]({url})"
|
|
507
507
|
unconnected_links.append(markdown_link)
|
|
508
508
|
for tool_id, tool_name in tool_entries:
|
|
509
509
|
if tool_name in available:
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/llm.py
RENAMED
|
@@ -9,7 +9,11 @@ from loguru import logger
|
|
|
9
9
|
|
|
10
10
|
@lru_cache(maxsize=8)
|
|
11
11
|
def load_chat_model(
|
|
12
|
-
fully_specified_name: str,
|
|
12
|
+
fully_specified_name: str,
|
|
13
|
+
temperature: float = 1.0,
|
|
14
|
+
tags: tuple[str, ...] | None = None,
|
|
15
|
+
thinking: bool = True,
|
|
16
|
+
disable_streaming: bool = False,
|
|
13
17
|
) -> BaseChatModel:
|
|
14
18
|
"""Load a chat model from a fully specified name.
|
|
15
19
|
Args:
|
|
@@ -30,7 +34,7 @@ def load_chat_model(
|
|
|
30
34
|
max_tokens=8096,
|
|
31
35
|
tags=tags,
|
|
32
36
|
stream_usage=True,
|
|
33
|
-
disable_streaming
|
|
37
|
+
disable_streaming=disable_streaming,
|
|
34
38
|
) # pyright: ignore[reportCallIssue]
|
|
35
39
|
elif provider == "azure":
|
|
36
40
|
return AzureChatOpenAI(
|
|
@@ -40,7 +44,7 @@ def load_chat_model(
|
|
|
40
44
|
temperature=temperature,
|
|
41
45
|
tags=tags,
|
|
42
46
|
stream_usage=True,
|
|
43
|
-
disable_streaming
|
|
47
|
+
disable_streaming=disable_streaming,
|
|
44
48
|
)
|
|
45
49
|
elif provider == "gemini":
|
|
46
50
|
return ChatGoogleGenerativeAI(model=model, temperature=temperature)
|
|
@@ -91,7 +91,7 @@ class LlmApp(BaseApplication):
|
|
|
91
91
|
|
|
92
92
|
full_prompt = f"{prompt}\n\nContext:\n{context_str}\n\n"
|
|
93
93
|
|
|
94
|
-
model = load_chat_model("azure/gpt-5-mini", disable_streaming
|
|
94
|
+
model = load_chat_model("azure/gpt-5-mini", disable_streaming=True, tags=("quiet",))
|
|
95
95
|
response = model.with_retry(stop_after_attempt=MAX_RETRIES).invoke(full_prompt, stream=False)
|
|
96
96
|
return str(response.content)
|
|
97
97
|
|
|
@@ -158,7 +158,7 @@ class LlmApp(BaseApplication):
|
|
|
158
158
|
reason: str = Field(..., description="The reasoning behind the classification.")
|
|
159
159
|
top_class: str = Field(..., description="The class with the highest probability.")
|
|
160
160
|
|
|
161
|
-
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming
|
|
161
|
+
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming=True, tags=("quiet",))
|
|
162
162
|
agent = create_agent(
|
|
163
163
|
model=model,
|
|
164
164
|
tools=[],
|
|
@@ -229,7 +229,7 @@ class LlmApp(BaseApplication):
|
|
|
229
229
|
"Return ONLY a valid JSON object that conforms to the provided schema, with no extra text."
|
|
230
230
|
)
|
|
231
231
|
|
|
232
|
-
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming
|
|
232
|
+
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming=True, tags=("quiet",))
|
|
233
233
|
|
|
234
234
|
response = (
|
|
235
235
|
model.with_structured_output(schema=output_schema, method="json_mode")
|
|
@@ -282,7 +282,7 @@ class LlmApp(BaseApplication):
|
|
|
282
282
|
|
|
283
283
|
prompt = f"{task_instructions}\n\nContext:\n{context_str}\n\nReturn ONLY a valid JSON object, no extra text."
|
|
284
284
|
|
|
285
|
-
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming
|
|
285
|
+
model = load_chat_model("azure/gpt-5-mini", temperature=0, disable_streaming=True, tags=("quiet",))
|
|
286
286
|
|
|
287
287
|
agent = create_agent(
|
|
288
288
|
model=model,
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/evals.yml
RENAMED
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/lint.yml
RENAMED
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/.github/workflows/tests.yml
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/exact.jsonl
RENAMED
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/tasks.jsonl
RENAMED
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/evals/datasets/test.jsonl
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/tests/test_sandbox.py
RENAMED
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/base.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/cli.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/hil.py
RENAMED
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/react.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{universal_mcp_agents-0.1.23rc12 → universal_mcp_agents-0.1.24rc1}/src/universal_mcp/agents/utils.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|