semantio 0.0.2__tar.gz → 0.0.4__tar.gz
Sign up to get free protection for your applications and to get access to all the features.
- {semantio-0.0.2 → semantio-0.0.4}/PKG-INFO +1 -1
- {semantio-0.0.2 → semantio-0.0.4}/semantio/agent.py +37 -75
- {semantio-0.0.2 → semantio-0.0.4}/semantio/cli/main.py +6 -6
- semantio-0.0.4/semantio/tools/web_browser.py +271 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/PKG-INFO +1 -1
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/SOURCES.txt +1 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/requires.txt +4 -0
- {semantio-0.0.2 → semantio-0.0.4}/setup.py +5 -1
- {semantio-0.0.2 → semantio-0.0.4}/LICENSE +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/README.md +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/api/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/api/api_generator.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/api/fastapi_app.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/cli/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/knowledge_base/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/knowledge_base/document_loader.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/knowledge_base/retriever.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/knowledge_base/vector_store.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/anthropic.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/base_llm.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/deepseek.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/gemini.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/groq.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/mistral.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/llm/openai.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/memory.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/rag.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/storage/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/storage/cloud_storage.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/storage/local_storage.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/tools/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/tools/base_tool.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/tools/crypto.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/tools/duckduckgo.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/tools/stocks.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/__init__.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/config.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/date_utils.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/file_utils.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/logger.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio/utils/validation_utils.py +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/dependency_links.txt +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/entry_points.txt +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/semantio.egg-info/top_level.txt +0 -0
- {semantio-0.0.2 → semantio-0.0.4}/setup.cfg +0 -0
@@ -21,24 +21,24 @@ import os
|
|
21
21
|
logging.basicConfig(level=logging.INFO)
|
22
22
|
logger = logging.getLogger(__name__)
|
23
23
|
|
24
|
-
class
|
24
|
+
class Agent(BaseModel):
|
25
25
|
# -*- Agent settings
|
26
|
-
name: Optional[str] = Field(None, description="Name of the
|
27
|
-
description: Optional[str] = Field(None, description="Description of the
|
28
|
-
instructions: Optional[List[str]] = Field(None, description="List of instructions for the
|
26
|
+
name: Optional[str] = Field(None, description="Name of the agent.")
|
27
|
+
description: Optional[str] = Field(None, description="Description of the agent's role.")
|
28
|
+
instructions: Optional[List[str]] = Field(None, description="List of instructions for the agent.")
|
29
29
|
model: Optional[str] = Field(None, description="This one is not in the use.")
|
30
30
|
show_tool_calls: bool = Field(False, description="Whether to show tool calls in the response.")
|
31
31
|
markdown: bool = Field(False, description="Whether to format the response in markdown.")
|
32
|
-
tools: Optional[List[BaseTool]] = Field(None, description="List of tools available to the
|
33
|
-
user_name: Optional[str] = Field("User", description="Name of the user interacting with the
|
34
|
-
emoji: Optional[str] = Field(":robot:", description="Emoji to represent the
|
32
|
+
tools: Optional[List[BaseTool]] = Field(None, description="List of tools available to the agent.")
|
33
|
+
user_name: Optional[str] = Field("User", description="Name of the user interacting with the agent.")
|
34
|
+
emoji: Optional[str] = Field(":robot:", description="Emoji to represent the agent in the CLI.")
|
35
35
|
rag: Optional[RAG] = Field(None, description="RAG instance for context retrieval.")
|
36
36
|
knowledge_base: Optional[Any] = Field(None, description="Knowledge base for domain-specific information.")
|
37
37
|
llm: Optional[str] = Field(None, description="The LLM provider to use (e.g., 'groq', 'openai', 'anthropic').")
|
38
38
|
llm_model: Optional[str] = Field(None, description="The specific model to use for the LLM provider.")
|
39
39
|
llm_instance: Optional[BaseLLM] = Field(None, description="The LLM instance to use.")
|
40
40
|
json_output: bool = Field(False, description="Whether to format the response as JSON.")
|
41
|
-
api: bool = Field(False, description="Whether to generate an API for the
|
41
|
+
api: bool = Field(False, description="Whether to generate an API for the agent.")
|
42
42
|
api_config: Optional[Dict] = Field(
|
43
43
|
None,
|
44
44
|
description="Configuration for the API (e.g., host, port, authentication).",
|
@@ -46,7 +46,7 @@ class Assistant(BaseModel):
|
|
46
46
|
api_generator: Optional[Any] = Field(None, description="The API generator instance.")
|
47
47
|
expected_output: Optional[Union[str, Dict]] = Field(None, description="The expected format or structure of the output.")
|
48
48
|
semantic_model: Optional[Any] = Field(None, description="SentenceTransformer model for semantic matching.")
|
49
|
-
team: Optional[List['
|
49
|
+
team: Optional[List['Agent']] = Field(None, description="List of assistants in the team.")
|
50
50
|
auto_tool: bool = Field(False, description="Whether to automatically detect and call tools.")
|
51
51
|
|
52
52
|
# Allow arbitrary types
|
@@ -56,8 +56,11 @@ class Assistant(BaseModel):
|
|
56
56
|
super().__init__(**kwargs)
|
57
57
|
# Initialize the model and tools here if needed
|
58
58
|
self._initialize_model()
|
59
|
-
#
|
59
|
+
# Initialize tools as an empty list if not provided
|
60
60
|
if self.tools is None:
|
61
|
+
self.tools = []
|
62
|
+
# Automatically discover and register tools if auto tool is enabled
|
63
|
+
if self.auto_tool and not self.tools:
|
61
64
|
self.tools = self._discover_tools()
|
62
65
|
# Pass the LLM instance to each tool
|
63
66
|
for tool in self.tools:
|
@@ -125,7 +128,7 @@ class Assistant(BaseModel):
|
|
125
128
|
try:
|
126
129
|
# Import the module
|
127
130
|
module_name = file.stem
|
128
|
-
module = importlib.import_module(f"
|
131
|
+
module = importlib.import_module(f"semantio.tools.{module_name}")
|
129
132
|
|
130
133
|
# Find all classes that inherit from BaseTool
|
131
134
|
for name, obj in module.__dict__.items():
|
@@ -197,7 +200,7 @@ class Assistant(BaseModel):
|
|
197
200
|
model_to_use = self.llm_model or default_model
|
198
201
|
|
199
202
|
# Dynamically import and initialize the LLM class
|
200
|
-
module_name = f"
|
203
|
+
module_name = f"semantio.llm.{llm_provider}"
|
201
204
|
llm_module = importlib.import_module(module_name)
|
202
205
|
llm_class = getattr(llm_module, llm_class_name)
|
203
206
|
self.llm_instance = llm_class(model=model_to_use, api_key=api_key)
|
@@ -213,11 +216,10 @@ class Assistant(BaseModel):
|
|
213
216
|
message: Optional[Union[str, Image, List, Dict]] = None,
|
214
217
|
stream: bool = False,
|
215
218
|
markdown: bool = False,
|
216
|
-
|
217
|
-
team: Optional[List['Assistant']] = None,
|
219
|
+
team: Optional[List['Agent']] = None,
|
218
220
|
**kwargs,
|
219
221
|
) -> Union[str, Dict]: # Add return type hint
|
220
|
-
"""Print the
|
222
|
+
"""Print the agent's response to the console and return it."""
|
221
223
|
|
222
224
|
if stream:
|
223
225
|
# Handle streaming response
|
@@ -228,60 +230,23 @@ class Assistant(BaseModel):
|
|
228
230
|
return response
|
229
231
|
else:
|
230
232
|
# Generate and return the response
|
231
|
-
response = self._generate_response(message, markdown=markdown,
|
233
|
+
response = self._generate_response(message, markdown=markdown, team=team, **kwargs)
|
232
234
|
print(response) # Print the response to the console
|
233
235
|
return response
|
234
236
|
|
235
237
|
|
236
238
|
def _stream_response(self, message: str, markdown: bool = False, **kwargs) -> Iterator[str]:
|
237
|
-
"""Stream the
|
239
|
+
"""Stream the agent's response."""
|
238
240
|
# Simulate streaming by yielding chunks of the response
|
239
241
|
response = self._generate_response(message, markdown=markdown, **kwargs)
|
240
242
|
for chunk in response.split():
|
241
243
|
yield chunk + " "
|
242
244
|
|
243
245
|
def register_tool(self, tool: BaseTool):
|
244
|
-
"""Register a tool for the
|
246
|
+
"""Register a tool for the agent."""
|
245
247
|
if self.tools is None:
|
246
248
|
self.tools = []
|
247
249
|
self.tools.append(tool)
|
248
|
-
|
249
|
-
def _detect_tool_call(self, message: str) -> Optional[Dict[str, Any]]:
|
250
|
-
"""
|
251
|
-
Use the LLM to detect which tool should be called based on the user's query.
|
252
|
-
"""
|
253
|
-
if not self.tools:
|
254
|
-
logger.warning("No tools available to detect.")
|
255
|
-
return None
|
256
|
-
|
257
|
-
# Create a prompt for the LLM
|
258
|
-
prompt = f"""
|
259
|
-
You are an AI assistant that helps users by selecting the most appropriate tool to answer their query. Below is a list of available tools and their functionalities:
|
260
|
-
|
261
|
-
{self._get_tool_descriptions()}
|
262
|
-
|
263
|
-
Based on the user's query, select the most appropriate tool. Respond with the name of the tool (e.g., "CryptoPriceChecker"). If no tool is suitable, respond with "None".
|
264
|
-
|
265
|
-
User Query: "{message}"
|
266
|
-
"""
|
267
|
-
|
268
|
-
try:
|
269
|
-
# Call the LLM to generate the response
|
270
|
-
response = self.llm_instance.generate(prompt=prompt)
|
271
|
-
tool_name = response.strip().replace('"', '').replace("'", "")
|
272
|
-
|
273
|
-
# Find the tool in the list of available tools
|
274
|
-
tool = next((t for t in self.tools if t.name.lower() == tool_name.lower()), None)
|
275
|
-
if tool:
|
276
|
-
logger.info(f"Detected tool call: {tool.name}")
|
277
|
-
return {
|
278
|
-
"tool": tool.name,
|
279
|
-
"input": {"query": message}
|
280
|
-
}
|
281
|
-
except Exception as e:
|
282
|
-
logger.error(f"Failed to detect tool call: {e}")
|
283
|
-
|
284
|
-
return None
|
285
250
|
|
286
251
|
def _analyze_query_and_select_tools(self, query: str) -> List[Dict[str, Any]]:
|
287
252
|
"""
|
@@ -290,7 +255,7 @@ class Assistant(BaseModel):
|
|
290
255
|
"""
|
291
256
|
# Create a prompt for the LLM to analyze the query and select tools
|
292
257
|
prompt = f"""
|
293
|
-
You are an AI
|
258
|
+
You are an AI agent that helps analyze user queries and select the most appropriate tools.
|
294
259
|
Below is a list of available tools and their functionalities:
|
295
260
|
|
296
261
|
{self._get_tool_descriptions()}
|
@@ -324,17 +289,16 @@ class Assistant(BaseModel):
|
|
324
289
|
return []
|
325
290
|
|
326
291
|
|
327
|
-
def _generate_response(self, message: str, markdown: bool = False,
|
328
|
-
"""Generate the
|
329
|
-
# Use the specified
|
330
|
-
if tools is not None:
|
331
|
-
self.tools = tools
|
292
|
+
def _generate_response(self, message: str, markdown: bool = False, team: Optional[List['Agent']] = None, **kwargs) -> str:
|
293
|
+
"""Generate the agent's response, including tool execution and context retrieval."""
|
294
|
+
# Use the specified team if provided
|
332
295
|
if team is not None:
|
333
296
|
return self._generate_team_response(message, team, markdown=markdown, **kwargs)
|
334
297
|
|
335
298
|
# Initialize tool_outputs as an empty dictionary
|
336
299
|
tool_outputs = {}
|
337
300
|
responses = []
|
301
|
+
tool_calls = []
|
338
302
|
|
339
303
|
# Use the LLM to analyze the query and dynamically select tools when auto_tool is enabled
|
340
304
|
if self.auto_tool:
|
@@ -344,7 +308,7 @@ class Assistant(BaseModel):
|
|
344
308
|
if self.tools:
|
345
309
|
tool_calls = [
|
346
310
|
{
|
347
|
-
"tool": tool.
|
311
|
+
"tool": tool.name,
|
348
312
|
"input": {
|
349
313
|
"query": message, # Use the message as the query
|
350
314
|
"context": None, # No context provided by default
|
@@ -352,10 +316,8 @@ class Assistant(BaseModel):
|
|
352
316
|
}
|
353
317
|
for tool in self.tools
|
354
318
|
]
|
355
|
-
else:
|
356
|
-
tool_calls = kwargs.get("tool_calls", [])
|
357
319
|
|
358
|
-
|
320
|
+
# Execute tools if any are detected
|
359
321
|
if tool_calls:
|
360
322
|
for tool_call in tool_calls:
|
361
323
|
tool_name = tool_call["tool"]
|
@@ -396,7 +358,7 @@ class Assistant(BaseModel):
|
|
396
358
|
except Exception as e:
|
397
359
|
logger.error(f"Failed to generate LLM response: {e}")
|
398
360
|
responses.append(f"An error occurred while generating the analysis: {e}")
|
399
|
-
if not tool_calls:
|
361
|
+
if not self.tools and not tool_calls:
|
400
362
|
# If no tools were executed, proceed with the original logic
|
401
363
|
# Retrieve relevant context using RAG
|
402
364
|
rag_context = self.rag.retrieve(message) if self.rag else None
|
@@ -435,12 +397,12 @@ class Assistant(BaseModel):
|
|
435
397
|
# Combine all responses into a single string
|
436
398
|
return "\n\n".join(responses)
|
437
399
|
|
438
|
-
def _generate_team_response(self, message: str, team: List['
|
400
|
+
def _generate_team_response(self, message: str, team: List['Agent'], markdown: bool = False, **kwargs) -> str:
|
439
401
|
"""Generate a response using a team of assistants."""
|
440
402
|
responses = []
|
441
|
-
for
|
442
|
-
response =
|
443
|
-
responses.append(f"**{
|
403
|
+
for agent in team:
|
404
|
+
response = agent.print_response(message, markdown=markdown, **kwargs)
|
405
|
+
responses.append(f"**{agent.name}:**\n\n{response}")
|
444
406
|
return "\n\n".join(responses)
|
445
407
|
|
446
408
|
def _build_prompt(self, message: str, context: Optional[List[Dict]]) -> str:
|
@@ -578,7 +540,7 @@ class Assistant(BaseModel):
|
|
578
540
|
exit_on: Optional[List[str]] = None,
|
579
541
|
**kwargs,
|
580
542
|
):
|
581
|
-
"""Run the
|
543
|
+
"""Run the agent in a CLI app."""
|
582
544
|
from rich.prompt import Prompt
|
583
545
|
|
584
546
|
if message:
|
@@ -593,15 +555,15 @@ class Assistant(BaseModel):
|
|
593
555
|
self.print_response(message=message, **kwargs)
|
594
556
|
|
595
557
|
def _generate_api(self):
|
596
|
-
"""Generate an API for the
|
558
|
+
"""Generate an API for the agent if api=True."""
|
597
559
|
from .api.api_generator import APIGenerator
|
598
560
|
self.api_generator = APIGenerator(self)
|
599
|
-
print(f"API generated for
|
561
|
+
print(f"API generated for agent '{self.name}'. Use `.run_api()` to start the API server.")
|
600
562
|
|
601
563
|
def run_api(self):
|
602
|
-
"""Run the API server for the
|
564
|
+
"""Run the API server for the agent."""
|
603
565
|
if not hasattr(self, 'api_generator'):
|
604
|
-
raise ValueError("API is not enabled for this
|
566
|
+
raise ValueError("API is not enabled for this agent. Set `api=True` when initializing the agent.")
|
605
567
|
|
606
568
|
# Get API configuration
|
607
569
|
host = self.api_config.get("host", "0.0.0.0") if self.api_config else "0.0.0.0"
|
@@ -1,7 +1,7 @@
|
|
1
1
|
import argparse
|
2
2
|
import warnings
|
3
|
-
from
|
4
|
-
from
|
3
|
+
from semantio.agent import Agent
|
4
|
+
from semantio.llm import get_llm
|
5
5
|
from urllib3.exceptions import NotOpenSSLWarning
|
6
6
|
|
7
7
|
# Suppress the NotOpenSSLWarning
|
@@ -9,7 +9,7 @@ warnings.filterwarnings("ignore", category=NotOpenSSLWarning)
|
|
9
9
|
|
10
10
|
def main():
|
11
11
|
parser = argparse.ArgumentParser(description="opAi CLI")
|
12
|
-
parser.add_argument("--message", type=str, required=True, help="Message to send to the
|
12
|
+
parser.add_argument("--message", type=str, required=True, help="Message to send to the agent")
|
13
13
|
parser.add_argument("--provider", type=str, required=True, help="LLM provider (e.g., groq, openai)")
|
14
14
|
parser.add_argument("--api-key", type=str, required=True, help="API key for the LLM provider")
|
15
15
|
parser.add_argument("--model", type=str, default=None, help="Model name (e.g., mixtral-8x7b-32768)")
|
@@ -22,9 +22,9 @@ def main():
|
|
22
22
|
|
23
23
|
llm = get_llm(provider=args.provider, **llm_config)
|
24
24
|
|
25
|
-
# Create an
|
26
|
-
|
27
|
-
|
25
|
+
# Create an agent
|
26
|
+
agent = Agent(model=args.provider, llm=llm)
|
27
|
+
agent.print_response(args.message)
|
28
28
|
|
29
29
|
|
30
30
|
if __name__ == "__main__":
|
@@ -0,0 +1,271 @@
|
|
1
|
+
# web_browser.py
|
2
|
+
from typing import Dict, Any, List, Optional
|
3
|
+
from pydantic import Field, BaseModel
|
4
|
+
from selenium import webdriver
|
5
|
+
from selenium.webdriver.common.by import By
|
6
|
+
from selenium.webdriver.support.ui import WebDriverWait
|
7
|
+
from selenium.webdriver.support import expected_conditions as EC
|
8
|
+
from selenium.webdriver.chrome.options import Options
|
9
|
+
from selenium.webdriver.chrome.service import Service
|
10
|
+
from webdriver_manager.chrome import ChromeDriverManager
|
11
|
+
from bs4 import BeautifulSoup
|
12
|
+
import json
|
13
|
+
import time
|
14
|
+
import re
|
15
|
+
import logging
|
16
|
+
from .base_tool import BaseTool
|
17
|
+
|
18
|
+
logger = logging.getLogger(__name__)
|
19
|
+
|
20
|
+
class BrowserPlan(BaseModel):
|
21
|
+
tasks: List[Dict[str, Any]] = Field(
|
22
|
+
...,
|
23
|
+
description="List of automation tasks to execute"
|
24
|
+
)
|
25
|
+
|
26
|
+
class WebBrowserTool(BaseTool):
|
27
|
+
name: str = Field("WebBrowser", description="Name of the tool")
|
28
|
+
description: str = Field(
|
29
|
+
"Universal web automation tool for dynamic website interactions",
|
30
|
+
description="Tool description"
|
31
|
+
)
|
32
|
+
|
33
|
+
def execute(self, input: Dict[str, Any]) -> Dict[str, Any]:
|
34
|
+
"""Execute dynamic web automation workflow"""
|
35
|
+
driver = None
|
36
|
+
try:
|
37
|
+
driver = self._init_browser(input.get("headless", False))
|
38
|
+
results = []
|
39
|
+
current_url = ""
|
40
|
+
|
41
|
+
# Generate initial plan
|
42
|
+
plan = self._generate_plan(input['query'], current_url)
|
43
|
+
|
44
|
+
for task in plan.tasks:
|
45
|
+
result = self._execute_safe_task(driver, task)
|
46
|
+
results.append(result)
|
47
|
+
|
48
|
+
if not result['success']:
|
49
|
+
break
|
50
|
+
|
51
|
+
# Update context for next tasks
|
52
|
+
current_url = driver.current_url
|
53
|
+
|
54
|
+
return {"status": "success", "results": results}
|
55
|
+
|
56
|
+
except Exception as e:
|
57
|
+
return {"status": "error", "message": str(e)}
|
58
|
+
finally:
|
59
|
+
if driver:
|
60
|
+
driver.quit()
|
61
|
+
|
62
|
+
def _init_browser(self, headless: bool) -> webdriver.Chrome:
|
63
|
+
"""Initialize browser with advanced options"""
|
64
|
+
options = Options()
|
65
|
+
options.add_argument("--start-maximized")
|
66
|
+
options.add_argument("--disable-blink-features=AutomationControlled")
|
67
|
+
options.add_experimental_option("excludeSwitches", ["enable-automation"])
|
68
|
+
|
69
|
+
if headless:
|
70
|
+
options.add_argument("--headless=new")
|
71
|
+
|
72
|
+
return webdriver.Chrome(
|
73
|
+
service=Service(ChromeDriverManager().install()),
|
74
|
+
options=options
|
75
|
+
)
|
76
|
+
|
77
|
+
def _generate_plan(self, query: str, current_url: str) -> BrowserPlan:
|
78
|
+
"""Generate adaptive execution plan using LLM"""
|
79
|
+
prompt = f"""Generate browser automation plan for: {query}
|
80
|
+
|
81
|
+
Current URL: {current_url or 'No page loaded yet'}
|
82
|
+
|
83
|
+
Required JSON format:
|
84
|
+
{{
|
85
|
+
"tasks": [
|
86
|
+
{{
|
87
|
+
"action": "navigate|click|type|wait|scroll",
|
88
|
+
"selector": "CSS selector (optional)",
|
89
|
+
"value": "input text/URL/seconds",
|
90
|
+
"description": "action purpose"
|
91
|
+
}}
|
92
|
+
]
|
93
|
+
}}
|
94
|
+
|
95
|
+
Guidelines:
|
96
|
+
1. Prefer IDs in selectors (#element-id)
|
97
|
+
2. Use semantic attributes (aria-label, name)
|
98
|
+
3. Include wait steps after navigation
|
99
|
+
4. Prioritize visible elements
|
100
|
+
5. Add scroll steps for hidden elements
|
101
|
+
"""
|
102
|
+
|
103
|
+
response = self.llm.generate(prompt=prompt)
|
104
|
+
return self._parse_plan(response)
|
105
|
+
|
106
|
+
def _parse_plan(self, response: str) -> BrowserPlan:
|
107
|
+
"""Robust JSON parsing with multiple fallback strategies"""
|
108
|
+
try:
|
109
|
+
# Try extracting JSON from markdown code block
|
110
|
+
json_match = re.search(r'```json\n?(.+?)\n?```', response, re.DOTALL)
|
111
|
+
if json_match:
|
112
|
+
plan_data = json.loads(json_match.group(1).strip())
|
113
|
+
else:
|
114
|
+
# Fallback to extract first JSON object
|
115
|
+
json_str = re.search(r'\{.*\}', response, re.DOTALL).group()
|
116
|
+
plan_data = json.loads(json_str)
|
117
|
+
|
118
|
+
# Validate tasks structure
|
119
|
+
validated_tasks = []
|
120
|
+
for task in plan_data.get("tasks", []):
|
121
|
+
if not all(key in task for key in ["action", "description"]):
|
122
|
+
continue
|
123
|
+
validated_tasks.append({
|
124
|
+
"action": task["action"],
|
125
|
+
"selector": task.get("selector", ""),
|
126
|
+
"value": task.get("value", ""),
|
127
|
+
"description": task["description"]
|
128
|
+
})
|
129
|
+
|
130
|
+
return BrowserPlan(tasks=validated_tasks)
|
131
|
+
|
132
|
+
except (json.JSONDecodeError, AttributeError) as e:
|
133
|
+
logger.error(f"Plan parsing failed: {e}")
|
134
|
+
return BrowserPlan(tasks=[])
|
135
|
+
|
136
|
+
def _execute_safe_task(self, driver, task: Dict) -> Dict[str, Any]:
|
137
|
+
"""Execute task with comprehensive error handling"""
|
138
|
+
try:
|
139
|
+
action = task["action"].lower()
|
140
|
+
selector = task.get("selector", "")
|
141
|
+
value = task.get("value", "")
|
142
|
+
|
143
|
+
if action == "navigate":
|
144
|
+
return self._handle_navigation(driver, value)
|
145
|
+
|
146
|
+
elif action == "click":
|
147
|
+
return self._handle_click(driver, selector)
|
148
|
+
|
149
|
+
elif action == "type":
|
150
|
+
return self._handle_typing(driver, selector, value)
|
151
|
+
|
152
|
+
elif action == "wait":
|
153
|
+
return self._handle_wait(value)
|
154
|
+
|
155
|
+
elif action == "scroll":
|
156
|
+
return self._handle_scroll(driver, selector)
|
157
|
+
|
158
|
+
return {
|
159
|
+
"action": action,
|
160
|
+
"success": False,
|
161
|
+
"message": f"Unsupported action: {action}"
|
162
|
+
}
|
163
|
+
|
164
|
+
except Exception as e:
|
165
|
+
return {
|
166
|
+
"action": action,
|
167
|
+
"success": False,
|
168
|
+
"message": f"Critical error: {str(e)}"
|
169
|
+
}
|
170
|
+
|
171
|
+
def _handle_navigation(self, driver, url: str) -> Dict[str, Any]:
|
172
|
+
"""Smart navigation handler"""
|
173
|
+
if not url.startswith(("http://", "https://")):
|
174
|
+
url = f"https://{url}"
|
175
|
+
|
176
|
+
try:
|
177
|
+
driver.get(url)
|
178
|
+
WebDriverWait(driver, 15).until(
|
179
|
+
EC.presence_of_element_located((By.TAG_NAME, "body"))
|
180
|
+
)
|
181
|
+
return {
|
182
|
+
"action": "navigate",
|
183
|
+
"success": True,
|
184
|
+
"message": f"Navigated to {url}"
|
185
|
+
}
|
186
|
+
except Exception as e:
|
187
|
+
return {
|
188
|
+
"action": "navigate",
|
189
|
+
"success": False,
|
190
|
+
"message": f"Navigation failed: {str(e)}"
|
191
|
+
}
|
192
|
+
|
193
|
+
def _handle_click(self, driver, selector: str) -> Dict[str, Any]:
|
194
|
+
"""Dynamic click handler"""
|
195
|
+
try:
|
196
|
+
element = WebDriverWait(driver, 15).until(
|
197
|
+
EC.element_to_be_clickable((By.CSS_SELECTOR, selector))
|
198
|
+
)
|
199
|
+
driver.execute_script("arguments[0].scrollIntoView({behavior: 'smooth'});", element)
|
200
|
+
element.click()
|
201
|
+
return {
|
202
|
+
"action": "click",
|
203
|
+
"success": True,
|
204
|
+
"message": f"Clicked element: {selector}"
|
205
|
+
}
|
206
|
+
except Exception as e:
|
207
|
+
return {
|
208
|
+
"action": "click",
|
209
|
+
"success": False,
|
210
|
+
"message": f"Click failed: {str(e)}"
|
211
|
+
}
|
212
|
+
|
213
|
+
def _handle_typing(self, driver, selector: str, text: str) -> Dict[str, Any]:
|
214
|
+
"""Universal typing handler"""
|
215
|
+
try:
|
216
|
+
element = WebDriverWait(driver, 15).until(
|
217
|
+
EC.presence_of_element_located((By.CSS_SELECTOR, selector))
|
218
|
+
)
|
219
|
+
element.clear()
|
220
|
+
element.send_keys(text)
|
221
|
+
return {
|
222
|
+
"action": "type",
|
223
|
+
"success": True,
|
224
|
+
"message": f"Typed '{text}' into {selector}"
|
225
|
+
}
|
226
|
+
except Exception as e:
|
227
|
+
return {
|
228
|
+
"action": "type",
|
229
|
+
"success": False,
|
230
|
+
"message": f"Typing failed: {str(e)}"
|
231
|
+
}
|
232
|
+
|
233
|
+
def _handle_wait(self, seconds: str) -> Dict[str, Any]:
|
234
|
+
"""Configurable wait handler"""
|
235
|
+
try:
|
236
|
+
wait_time = float(seconds)
|
237
|
+
time.sleep(wait_time)
|
238
|
+
return {
|
239
|
+
"action": "wait",
|
240
|
+
"success": True,
|
241
|
+
"message": f"Waited {wait_time} seconds"
|
242
|
+
}
|
243
|
+
except ValueError:
|
244
|
+
return {
|
245
|
+
"action": "wait",
|
246
|
+
"success": False,
|
247
|
+
"message": "Invalid wait time"
|
248
|
+
}
|
249
|
+
|
250
|
+
def _handle_scroll(self, driver, selector: str) -> Dict[str, Any]:
|
251
|
+
"""Smart scroll handler"""
|
252
|
+
try:
|
253
|
+
if selector:
|
254
|
+
element = WebDriverWait(driver, 15).until(
|
255
|
+
EC.presence_of_element_located((By.CSS_SELECTOR, selector))
|
256
|
+
)
|
257
|
+
driver.execute_script("arguments[0].scrollIntoView({behavior: 'smooth'});", element)
|
258
|
+
else:
|
259
|
+
driver.execute_script("window.scrollTo(0, document.body.scrollHeight);")
|
260
|
+
|
261
|
+
return {
|
262
|
+
"action": "scroll",
|
263
|
+
"success": True,
|
264
|
+
"message": f"Scrolled to {selector or 'page bottom'}"
|
265
|
+
}
|
266
|
+
except Exception as e:
|
267
|
+
return {
|
268
|
+
"action": "scroll",
|
269
|
+
"success": False,
|
270
|
+
"message": f"Scroll failed: {str(e)}"
|
271
|
+
}
|
@@ -2,7 +2,7 @@ from setuptools import setup, find_packages
|
|
2
2
|
|
3
3
|
setup(
|
4
4
|
name="semantio",
|
5
|
-
version="0.0.
|
5
|
+
version="0.0.4",
|
6
6
|
description="A powerful SDK for building AI agents",
|
7
7
|
long_description=open("README.md").read(),
|
8
8
|
long_description_content_type="text/markdown",
|
@@ -28,6 +28,10 @@ setup(
|
|
28
28
|
"fuzzywuzzy", # For fuzzy string matching
|
29
29
|
"duckduckgo-search", # For DuckDuckGo search
|
30
30
|
"yfinance", # For stock/crypto prices
|
31
|
+
"selenium", # For web scraping
|
32
|
+
"beautifulsoup4", # For HTML parsing
|
33
|
+
"webdriver-manager", # For browser automation
|
34
|
+
"validators", # For URL validation
|
31
35
|
|
32
36
|
],
|
33
37
|
classifiers=[
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|