webscout 6.2b0__py3-none-any.whl → 6.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of webscout might be problematic. Click here for more details.

Files changed (97) hide show
  1. webscout/AIauto.py +191 -176
  2. webscout/AIbase.py +112 -239
  3. webscout/AIutel.py +488 -1130
  4. webscout/Agents/functioncall.py +248 -198
  5. webscout/Bing_search.py +250 -153
  6. webscout/DWEBS.py +454 -178
  7. webscout/Extra/__init__.py +2 -1
  8. webscout/Extra/autocoder/__init__.py +9 -0
  9. webscout/Extra/autocoder/autocoder_utiles.py +121 -0
  10. webscout/Extra/autocoder/rawdog.py +681 -0
  11. webscout/Extra/autollama.py +246 -195
  12. webscout/Extra/gguf.py +441 -226
  13. webscout/Extra/weather.py +172 -67
  14. webscout/LLM.py +442 -100
  15. webscout/Litlogger/__init__.py +681 -0
  16. webscout/Local/formats.py +4 -2
  17. webscout/Provider/Amigo.py +19 -10
  18. webscout/Provider/Andi.py +0 -33
  19. webscout/Provider/Blackboxai.py +4 -204
  20. webscout/Provider/DARKAI.py +1 -1
  21. webscout/Provider/EDITEE.py +1 -1
  22. webscout/Provider/Llama3.py +1 -1
  23. webscout/Provider/Marcus.py +137 -0
  24. webscout/Provider/NinjaChat.py +1 -1
  25. webscout/Provider/PI.py +221 -207
  26. webscout/Provider/Perplexity.py +598 -598
  27. webscout/Provider/RoboCoders.py +206 -0
  28. webscout/Provider/TTI/AiForce/__init__.py +22 -0
  29. webscout/Provider/TTI/AiForce/async_aiforce.py +257 -0
  30. webscout/Provider/TTI/AiForce/sync_aiforce.py +242 -0
  31. webscout/Provider/TTI/Nexra/__init__.py +22 -0
  32. webscout/Provider/TTI/Nexra/async_nexra.py +286 -0
  33. webscout/Provider/TTI/Nexra/sync_nexra.py +258 -0
  34. webscout/Provider/TTI/PollinationsAI/__init__.py +23 -0
  35. webscout/Provider/TTI/PollinationsAI/async_pollinations.py +330 -0
  36. webscout/Provider/TTI/PollinationsAI/sync_pollinations.py +285 -0
  37. webscout/Provider/TTI/__init__.py +3 -4
  38. webscout/Provider/TTI/artbit/__init__.py +22 -0
  39. webscout/Provider/TTI/artbit/async_artbit.py +184 -0
  40. webscout/Provider/TTI/artbit/sync_artbit.py +176 -0
  41. webscout/Provider/TTI/blackbox/__init__.py +4 -0
  42. webscout/Provider/TTI/blackbox/async_blackbox.py +212 -0
  43. webscout/Provider/TTI/{blackboximage.py → blackbox/sync_blackbox.py} +199 -153
  44. webscout/Provider/TTI/deepinfra/__init__.py +4 -0
  45. webscout/Provider/TTI/deepinfra/async_deepinfra.py +227 -0
  46. webscout/Provider/TTI/deepinfra/sync_deepinfra.py +199 -0
  47. webscout/Provider/TTI/huggingface/__init__.py +22 -0
  48. webscout/Provider/TTI/huggingface/async_huggingface.py +199 -0
  49. webscout/Provider/TTI/huggingface/sync_huggingface.py +195 -0
  50. webscout/Provider/TTI/imgninza/__init__.py +4 -0
  51. webscout/Provider/TTI/imgninza/async_ninza.py +214 -0
  52. webscout/Provider/TTI/{imgninza.py → imgninza/sync_ninza.py} +209 -136
  53. webscout/Provider/TTI/talkai/__init__.py +4 -0
  54. webscout/Provider/TTI/talkai/async_talkai.py +229 -0
  55. webscout/Provider/TTI/talkai/sync_talkai.py +207 -0
  56. webscout/Provider/__init__.py +146 -132
  57. webscout/Provider/askmyai.py +158 -0
  58. webscout/Provider/cerebras.py +227 -206
  59. webscout/Provider/geminiapi.py +208 -198
  60. webscout/Provider/llama3mitril.py +180 -0
  61. webscout/Provider/llmchat.py +203 -0
  62. webscout/Provider/mhystical.py +176 -0
  63. webscout/Provider/perplexitylabs.py +265 -0
  64. webscout/Provider/talkai.py +196 -0
  65. webscout/Provider/twitterclone.py +251 -244
  66. webscout/Provider/typegpt.py +359 -0
  67. webscout/__init__.py +28 -23
  68. webscout/__main__.py +5 -5
  69. webscout/cli.py +327 -347
  70. webscout/conversation.py +227 -0
  71. webscout/exceptions.py +161 -29
  72. webscout/litagent/__init__.py +172 -0
  73. webscout/litprinter/__init__.py +831 -0
  74. webscout/optimizers.py +270 -0
  75. webscout/prompt_manager.py +279 -0
  76. webscout/swiftcli/__init__.py +810 -0
  77. webscout/transcriber.py +479 -551
  78. webscout/update_checker.py +125 -0
  79. webscout/version.py +1 -1
  80. webscout-6.4.dist-info/LICENSE.md +211 -0
  81. {webscout-6.2b0.dist-info → webscout-6.4.dist-info}/METADATA +34 -55
  82. webscout-6.4.dist-info/RECORD +154 -0
  83. webscout/Provider/TTI/AIuncensored.py +0 -103
  84. webscout/Provider/TTI/Nexra.py +0 -120
  85. webscout/Provider/TTI/PollinationsAI.py +0 -138
  86. webscout/Provider/TTI/WebSimAI.py +0 -142
  87. webscout/Provider/TTI/aiforce.py +0 -160
  88. webscout/Provider/TTI/artbit.py +0 -141
  89. webscout/Provider/TTI/deepinfra.py +0 -148
  90. webscout/Provider/TTI/huggingface.py +0 -155
  91. webscout/models.py +0 -23
  92. webscout-6.2b0.dist-info/LICENSE.md +0 -50
  93. webscout-6.2b0.dist-info/RECORD +0 -118
  94. /webscout/{g4f.py → gpt4free.py} +0 -0
  95. {webscout-6.2b0.dist-info → webscout-6.4.dist-info}/WHEEL +0 -0
  96. {webscout-6.2b0.dist-info → webscout-6.4.dist-info}/entry_points.txt +0 -0
  97. {webscout-6.2b0.dist-info → webscout-6.4.dist-info}/top_level.txt +0 -0
webscout/Bing_search.py CHANGED
@@ -1,154 +1,251 @@
1
- from bs4 import BeautifulSoup
2
- import requests
3
- from typing import Dict, List, Optional, Union
4
- from concurrent.futures import ThreadPoolExecutor, as_completed
5
- from urllib.parse import quote, urlparse, parse_qs
6
- import base64
7
- import urllib3
8
-
9
- # Disable SSL warnings
10
- urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
11
-
12
- class BingS:
13
- """Bing search class to get search results from bing.com."""
14
-
15
- _executor: ThreadPoolExecutor = ThreadPoolExecutor(max_workers=10)
16
-
17
- def __init__(
18
- self,
19
- headers: Optional[Dict[str, str]] = None,
20
- proxy: Optional[str] = None,
21
- timeout: Optional[int] = 10,
22
- ) -> None:
23
- """Initialize the BingS object."""
24
- self.proxy: Optional[str] = proxy
25
- self.headers = headers if headers else {
26
- "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36"
27
- }
28
- self.headers["Referer"] = "https://www.bing.com/"
29
- self.client = requests.Session()
30
- self.client.headers.update(self.headers)
31
- self.client.proxies.update({"http": self.proxy, "https": self.proxy})
32
- self.timeout = timeout
33
-
34
- def __enter__(self) -> "BingS":
35
- return self
36
-
37
- def __exit__(self, exc_type, exc_val, exc_tb):
38
- self.client.close()
39
-
40
- def _get_url(
41
- self,
42
- method: str,
43
- url: str,
44
- params: Optional[Dict[str, str]] = None,
45
- data: Optional[Union[Dict[str, str], bytes]] = None,
46
- ) -> bytes:
47
- try:
48
- resp = self.client.request(method, url, params=params, data=data, timeout=self.timeout, verify=False)
49
- except Exception as ex:
50
- raise Exception(f"{url} {type(ex).__name__}: {ex}") from ex
51
- if resp.status_code == 200:
52
- return resp.content
53
- raise Exception(f"{resp.url} returned status code {resp.status_code}. {params=} {data=}")
54
-
55
- def extract_text_from_webpage(self, html_content, max_characters=None):
56
- """Extracts visible text from HTML content using BeautifulSoup."""
57
- soup = BeautifulSoup(html_content, "html.parser")
58
- # Remove unwanted tags
59
- for tag in soup(["script", "style", "header", "footer", "nav"]):
60
- tag.extract()
61
- # Get the remaining visible text
62
- visible_text = soup.get_text(separator=' ', strip=True)
63
- if max_characters:
64
- visible_text = visible_text[:max_characters]
65
- return visible_text
66
-
67
- def search(
68
- self,
69
- keywords: str,
70
- max_results: Optional[int] = 10,
71
- extract_webpage_text: bool = False,
72
- max_extract_characters: Optional[int] = 100,
73
- ) -> List[Dict[str, str]]:
74
- """Bing text search."""
75
- assert keywords, "keywords is mandatory"
76
-
77
- results = []
78
- futures = []
79
- start = 1
80
- while len(results) < max_results:
81
- params = {
82
- "q": keywords,
83
- "first": start
84
- }
85
- futures.append(self._executor.submit(self._get_url, "GET", "https://www.bing.com/search", params=params))
86
- start += 10
87
-
88
- for future in as_completed(futures):
89
- try:
90
- resp_content = future.result()
91
- soup = BeautifulSoup(resp_content, "html.parser")
92
- result_block = soup.select('li.b_algo')
93
-
94
- if not result_block:
95
- break
96
-
97
- for result in result_block:
98
- try:
99
- link = result.select_one('h2 a')
100
- title = link.text if link else ""
101
- url = link['href'] if link else ""
102
- abstract = result.select_one('.b_caption p')
103
- description = abstract.text if abstract else ""
104
-
105
- # Remove "WEB" from the beginning of the description if it exists
106
- if description.startswith("WEB"):
107
- description = description[3:].strip()
108
-
109
- visible_text = ""
110
- if extract_webpage_text:
111
- try:
112
- actual_url = self._decode_bing_url(url)
113
- page_content = self._get_url("GET", actual_url)
114
- visible_text = self.extract_text_from_webpage(
115
- page_content, max_characters=max_extract_characters
116
- )
117
- except Exception as e:
118
- print(f"Error extracting text from {url}: {e}")
119
-
120
- results.append({
121
- "title": title,
122
- "href": url,
123
- "abstract": description,
124
- "index": len(results),
125
- "type": "web",
126
- "visible_text": visible_text,
127
- })
128
-
129
- if len(results) >= max_results:
130
- return results
131
-
132
- except Exception as e:
133
- print(f"Error extracting result: {e}")
134
-
135
- except Exception as e:
136
- print(f"Error fetching URL: {e}")
137
-
138
- return results
139
-
140
- def _decode_bing_url(self, url):
141
- if 'bing.com/ck/a' in url:
142
- parsed_url = urlparse(url)
143
- query_params = parse_qs(parsed_url.query)
144
- if 'u' in query_params:
145
- encoded_url = query_params['u'][0]
146
- return base64.b64decode(encoded_url).decode('utf-8')
147
- return url
148
-
149
- if __name__ == "__main__":
150
- from rich import print
151
- searcher = BingS()
152
- results = searcher.search("Python development tools", max_results=5, extract_webpage_text=True, max_extract_characters=2000)
153
- for result in results:
1
+ from bs4 import BeautifulSoup
2
+ import requests
3
+ from typing import Dict, List, Optional, Union
4
+ from concurrent.futures import ThreadPoolExecutor, as_completed
5
+ from urllib.parse import quote, urlparse, parse_qs
6
+ import base64
7
+ import urllib3
8
+
9
+ # Disable SSL warnings
10
+ urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
11
+
12
+ class BingS:
13
+ """A Python interface for Bing search engine.
14
+
15
+ The BingS class provides a simple interface to perform searches on Bing.com
16
+ and extract search results programmatically.
17
+
18
+ Basic Usage:
19
+ >>> from webscout.Bing_search import BingS
20
+ >>> searcher = BingS()
21
+ >>> results = searcher.search("Python programming")
22
+ >>> for result in results:
23
+ ... print(result['title'], result['href'])
24
+
25
+ Advanced Usage:
26
+ >>> # With custom headers and proxy
27
+ >>> headers = {'User-Agent': 'Custom User Agent'}
28
+ >>> proxy = 'http://proxy.example.com:8080'
29
+ >>> searcher = BingS(headers=headers, proxy=proxy)
30
+ >>> results = searcher.search(
31
+ ... "AI developments",
32
+ ... max_results=5,
33
+ ... extract_webpage_text=True,
34
+ ... max_extract_characters=1000
35
+ ... )
36
+ >>> # Access result fields
37
+ >>> for result in results:
38
+ ... print(f"Title: {result['title']}")
39
+ ... print(f"URL: {result['href']}")
40
+ ... print(f"Description: {result['abstract']}")
41
+ ... if result['visible_text']:
42
+ ... print(f"Page Content: {result['visible_text'][:100]}...")
43
+
44
+ The class supports context management protocol:
45
+ >>> with BingS() as searcher:
46
+ ... results = searcher.search("Python tutorials")
47
+
48
+ Return Dictionary Format:
49
+ {
50
+ 'title': str, # The title of the search result
51
+ 'href': str, # The URL of the search result
52
+ 'abstract': str, # Brief description or snippet
53
+ 'index': int, # Position in search results
54
+ 'type': str, # Type of result (always 'web')
55
+ 'visible_text': str # Extracted webpage text (if requested)
56
+ }
57
+ """
58
+
59
+ _executor: ThreadPoolExecutor = ThreadPoolExecutor(max_workers=10)
60
+
61
+ def __init__(
62
+ self,
63
+ headers: Optional[Dict[str, str]] = None,
64
+ proxy: Optional[str] = None,
65
+ timeout: Optional[int] = 10,
66
+ ) -> None:
67
+ """Initialize a new BingS instance.
68
+
69
+ Args:
70
+ headers (Optional[Dict[str, str]]): Custom HTTP headers for requests.
71
+ Defaults to a standard User-Agent if not provided.
72
+ proxy (Optional[str]): Proxy URL to use for requests.
73
+ Example: 'http://proxy.example.com:8080'
74
+ timeout (Optional[int]): Request timeout in seconds. Defaults to 10.
75
+
76
+ Example:
77
+ >>> searcher = BingS(
78
+ ... headers={'User-Agent': 'Custom UA'},
79
+ ... proxy='http://proxy.example.com:8080',
80
+ ... timeout=15
81
+ ... )
82
+ """
83
+ self.proxy: Optional[str] = proxy
84
+ self.headers = headers if headers else {
85
+ "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36"
86
+ }
87
+ self.headers["Referer"] = "https://www.bing.com/"
88
+ self.client = requests.Session()
89
+ self.client.headers.update(self.headers)
90
+ self.client.proxies.update({"http": self.proxy, "https": self.proxy})
91
+ self.timeout = timeout
92
+
93
+ def __enter__(self) -> "BingS":
94
+ return self
95
+
96
+ def __exit__(self, exc_type, exc_val, exc_tb):
97
+ self.client.close()
98
+
99
+ def _get_url(
100
+ self,
101
+ method: str,
102
+ url: str,
103
+ params: Optional[Dict[str, str]] = None,
104
+ data: Optional[Union[Dict[str, str], bytes]] = None,
105
+ ) -> bytes:
106
+ try:
107
+ resp = self.client.request(method, url, params=params, data=data, timeout=self.timeout, verify=False)
108
+ except Exception as ex:
109
+ raise Exception(f"{url} {type(ex).__name__}: {ex}") from ex
110
+ if resp.status_code == 200:
111
+ return resp.content
112
+ raise Exception(f"{resp.url} returned status code {resp.status_code}. {params=} {data=}")
113
+
114
+ def extract_text_from_webpage(self, html_content, max_characters=None):
115
+ """Extracts visible text from HTML content using BeautifulSoup."""
116
+ soup = BeautifulSoup(html_content, "html.parser")
117
+ # Remove unwanted tags
118
+ for tag in soup(["script", "style", "header", "footer", "nav"]):
119
+ tag.extract()
120
+ # Get the remaining visible text
121
+ visible_text = soup.get_text(separator=' ', strip=True)
122
+ if max_characters:
123
+ visible_text = visible_text[:max_characters]
124
+ return visible_text
125
+
126
+ def search(
127
+ self,
128
+ keywords: str,
129
+ max_results: Optional[int] = 10,
130
+ extract_webpage_text: bool = False,
131
+ max_extract_characters: Optional[int] = 100,
132
+ ) -> List[Dict[str, str]]:
133
+ """Perform a Bing search and return results.
134
+
135
+ Args:
136
+ keywords (str): Search query string.
137
+ max_results (Optional[int]): Maximum number of results to return.
138
+ Defaults to 10.
139
+ extract_webpage_text (bool): If True, fetches and extracts text from
140
+ each result webpage. Defaults to False.
141
+ max_extract_characters (Optional[int]): Maximum number of characters
142
+ to extract from each webpage. Only used if extract_webpage_text
143
+ is True. Defaults to 100.
144
+
145
+ Returns:
146
+ List[Dict[str, str]]: List of search results. Each result contains:
147
+ - title: The title of the search result
148
+ - href: The URL of the search result
149
+ - abstract: Brief description or snippet
150
+ - index: Position in search results
151
+ - type: Type of result (always 'web')
152
+ - visible_text: Extracted webpage text (if extract_webpage_text=True)
153
+
154
+ Raises:
155
+ AssertionError: If keywords is empty.
156
+ Exception: If request fails or returns non-200 status code.
157
+
158
+ Example:
159
+ >>> searcher = BingS()
160
+ >>> results = searcher.search(
161
+ ... "Python tutorials",
162
+ ... max_results=5,
163
+ ... extract_webpage_text=True
164
+ ... )
165
+ >>> for result in results:
166
+ ... print(f"Title: {result['title']}")
167
+ ... print(f"URL: {result['href']}")
168
+ ... print(f"Description: {result['abstract']}")
169
+ ... if result['visible_text']:
170
+ ... print(f"Content: {result['visible_text'][:100]}...")
171
+ """
172
+ assert keywords, "keywords is mandatory"
173
+
174
+ results = []
175
+ futures = []
176
+ start = 1
177
+ while len(results) < max_results:
178
+ params = {
179
+ "q": keywords,
180
+ "first": start
181
+ }
182
+ futures.append(self._executor.submit(self._get_url, "GET", "https://www.bing.com/search", params=params))
183
+ start += 10
184
+
185
+ for future in as_completed(futures):
186
+ try:
187
+ resp_content = future.result()
188
+ soup = BeautifulSoup(resp_content, "html.parser")
189
+ result_block = soup.select('li.b_algo')
190
+
191
+ if not result_block:
192
+ break
193
+
194
+ for result in result_block:
195
+ try:
196
+ link = result.select_one('h2 a')
197
+ title = link.text if link else ""
198
+ url = link['href'] if link else ""
199
+ abstract = result.select_one('.b_caption p')
200
+ description = abstract.text if abstract else ""
201
+
202
+ # Remove "WEB" from the beginning of the description if it exists
203
+ if description.startswith("WEB"):
204
+ description = description[3:].strip()
205
+
206
+ visible_text = ""
207
+ if extract_webpage_text:
208
+ try:
209
+ actual_url = self._decode_bing_url(url)
210
+ page_content = self._get_url("GET", actual_url)
211
+ visible_text = self.extract_text_from_webpage(
212
+ page_content, max_characters=max_extract_characters
213
+ )
214
+ except Exception as e:
215
+ print(f"Error extracting text from {url}: {e}")
216
+
217
+ results.append({
218
+ "title": title,
219
+ "href": url,
220
+ "abstract": description,
221
+ "index": len(results),
222
+ "type": "web",
223
+ "visible_text": visible_text,
224
+ })
225
+
226
+ if len(results) >= max_results:
227
+ return results
228
+
229
+ except Exception as e:
230
+ print(f"Error extracting result: {e}")
231
+
232
+ except Exception as e:
233
+ print(f"Error fetching URL: {e}")
234
+
235
+ return results
236
+
237
+ def _decode_bing_url(self, url):
238
+ if 'bing.com/ck/a' in url:
239
+ parsed_url = urlparse(url)
240
+ query_params = parse_qs(parsed_url.query)
241
+ if 'u' in query_params:
242
+ encoded_url = query_params['u'][0]
243
+ return base64.b64decode(encoded_url).decode('utf-8')
244
+ return url
245
+
246
+ if __name__ == "__main__":
247
+ from rich import print
248
+ searcher = BingS()
249
+ results = searcher.search("Python development tools", max_results=5, extract_webpage_text=True, max_extract_characters=2000)
250
+ for result in results:
154
251
  print(result)