webscout 7.7__py3-none-any.whl → 7.8__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of webscout might be problematic. Click here for more details.
- webscout/AIutel.py +2 -1
- webscout/Bard.py +14 -11
- webscout/DWEBS.py +431 -415
- webscout/Extra/autocoder/__init__.py +9 -9
- webscout/Extra/autocoder/autocoder_utiles.py +332 -194
- webscout/Extra/autocoder/rawdog.py +68 -9
- webscout/Extra/gguf.py +682 -682
- webscout/Provider/AI21.py +1 -1
- webscout/Provider/AISEARCH/DeepFind.py +2 -2
- webscout/Provider/AISEARCH/ISou.py +2 -2
- webscout/Provider/AISEARCH/felo_search.py +6 -6
- webscout/Provider/AISEARCH/genspark_search.py +1 -1
- webscout/Provider/Aitopia.py +292 -0
- webscout/Provider/AllenAI.py +1 -1
- webscout/Provider/Andi.py +3 -3
- webscout/Provider/C4ai.py +1 -1
- webscout/Provider/ChatGPTES.py +3 -5
- webscout/Provider/ChatGPTGratis.py +4 -4
- webscout/Provider/Chatify.py +2 -2
- webscout/Provider/Cloudflare.py +3 -2
- webscout/Provider/DARKAI.py +3 -2
- webscout/Provider/DeepSeek.py +2 -2
- webscout/Provider/Deepinfra.py +1 -1
- webscout/Provider/EDITEE.py +1 -1
- webscout/Provider/ElectronHub.py +178 -96
- webscout/Provider/ExaChat.py +310 -0
- webscout/Provider/Free2GPT.py +2 -2
- webscout/Provider/Gemini.py +5 -19
- webscout/Provider/GithubChat.py +1 -1
- webscout/Provider/Glider.py +4 -4
- webscout/Provider/Groq.py +3 -3
- webscout/Provider/HF_space/qwen_qwen2.py +1 -1
- webscout/Provider/HeckAI.py +1 -1
- webscout/Provider/HuggingFaceChat.py +1 -1
- webscout/Provider/Hunyuan.py +1 -1
- webscout/Provider/Jadve.py +3 -3
- webscout/Provider/Koboldai.py +3 -3
- webscout/Provider/LambdaChat.py +1 -1
- webscout/Provider/Llama.py +3 -5
- webscout/Provider/Llama3.py +4 -12
- webscout/Provider/Marcus.py +3 -3
- webscout/Provider/OLLAMA.py +8 -8
- webscout/Provider/Openai.py +7 -3
- webscout/Provider/PI.py +1 -1
- webscout/Provider/Perplexitylabs.py +1 -1
- webscout/Provider/Phind.py +1 -1
- webscout/Provider/PizzaGPT.py +1 -1
- webscout/Provider/QwenLM.py +4 -7
- webscout/Provider/TTI/FreeAIPlayground/async_freeaiplayground.py +3 -1
- webscout/Provider/TTI/FreeAIPlayground/sync_freeaiplayground.py +3 -3
- webscout/Provider/TTI/ImgSys/__init__.py +23 -0
- webscout/Provider/TTI/ImgSys/async_imgsys.py +202 -0
- webscout/Provider/TTI/ImgSys/sync_imgsys.py +195 -0
- webscout/Provider/TTI/__init__.py +3 -1
- webscout/Provider/TTI/artbit/async_artbit.py +1 -1
- webscout/Provider/TTI/artbit/sync_artbit.py +1 -1
- webscout/Provider/TTI/huggingface/async_huggingface.py +1 -1
- webscout/Provider/TTI/huggingface/sync_huggingface.py +1 -1
- webscout/Provider/TTI/piclumen/__init__.py +22 -22
- webscout/Provider/TTI/piclumen/sync_piclumen.py +232 -232
- webscout/Provider/TTI/pixelmuse/__init__.py +4 -0
- webscout/Provider/TTI/pixelmuse/async_pixelmuse.py +249 -0
- webscout/Provider/TTI/pixelmuse/sync_pixelmuse.py +182 -0
- webscout/Provider/TTI/talkai/sync_talkai.py +1 -1
- webscout/Provider/TTS/utils.py +1 -1
- webscout/Provider/TeachAnything.py +1 -1
- webscout/Provider/TextPollinationsAI.py +4 -4
- webscout/Provider/TwoAI.py +1 -2
- webscout/Provider/Venice.py +4 -2
- webscout/Provider/VercelAI.py +234 -0
- webscout/Provider/WebSim.py +3 -2
- webscout/Provider/WiseCat.py +10 -12
- webscout/Provider/Youchat.py +1 -1
- webscout/Provider/__init__.py +10 -0
- webscout/Provider/ai4chat.py +1 -1
- webscout/Provider/aimathgpt.py +2 -6
- webscout/Provider/akashgpt.py +1 -1
- webscout/Provider/askmyai.py +4 -4
- webscout/Provider/asksteve.py +203 -0
- webscout/Provider/bagoodex.py +2 -2
- webscout/Provider/cerebras.py +1 -1
- webscout/Provider/chatglm.py +4 -4
- webscout/Provider/cleeai.py +1 -0
- webscout/Provider/copilot.py +427 -415
- webscout/Provider/elmo.py +1 -1
- webscout/Provider/flowith.py +1 -1
- webscout/Provider/freeaichat.py +57 -31
- webscout/Provider/gaurish.py +3 -5
- webscout/Provider/geminiprorealtime.py +1 -1
- webscout/Provider/granite.py +4 -4
- webscout/Provider/hermes.py +5 -5
- webscout/Provider/julius.py +1 -1
- webscout/Provider/koala.py +1 -1
- webscout/Provider/lepton.py +1 -1
- webscout/Provider/llama3mitril.py +4 -4
- webscout/Provider/llamatutor.py +1 -1
- webscout/Provider/llmchat.py +3 -3
- webscout/Provider/meta.py +1 -1
- webscout/Provider/multichat.py +10 -10
- webscout/Provider/promptrefine.py +1 -1
- webscout/Provider/searchchat.py +293 -0
- webscout/Provider/sonus.py +2 -2
- webscout/Provider/talkai.py +2 -2
- webscout/Provider/turboseek.py +1 -1
- webscout/Provider/tutorai.py +1 -1
- webscout/Provider/typegpt.py +5 -42
- webscout/Provider/uncovr.py +4 -2
- webscout/Provider/x0gpt.py +1 -1
- webscout/__init__.py +36 -36
- webscout/cli.py +293 -332
- webscout/tempid.py +11 -11
- webscout/utils.py +2 -2
- webscout/version.py +1 -1
- webscout/webscout_search.py +1282 -1223
- webscout/webscout_search_async.py +813 -692
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/METADATA +50 -29
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/RECORD +121 -110
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/LICENSE.md +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/WHEEL +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/entry_points.txt +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/top_level.txt +0 -0
webscout/Provider/AI21.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
from uuid import uuid4
|
|
2
2
|
import requests
|
|
3
3
|
import re
|
|
4
|
-
from typing import Any, Dict, Generator, Optional
|
|
4
|
+
from typing import Any, Dict, Generator, Optional, Union
|
|
5
5
|
|
|
6
6
|
from webscout.AIbase import AISearch
|
|
7
7
|
from webscout import exceptions
|
|
@@ -114,7 +114,7 @@ class DeepFind(AISearch):
|
|
|
114
114
|
prompt: str,
|
|
115
115
|
stream: bool = False,
|
|
116
116
|
raw: bool = False,
|
|
117
|
-
) -> Dict[str,
|
|
117
|
+
) -> Union[Response, Generator[Union[Dict[str, str], Response], None, None]]:
|
|
118
118
|
"""Search using the DeepFind API and get AI-generated responses.
|
|
119
119
|
|
|
120
120
|
This method sends a search query to DeepFind and returns the AI-generated response.
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import requests
|
|
2
2
|
import json
|
|
3
3
|
import re
|
|
4
|
-
from typing import Dict, Optional, Generator, Any
|
|
4
|
+
from typing import Dict, Optional, Generator, Any, Union
|
|
5
5
|
from webscout import LitAgent
|
|
6
6
|
from webscout import exceptions
|
|
7
7
|
from webscout.AIbase import AISearch
|
|
@@ -118,7 +118,7 @@ class Isou(AISearch):
|
|
|
118
118
|
prompt: str,
|
|
119
119
|
stream: bool = False,
|
|
120
120
|
raw: bool = False,
|
|
121
|
-
) ->
|
|
121
|
+
) -> Union[Response, Generator[Union[Dict[str, str], Response], None, None]]:
|
|
122
122
|
"""Search using the Isou API and get AI-generated responses.
|
|
123
123
|
|
|
124
124
|
Args:
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import requests
|
|
2
2
|
from uuid import uuid4
|
|
3
3
|
import json
|
|
4
|
-
from typing import Any, Dict, Generator, Optional
|
|
4
|
+
from typing import Any, Dict, Generator, Optional, Union
|
|
5
5
|
|
|
6
6
|
from webscout.AIbase import AISearch
|
|
7
7
|
from webscout import exceptions
|
|
@@ -9,7 +9,7 @@ from webscout import LitAgent
|
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class Response:
|
|
12
|
-
"""A wrapper class for
|
|
12
|
+
"""A wrapper class for API responses.
|
|
13
13
|
|
|
14
14
|
This class automatically converts response objects to their text representation
|
|
15
15
|
when printed or converted to string.
|
|
@@ -115,7 +115,7 @@ class Felo(AISearch):
|
|
|
115
115
|
prompt: str,
|
|
116
116
|
stream: bool = False,
|
|
117
117
|
raw: bool = False,
|
|
118
|
-
) -> Dict[str,
|
|
118
|
+
) -> Union[Response, Generator[Union[Dict[str, str], Response], None, None]]:
|
|
119
119
|
"""Search using the Felo API and get AI-generated responses.
|
|
120
120
|
|
|
121
121
|
This method sends a search query to Felo and returns the AI-generated response.
|
|
@@ -130,9 +130,9 @@ class Felo(AISearch):
|
|
|
130
130
|
Defaults to False.
|
|
131
131
|
|
|
132
132
|
Returns:
|
|
133
|
-
Union[Dict[str,
|
|
134
|
-
- If stream=False: Returns complete response
|
|
135
|
-
- If stream=True: Yields response chunks as
|
|
133
|
+
Union[Response, Generator[Union[Dict[str, str], Response], None, None]]:
|
|
134
|
+
- If stream=False: Returns complete response as Response object
|
|
135
|
+
- If stream=True: Yields response chunks as either Dict or Response objects
|
|
136
136
|
|
|
137
137
|
Raises:
|
|
138
138
|
APIConnectionError: If the API request fails
|
|
@@ -117,7 +117,7 @@ class Genspark(AISearch):
|
|
|
117
117
|
prompt: str,
|
|
118
118
|
stream: bool = False,
|
|
119
119
|
raw: bool = False,
|
|
120
|
-
) -> Union[
|
|
120
|
+
) -> Union[Response, Generator[Union[Dict[str, str], Response], None, None]]:
|
|
121
121
|
"""Search using the Genspark API and get AI-generated responses.
|
|
122
122
|
|
|
123
123
|
Args:
|
|
@@ -0,0 +1,292 @@
|
|
|
1
|
+
import requests
|
|
2
|
+
import json
|
|
3
|
+
import uuid
|
|
4
|
+
import time
|
|
5
|
+
import hashlib
|
|
6
|
+
from typing import Any, Dict, Optional, Generator, Union
|
|
7
|
+
|
|
8
|
+
from webscout.AIutel import Optimizers
|
|
9
|
+
from webscout.AIutel import Conversation
|
|
10
|
+
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
11
|
+
from webscout.AIbase import Provider, AsyncProvider
|
|
12
|
+
from webscout import exceptions
|
|
13
|
+
from webscout.litagent import LitAgent
|
|
14
|
+
|
|
15
|
+
class Aitopia(Provider):
|
|
16
|
+
"""
|
|
17
|
+
A class to interact with the Aitopia API with LitAgent user-agent.
|
|
18
|
+
"""
|
|
19
|
+
|
|
20
|
+
AVAILABLE_MODELS = [
|
|
21
|
+
"Claude 3 Haiku",
|
|
22
|
+
"GPT-4o Mini",
|
|
23
|
+
"Gemini 1.5 Flash",
|
|
24
|
+
"Llama 3.1 70B"
|
|
25
|
+
]
|
|
26
|
+
|
|
27
|
+
def __init__(
|
|
28
|
+
self,
|
|
29
|
+
is_conversation: bool = True,
|
|
30
|
+
max_tokens: int = 2049,
|
|
31
|
+
timeout: int = 30,
|
|
32
|
+
intro: str = None,
|
|
33
|
+
filepath: str = None,
|
|
34
|
+
update_file: bool = True,
|
|
35
|
+
proxies: dict = {},
|
|
36
|
+
history_offset: int = 10250,
|
|
37
|
+
act: str = None,
|
|
38
|
+
model: str = "Claude 3 Haiku",
|
|
39
|
+
browser: str = "chrome"
|
|
40
|
+
):
|
|
41
|
+
"""Initializes the Aitopia API client."""
|
|
42
|
+
if model not in self.AVAILABLE_MODELS:
|
|
43
|
+
raise ValueError(f"Invalid model: {model}. Choose from: {self.AVAILABLE_MODELS}")
|
|
44
|
+
|
|
45
|
+
self.url = "https://extensions.aitopia.ai/ai/send"
|
|
46
|
+
|
|
47
|
+
# Initialize LitAgent for user agent generation
|
|
48
|
+
self.agent = LitAgent()
|
|
49
|
+
# Use fingerprinting to create a consistent browser identity
|
|
50
|
+
self.fingerprint = self.agent.generate_fingerprint(browser)
|
|
51
|
+
|
|
52
|
+
# Use the fingerprint for headers
|
|
53
|
+
self.headers = {
|
|
54
|
+
"accept": "text/plain",
|
|
55
|
+
"accept-language": self.fingerprint["accept_language"],
|
|
56
|
+
"content-type": "text/plain;charset=UTF-8",
|
|
57
|
+
"dnt": "1",
|
|
58
|
+
"origin": "https://chat.aitopia.ai",
|
|
59
|
+
"priority": "u=1, i",
|
|
60
|
+
"referer": "https://chat.aitopia.ai/",
|
|
61
|
+
"sec-ch-ua": self.fingerprint["sec_ch_ua"] or '"Chromium";v="134", "Not:A-Brand";v="24", "Microsoft Edge";v="134"',
|
|
62
|
+
"sec-ch-ua-mobile": "?0",
|
|
63
|
+
"sec-ch-ua-platform": f'"{self.fingerprint["platform"]}"',
|
|
64
|
+
"sec-fetch-dest": "empty",
|
|
65
|
+
"sec-fetch-mode": "cors",
|
|
66
|
+
"sec-fetch-site": "same-site",
|
|
67
|
+
"user-agent": self.fingerprint["user_agent"]
|
|
68
|
+
}
|
|
69
|
+
|
|
70
|
+
self.session = requests.Session()
|
|
71
|
+
self.session.headers.update(self.headers)
|
|
72
|
+
self.session.proxies.update(proxies)
|
|
73
|
+
|
|
74
|
+
self.is_conversation = is_conversation
|
|
75
|
+
self.max_tokens_to_sample = max_tokens
|
|
76
|
+
self.timeout = timeout
|
|
77
|
+
self.last_response = {}
|
|
78
|
+
self.model = model
|
|
79
|
+
|
|
80
|
+
self.__available_optimizers = (
|
|
81
|
+
method
|
|
82
|
+
for method in dir(Optimizers)
|
|
83
|
+
if callable(getattr(Optimizers, method)) and not method.startswith("__")
|
|
84
|
+
)
|
|
85
|
+
Conversation.intro = (
|
|
86
|
+
AwesomePrompts().get_act(
|
|
87
|
+
act, raise_not_found=True, default=None, case_insensitive=True
|
|
88
|
+
)
|
|
89
|
+
if act
|
|
90
|
+
else intro or Conversation.intro
|
|
91
|
+
)
|
|
92
|
+
|
|
93
|
+
self.conversation = Conversation(
|
|
94
|
+
is_conversation, self.max_tokens_to_sample, filepath, update_file
|
|
95
|
+
)
|
|
96
|
+
self.conversation.history_offset = history_offset
|
|
97
|
+
|
|
98
|
+
def refresh_identity(self, browser: str = None):
|
|
99
|
+
"""
|
|
100
|
+
Refreshes the browser identity fingerprint.
|
|
101
|
+
|
|
102
|
+
Args:
|
|
103
|
+
browser: Specific browser to use for the new fingerprint
|
|
104
|
+
"""
|
|
105
|
+
browser = browser or self.fingerprint.get("browser_type", "chrome")
|
|
106
|
+
self.fingerprint = self.agent.generate_fingerprint(browser)
|
|
107
|
+
|
|
108
|
+
# Update headers with new fingerprint
|
|
109
|
+
self.headers.update({
|
|
110
|
+
"accept-language": self.fingerprint["accept_language"],
|
|
111
|
+
"sec-ch-ua": self.fingerprint["sec_ch_ua"] or self.headers["sec-ch-ua"],
|
|
112
|
+
"sec-ch-ua-platform": f'"{self.fingerprint["platform"]}"',
|
|
113
|
+
"user-agent": self.fingerprint["user_agent"],
|
|
114
|
+
})
|
|
115
|
+
|
|
116
|
+
# Update session headers
|
|
117
|
+
for header, value in self.headers.items():
|
|
118
|
+
self.session.headers[header] = value
|
|
119
|
+
|
|
120
|
+
return self.fingerprint
|
|
121
|
+
|
|
122
|
+
def generate_uuid_search(self):
|
|
123
|
+
"""Generate a UUID and convert to base64-like string."""
|
|
124
|
+
uuid_str = str(uuid.uuid4())
|
|
125
|
+
return uuid_str.replace('-', '')
|
|
126
|
+
|
|
127
|
+
def generate_hopekey(self):
|
|
128
|
+
"""Generate a random string and hash it."""
|
|
129
|
+
random_str = str(uuid.uuid4()) + str(time.time())
|
|
130
|
+
return hashlib.md5(random_str.encode()).hexdigest()
|
|
131
|
+
|
|
132
|
+
def ask(
|
|
133
|
+
self,
|
|
134
|
+
prompt: str,
|
|
135
|
+
stream: bool = False,
|
|
136
|
+
raw: bool = False,
|
|
137
|
+
optimizer: str = None,
|
|
138
|
+
conversationally: bool = False,
|
|
139
|
+
) -> Union[Dict[str, Any], Generator]:
|
|
140
|
+
conversation_prompt = self.conversation.gen_complete_prompt(prompt)
|
|
141
|
+
if optimizer:
|
|
142
|
+
if optimizer in self.__available_optimizers:
|
|
143
|
+
conversation_prompt = getattr(Optimizers, optimizer)(
|
|
144
|
+
conversation_prompt if conversationally else prompt
|
|
145
|
+
)
|
|
146
|
+
else:
|
|
147
|
+
raise Exception(f"Optimizer is not one of {self.__available_optimizers}")
|
|
148
|
+
|
|
149
|
+
# Generate hopekey and update headers
|
|
150
|
+
hopekey = self.generate_hopekey()
|
|
151
|
+
self.headers["hopekey"] = hopekey
|
|
152
|
+
|
|
153
|
+
# Default history if none provided
|
|
154
|
+
history = [
|
|
155
|
+
{
|
|
156
|
+
"item": "Hello, how can I help you today?",
|
|
157
|
+
"role": "assistant",
|
|
158
|
+
# "model": "GPT-4o Mini"
|
|
159
|
+
}
|
|
160
|
+
]
|
|
161
|
+
|
|
162
|
+
# Generate current timestamp for chat_id
|
|
163
|
+
current_time = int(time.time() * 1000)
|
|
164
|
+
|
|
165
|
+
# Request payload
|
|
166
|
+
payload = {
|
|
167
|
+
"history": history,
|
|
168
|
+
"text": conversation_prompt,
|
|
169
|
+
"model": self.model,
|
|
170
|
+
"stream": stream,
|
|
171
|
+
"uuid_search": self.generate_uuid_search(),
|
|
172
|
+
"mode": "ai_chat",
|
|
173
|
+
"prompt_mode": False,
|
|
174
|
+
"extra_key": "__all",
|
|
175
|
+
"extra_data": {"prompt_mode": False},
|
|
176
|
+
"chat_id": current_time,
|
|
177
|
+
"language_detail": {
|
|
178
|
+
"lang_code": "en",
|
|
179
|
+
"name": "English",
|
|
180
|
+
"title": "English"
|
|
181
|
+
},
|
|
182
|
+
"is_continue": False,
|
|
183
|
+
"lang_code": "en"
|
|
184
|
+
}
|
|
185
|
+
|
|
186
|
+
def for_stream():
|
|
187
|
+
try:
|
|
188
|
+
with requests.post(self.url, headers=self.headers, json=payload, stream=True, timeout=self.timeout) as response:
|
|
189
|
+
if response.status_code != 200:
|
|
190
|
+
raise exceptions.FailedToGenerateResponseError(
|
|
191
|
+
f"Request failed with status code {response.status_code}"
|
|
192
|
+
)
|
|
193
|
+
|
|
194
|
+
streaming_text = ""
|
|
195
|
+
for line in response.iter_lines():
|
|
196
|
+
if line:
|
|
197
|
+
line = line.decode('utf-8')
|
|
198
|
+
if line.startswith('data: '):
|
|
199
|
+
data = line[6:]
|
|
200
|
+
if data == '[DONE]':
|
|
201
|
+
break
|
|
202
|
+
try:
|
|
203
|
+
json_data = json.loads(data)
|
|
204
|
+
|
|
205
|
+
# Handle Claude 3 Haiku response format
|
|
206
|
+
if "delta" in json_data and "text" in json_data["delta"]:
|
|
207
|
+
content = json_data["delta"]["text"]
|
|
208
|
+
if content:
|
|
209
|
+
streaming_text += content
|
|
210
|
+
resp = dict(text=content)
|
|
211
|
+
yield resp if raw else resp
|
|
212
|
+
# Handle GPT-4o Mini response format
|
|
213
|
+
elif "choices" in json_data and "0" in json_data["choices"]:
|
|
214
|
+
content = json_data["choices"]["0"]["delta"].get("content", "")
|
|
215
|
+
if content:
|
|
216
|
+
streaming_text += content
|
|
217
|
+
resp = dict(text=content)
|
|
218
|
+
yield resp if raw else resp
|
|
219
|
+
except json.JSONDecodeError:
|
|
220
|
+
continue
|
|
221
|
+
|
|
222
|
+
self.last_response = {"text": streaming_text}
|
|
223
|
+
self.conversation.update_chat_history(prompt, streaming_text)
|
|
224
|
+
|
|
225
|
+
except requests.RequestException as e:
|
|
226
|
+
raise exceptions.FailedToGenerateResponseError(f"Request failed: {str(e)}")
|
|
227
|
+
|
|
228
|
+
def for_non_stream():
|
|
229
|
+
try:
|
|
230
|
+
response = requests.post(self.url, headers=self.headers, json=payload, timeout=self.timeout)
|
|
231
|
+
if response.status_code != 200:
|
|
232
|
+
raise exceptions.FailedToGenerateResponseError(
|
|
233
|
+
f"Request failed with status code {response.status_code}"
|
|
234
|
+
)
|
|
235
|
+
|
|
236
|
+
response_data = response.json()
|
|
237
|
+
if 'choices' in response_data and len(response_data['choices']) > 0:
|
|
238
|
+
content = response_data['choices'][0].get('message', {}).get('content', '')
|
|
239
|
+
self.last_response = {"text": content}
|
|
240
|
+
self.conversation.update_chat_history(prompt, content)
|
|
241
|
+
return {"text": content}
|
|
242
|
+
else:
|
|
243
|
+
raise exceptions.FailedToGenerateResponseError("No response content found")
|
|
244
|
+
except Exception as e:
|
|
245
|
+
raise exceptions.FailedToGenerateResponseError(f"Request failed: {e}")
|
|
246
|
+
|
|
247
|
+
return for_stream() if stream else for_non_stream()
|
|
248
|
+
|
|
249
|
+
def chat(
|
|
250
|
+
self,
|
|
251
|
+
prompt: str,
|
|
252
|
+
stream: bool = False,
|
|
253
|
+
optimizer: str = None,
|
|
254
|
+
conversationally: bool = False,
|
|
255
|
+
) -> Union[str, Generator[str, None, None]]:
|
|
256
|
+
def for_stream():
|
|
257
|
+
for response in self.ask(prompt, True, optimizer=optimizer, conversationally=conversationally):
|
|
258
|
+
yield self.get_message(response)
|
|
259
|
+
def for_non_stream():
|
|
260
|
+
return self.get_message(
|
|
261
|
+
self.ask(prompt, False, optimizer=optimizer, conversationally=conversationally)
|
|
262
|
+
)
|
|
263
|
+
return for_stream() if stream else for_non_stream()
|
|
264
|
+
|
|
265
|
+
def get_message(self, response: dict) -> str:
|
|
266
|
+
assert isinstance(response, dict), "Response should be of dict data-type only"
|
|
267
|
+
return response["text"]
|
|
268
|
+
|
|
269
|
+
if __name__ == "__main__":
|
|
270
|
+
print("-" * 80)
|
|
271
|
+
print(f"{'Model':<50} {'Status':<10} {'Response'}")
|
|
272
|
+
print("-" * 80)
|
|
273
|
+
|
|
274
|
+
for model in Aitopia.AVAILABLE_MODELS:
|
|
275
|
+
try:
|
|
276
|
+
test_ai = Aitopia(model=model, timeout=60)
|
|
277
|
+
response = test_ai.chat("Say 'Hello' in one word", stream=True)
|
|
278
|
+
response_text = ""
|
|
279
|
+
for chunk in response:
|
|
280
|
+
response_text += chunk
|
|
281
|
+
|
|
282
|
+
if response_text and len(response_text.strip()) > 0:
|
|
283
|
+
status = "✓"
|
|
284
|
+
# Clean and truncate response
|
|
285
|
+
clean_text = response_text.strip().encode('utf-8', errors='ignore').decode('utf-8')
|
|
286
|
+
display_text = clean_text[:50] + "..." if len(clean_text) > 50 else clean_text
|
|
287
|
+
else:
|
|
288
|
+
status = "✗"
|
|
289
|
+
display_text = "Empty or invalid response"
|
|
290
|
+
print(f"\r{model:<50} {status:<10} {display_text}")
|
|
291
|
+
except Exception as e:
|
|
292
|
+
print(f"\r{model:<50} {'✗':<10} {str(e)}")
|
webscout/Provider/AllenAI.py
CHANGED
|
@@ -9,7 +9,7 @@ from webscout.AIutel import Conversation
|
|
|
9
9
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
10
10
|
from webscout.AIbase import Provider, AsyncProvider
|
|
11
11
|
from webscout import exceptions
|
|
12
|
-
from webscout import LitAgent
|
|
12
|
+
from webscout.litagent import LitAgent
|
|
13
13
|
|
|
14
14
|
class AllenAI(Provider):
|
|
15
15
|
"""
|
webscout/Provider/Andi.py
CHANGED
|
@@ -6,9 +6,9 @@ from webscout.AIutel import Conversation
|
|
|
6
6
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
7
7
|
from webscout.AIbase import Provider, AsyncProvider
|
|
8
8
|
from webscout import exceptions
|
|
9
|
-
from typing import Any, AsyncGenerator, Dict
|
|
9
|
+
from typing import Union, Any, AsyncGenerator, Dict
|
|
10
10
|
from webscout import WEBS
|
|
11
|
-
from
|
|
11
|
+
from webscout.litagent import LitAgent
|
|
12
12
|
|
|
13
13
|
class AndiSearch(Provider):
|
|
14
14
|
def __init__(
|
|
@@ -60,7 +60,7 @@ class AndiSearch(Provider):
|
|
|
60
60
|
"sec-fetch-dest": "empty",
|
|
61
61
|
"sec-fetch-mode": "cors",
|
|
62
62
|
"sec-fetch-site": "same-site",
|
|
63
|
-
"user-agent":
|
|
63
|
+
"user-agent": LitAgent().random(),
|
|
64
64
|
"x-amz-date": "20240730T031106Z",
|
|
65
65
|
"x-amz-security-token": str(uuid4()),
|
|
66
66
|
}
|
webscout/Provider/C4ai.py
CHANGED
|
@@ -9,7 +9,7 @@ from typing import Any, Dict, List, Optional, Union, Generator
|
|
|
9
9
|
from webscout.AIutel import Conversation
|
|
10
10
|
from webscout.AIbase import Provider
|
|
11
11
|
from webscout import exceptions
|
|
12
|
-
from webscout import LitAgent
|
|
12
|
+
from webscout.litagent import LitAgent
|
|
13
13
|
|
|
14
14
|
class C4ai(Provider):
|
|
15
15
|
"""
|
webscout/Provider/ChatGPTES.py
CHANGED
|
@@ -2,12 +2,11 @@ import requests
|
|
|
2
2
|
import re
|
|
3
3
|
import json
|
|
4
4
|
import os
|
|
5
|
-
from typing import List, Dict
|
|
5
|
+
from typing import Union, List, Dict
|
|
6
6
|
from webscout.AIutel import Optimizers, Conversation, AwesomePrompts
|
|
7
7
|
from webscout.AIbase import Provider
|
|
8
8
|
from webscout import exceptions
|
|
9
|
-
from
|
|
10
|
-
|
|
9
|
+
from webscout.litagent import LitAgent
|
|
11
10
|
class ChatGPTES(Provider):
|
|
12
11
|
"""
|
|
13
12
|
A class to interact with the ChatGPT.es API.
|
|
@@ -47,8 +46,7 @@ class ChatGPTES(Provider):
|
|
|
47
46
|
self.system_prompt = system_prompt
|
|
48
47
|
self.model = model
|
|
49
48
|
self.initial_headers = {
|
|
50
|
-
'User-Agent':
|
|
51
|
-
'AppleWebKit/537.36 (KHTML, like Gecko) Chrome/129.0.0.0 Safari/537.36',
|
|
49
|
+
'User-Agent': LitAgent().random(),
|
|
52
50
|
'Referer': 'https://www.google.com/',
|
|
53
51
|
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,'
|
|
54
52
|
'image/avif,image/webp,image/apng,*/*;q=0.8,'
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
from typing import Any, Dict, Generator, Optional
|
|
1
|
+
from typing import Union, Any, Dict, Generator, Optional
|
|
2
2
|
import requests
|
|
3
3
|
import json
|
|
4
4
|
|
|
5
5
|
from webscout.AIutel import Optimizers, Conversation, AwesomePrompts
|
|
6
6
|
from webscout.AIbase import Provider
|
|
7
7
|
from webscout import exceptions
|
|
8
|
-
from webscout import LitAgent as Lit
|
|
8
|
+
from webscout.litagent import LitAgent as Lit
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class ChatGPTGratis(Provider):
|
|
@@ -74,7 +74,7 @@ class ChatGPTGratis(Provider):
|
|
|
74
74
|
raw: bool = False,
|
|
75
75
|
optimizer: Optional[str] = None,
|
|
76
76
|
conversationally: bool = False,
|
|
77
|
-
) -> Dict[str, Any]
|
|
77
|
+
) -> Union[Dict[str, Any], Generator[Any, None, None]]:
|
|
78
78
|
"""
|
|
79
79
|
Sends a request to the API and returns the response.
|
|
80
80
|
If stream is True, yields response chunks as they are received.
|
|
@@ -144,7 +144,7 @@ class ChatGPTGratis(Provider):
|
|
|
144
144
|
stream: bool = False,
|
|
145
145
|
optimizer: Optional[str] = None,
|
|
146
146
|
conversationally: bool = False,
|
|
147
|
-
) -> str
|
|
147
|
+
) -> Union[str, Generator[str, None, None]]:
|
|
148
148
|
"""
|
|
149
149
|
Returns the response as a string.
|
|
150
150
|
For streaming requests, yields each response chunk as a string.
|
webscout/Provider/Chatify.py
CHANGED
|
@@ -4,7 +4,7 @@ from webscout.AIutel import AwesomePrompts
|
|
|
4
4
|
from webscout.AIbase import Provider
|
|
5
5
|
from webscout import exceptions
|
|
6
6
|
import requests
|
|
7
|
-
|
|
7
|
+
from webscout.litagent import LitAgent
|
|
8
8
|
class Chatify(Provider):
|
|
9
9
|
"""
|
|
10
10
|
A class to interact with the Chatify AI API.
|
|
@@ -46,7 +46,7 @@ class Chatify(Provider):
|
|
|
46
46
|
'Sec-Fetch-Dest': 'empty',
|
|
47
47
|
'Sec-Fetch-Mode': 'cors',
|
|
48
48
|
'Sec-Fetch-Site': 'same-origin',
|
|
49
|
-
'User-Agent':
|
|
49
|
+
'User-Agent': LitAgent().random(),
|
|
50
50
|
}
|
|
51
51
|
|
|
52
52
|
self.__available_optimizers = (
|
webscout/Provider/Cloudflare.py
CHANGED
|
@@ -6,9 +6,9 @@ from webscout.AIutel import Conversation
|
|
|
6
6
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
7
7
|
from webscout.AIbase import Provider, AsyncProvider
|
|
8
8
|
from webscout import exceptions
|
|
9
|
-
from typing import Any, AsyncGenerator, Dict
|
|
9
|
+
from typing import Union, Any, AsyncGenerator, Dict
|
|
10
10
|
import cloudscraper
|
|
11
|
-
from webscout import LitAgent
|
|
11
|
+
from webscout.litagent import LitAgent
|
|
12
12
|
|
|
13
13
|
class Cloudflare(Provider):
|
|
14
14
|
"""
|
|
@@ -37,6 +37,7 @@ class Cloudflare(Provider):
|
|
|
37
37
|
"@cf/meta/llama-3.2-1b-instruct",
|
|
38
38
|
"@cf/meta/llama-3.2-3b-instruct",
|
|
39
39
|
"@cf/meta/llama-3.3-70b-instruct-fp8-fast",
|
|
40
|
+
"@cf/meta/llama-guard-3-8b",
|
|
40
41
|
"@hf/thebloke/llamaguard-7b-awq",
|
|
41
42
|
"@hf/meta-llama/meta-llama-3-8b-instruct",
|
|
42
43
|
"@cf/mistral/mistral-7b-instruct-v0.1",
|
webscout/Provider/DARKAI.py
CHANGED
|
@@ -1,10 +1,11 @@
|
|
|
1
1
|
import json
|
|
2
|
-
from typing import Any, Dict, Optional
|
|
2
|
+
from typing import Union, Any, Dict, Optional
|
|
3
3
|
from webscout.AIutel import Optimizers
|
|
4
4
|
from webscout.AIutel import Conversation
|
|
5
5
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
6
6
|
from webscout.AIbase import Provider
|
|
7
|
-
from webscout import exceptions
|
|
7
|
+
from webscout import exceptions
|
|
8
|
+
from webscout.litagent import LitAgent
|
|
8
9
|
import requests
|
|
9
10
|
|
|
10
11
|
class DARKAI(Provider):
|
webscout/Provider/DeepSeek.py
CHANGED
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import requests
|
|
2
2
|
import json
|
|
3
|
-
from typing import Any, Dict, Generator
|
|
3
|
+
from typing import Union, Any, Dict, Generator
|
|
4
4
|
from webscout.AIutel import Optimizers
|
|
5
5
|
from webscout.AIutel import Conversation
|
|
6
6
|
from webscout.AIutel import AwesomePrompts
|
|
7
7
|
from webscout.AIbase import Provider
|
|
8
8
|
from webscout import exceptions
|
|
9
|
-
from webscout import LitAgent as Lit
|
|
9
|
+
from webscout.litagent import LitAgent as Lit
|
|
10
10
|
|
|
11
11
|
class DeepSeek(Provider):
|
|
12
12
|
"""
|
webscout/Provider/Deepinfra.py
CHANGED
|
@@ -8,7 +8,7 @@ from webscout.AIutel import Conversation
|
|
|
8
8
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
9
9
|
from webscout.AIbase import Provider, AsyncProvider
|
|
10
10
|
from webscout import exceptions
|
|
11
|
-
from webscout import LitAgent
|
|
11
|
+
from webscout.litagent import LitAgent
|
|
12
12
|
|
|
13
13
|
class DeepInfra(Provider):
|
|
14
14
|
"""
|
webscout/Provider/EDITEE.py
CHANGED
|
@@ -4,7 +4,7 @@ from webscout.AIutel import Conversation
|
|
|
4
4
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
5
5
|
from webscout.AIbase import Provider, AsyncProvider
|
|
6
6
|
from webscout import exceptions
|
|
7
|
-
from typing import Any, AsyncGenerator, Dict
|
|
7
|
+
from typing import Union, Any, AsyncGenerator, Dict
|
|
8
8
|
|
|
9
9
|
|
|
10
10
|
class Editee(Provider):
|