aient 1.2.28__tar.gz → 1.2.30__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {aient-1.2.28 → aient-1.2.30}/PKG-INFO +1 -1
- {aient-1.2.28 → aient-1.2.30}/aient/core/request.py +4 -2
- {aient-1.2.28 → aient-1.2.30}/aient/core/utils.py +6 -1
- {aient-1.2.28 → aient-1.2.30}/aient/models/audio.py +2 -2
- {aient-1.2.28 → aient-1.2.30}/aient/models/base.py +2 -2
- {aient-1.2.28 → aient-1.2.30}/aient/models/chatgpt.py +2 -2
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/image.py +2 -2
- {aient-1.2.28 → aient-1.2.30}/aient/utils/prompt.py +0 -4
- {aient-1.2.28 → aient-1.2.30}/aient/utils/scripts.py +0 -8
- {aient-1.2.28 → aient-1.2.30}/aient.egg-info/PKG-INFO +1 -1
- {aient-1.2.28 → aient-1.2.30}/aient.egg-info/SOURCES.txt +1 -3
- {aient-1.2.28 → aient-1.2.30}/pyproject.toml +1 -1
- {aient-1.2.28 → aient-1.2.30}/test/test_whisper.py +1 -1
- aient-1.2.28/test/test_search.py +0 -18
- aient-1.2.28/test/test_yjh.py +0 -21
- {aient-1.2.28 → aient-1.2.30}/LICENSE +0 -0
- {aient-1.2.28 → aient-1.2.30}/README.md +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/architext/architext/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/architext/architext/core.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/architext/test/openai_client.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/architext/test/test.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/architext/test/test_save_load.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/log_config.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/models.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/response.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/test/test_base_api.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/test/test_geminimask.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/test/test_image.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/core/test/test_payload.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/models/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/arXiv.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/config.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/excute_command.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/get_time.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/list_directory.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/read_file.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/read_image.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/readonly.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/registry.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/run_python.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/websearch.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/plugins/write_file.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient/utils/__init__.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient.egg-info/dependency_links.txt +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient.egg-info/requires.txt +0 -0
- {aient-1.2.28 → aient-1.2.30}/aient.egg-info/top_level.txt +0 -0
- {aient-1.2.28 → aient-1.2.30}/setup.cfg +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_Web_crawler.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_ddg_search.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_google_search.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_ollama.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_plugin.py +0 -0
- {aient-1.2.28 → aient-1.2.30}/test/test_url.py +0 -0
@@ -1,5 +1,6 @@
|
|
1
1
|
import re
|
2
2
|
import json
|
3
|
+
import copy
|
3
4
|
import httpx
|
4
5
|
import base64
|
5
6
|
import asyncio
|
@@ -57,7 +58,7 @@ async def get_gemini_payload(request, engine, provider, api_key=None):
|
|
57
58
|
try:
|
58
59
|
request_messages = [Message(role="user", content=request.prompt)]
|
59
60
|
except:
|
60
|
-
request_messages = request.messages
|
61
|
+
request_messages = copy.deepcopy(request.messages)
|
61
62
|
for msg in request_messages:
|
62
63
|
if msg.role == "assistant":
|
63
64
|
msg.role = "model"
|
@@ -399,7 +400,8 @@ async def get_vertex_gemini_payload(request, engine, provider, api_key=None):
|
|
399
400
|
systemInstruction = None
|
400
401
|
system_prompt = ""
|
401
402
|
function_arguments = None
|
402
|
-
|
403
|
+
request_messages = copy.deepcopy(request.messages)
|
404
|
+
for msg in request_messages:
|
403
405
|
if msg.role == "assistant":
|
404
406
|
msg.role = "model"
|
405
407
|
tool_calls = None
|
@@ -228,7 +228,12 @@ async def update_initial_model(provider):
|
|
228
228
|
def safe_get(data, *keys, default=None):
|
229
229
|
for key in keys:
|
230
230
|
try:
|
231
|
-
|
231
|
+
if isinstance(data, (dict, list)):
|
232
|
+
data = data[key]
|
233
|
+
elif isinstance(key, str) and hasattr(data, key):
|
234
|
+
data = getattr(data, key)
|
235
|
+
else:
|
236
|
+
data = data.get(key)
|
232
237
|
except (KeyError, IndexError, AttributeError, TypeError):
|
233
238
|
return default
|
234
239
|
if not data:
|
@@ -4,13 +4,13 @@ import json
|
|
4
4
|
from .base import BaseLLM
|
5
5
|
|
6
6
|
API = os.environ.get('API', None)
|
7
|
-
|
7
|
+
BASE_URL = os.environ.get('BASE_URL', None)
|
8
8
|
|
9
9
|
class whisper(BaseLLM):
|
10
10
|
def __init__(
|
11
11
|
self,
|
12
12
|
api_key: str,
|
13
|
-
api_url: str = (os.environ.get("
|
13
|
+
api_url: str = (os.environ.get("BASE_URL") or "https://api.openai.com/v1/audio/transcriptions"),
|
14
14
|
timeout: float = 20,
|
15
15
|
):
|
16
16
|
super().__init__(api_key, api_url=api_url, timeout=timeout)
|
@@ -11,8 +11,8 @@ class BaseLLM:
|
|
11
11
|
def __init__(
|
12
12
|
self,
|
13
13
|
api_key: str = None,
|
14
|
-
engine: str = os.environ.get("
|
15
|
-
api_url: str = (os.environ.get("
|
14
|
+
engine: str = os.environ.get("MODEL") or "gpt-3.5-turbo",
|
15
|
+
api_url: str = (os.environ.get("BASE_URL", None) or "https://api.openai.com/v1/chat/completions"),
|
16
16
|
system_prompt: str = prompt.chatgpt_system_prompt,
|
17
17
|
proxy: str = None,
|
18
18
|
timeout: float = 600,
|
@@ -74,8 +74,8 @@ class chatgpt(BaseLLM):
|
|
74
74
|
def __init__(
|
75
75
|
self,
|
76
76
|
api_key: str = None,
|
77
|
-
engine: str = os.environ.get("
|
78
|
-
api_url: str = (os.environ.get("
|
77
|
+
engine: str = os.environ.get("MODEL") or "gpt-4o",
|
78
|
+
api_url: str = (os.environ.get("BASE_URL") or "https://api.openai.com/v1/chat/completions"),
|
79
79
|
system_prompt: str = "You are ChatGPT, a large language model trained by OpenAI. Respond conversationally",
|
80
80
|
proxy: str = None,
|
81
81
|
timeout: float = 600,
|
@@ -5,13 +5,13 @@ from ..models.base import BaseLLM
|
|
5
5
|
from .registry import register_tool
|
6
6
|
|
7
7
|
API = os.environ.get('API', None)
|
8
|
-
|
8
|
+
BASE_URL = os.environ.get('BASE_URL', None)
|
9
9
|
|
10
10
|
class dalle3(BaseLLM):
|
11
11
|
def __init__(
|
12
12
|
self,
|
13
13
|
api_key: str,
|
14
|
-
api_url: str = (os.environ.get("
|
14
|
+
api_url: str = (os.environ.get("BASE_URL") or "https://api.openai.com/v1/images/generations"),
|
15
15
|
timeout: float = 20,
|
16
16
|
):
|
17
17
|
super().__init__(api_key, api_url=api_url, timeout=timeout)
|
@@ -90,10 +90,6 @@ chatgpt_system_prompt = (
|
|
90
90
|
"You are ChatGPT, a large language model trained by OpenAI. Use simple characters to represent mathematical symbols. Do not use LaTeX commands. Respond conversationally"
|
91
91
|
)
|
92
92
|
|
93
|
-
claude_system_prompt = (
|
94
|
-
"You are Claude, a large language model trained by Anthropic. Use simple characters to represent mathematical symbols. Do not use LaTeX commands. Respond conversationally in {}."
|
95
|
-
)
|
96
|
-
|
97
93
|
search_system_prompt = (
|
98
94
|
"You are ChatGPT, a large language model trained by OpenAI. Respond conversationally in {}."
|
99
95
|
"You can break down the task into multiple steps and search the web to answer my questions one by one."
|
@@ -134,14 +134,6 @@ def is_surrounded_by_chinese(text, index):
|
|
134
134
|
def replace_char(string, index, new_char):
|
135
135
|
return string[:index] + new_char + string[index+1:]
|
136
136
|
|
137
|
-
def claude_replace(text):
|
138
|
-
Punctuation_mapping = {",": ",", ":": ":", "!": "!", "?": "?", ";": ";"}
|
139
|
-
key_list = list(Punctuation_mapping.keys())
|
140
|
-
for i in range(len(text)):
|
141
|
-
if is_surrounded_by_chinese(text, i) and (text[i] in key_list):
|
142
|
-
text = replace_char(text, i, Punctuation_mapping[text[i]])
|
143
|
-
return text
|
144
|
-
|
145
137
|
def safe_get(data, *keys, default=None):
|
146
138
|
for key in keys:
|
147
139
|
try:
|
aient-1.2.28/test/test_search.py
DELETED
@@ -1,18 +0,0 @@
|
|
1
|
-
import os
|
2
|
-
from aient.models import chatgpt
|
3
|
-
|
4
|
-
API = os.environ.get('API', None)
|
5
|
-
API_URL = os.environ.get('API_URL', None)
|
6
|
-
GPT_ENGINE = os.environ.get('GPT_ENGINE', 'gpt-4o')
|
7
|
-
|
8
|
-
systemprompt = (
|
9
|
-
"You are ChatGPT, a large language model trained by OpenAI. Respond conversationally"
|
10
|
-
)
|
11
|
-
bot = chatgpt(api_key=API, api_url=API_URL, engine=GPT_ENGINE, system_prompt=systemprompt, print_log=True)
|
12
|
-
for text in bot.ask_stream("搜索上海的天气"):
|
13
|
-
# for text in bot.ask_stream("我在广州市,想周一去香港,周四早上回来,是去游玩,请你帮我规划整个行程。包括细节,如交通,住宿,餐饮,价格,等等,最好细节到每天各个部分的时间,花费,等等,尽量具体,用户一看就能直接执行的那种"):
|
14
|
-
# for text in bot.ask_stream("上海有哪些好玩的地方?"):
|
15
|
-
# for text in bot.ask_stream("just say test"):
|
16
|
-
# for text in bot.ask_stream("我在上海想去重庆旅游,我只有2000元预算,我想在重庆玩一周,你能帮我规划一下吗?"):
|
17
|
-
# for text in bot.ask_stream("我在上海想去重庆旅游,我有一天的时间。你能帮我规划一下吗?"):
|
18
|
-
print(text, end="")
|
aient-1.2.28/test/test_yjh.py
DELETED
@@ -1,21 +0,0 @@
|
|
1
|
-
import os
|
2
|
-
from datetime import datetime
|
3
|
-
|
4
|
-
from aient.models import chatgpt
|
5
|
-
from aient.utils import prompt
|
6
|
-
|
7
|
-
API = os.environ.get('API', None)
|
8
|
-
API_URL = os.environ.get('API_URL', None)
|
9
|
-
GPT_ENGINE = os.environ.get('GPT_ENGINE', 'gpt-4o')
|
10
|
-
LANGUAGE = os.environ.get('LANGUAGE', 'Simplified Chinese')
|
11
|
-
|
12
|
-
current_date = datetime.now()
|
13
|
-
Current_Date = current_date.strftime("%Y-%m-%d")
|
14
|
-
|
15
|
-
systemprompt = os.environ.get('SYSTEMPROMPT', prompt.system_prompt.format(LANGUAGE, Current_Date))
|
16
|
-
|
17
|
-
bot = chatgpt(api_key=API, api_url=API_URL, engine=GPT_ENGINE, system_prompt=systemprompt)
|
18
|
-
for text in bot.ask_stream("arXiv:2210.10716 这篇文章讲了啥"):
|
19
|
-
# for text in bot.ask_stream("今天的微博热搜有哪些?"):
|
20
|
-
# for text in bot.ask_stream("你现在是什么版本?"):
|
21
|
-
print(text, end="")
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|