lm-deluge 0.0.4__tar.gz → 0.0.5__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of lm-deluge might be problematic. Click here for more details.
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/PKG-INFO +1 -1
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/pyproject.toml +1 -1
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/anthropic.py +0 -2
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/cohere.py +2 -6
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge.egg-info/PKG-INFO +1 -1
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/README.md +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/setup.cfg +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/__init__.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/__init__.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/base.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/common.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/deprecated/bedrock.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/deprecated/deepseek.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/deprecated/mistral.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/google.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/openai.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/api_requests/vertex.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/cache.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/client.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/embed.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/errors.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/gemini_limits.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/image.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/llm_tools/__init__.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/llm_tools/extract.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/llm_tools/score.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/llm_tools/translate.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/models.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/prompt.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/rerank.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/sampling_params.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/tool.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/tracker.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/util/json.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/util/logprobs.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/util/pdf.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/util/validation.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge/util/xml.py +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge.egg-info/SOURCES.txt +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge.egg-info/dependency_links.txt +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge.egg-info/requires.txt +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/src/lm_deluge.egg-info/top_level.txt +0 -0
- {lm_deluge-0.0.4 → lm_deluge-0.0.5}/tests/test_heal_json.py +0 -0
|
@@ -119,9 +119,7 @@ class AnthropicRequest(APIRequestBase):
|
|
|
119
119
|
if status_code >= 200 and status_code < 300:
|
|
120
120
|
try:
|
|
121
121
|
data = await http_response.json()
|
|
122
|
-
print("response data:", data)
|
|
123
122
|
content = data["content"] # [0]["text"]
|
|
124
|
-
print("content is length", len(content))
|
|
125
123
|
for item in content:
|
|
126
124
|
if item["type"] == "text":
|
|
127
125
|
completion = item["text"]
|
|
@@ -55,7 +55,7 @@ class CohereRequest(APIRequestBase):
|
|
|
55
55
|
|
|
56
56
|
self.model = APIModel.from_registry(model_name)
|
|
57
57
|
self.url = f"{self.model.api_base}/chat"
|
|
58
|
-
|
|
58
|
+
messages = prompt.to_cohere()
|
|
59
59
|
|
|
60
60
|
self.request_header = {
|
|
61
61
|
"Authorization": f"bearer {os.getenv(self.model.api_key_env_var)}",
|
|
@@ -65,16 +65,12 @@ class CohereRequest(APIRequestBase):
|
|
|
65
65
|
|
|
66
66
|
self.request_json = {
|
|
67
67
|
"model": self.model.name,
|
|
68
|
-
"
|
|
69
|
-
"message": last_user_message,
|
|
68
|
+
"messages": messages,
|
|
70
69
|
"temperature": sampling_params.temperature,
|
|
71
70
|
"top_p": sampling_params.top_p,
|
|
72
71
|
"max_tokens": sampling_params.max_new_tokens,
|
|
73
72
|
}
|
|
74
73
|
|
|
75
|
-
if self.system_message:
|
|
76
|
-
self.request_json["preamble"] = self.system_message
|
|
77
|
-
|
|
78
74
|
async def handle_response(self, http_response: ClientResponse) -> APIResponse:
|
|
79
75
|
is_error = False
|
|
80
76
|
error_message = None
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|