vectorvein 0.1.18__tar.gz → 0.1.20__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {vectorvein-0.1.18 → vectorvein-0.1.20}/PKG-INFO +1 -1
- {vectorvein-0.1.18 → vectorvein-0.1.20}/pyproject.toml +1 -1
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/gemini_client.py +8 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/minimax_client.py +8 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -38
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/types/defaults.py +28 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/README.md +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/__init__.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/__init__.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/base_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/groq_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/local_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/mistral_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/openai_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/qwen_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/utils.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/yi_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/settings/__init__.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/types/enums.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/types/llm_parameters.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/utilities/media_processing.py +0 -0
- {vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/utilities/retry.py +0 -0
@@ -39,6 +39,10 @@ class GeminiChatClient(BaseChatClient):
|
|
39
39
|
**kwargs,
|
40
40
|
)
|
41
41
|
|
42
|
+
@property
|
43
|
+
def raw_client(self):
|
44
|
+
return self.http_client
|
45
|
+
|
42
46
|
def create_completion(
|
43
47
|
self,
|
44
48
|
messages: list = list,
|
@@ -209,6 +213,10 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
|
|
209
213
|
**kwargs,
|
210
214
|
)
|
211
215
|
|
216
|
+
@property
|
217
|
+
def raw_client(self):
|
218
|
+
return self.http_client
|
219
|
+
|
212
220
|
async def create_completion(
|
213
221
|
self,
|
214
222
|
messages: list = list,
|
@@ -66,6 +66,10 @@ class MiniMaxChatClient(BaseChatClient):
|
|
66
66
|
else:
|
67
67
|
self.http_client = httpx.Client()
|
68
68
|
|
69
|
+
@property
|
70
|
+
def raw_client(self):
|
71
|
+
return self.http_client
|
72
|
+
|
69
73
|
def create_completion(
|
70
74
|
self,
|
71
75
|
messages: list = list,
|
@@ -231,6 +235,10 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
|
|
231
235
|
else:
|
232
236
|
self.http_client = httpx.AsyncClient()
|
233
237
|
|
238
|
+
@property
|
239
|
+
def raw_client(self):
|
240
|
+
return self.http_client
|
241
|
+
|
234
242
|
async def create_completion(
|
235
243
|
self,
|
236
244
|
messages: list = list,
|
{vectorvein-0.1.18 → vectorvein-0.1.20}/src/vectorvein/chat_clients/openai_compatible_client.py
RENAMED
@@ -90,25 +90,6 @@ class OpenAICompatibleChatClient(BaseChatClient):
|
|
90
90
|
|
91
91
|
self.model_setting = self.backend_settings.models[self.model]
|
92
92
|
|
93
|
-
# if self.random_endpoint:
|
94
|
-
# self.random_endpoint = True
|
95
|
-
# self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
|
96
|
-
# self.endpoint = settings.get_endpoint(self.endpoint_id)
|
97
|
-
|
98
|
-
# if self.endpoint.is_azure:
|
99
|
-
# self._client = AzureOpenAI(
|
100
|
-
# azure_endpoint=self.endpoint.api_base,
|
101
|
-
# api_key=self.endpoint.api_key,
|
102
|
-
# api_version="2024-08-01-preview",
|
103
|
-
# http_client=self.http_client,
|
104
|
-
# )
|
105
|
-
# else:
|
106
|
-
# self._client = OpenAI(
|
107
|
-
# api_key=self.endpoint.api_key,
|
108
|
-
# base_url=self.endpoint.api_base,
|
109
|
-
# http_client=self.http_client,
|
110
|
-
# )
|
111
|
-
|
112
93
|
if self.context_length_control == ContextLengthControlType.Latest:
|
113
94
|
messages = cutoff_messages(
|
114
95
|
messages,
|
@@ -269,25 +250,6 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
|
|
269
250
|
|
270
251
|
self.model_setting = self.backend_settings.models[self.model]
|
271
252
|
|
272
|
-
# if self.random_endpoint:
|
273
|
-
# self.random_endpoint = True
|
274
|
-
# self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
|
275
|
-
# self.endpoint = settings.get_endpoint(self.endpoint_id)
|
276
|
-
|
277
|
-
# if self.endpoint.is_azure:
|
278
|
-
# self._client = AsyncAzureOpenAI(
|
279
|
-
# azure_endpoint=self.endpoint.api_base,
|
280
|
-
# api_key=self.endpoint.api_key,
|
281
|
-
# api_version="2024-08-01-preview",
|
282
|
-
# http_client=self.http_client,
|
283
|
-
# )
|
284
|
-
# else:
|
285
|
-
# self._client = AsyncOpenAI(
|
286
|
-
# api_key=self.endpoint.api_key,
|
287
|
-
# base_url=self.endpoint.api_base,
|
288
|
-
# http_client=self.http_client,
|
289
|
-
# )
|
290
|
-
|
291
253
|
if self.context_length_control == ContextLengthControlType.Latest:
|
292
254
|
messages = cutoff_messages(
|
293
255
|
messages,
|
@@ -119,6 +119,34 @@ GROQ_MODELS = {
|
|
119
119
|
"function_call_available": True,
|
120
120
|
"response_format_available": True,
|
121
121
|
},
|
122
|
+
"gemma2-9b-it": {
|
123
|
+
"id": "gemma2-9b-it",
|
124
|
+
"context_length": 8192,
|
125
|
+
},
|
126
|
+
"llama3-groq-70b-8192-tool-use-preview": {
|
127
|
+
"id": "llama3-groq-70b-8192-tool-use-preview",
|
128
|
+
"context_length": 8192,
|
129
|
+
"function_call_available": True,
|
130
|
+
"max_output_tokens": 8000,
|
131
|
+
},
|
132
|
+
"llama3-groq-8b-8192-tool-use-preview": {
|
133
|
+
"id": "llama3-groq-8b-8192-tool-use-preview",
|
134
|
+
"context_length": 8192,
|
135
|
+
"function_call_available": True,
|
136
|
+
"max_output_tokens": 8000,
|
137
|
+
},
|
138
|
+
"llama-3.1-70b-versatile": {
|
139
|
+
"id": "llama-3.1-70b-versatile",
|
140
|
+
"context_length": 131072,
|
141
|
+
"function_call_available": True,
|
142
|
+
"max_output_tokens": 8000,
|
143
|
+
},
|
144
|
+
"llama-3.1-8b-instant": {
|
145
|
+
"id": "llama-3.1-8b-instant",
|
146
|
+
"context_length": 131072,
|
147
|
+
"function_call_available": True,
|
148
|
+
"max_output_tokens": 8000,
|
149
|
+
},
|
122
150
|
}
|
123
151
|
|
124
152
|
# Qwen models
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|