vectorvein 0.1.17__tar.gz → 0.1.19__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {vectorvein-0.1.17 → vectorvein-0.1.19}/PKG-INFO +1 -1
- {vectorvein-0.1.17 → vectorvein-0.1.19}/pyproject.toml +1 -1
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -38
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/types/defaults.py +32 -4
- {vectorvein-0.1.17 → vectorvein-0.1.19}/README.md +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/__init__.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/__init__.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/base_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/gemini_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/groq_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/local_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/minimax_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/mistral_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/openai_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/qwen_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/utils.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/yi_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/settings/__init__.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/types/enums.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/types/llm_parameters.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/utilities/media_processing.py +0 -0
- {vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/utilities/retry.py +0 -0
{vectorvein-0.1.17 → vectorvein-0.1.19}/src/vectorvein/chat_clients/openai_compatible_client.py
RENAMED
@@ -90,25 +90,6 @@ class OpenAICompatibleChatClient(BaseChatClient):
|
|
90
90
|
|
91
91
|
self.model_setting = self.backend_settings.models[self.model]
|
92
92
|
|
93
|
-
# if self.random_endpoint:
|
94
|
-
# self.random_endpoint = True
|
95
|
-
# self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
|
96
|
-
# self.endpoint = settings.get_endpoint(self.endpoint_id)
|
97
|
-
|
98
|
-
# if self.endpoint.is_azure:
|
99
|
-
# self._client = AzureOpenAI(
|
100
|
-
# azure_endpoint=self.endpoint.api_base,
|
101
|
-
# api_key=self.endpoint.api_key,
|
102
|
-
# api_version="2024-08-01-preview",
|
103
|
-
# http_client=self.http_client,
|
104
|
-
# )
|
105
|
-
# else:
|
106
|
-
# self._client = OpenAI(
|
107
|
-
# api_key=self.endpoint.api_key,
|
108
|
-
# base_url=self.endpoint.api_base,
|
109
|
-
# http_client=self.http_client,
|
110
|
-
# )
|
111
|
-
|
112
93
|
if self.context_length_control == ContextLengthControlType.Latest:
|
113
94
|
messages = cutoff_messages(
|
114
95
|
messages,
|
@@ -269,25 +250,6 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
|
|
269
250
|
|
270
251
|
self.model_setting = self.backend_settings.models[self.model]
|
271
252
|
|
272
|
-
# if self.random_endpoint:
|
273
|
-
# self.random_endpoint = True
|
274
|
-
# self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
|
275
|
-
# self.endpoint = settings.get_endpoint(self.endpoint_id)
|
276
|
-
|
277
|
-
# if self.endpoint.is_azure:
|
278
|
-
# self._client = AsyncAzureOpenAI(
|
279
|
-
# azure_endpoint=self.endpoint.api_base,
|
280
|
-
# api_key=self.endpoint.api_key,
|
281
|
-
# api_version="2024-08-01-preview",
|
282
|
-
# http_client=self.http_client,
|
283
|
-
# )
|
284
|
-
# else:
|
285
|
-
# self._client = AsyncOpenAI(
|
286
|
-
# api_key=self.endpoint.api_key,
|
287
|
-
# base_url=self.endpoint.api_base,
|
288
|
-
# http_client=self.http_client,
|
289
|
-
# )
|
290
|
-
|
291
253
|
if self.context_length_control == ContextLengthControlType.Latest:
|
292
254
|
messages = cutoff_messages(
|
293
255
|
messages,
|
@@ -119,6 +119,34 @@ GROQ_MODELS = {
|
|
119
119
|
"function_call_available": True,
|
120
120
|
"response_format_available": True,
|
121
121
|
},
|
122
|
+
"gemma2-9b-it": {
|
123
|
+
"id": "gemma2-9b-it",
|
124
|
+
"context_length": 8192,
|
125
|
+
},
|
126
|
+
"llama3-groq-70b-8192-tool-use-preview": {
|
127
|
+
"id": "llama3-groq-70b-8192-tool-use-preview",
|
128
|
+
"context_length": 8192,
|
129
|
+
"function_call_available": True,
|
130
|
+
"max_output_tokens": 8000,
|
131
|
+
},
|
132
|
+
"llama3-groq-8b-8192-tool-use-preview": {
|
133
|
+
"id": "llama3-groq-8b-8192-tool-use-preview",
|
134
|
+
"context_length": 8192,
|
135
|
+
"function_call_available": True,
|
136
|
+
"max_output_tokens": 8000,
|
137
|
+
},
|
138
|
+
"llama-3.1-70b-versatile": {
|
139
|
+
"id": "llama-3.1-70b-versatile",
|
140
|
+
"context_length": 131072,
|
141
|
+
"function_call_available": True,
|
142
|
+
"max_output_tokens": 8000,
|
143
|
+
},
|
144
|
+
"llama-3.1-8b-instant": {
|
145
|
+
"id": "llama-3.1-8b-instant",
|
146
|
+
"context_length": 131072,
|
147
|
+
"function_call_available": True,
|
148
|
+
"max_output_tokens": 8000,
|
149
|
+
},
|
122
150
|
}
|
123
151
|
|
124
152
|
# Qwen models
|
@@ -424,7 +452,7 @@ ANTHROPIC_MODELS = {
|
|
424
452
|
"context_length": 200000,
|
425
453
|
"max_output_tokens": 4096,
|
426
454
|
"function_call_available": True,
|
427
|
-
"response_format_available":
|
455
|
+
"response_format_available": False,
|
428
456
|
"native_multimodal": True,
|
429
457
|
},
|
430
458
|
"claude-3-sonnet-20240229": {
|
@@ -433,14 +461,14 @@ ANTHROPIC_MODELS = {
|
|
433
461
|
"max_output_tokens": 4096,
|
434
462
|
"function_call_available": True,
|
435
463
|
"native_multimodal": True,
|
436
|
-
"response_format_available":
|
464
|
+
"response_format_available": False,
|
437
465
|
},
|
438
466
|
"claude-3-haiku-20240307": {
|
439
467
|
"id": "claude-3-haiku-20240307",
|
440
468
|
"context_length": 200000,
|
441
469
|
"max_output_tokens": 4096,
|
442
470
|
"function_call_available": True,
|
443
|
-
"response_format_available":
|
471
|
+
"response_format_available": False,
|
444
472
|
"native_multimodal": True,
|
445
473
|
},
|
446
474
|
"claude-3-5-sonnet-20240620": {
|
@@ -448,7 +476,7 @@ ANTHROPIC_MODELS = {
|
|
448
476
|
"context_length": 200000,
|
449
477
|
"max_output_tokens": 4096,
|
450
478
|
"function_call_available": True,
|
451
|
-
"response_format_available":
|
479
|
+
"response_format_available": False,
|
452
480
|
"native_multimodal": True,
|
453
481
|
},
|
454
482
|
}
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|