vectorvein 0.1.18__tar.gz → 0.1.19__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (27) hide show
  1. {vectorvein-0.1.18 → vectorvein-0.1.19}/PKG-INFO +1 -1
  2. {vectorvein-0.1.18 → vectorvein-0.1.19}/pyproject.toml +1 -1
  3. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -38
  4. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/types/defaults.py +28 -0
  5. {vectorvein-0.1.18 → vectorvein-0.1.19}/README.md +0 -0
  6. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/__init__.py +0 -0
  7. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/__init__.py +0 -0
  8. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
  9. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
  10. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/base_client.py +0 -0
  11. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
  12. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/gemini_client.py +0 -0
  13. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/groq_client.py +0 -0
  14. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/local_client.py +0 -0
  15. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/minimax_client.py +0 -0
  16. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/mistral_client.py +0 -0
  17. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
  18. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/openai_client.py +0 -0
  19. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/qwen_client.py +0 -0
  20. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/utils.py +0 -0
  21. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/yi_client.py +0 -0
  22. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
  23. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/settings/__init__.py +0 -0
  24. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/types/enums.py +0 -0
  25. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/types/llm_parameters.py +0 -0
  26. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/utilities/media_processing.py +0 -0
  27. {vectorvein-0.1.18 → vectorvein-0.1.19}/src/vectorvein/utilities/retry.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vectorvein
3
- Version: 0.1.18
3
+ Version: 0.1.19
4
4
  Summary: Default template for PDM package
5
5
  Author-Email: Anderson <andersonby@163.com>
6
6
  License: MIT
@@ -16,7 +16,7 @@ description = "Default template for PDM package"
16
16
  name = "vectorvein"
17
17
  readme = "README.md"
18
18
  requires-python = ">=3.10"
19
- version = "0.1.18"
19
+ version = "0.1.19"
20
20
 
21
21
  [project.license]
22
22
  text = "MIT"
@@ -90,25 +90,6 @@ class OpenAICompatibleChatClient(BaseChatClient):
90
90
 
91
91
  self.model_setting = self.backend_settings.models[self.model]
92
92
 
93
- # if self.random_endpoint:
94
- # self.random_endpoint = True
95
- # self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
96
- # self.endpoint = settings.get_endpoint(self.endpoint_id)
97
-
98
- # if self.endpoint.is_azure:
99
- # self._client = AzureOpenAI(
100
- # azure_endpoint=self.endpoint.api_base,
101
- # api_key=self.endpoint.api_key,
102
- # api_version="2024-08-01-preview",
103
- # http_client=self.http_client,
104
- # )
105
- # else:
106
- # self._client = OpenAI(
107
- # api_key=self.endpoint.api_key,
108
- # base_url=self.endpoint.api_base,
109
- # http_client=self.http_client,
110
- # )
111
-
112
93
  if self.context_length_control == ContextLengthControlType.Latest:
113
94
  messages = cutoff_messages(
114
95
  messages,
@@ -269,25 +250,6 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
269
250
 
270
251
  self.model_setting = self.backend_settings.models[self.model]
271
252
 
272
- # if self.random_endpoint:
273
- # self.random_endpoint = True
274
- # self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
275
- # self.endpoint = settings.get_endpoint(self.endpoint_id)
276
-
277
- # if self.endpoint.is_azure:
278
- # self._client = AsyncAzureOpenAI(
279
- # azure_endpoint=self.endpoint.api_base,
280
- # api_key=self.endpoint.api_key,
281
- # api_version="2024-08-01-preview",
282
- # http_client=self.http_client,
283
- # )
284
- # else:
285
- # self._client = AsyncOpenAI(
286
- # api_key=self.endpoint.api_key,
287
- # base_url=self.endpoint.api_base,
288
- # http_client=self.http_client,
289
- # )
290
-
291
253
  if self.context_length_control == ContextLengthControlType.Latest:
292
254
  messages = cutoff_messages(
293
255
  messages,
@@ -119,6 +119,34 @@ GROQ_MODELS = {
119
119
  "function_call_available": True,
120
120
  "response_format_available": True,
121
121
  },
122
+ "gemma2-9b-it": {
123
+ "id": "gemma2-9b-it",
124
+ "context_length": 8192,
125
+ },
126
+ "llama3-groq-70b-8192-tool-use-preview": {
127
+ "id": "llama3-groq-70b-8192-tool-use-preview",
128
+ "context_length": 8192,
129
+ "function_call_available": True,
130
+ "max_output_tokens": 8000,
131
+ },
132
+ "llama3-groq-8b-8192-tool-use-preview": {
133
+ "id": "llama3-groq-8b-8192-tool-use-preview",
134
+ "context_length": 8192,
135
+ "function_call_available": True,
136
+ "max_output_tokens": 8000,
137
+ },
138
+ "llama-3.1-70b-versatile": {
139
+ "id": "llama-3.1-70b-versatile",
140
+ "context_length": 131072,
141
+ "function_call_available": True,
142
+ "max_output_tokens": 8000,
143
+ },
144
+ "llama-3.1-8b-instant": {
145
+ "id": "llama-3.1-8b-instant",
146
+ "context_length": 131072,
147
+ "function_call_available": True,
148
+ "max_output_tokens": 8000,
149
+ },
122
150
  }
123
151
 
124
152
  # Qwen models
File without changes