vectorvein 0.1.8__tar.gz → 0.1.9__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (34) hide show
  1. {vectorvein-0.1.8 → vectorvein-0.1.9}/PKG-INFO +1 -1
  2. {vectorvein-0.1.8 → vectorvein-0.1.9}/pyproject.toml +1 -1
  3. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/minimax_client.py +20 -30
  4. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/sample_settings.py +0 -1
  5. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/test_create_chat_client.py +6 -9
  6. {vectorvein-0.1.8 → vectorvein-0.1.9}/README.md +0 -0
  7. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/__init__.py +0 -0
  8. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/__init__.py +0 -0
  9. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
  10. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/base_client.py +0 -0
  11. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
  12. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/gemini_client.py +0 -0
  13. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/groq_client.py +0 -0
  14. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/local_client.py +0 -0
  15. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/mistral_client.py +0 -0
  16. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
  17. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/openai_client.py +0 -0
  18. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -0
  19. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/qwen_client.py +0 -0
  20. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/utils.py +0 -0
  21. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/yi_client.py +0 -0
  22. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
  23. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/settings/__init__.py +0 -0
  24. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/types/defaults.py +0 -0
  25. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/types/enums.py +0 -0
  26. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/types/llm_parameters.py +0 -0
  27. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/utilities/media_processing.py +0 -0
  28. {vectorvein-0.1.8 → vectorvein-0.1.9}/src/vectorvein/utilities/retry.py +0 -0
  29. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/__init__.py +0 -0
  30. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/cat.png +0 -0
  31. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/test_format_messages.py +0 -0
  32. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/test_image_input_chat_client.py +0 -0
  33. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/test_tokens_count.py +0 -0
  34. {vectorvein-0.1.8 → vectorvein-0.1.9}/tests/test_tool_use_multi_turns.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vectorvein
3
- Version: 0.1.8
3
+ Version: 0.1.9
4
4
  Summary: Default template for PDM package
5
5
  Author-Email: Anderson <andersonby@163.com>
6
6
  License: MIT
@@ -16,7 +16,7 @@ description = "Default template for PDM package"
16
16
  name = "vectorvein"
17
17
  readme = "README.md"
18
18
  requires-python = ">=3.10"
19
- version = "0.1.8"
19
+ version = "0.1.9"
20
20
 
21
21
  [project.license]
22
22
  text = "MIT"
@@ -93,7 +93,7 @@ class MiniMaxChatClient(BaseChatClient):
93
93
  model=self.model_setting.id,
94
94
  )
95
95
 
96
- if tools is not None:
96
+ if tools:
97
97
  tools_params = {
98
98
  "tools": [
99
99
  {
@@ -149,24 +149,19 @@ class MiniMaxChatClient(BaseChatClient):
149
149
  for chunk in response.iter_lines():
150
150
  if chunk:
151
151
  chunk_data = json.loads(chunk[6:])
152
+ if chunk_data["object"] != "chat.completion.chunk":
153
+ continue
152
154
  tool_calls_params = extract_tool_calls(chunk_data)
153
155
  has_tool_calls = True if tool_calls_params else False
154
156
  if has_tool_calls:
155
- if "usage" not in chunk_data:
156
- continue
157
- else:
158
- if chunk_data["object"] != "chat.completion.chunk":
159
- continue
160
- yield ChatCompletionDeltaMessage(
161
- **{
162
- "content": chunk_data["choices"][0]["delta"].get("content"),
163
- "role": "assistant",
164
- **tool_calls_params,
165
- }
166
- )
157
+ yield ChatCompletionDeltaMessage(
158
+ **{
159
+ "content": chunk_data["choices"][0]["delta"].get("content"),
160
+ "role": "assistant",
161
+ **tool_calls_params,
162
+ }
163
+ )
167
164
  else:
168
- if "usage" in chunk_data:
169
- continue
170
165
  yield ChatCompletionDeltaMessage(
171
166
  **{
172
167
  "content": chunk_data["choices"][0]["delta"]["content"],
@@ -250,7 +245,7 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
250
245
  model=self.model_setting.id,
251
246
  )
252
247
 
253
- if tools is not None:
248
+ if tools:
254
249
  tools_params = {
255
250
  "tools": [
256
251
  {
@@ -305,24 +300,19 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
305
300
  async for chunk in response.aiter_lines():
306
301
  if chunk:
307
302
  chunk_data = json.loads(chunk[6:])
303
+ if chunk_data["object"] != "chat.completion.chunk":
304
+ continue
308
305
  tool_calls_params = extract_tool_calls(chunk_data)
309
306
  has_tool_calls = True if tool_calls_params else False
310
307
  if has_tool_calls:
311
- if "usage" not in chunk_data:
312
- continue
313
- else:
314
- if chunk_data["object"] != "chat.completion.chunk":
315
- continue
316
- yield ChatCompletionDeltaMessage(
317
- **{
318
- "content": chunk_data["choices"][0]["delta"].get("content"),
319
- "role": "assistant",
320
- **tool_calls_params,
321
- }
322
- )
308
+ yield ChatCompletionDeltaMessage(
309
+ **{
310
+ "content": chunk_data["choices"][0]["delta"].get("content"),
311
+ "role": "assistant",
312
+ **tool_calls_params,
313
+ }
314
+ )
323
315
  else:
324
- if "usage" in chunk_data:
325
- continue
326
316
  yield ChatCompletionDeltaMessage(
327
317
  **{
328
318
  "content": chunk_data["choices"][0]["delta"]["content"],
@@ -425,7 +425,6 @@ sample_settings = {
425
425
  "id": "gpt-35-turbo",
426
426
  "endpoints": [
427
427
  "azure-openai-vectorvein-east-us",
428
- "azure-openai-vectorvein-east-us-2",
429
428
  "azure-openai-vectorvein-au-east",
430
429
  ],
431
430
  },
@@ -192,9 +192,6 @@ model = "deepseek-chat"
192
192
  # model = "moonshot-v1-8k"
193
193
  # backend = BackendType.OpenAI
194
194
  # model = "gpt-4o"
195
- # backend = BackendType.MiniMax
196
- # model = "abab6.5s-chat"
197
- # backend = BackendType.OpenAI
198
195
  # model = "gpt-35-turbo"
199
196
  # backend = BackendType.MiniMax
200
197
  # model = "abab6.5s-chat"
@@ -225,13 +222,13 @@ model = "deepseek-chat"
225
222
  # model = "mistral-large"
226
223
 
227
224
  start_time = time.perf_counter()
228
- # test_sync(backend=backend, model=model, stream=False, use_tool=False)
225
+ test_sync(backend=backend, model=model, stream=False, use_tool=False)
229
226
  test_sync(backend=backend, model=model, stream=False, use_tool=True)
230
- # test_sync(backend=backend, model=model, stream=True, use_tool=False)
227
+ test_sync(backend=backend, model=model, stream=True, use_tool=False)
231
228
  test_sync(backend=backend, model=model, stream=True, use_tool=True)
232
- # asyncio.run(test_async(backend=backend, model=model, stream=False, use_tool=False))
233
- # asyncio.run(test_async(backend=backend, model=model, stream=False, use_tool=True))
234
- # asyncio.run(test_async(backend=backend, model=model, stream=True, use_tool=False))
235
- # asyncio.run(test_async(backend=backend, model=model, stream=True, use_tool=True))
229
+ asyncio.run(test_async(backend=backend, model=model, stream=False, use_tool=False))
230
+ asyncio.run(test_async(backend=backend, model=model, stream=False, use_tool=True))
231
+ asyncio.run(test_async(backend=backend, model=model, stream=True, use_tool=False))
232
+ asyncio.run(test_async(backend=backend, model=model, stream=True, use_tool=True))
236
233
  end_time = time.perf_counter()
237
234
  print(f"Stream time elapsed: {end_time - start_time} seconds")
File without changes
File without changes
File without changes