vectorvein 0.1.47__tar.gz → 0.1.49__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (34) hide show
  1. {vectorvein-0.1.47 → vectorvein-0.1.49}/PKG-INFO +1 -1
  2. {vectorvein-0.1.47 → vectorvein-0.1.49}/pyproject.toml +1 -1
  3. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/openai_compatible_client.py +11 -4
  4. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/types/defaults.py +10 -2
  5. {vectorvein-0.1.47 → vectorvein-0.1.49}/README.md +0 -0
  6. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/__init__.py +0 -0
  7. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/__init__.py +0 -0
  8. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
  9. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
  10. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/base_client.py +0 -0
  11. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
  12. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/gemini_client.py +0 -0
  13. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/groq_client.py +0 -0
  14. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/local_client.py +0 -0
  15. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/minimax_client.py +0 -0
  16. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/mistral_client.py +0 -0
  17. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
  18. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/openai_client.py +0 -0
  19. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/py.typed +0 -0
  20. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/qwen_client.py +0 -0
  21. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/stepfun_client.py +0 -0
  22. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/utils.py +0 -0
  23. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/yi_client.py +0 -0
  24. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
  25. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/py.typed +0 -0
  26. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/server/token_server.py +0 -0
  27. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/settings/__init__.py +0 -0
  28. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/settings/py.typed +0 -0
  29. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/types/enums.py +0 -0
  30. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/types/exception.py +0 -0
  31. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/types/llm_parameters.py +0 -0
  32. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/types/py.typed +0 -0
  33. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/utilities/media_processing.py +0 -0
  34. {vectorvein-0.1.47 → vectorvein-0.1.49}/src/vectorvein/utilities/retry.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vectorvein
3
- Version: 0.1.47
3
+ Version: 0.1.49
4
4
  Summary: Default template for PDM package
5
5
  Author-Email: Anderson <andersonby@163.com>
6
6
  License: MIT
@@ -17,7 +17,7 @@ description = "Default template for PDM package"
17
17
  name = "vectorvein"
18
18
  readme = "README.md"
19
19
  requires-python = ">=3.10"
20
- version = "0.1.47"
20
+ version = "0.1.49"
21
21
 
22
22
  [project.license]
23
23
  text = "MIT"
@@ -223,11 +223,14 @@ class OpenAICompatibleChatClient(BaseChatClient):
223
223
  prompt_tokens=chunk.usage.prompt_tokens or 0,
224
224
  total_tokens=chunk.usage.total_tokens or 0,
225
225
  )
226
- else:
227
- usage = None
226
+
228
227
  if len(chunk.choices) == 0:
228
+ if usage:
229
+ yield ChatCompletionDeltaMessage(usage=usage)
229
230
  continue
230
231
  if not chunk.choices[0].delta:
232
+ if usage:
233
+ yield ChatCompletionDeltaMessage(usage=usage)
231
234
  continue
232
235
  if self.model_setting.function_call_available:
233
236
  if chunk.choices[0].delta.tool_calls:
@@ -474,12 +477,16 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
474
477
  prompt_tokens=chunk.usage.prompt_tokens or 0,
475
478
  total_tokens=chunk.usage.total_tokens or 0,
476
479
  )
477
- else:
478
- usage = None
480
+
479
481
  if len(chunk.choices) == 0:
482
+ if usage:
483
+ yield ChatCompletionDeltaMessage(usage=usage)
480
484
  continue
481
485
  if not chunk.choices[0].delta:
486
+ if usage:
487
+ yield ChatCompletionDeltaMessage(usage=usage)
482
488
  continue
489
+
483
490
  if self.model_setting.function_call_available:
484
491
  if chunk.choices[0].delta.tool_calls:
485
492
  for index, tool_call in enumerate(chunk.choices[0].delta.tool_calls):
@@ -552,10 +552,18 @@ ANTHROPIC_MODELS: Final[Dict[str, Dict[str, Any]]] = {
552
552
  "response_format_available": False,
553
553
  "native_multimodal": True,
554
554
  },
555
+ "claude-3-5-haiku-20241022": {
556
+ "id": "claude-3-5-haiku-20241022",
557
+ "context_length": 200000,
558
+ "max_output_tokens": 8192,
559
+ "function_call_available": True,
560
+ "response_format_available": False,
561
+ "native_multimodal": True,
562
+ },
555
563
  "claude-3-5-sonnet-20240620": {
556
564
  "id": "claude-3-5-sonnet-20240620",
557
565
  "context_length": 200000,
558
- "max_output_tokens": 4096,
566
+ "max_output_tokens": 8192,
559
567
  "function_call_available": True,
560
568
  "response_format_available": False,
561
569
  "native_multimodal": True,
@@ -563,7 +571,7 @@ ANTHROPIC_MODELS: Final[Dict[str, Dict[str, Any]]] = {
563
571
  "claude-3-5-sonnet-20241022": {
564
572
  "id": "claude-3-5-sonnet-20241022",
565
573
  "context_length": 200000,
566
- "max_output_tokens": 4096,
574
+ "max_output_tokens": 8192,
567
575
  "function_call_available": True,
568
576
  "response_format_available": False,
569
577
  "native_multimodal": True,
File without changes