vectorvein 0.1.20__tar.gz → 0.1.21__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (27) hide show
  1. {vectorvein-0.1.20 → vectorvein-0.1.21}/PKG-INFO +1 -1
  2. {vectorvein-0.1.20 → vectorvein-0.1.21}/pyproject.toml +1 -1
  3. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/gemini_client.py +4 -0
  4. {vectorvein-0.1.20 → vectorvein-0.1.21}/README.md +0 -0
  5. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/__init__.py +0 -0
  6. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/__init__.py +0 -0
  7. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
  8. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
  9. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/base_client.py +0 -0
  10. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
  11. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/groq_client.py +0 -0
  12. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/local_client.py +0 -0
  13. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/minimax_client.py +0 -0
  14. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/mistral_client.py +0 -0
  15. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
  16. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/openai_client.py +0 -0
  17. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -0
  18. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/qwen_client.py +0 -0
  19. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/utils.py +0 -0
  20. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/yi_client.py +0 -0
  21. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
  22. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/settings/__init__.py +0 -0
  23. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/types/defaults.py +0 -0
  24. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/types/enums.py +0 -0
  25. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/types/llm_parameters.py +0 -0
  26. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/utilities/media_processing.py +0 -0
  27. {vectorvein-0.1.20 → vectorvein-0.1.21}/src/vectorvein/utilities/retry.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vectorvein
3
- Version: 0.1.20
3
+ Version: 0.1.21
4
4
  Summary: Default template for PDM package
5
5
  Author-Email: Anderson <andersonby@163.com>
6
6
  License: MIT
@@ -16,7 +16,7 @@ description = "Default template for PDM package"
16
16
  name = "vectorvein"
17
17
  readme = "README.md"
18
18
  requires-python = ">=3.10"
19
- version = "0.1.20"
19
+ version = "0.1.21"
20
20
 
21
21
  [project.license]
22
22
  text = "MIT"
@@ -52,6 +52,7 @@ class GeminiChatClient(BaseChatClient):
52
52
  max_tokens: int | None = None,
53
53
  tools: list | None = None,
54
54
  tool_choice: str | None = None,
55
+ **kwargs,
55
56
  ):
56
57
  if model is not None:
57
58
  self.model = model
@@ -99,6 +100,7 @@ class GeminiChatClient(BaseChatClient):
99
100
  "maxOutputTokens": max_tokens,
100
101
  },
101
102
  **tools_params,
103
+ **kwargs,
102
104
  }
103
105
  if system_prompt:
104
106
  request_body["systemInstruction"] = {"parts": [{"text": system_prompt}]}
@@ -226,6 +228,7 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
226
228
  max_tokens: int | None = None,
227
229
  tools: list | None = None,
228
230
  tool_choice: str | None = None,
231
+ **kwargs,
229
232
  ):
230
233
  if model is not None:
231
234
  self.model = model
@@ -273,6 +276,7 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
273
276
  "maxOutputTokens": max_tokens,
274
277
  },
275
278
  **tools_params,
279
+ **kwargs,
276
280
  }
277
281
  if system_prompt:
278
282
  request_body["systemInstruction"] = {"parts": [{"text": system_prompt}]}
File without changes