vectorvein 0.1.10__py3-none-any.whl → 0.1.11__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- vectorvein/chat_clients/anthropic_client.py +4 -0
- vectorvein/chat_clients/gemini_client.py +6 -6
- vectorvein/chat_clients/minimax_client.py +4 -0
- vectorvein/chat_clients/openai_compatible_client.py +4 -0
- vectorvein/types/defaults.py +14 -0
- {vectorvein-0.1.10.dist-info → vectorvein-0.1.11.dist-info}/METADATA +1 -1
- {vectorvein-0.1.10.dist-info → vectorvein-0.1.11.dist-info}/RECORD +8 -8
- {vectorvein-0.1.10.dist-info → vectorvein-0.1.11.dist-info}/WHEEL +0 -0
@@ -118,6 +118,7 @@ class AnthropicChatClient(BaseChatClient):
|
|
118
118
|
max_tokens: int | None = None,
|
119
119
|
tools: list | NotGiven = NOT_GIVEN,
|
120
120
|
tool_choice: str | NotGiven = NOT_GIVEN,
|
121
|
+
**kwargs,
|
121
122
|
):
|
122
123
|
if model is not None:
|
123
124
|
self.model = model
|
@@ -210,6 +211,7 @@ class AnthropicChatClient(BaseChatClient):
|
|
210
211
|
max_tokens=max_tokens,
|
211
212
|
tools=tools_params,
|
212
213
|
tool_choice=tool_choice,
|
214
|
+
**kwargs,
|
213
215
|
)
|
214
216
|
|
215
217
|
if self.stream:
|
@@ -318,6 +320,7 @@ class AsyncAnthropicChatClient(BaseAsyncChatClient):
|
|
318
320
|
max_tokens: int | None = None,
|
319
321
|
tools: list | NotGiven = NOT_GIVEN,
|
320
322
|
tool_choice: str | NotGiven = NOT_GIVEN,
|
323
|
+
**kwargs,
|
321
324
|
):
|
322
325
|
if model is not None:
|
323
326
|
self.model = model
|
@@ -410,6 +413,7 @@ class AsyncAnthropicChatClient(BaseAsyncChatClient):
|
|
410
413
|
max_tokens=max_tokens,
|
411
414
|
tools=tools_params,
|
412
415
|
tool_choice=tool_choice,
|
416
|
+
**kwargs,
|
413
417
|
)
|
414
418
|
|
415
419
|
if self.stream:
|
@@ -146,9 +146,9 @@ class GeminiChatClient(BaseChatClient):
|
|
146
146
|
result = {
|
147
147
|
"content": "",
|
148
148
|
"usage": {
|
149
|
-
"prompt_tokens": response
|
150
|
-
"completion_tokens": response
|
151
|
-
"total_tokens": response
|
149
|
+
"prompt_tokens": response.get("usageMetadata", {}).get("promptTokenCount", 0),
|
150
|
+
"completion_tokens": response.get("usageMetadata", {}).get("candidatesTokenCount", 0),
|
151
|
+
"total_tokens": response.get("usageMetadata", {}).get("totalTokenCount", 0),
|
152
152
|
},
|
153
153
|
}
|
154
154
|
tool_calls = []
|
@@ -309,9 +309,9 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
|
|
309
309
|
result = {
|
310
310
|
"content": "",
|
311
311
|
"usage": {
|
312
|
-
"prompt_tokens": response
|
313
|
-
"completion_tokens": response
|
314
|
-
"total_tokens": response
|
312
|
+
"prompt_tokens": response.get("usageMetadata", {}).get("promptTokenCount", 0),
|
313
|
+
"completion_tokens": response.get("usageMetadata", {}).get("candidatesTokenCount", 0),
|
314
|
+
"total_tokens": response.get("usageMetadata", {}).get("totalTokenCount", 0),
|
315
315
|
},
|
316
316
|
}
|
317
317
|
tool_calls = []
|
@@ -70,6 +70,7 @@ class MiniMaxChatClient(BaseChatClient):
|
|
70
70
|
max_tokens: int | None = None,
|
71
71
|
tools: list | None = None,
|
72
72
|
tool_choice: str = "auto",
|
73
|
+
**kwargs,
|
73
74
|
):
|
74
75
|
if model is not None:
|
75
76
|
self.model = model
|
@@ -135,6 +136,7 @@ class MiniMaxChatClient(BaseChatClient):
|
|
135
136
|
"stream": self.stream,
|
136
137
|
"mask_sensitive_info": False,
|
137
138
|
**tools_params,
|
139
|
+
**kwargs,
|
138
140
|
}
|
139
141
|
|
140
142
|
if self.stream:
|
@@ -228,6 +230,7 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
|
|
228
230
|
max_tokens: int | None = None,
|
229
231
|
tools: list | None = None,
|
230
232
|
tool_choice: str = "auto",
|
233
|
+
**kwargs,
|
231
234
|
):
|
232
235
|
if model is not None:
|
233
236
|
self.model = model
|
@@ -291,6 +294,7 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
|
|
291
294
|
"stream": self.stream,
|
292
295
|
"mask_sensitive_info": False,
|
293
296
|
**tools_params,
|
297
|
+
**kwargs,
|
294
298
|
}
|
295
299
|
|
296
300
|
if self.stream:
|
@@ -54,6 +54,7 @@ class OpenAICompatibleChatClient(BaseChatClient):
|
|
54
54
|
max_tokens: int | None = None,
|
55
55
|
tools: list | NotGiven = NOT_GIVEN,
|
56
56
|
tool_choice: str | NotGiven = NOT_GIVEN,
|
57
|
+
**kwargs,
|
57
58
|
):
|
58
59
|
if model is not None:
|
59
60
|
self.model = model
|
@@ -120,6 +121,7 @@ class OpenAICompatibleChatClient(BaseChatClient):
|
|
120
121
|
temperature=self.temperature,
|
121
122
|
max_tokens=max_tokens,
|
122
123
|
**tools_params,
|
124
|
+
**kwargs,
|
123
125
|
)
|
124
126
|
|
125
127
|
if self.stream:
|
@@ -207,6 +209,7 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
|
|
207
209
|
max_tokens: int | None = None,
|
208
210
|
tools: list | NotGiven = NOT_GIVEN,
|
209
211
|
tool_choice: str | NotGiven = NOT_GIVEN,
|
212
|
+
**kwargs,
|
210
213
|
):
|
211
214
|
if model is not None:
|
212
215
|
self.model = model
|
@@ -273,6 +276,7 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
|
|
273
276
|
temperature=self.temperature,
|
274
277
|
max_tokens=max_tokens,
|
275
278
|
**tools_params,
|
279
|
+
**kwargs,
|
276
280
|
)
|
277
281
|
|
278
282
|
if self.stream:
|
vectorvein/types/defaults.py
CHANGED
@@ -213,6 +213,13 @@ ZHIPUAI_MODELS = {
|
|
213
213
|
"response_format_available": False,
|
214
214
|
"max_output_tokens": 4095,
|
215
215
|
},
|
216
|
+
"glm-4-plus": {
|
217
|
+
"id": "glm-4-plus",
|
218
|
+
"context_length": 128000,
|
219
|
+
"function_call_available": True,
|
220
|
+
"response_format_available": False,
|
221
|
+
"max_output_tokens": 4095,
|
222
|
+
},
|
216
223
|
"glm-4-0520": {
|
217
224
|
"id": "glm-4-0520",
|
218
225
|
"context_length": 128000,
|
@@ -255,6 +262,13 @@ ZHIPUAI_MODELS = {
|
|
255
262
|
"response_format_available": False,
|
256
263
|
"max_output_tokens": 1024,
|
257
264
|
},
|
265
|
+
"glm-4v-plus": {
|
266
|
+
"id": "glm-4v-plus",
|
267
|
+
"context_length": 2000,
|
268
|
+
"function_call_available": False,
|
269
|
+
"response_format_available": False,
|
270
|
+
"max_output_tokens": 1024,
|
271
|
+
},
|
258
272
|
}
|
259
273
|
|
260
274
|
# Mistral models
|
@@ -1,26 +1,26 @@
|
|
1
|
-
vectorvein-0.1.
|
2
|
-
vectorvein-0.1.
|
1
|
+
vectorvein-0.1.11.dist-info/METADATA,sha256=rw0fFY-7V-teJUIt4t4KlvLIZHp4hjgaQBRpQ816eP0,502
|
2
|
+
vectorvein-0.1.11.dist-info/WHEEL,sha256=rSwsxJWe3vzyR5HCwjWXQruDgschpei4h_giTm0dJVE,90
|
3
3
|
vectorvein/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
4
4
|
vectorvein/chat_clients/__init__.py,sha256=5j7W--jr-l2cqDJp38uXYvkydDK0rnzm7MYGSACHKmU,3976
|
5
|
-
vectorvein/chat_clients/anthropic_client.py,sha256=
|
5
|
+
vectorvein/chat_clients/anthropic_client.py,sha256=MHpLywbIIgFsAkj8EiKEoxUgxcbx3njFxLTN2GhEjys,20088
|
6
6
|
vectorvein/chat_clients/base_client.py,sha256=wMXpQ1L1KDb2Hg6va3H3GmcVeQB6r6sh7F4IS0DBQWI,4275
|
7
7
|
vectorvein/chat_clients/deepseek_client.py,sha256=3qWu01NlJAP2N-Ff62d5-CZXZitlizE1fzb20LNetig,526
|
8
|
-
vectorvein/chat_clients/gemini_client.py,sha256=
|
8
|
+
vectorvein/chat_clients/gemini_client.py,sha256=jUu-wwpaChP8_lxLdWnYkgVr_aSegCHbmBgnpFztqAw,13743
|
9
9
|
vectorvein/chat_clients/groq_client.py,sha256=Uow4pgdmFi93ZQSoOol2-0PhhqkW-S0XuSldvppz5U4,498
|
10
10
|
vectorvein/chat_clients/local_client.py,sha256=55nOsxzqUf79q3Y14MKROA71zxhsT7p7FsDZ89rts2M,422
|
11
|
-
vectorvein/chat_clients/minimax_client.py,sha256=
|
11
|
+
vectorvein/chat_clients/minimax_client.py,sha256=ZO-7WNsvsct3cpuo0Qc6WL3A0kkijn0J8o0ahlxmUyE,13478
|
12
12
|
vectorvein/chat_clients/mistral_client.py,sha256=1aKSylzBDaLYcFnaBIL4-sXSzWmXfBeON9Q0rq-ziWw,534
|
13
13
|
vectorvein/chat_clients/moonshot_client.py,sha256=gbu-6nGxx8uM_U2WlI4Wus881rFRotzHtMSoYOcruGU,526
|
14
14
|
vectorvein/chat_clients/openai_client.py,sha256=Nz6tV45pWcsOupxjnsRsGTicbQNJWIZyxuJoJ5DGMpg,527
|
15
|
-
vectorvein/chat_clients/openai_compatible_client.py,sha256=
|
15
|
+
vectorvein/chat_clients/openai_compatible_client.py,sha256=rPxIPj4STm1sy7T6bqyZpO-FPwMyjyQmThyJqd8_jYs,13822
|
16
16
|
vectorvein/chat_clients/qwen_client.py,sha256=-ryh-m9PgsO0fc4ulcCmPTy1155J8YUy15uPoJQOHA0,513
|
17
17
|
vectorvein/chat_clients/utils.py,sha256=mnAew2Ie3nQHdEyDLKuJvXkQ5QdcSAJ6SpYk5JPbR1Q,20888
|
18
18
|
vectorvein/chat_clients/yi_client.py,sha256=RNf4CRuPJfixrwLZ3-DEc3t25QDe1mvZeb9sku2f8Bc,484
|
19
19
|
vectorvein/chat_clients/zhipuai_client.py,sha256=Ys5DSeLCuedaDXr3PfG1EW2zKXopt-awO2IylWSwY0s,519
|
20
20
|
vectorvein/settings/__init__.py,sha256=4mpccT7eZC3yI1vVnVViW4wHBnDEH9D2R5EsIP34VgU,3218
|
21
|
-
vectorvein/types/defaults.py,sha256=
|
21
|
+
vectorvein/types/defaults.py,sha256=lCzGOLybX8FzHX-Cv32BaQFZ8sHvPhGIIwDD-VksP20,13460
|
22
22
|
vectorvein/types/enums.py,sha256=vzOenCnRlFXBwPh-lfFhjGfM-6yfDj7wZColHODqocI,1550
|
23
23
|
vectorvein/types/llm_parameters.py,sha256=nBjStC2zndTY__yhD2WFXB09taxEhDLE3OHA6MICfgE,3494
|
24
24
|
vectorvein/utilities/media_processing.py,sha256=BujciRmw1GMmc3ELRvafL8STcy6r5b2rVnh27-uA7so,2256
|
25
25
|
vectorvein/utilities/retry.py,sha256=9ePuJdeUUGx-qMWfaFxmlOvG_lQPwCQ4UB1z3Edlo34,993
|
26
|
-
vectorvein-0.1.
|
26
|
+
vectorvein-0.1.11.dist-info/RECORD,,
|
File without changes
|