vectorvein 0.1.14__tar.gz → 0.1.15__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {vectorvein-0.1.14 → vectorvein-0.1.15}/PKG-INFO +1 -1
- {vectorvein-0.1.14 → vectorvein-0.1.15}/pyproject.toml +6 -1
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/types/defaults.py +11 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/types/llm_parameters.py +1 -0
- vectorvein-0.1.14/tests/__init__.py +0 -0
- vectorvein-0.1.14/tests/cat.png +0 -0
- vectorvein-0.1.14/tests/sample_settings.py +0 -614
- vectorvein-0.1.14/tests/test_chat_prefix.py +0 -23
- vectorvein-0.1.14/tests/test_create_chat_client.py +0 -234
- vectorvein-0.1.14/tests/test_format_messages.py +0 -41
- vectorvein-0.1.14/tests/test_http_client.py +0 -24
- vectorvein-0.1.14/tests/test_image_input_chat_client.py +0 -45
- vectorvein-0.1.14/tests/test_stop.py +0 -25
- vectorvein-0.1.14/tests/test_tokens_count.py +0 -46
- vectorvein-0.1.14/tests/test_tool_use_multi_turns.py +0 -159
- {vectorvein-0.1.14 → vectorvein-0.1.15}/README.md +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/__init__.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/__init__.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/base_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/gemini_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/groq_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/local_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/minimax_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/mistral_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/openai_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/openai_compatible_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/qwen_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/utils.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/yi_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/settings/__init__.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/types/enums.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/utilities/media_processing.py +0 -0
- {vectorvein-0.1.14 → vectorvein-0.1.15}/src/vectorvein/utilities/retry.py +0 -0
@@ -16,7 +16,7 @@ description = "Default template for PDM package"
|
|
16
16
|
name = "vectorvein"
|
17
17
|
readme = "README.md"
|
18
18
|
requires-python = ">=3.10"
|
19
|
-
version = "0.1.
|
19
|
+
version = "0.1.15"
|
20
20
|
|
21
21
|
[project.license]
|
22
22
|
text = "MIT"
|
@@ -29,3 +29,8 @@ requires = [
|
|
29
29
|
|
30
30
|
[tool.pdm]
|
31
31
|
distribution = true
|
32
|
+
|
33
|
+
[tool.pdm.build]
|
34
|
+
excludes = [
|
35
|
+
"tests",
|
36
|
+
]
|
@@ -233,6 +233,7 @@ YI_MODELS = {
|
|
233
233
|
"max_output_tokens": 2000,
|
234
234
|
"function_call_available": False,
|
235
235
|
"response_format_available": False,
|
236
|
+
"native_multimodal": True,
|
236
237
|
},
|
237
238
|
}
|
238
239
|
|
@@ -301,6 +302,7 @@ ZHIPUAI_MODELS = {
|
|
301
302
|
"function_call_available": False,
|
302
303
|
"response_format_available": False,
|
303
304
|
"max_output_tokens": 1024,
|
305
|
+
"native_multimodal": True,
|
304
306
|
},
|
305
307
|
"glm-4v-plus": {
|
306
308
|
"id": "glm-4v-plus",
|
@@ -308,6 +310,7 @@ ZHIPUAI_MODELS = {
|
|
308
310
|
"function_call_available": False,
|
309
311
|
"response_format_available": False,
|
310
312
|
"max_output_tokens": 1024,
|
313
|
+
"native_multimodal": True,
|
311
314
|
},
|
312
315
|
}
|
313
316
|
|
@@ -394,6 +397,7 @@ OPENAI_MODELS = {
|
|
394
397
|
"max_output_tokens": 4096,
|
395
398
|
"function_call_available": True,
|
396
399
|
"response_format_available": True,
|
400
|
+
"native_multimodal": True,
|
397
401
|
},
|
398
402
|
"gpt-4o-mini": {
|
399
403
|
"id": "gpt-4o-mini",
|
@@ -401,6 +405,7 @@ OPENAI_MODELS = {
|
|
401
405
|
"max_output_tokens": 16384,
|
402
406
|
"function_call_available": True,
|
403
407
|
"response_format_available": True,
|
408
|
+
"native_multimodal": True,
|
404
409
|
},
|
405
410
|
"gpt-4v": {
|
406
411
|
"id": "gpt-4v",
|
@@ -420,12 +425,14 @@ ANTHROPIC_MODELS = {
|
|
420
425
|
"max_output_tokens": 4096,
|
421
426
|
"function_call_available": True,
|
422
427
|
"response_format_available": True,
|
428
|
+
"native_multimodal": True,
|
423
429
|
},
|
424
430
|
"claude-3-sonnet-20240229": {
|
425
431
|
"id": "claude-3-sonnet-20240229",
|
426
432
|
"context_length": 200000,
|
427
433
|
"max_output_tokens": 4096,
|
428
434
|
"function_call_available": True,
|
435
|
+
"native_multimodal": True,
|
429
436
|
"response_format_available": True,
|
430
437
|
},
|
431
438
|
"claude-3-haiku-20240307": {
|
@@ -434,6 +441,7 @@ ANTHROPIC_MODELS = {
|
|
434
441
|
"max_output_tokens": 4096,
|
435
442
|
"function_call_available": True,
|
436
443
|
"response_format_available": True,
|
444
|
+
"native_multimodal": True,
|
437
445
|
},
|
438
446
|
"claude-3-5-sonnet-20240620": {
|
439
447
|
"id": "claude-3-5-sonnet-20240620",
|
@@ -441,6 +449,7 @@ ANTHROPIC_MODELS = {
|
|
441
449
|
"max_output_tokens": 4096,
|
442
450
|
"function_call_available": True,
|
443
451
|
"response_format_available": True,
|
452
|
+
"native_multimodal": True,
|
444
453
|
},
|
445
454
|
}
|
446
455
|
|
@@ -485,11 +494,13 @@ GEMINI_MODELS = {
|
|
485
494
|
"context_length": 1048576,
|
486
495
|
"function_call_available": True,
|
487
496
|
"response_format_available": True,
|
497
|
+
"native_multimodal": True,
|
488
498
|
},
|
489
499
|
"gemini-1.5-flash": {
|
490
500
|
"id": "gemini-1.5-flash",
|
491
501
|
"context_length": 1048576,
|
492
502
|
"function_call_available": True,
|
493
503
|
"response_format_available": True,
|
504
|
+
"native_multimodal": True,
|
494
505
|
},
|
495
506
|
}
|
@@ -30,6 +30,7 @@ class ModelSetting(BaseModel):
|
|
30
30
|
endpoints: List[str] = Field(default_factory=list, description="Available endpoints for the model.")
|
31
31
|
function_call_available: bool = Field(False, description="Indicates if function call is available.")
|
32
32
|
response_format_available: bool = Field(False, description="Indicates if response format is available.")
|
33
|
+
native_multimodal: bool = Field(False, description="Indicates if the model is a native multimodal model.")
|
33
34
|
context_length: int = Field(32768, description="The context length for the model.")
|
34
35
|
max_output_tokens: Optional[int] = Field(None, description="Maximum number of output tokens allowed.")
|
35
36
|
|
File without changes
|
vectorvein-0.1.14/tests/cat.png
DELETED
Binary file
|