vectorvein 0.1.26__tar.gz → 0.1.27__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {vectorvein-0.1.26 → vectorvein-0.1.27}/PKG-INFO +1 -1
- {vectorvein-0.1.26 → vectorvein-0.1.27}/pyproject.toml +1 -1
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/base_client.py +32 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/gemini_client.py +30 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/minimax_client.py +30 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/openai_compatible_client.py +30 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/README.md +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/__init__.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/__init__.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/anthropic_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/baichuan_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/deepseek_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/groq_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/local_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/mistral_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/moonshot_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/openai_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/qwen_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/utils.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/yi_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/zhipuai_client.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/settings/__init__.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/types/defaults.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/types/enums.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/types/exception.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/types/llm_parameters.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/utilities/media_processing.py +0 -0
- {vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/utilities/retry.py +0 -0
@@ -88,6 +88,22 @@ class BaseChatClient(ABC):
|
|
88
88
|
) -> Generator[ChatCompletionDeltaMessage, Any, None]:
|
89
89
|
pass
|
90
90
|
|
91
|
+
@overload
|
92
|
+
@abstractmethod
|
93
|
+
def create_completion(
|
94
|
+
self,
|
95
|
+
messages: list,
|
96
|
+
model: str | None = None,
|
97
|
+
stream: bool = False,
|
98
|
+
temperature: float = 0.7,
|
99
|
+
max_tokens: int | None = None,
|
100
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
101
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
102
|
+
response_format: dict | None = None,
|
103
|
+
**kwargs,
|
104
|
+
) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
|
105
|
+
pass
|
106
|
+
|
91
107
|
@abstractmethod
|
92
108
|
def create_completion(
|
93
109
|
self,
|
@@ -196,6 +212,22 @@ class BaseAsyncChatClient(ABC):
|
|
196
212
|
) -> AsyncGenerator[ChatCompletionDeltaMessage, None]:
|
197
213
|
pass
|
198
214
|
|
215
|
+
@overload
|
216
|
+
@abstractmethod
|
217
|
+
async def create_completion(
|
218
|
+
self,
|
219
|
+
messages: list,
|
220
|
+
model: str | None = None,
|
221
|
+
stream: bool = False,
|
222
|
+
temperature: float = 0.7,
|
223
|
+
max_tokens: int | None = None,
|
224
|
+
tools: list | NotGiven = NOT_GIVEN,
|
225
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
226
|
+
response_format: dict | None = None,
|
227
|
+
**kwargs,
|
228
|
+
) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, None]:
|
229
|
+
pass
|
230
|
+
|
199
231
|
@abstractmethod
|
200
232
|
async def create_completion(
|
201
233
|
self,
|
@@ -82,6 +82,21 @@ class GeminiChatClient(BaseChatClient):
|
|
82
82
|
) -> Generator[ChatCompletionDeltaMessage, None, None]:
|
83
83
|
pass
|
84
84
|
|
85
|
+
@overload
|
86
|
+
def create_completion(
|
87
|
+
self,
|
88
|
+
messages: list,
|
89
|
+
model: str | None = None,
|
90
|
+
stream: bool | None = None,
|
91
|
+
temperature: float | None = None,
|
92
|
+
max_tokens: int | None = None,
|
93
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
94
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
95
|
+
response_format: dict | None = None,
|
96
|
+
**kwargs,
|
97
|
+
) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
|
98
|
+
pass
|
99
|
+
|
85
100
|
def create_completion(
|
86
101
|
self,
|
87
102
|
messages: list,
|
@@ -295,6 +310,21 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
|
|
295
310
|
) -> AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
296
311
|
pass
|
297
312
|
|
313
|
+
@overload
|
314
|
+
async def create_completion(
|
315
|
+
self,
|
316
|
+
messages: list,
|
317
|
+
model: str | None = None,
|
318
|
+
stream: bool | None = None,
|
319
|
+
temperature: float | None = None,
|
320
|
+
max_tokens: int | None = None,
|
321
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
322
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
323
|
+
response_format: dict | None = None,
|
324
|
+
**kwargs,
|
325
|
+
) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
326
|
+
pass
|
327
|
+
|
298
328
|
async def create_completion(
|
299
329
|
self,
|
300
330
|
messages: list,
|
@@ -107,6 +107,21 @@ class MiniMaxChatClient(BaseChatClient):
|
|
107
107
|
) -> Generator[ChatCompletionDeltaMessage, None, None]:
|
108
108
|
pass
|
109
109
|
|
110
|
+
@overload
|
111
|
+
def create_completion(
|
112
|
+
self,
|
113
|
+
messages: list,
|
114
|
+
model: str | None = None,
|
115
|
+
stream: bool | None = None,
|
116
|
+
temperature: float | None = None,
|
117
|
+
max_tokens: int | None = None,
|
118
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
119
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
120
|
+
response_format: dict | None = None,
|
121
|
+
**kwargs,
|
122
|
+
) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
|
123
|
+
pass
|
124
|
+
|
110
125
|
def create_completion(
|
111
126
|
self,
|
112
127
|
messages: list,
|
@@ -307,6 +322,21 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
|
|
307
322
|
) -> AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
308
323
|
pass
|
309
324
|
|
325
|
+
@overload
|
326
|
+
async def create_completion(
|
327
|
+
self,
|
328
|
+
messages: list,
|
329
|
+
model: str | None = None,
|
330
|
+
stream: bool | None = None,
|
331
|
+
temperature: float | None = None,
|
332
|
+
max_tokens: int | None = None,
|
333
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
334
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
335
|
+
response_format: dict | None = None,
|
336
|
+
**kwargs,
|
337
|
+
) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
338
|
+
pass
|
339
|
+
|
310
340
|
async def create_completion(
|
311
341
|
self,
|
312
342
|
messages: list,
|
{vectorvein-0.1.26 → vectorvein-0.1.27}/src/vectorvein/chat_clients/openai_compatible_client.py
RENAMED
@@ -107,6 +107,21 @@ class OpenAICompatibleChatClient(BaseChatClient):
|
|
107
107
|
) -> Generator[ChatCompletionDeltaMessage, None, None]:
|
108
108
|
pass
|
109
109
|
|
110
|
+
@overload
|
111
|
+
def create_completion(
|
112
|
+
self,
|
113
|
+
messages: list,
|
114
|
+
model: str | None = None,
|
115
|
+
stream: bool | None = None,
|
116
|
+
temperature: float | None = None,
|
117
|
+
max_tokens: int | None = None,
|
118
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
119
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
120
|
+
response_format: dict | None = None,
|
121
|
+
**kwargs,
|
122
|
+
) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
|
123
|
+
pass
|
124
|
+
|
110
125
|
def create_completion(
|
111
126
|
self,
|
112
127
|
messages: list,
|
@@ -314,6 +329,21 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
|
|
314
329
|
) -> AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
315
330
|
pass
|
316
331
|
|
332
|
+
@overload
|
333
|
+
async def create_completion(
|
334
|
+
self,
|
335
|
+
messages: list,
|
336
|
+
model: str | None = None,
|
337
|
+
stream: bool | None = None,
|
338
|
+
temperature: float | None = None,
|
339
|
+
max_tokens: int | None = None,
|
340
|
+
tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
|
341
|
+
tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
|
342
|
+
response_format: dict | None = None,
|
343
|
+
**kwargs,
|
344
|
+
) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, Any]:
|
345
|
+
pass
|
346
|
+
|
317
347
|
async def create_completion(
|
318
348
|
self,
|
319
349
|
messages: list,
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|