vectorvein 0.1.18__py3-none-any.whl → 0.1.20__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -39,6 +39,10 @@ class GeminiChatClient(BaseChatClient):
39
39
  **kwargs,
40
40
  )
41
41
 
42
+ @property
43
+ def raw_client(self):
44
+ return self.http_client
45
+
42
46
  def create_completion(
43
47
  self,
44
48
  messages: list = list,
@@ -209,6 +213,10 @@ class AsyncGeminiChatClient(BaseAsyncChatClient):
209
213
  **kwargs,
210
214
  )
211
215
 
216
+ @property
217
+ def raw_client(self):
218
+ return self.http_client
219
+
212
220
  async def create_completion(
213
221
  self,
214
222
  messages: list = list,
@@ -66,6 +66,10 @@ class MiniMaxChatClient(BaseChatClient):
66
66
  else:
67
67
  self.http_client = httpx.Client()
68
68
 
69
+ @property
70
+ def raw_client(self):
71
+ return self.http_client
72
+
69
73
  def create_completion(
70
74
  self,
71
75
  messages: list = list,
@@ -231,6 +235,10 @@ class AsyncMiniMaxChatClient(BaseAsyncChatClient):
231
235
  else:
232
236
  self.http_client = httpx.AsyncClient()
233
237
 
238
+ @property
239
+ def raw_client(self):
240
+ return self.http_client
241
+
234
242
  async def create_completion(
235
243
  self,
236
244
  messages: list = list,
@@ -90,25 +90,6 @@ class OpenAICompatibleChatClient(BaseChatClient):
90
90
 
91
91
  self.model_setting = self.backend_settings.models[self.model]
92
92
 
93
- # if self.random_endpoint:
94
- # self.random_endpoint = True
95
- # self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
96
- # self.endpoint = settings.get_endpoint(self.endpoint_id)
97
-
98
- # if self.endpoint.is_azure:
99
- # self._client = AzureOpenAI(
100
- # azure_endpoint=self.endpoint.api_base,
101
- # api_key=self.endpoint.api_key,
102
- # api_version="2024-08-01-preview",
103
- # http_client=self.http_client,
104
- # )
105
- # else:
106
- # self._client = OpenAI(
107
- # api_key=self.endpoint.api_key,
108
- # base_url=self.endpoint.api_base,
109
- # http_client=self.http_client,
110
- # )
111
-
112
93
  if self.context_length_control == ContextLengthControlType.Latest:
113
94
  messages = cutoff_messages(
114
95
  messages,
@@ -269,25 +250,6 @@ class AsyncOpenAICompatibleChatClient(BaseAsyncChatClient):
269
250
 
270
251
  self.model_setting = self.backend_settings.models[self.model]
271
252
 
272
- # if self.random_endpoint:
273
- # self.random_endpoint = True
274
- # self.endpoint_id = random.choice(self.backend_settings.models[self.model].endpoints)
275
- # self.endpoint = settings.get_endpoint(self.endpoint_id)
276
-
277
- # if self.endpoint.is_azure:
278
- # self._client = AsyncAzureOpenAI(
279
- # azure_endpoint=self.endpoint.api_base,
280
- # api_key=self.endpoint.api_key,
281
- # api_version="2024-08-01-preview",
282
- # http_client=self.http_client,
283
- # )
284
- # else:
285
- # self._client = AsyncOpenAI(
286
- # api_key=self.endpoint.api_key,
287
- # base_url=self.endpoint.api_base,
288
- # http_client=self.http_client,
289
- # )
290
-
291
253
  if self.context_length_control == ContextLengthControlType.Latest:
292
254
  messages = cutoff_messages(
293
255
  messages,
@@ -119,6 +119,34 @@ GROQ_MODELS = {
119
119
  "function_call_available": True,
120
120
  "response_format_available": True,
121
121
  },
122
+ "gemma2-9b-it": {
123
+ "id": "gemma2-9b-it",
124
+ "context_length": 8192,
125
+ },
126
+ "llama3-groq-70b-8192-tool-use-preview": {
127
+ "id": "llama3-groq-70b-8192-tool-use-preview",
128
+ "context_length": 8192,
129
+ "function_call_available": True,
130
+ "max_output_tokens": 8000,
131
+ },
132
+ "llama3-groq-8b-8192-tool-use-preview": {
133
+ "id": "llama3-groq-8b-8192-tool-use-preview",
134
+ "context_length": 8192,
135
+ "function_call_available": True,
136
+ "max_output_tokens": 8000,
137
+ },
138
+ "llama-3.1-70b-versatile": {
139
+ "id": "llama-3.1-70b-versatile",
140
+ "context_length": 131072,
141
+ "function_call_available": True,
142
+ "max_output_tokens": 8000,
143
+ },
144
+ "llama-3.1-8b-instant": {
145
+ "id": "llama-3.1-8b-instant",
146
+ "context_length": 131072,
147
+ "function_call_available": True,
148
+ "max_output_tokens": 8000,
149
+ },
122
150
  }
123
151
 
124
152
  # Qwen models
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vectorvein
3
- Version: 0.1.18
3
+ Version: 0.1.20
4
4
  Summary: Default template for PDM package
5
5
  Author-Email: Anderson <andersonby@163.com>
6
6
  License: MIT
@@ -1,27 +1,28 @@
1
- vectorvein-0.1.18.dist-info/METADATA,sha256=cJFS7kZ5JNH96mpTO8IBBzzj-fuXi5jVpHs5EN1aXPc,502
2
- vectorvein-0.1.18.dist-info/WHEEL,sha256=rSwsxJWe3vzyR5HCwjWXQruDgschpei4h_giTm0dJVE,90
1
+ vectorvein-0.1.20.dist-info/METADATA,sha256=QSFFqZbpsbsi8xtCN7QAMLplu5iNXxekQfLixaGyc-g,502
2
+ vectorvein-0.1.20.dist-info/WHEEL,sha256=Yaoh9rlmnhBUmykFXwmFW_cmVAW3ZIWH8QkQVngJtyw,90
3
+ vectorvein-0.1.20.dist-info/entry_points.txt,sha256=6OYgBcLyFCUgeqLgnvMyOJxPCWzgy7se4rLPKtNonMs,34
3
4
  vectorvein/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
5
  vectorvein/chat_clients/__init__.py,sha256=lOGrIEBGN-EoxJ-dF5uMsO6viNCIFIeNL8whDwE6x3g,4657
5
6
  vectorvein/chat_clients/anthropic_client.py,sha256=VJQi7cKLbbLkJtmEFN9M41JUaGjwEuJaXIuQLK-3rdw,20332
6
7
  vectorvein/chat_clients/baichuan_client.py,sha256=CVMvpgjdrZGv0BWnTOBD-f2ufZ3wq3496wqukumsAr4,526
7
8
  vectorvein/chat_clients/base_client.py,sha256=AnzEmNfT4XW4lKcwf8fv2NlSk1EHK0evr4-1EH0KkwI,5018
8
9
  vectorvein/chat_clients/deepseek_client.py,sha256=3qWu01NlJAP2N-Ff62d5-CZXZitlizE1fzb20LNetig,526
9
- vectorvein/chat_clients/gemini_client.py,sha256=W-9Vu-GTE9wxStPznyNR0rBEgDG3LYBu2uQXd4sh1YQ,14425
10
+ vectorvein/chat_clients/gemini_client.py,sha256=YajhQdXnJ6h6Y9sA2dsxH7GF5-3c67QATU0UEyJP5eQ,14579
10
11
  vectorvein/chat_clients/groq_client.py,sha256=Uow4pgdmFi93ZQSoOol2-0PhhqkW-S0XuSldvppz5U4,498
11
12
  vectorvein/chat_clients/local_client.py,sha256=55nOsxzqUf79q3Y14MKROA71zxhsT7p7FsDZ89rts2M,422
12
- vectorvein/chat_clients/minimax_client.py,sha256=iNq87zWx79g8tGo784c67xUt-YQ4TyMoFWUIjDVEMGw,13801
13
+ vectorvein/chat_clients/minimax_client.py,sha256=xykptVg0qzOrJKuswHLG3M7r8H37jEMOb0gAnYAUdKQ,13947
13
14
  vectorvein/chat_clients/mistral_client.py,sha256=1aKSylzBDaLYcFnaBIL4-sXSzWmXfBeON9Q0rq-ziWw,534
14
15
  vectorvein/chat_clients/moonshot_client.py,sha256=gbu-6nGxx8uM_U2WlI4Wus881rFRotzHtMSoYOcruGU,526
15
16
  vectorvein/chat_clients/openai_client.py,sha256=Nz6tV45pWcsOupxjnsRsGTicbQNJWIZyxuJoJ5DGMpg,527
16
- vectorvein/chat_clients/openai_compatible_client.py,sha256=Rq_VTEddx6bXmr9TDKeI55mGUIos4THpRUgw1h3R4w4,15710
17
+ vectorvein/chat_clients/openai_compatible_client.py,sha256=NR88EgyLtXt1XiOKibZFtgMPb3_haTlKmsxoPvcYnb0,14130
17
18
  vectorvein/chat_clients/qwen_client.py,sha256=-ryh-m9PgsO0fc4ulcCmPTy1155J8YUy15uPoJQOHA0,513
18
19
  vectorvein/chat_clients/utils.py,sha256=EbSodMwLCR8wnzyt2J6iOsL8WvSzlwdu71xZikzNdlg,22483
19
20
  vectorvein/chat_clients/yi_client.py,sha256=RNf4CRuPJfixrwLZ3-DEc3t25QDe1mvZeb9sku2f8Bc,484
20
21
  vectorvein/chat_clients/zhipuai_client.py,sha256=Ys5DSeLCuedaDXr3PfG1EW2zKXopt-awO2IylWSwY0s,519
21
22
  vectorvein/settings/__init__.py,sha256=jVHbhHn1BuMcyfZGXrxWKiI4NdY9wzvYyGMvKYmUtqg,3378
22
- vectorvein/types/defaults.py,sha256=ljlRItl0QI9zQljPYPycE4FEpq7rb9FQ18VZt1hXbu0,15067
23
+ vectorvein/types/defaults.py,sha256=GaloEG4rNCcHz8DGRAF9XV0wHmIOiIqwI6sIzJHSY8s,15994
23
24
  vectorvein/types/enums.py,sha256=PNK_pTIyjJFy-yAG2PHaMIO1ey3W6fReMCkH8M8VRW4,1595
24
25
  vectorvein/types/llm_parameters.py,sha256=mmJjJZz4bPRi0nHzYNUNdWsQLHa9lbf3-MNVnU78vaY,3608
25
26
  vectorvein/utilities/media_processing.py,sha256=BujciRmw1GMmc3ELRvafL8STcy6r5b2rVnh27-uA7so,2256
26
27
  vectorvein/utilities/retry.py,sha256=9ePuJdeUUGx-qMWfaFxmlOvG_lQPwCQ4UB1z3Edlo34,993
27
- vectorvein-0.1.18.dist-info/RECORD,,
28
+ vectorvein-0.1.20.dist-info/RECORD,,
@@ -1,4 +1,4 @@
1
1
  Wheel-Version: 1.0
2
- Generator: pdm-backend (2.3.3)
2
+ Generator: pdm-backend (2.4.0)
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any
@@ -0,0 +1,4 @@
1
+ [console_scripts]
2
+
3
+ [gui_scripts]
4
+