webscout 7.7__py3-none-any.whl → 7.8__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of webscout might be problematic. Click here for more details.
- webscout/AIutel.py +2 -1
- webscout/Bard.py +14 -11
- webscout/DWEBS.py +431 -415
- webscout/Extra/autocoder/__init__.py +9 -9
- webscout/Extra/autocoder/autocoder_utiles.py +332 -194
- webscout/Extra/autocoder/rawdog.py +68 -9
- webscout/Extra/gguf.py +682 -682
- webscout/Provider/AI21.py +1 -1
- webscout/Provider/AISEARCH/DeepFind.py +2 -2
- webscout/Provider/AISEARCH/ISou.py +2 -2
- webscout/Provider/AISEARCH/felo_search.py +6 -6
- webscout/Provider/AISEARCH/genspark_search.py +1 -1
- webscout/Provider/Aitopia.py +292 -0
- webscout/Provider/AllenAI.py +1 -1
- webscout/Provider/Andi.py +3 -3
- webscout/Provider/C4ai.py +1 -1
- webscout/Provider/ChatGPTES.py +3 -5
- webscout/Provider/ChatGPTGratis.py +4 -4
- webscout/Provider/Chatify.py +2 -2
- webscout/Provider/Cloudflare.py +3 -2
- webscout/Provider/DARKAI.py +3 -2
- webscout/Provider/DeepSeek.py +2 -2
- webscout/Provider/Deepinfra.py +1 -1
- webscout/Provider/EDITEE.py +1 -1
- webscout/Provider/ElectronHub.py +178 -96
- webscout/Provider/ExaChat.py +310 -0
- webscout/Provider/Free2GPT.py +2 -2
- webscout/Provider/Gemini.py +5 -19
- webscout/Provider/GithubChat.py +1 -1
- webscout/Provider/Glider.py +4 -4
- webscout/Provider/Groq.py +3 -3
- webscout/Provider/HF_space/qwen_qwen2.py +1 -1
- webscout/Provider/HeckAI.py +1 -1
- webscout/Provider/HuggingFaceChat.py +1 -1
- webscout/Provider/Hunyuan.py +1 -1
- webscout/Provider/Jadve.py +3 -3
- webscout/Provider/Koboldai.py +3 -3
- webscout/Provider/LambdaChat.py +1 -1
- webscout/Provider/Llama.py +3 -5
- webscout/Provider/Llama3.py +4 -12
- webscout/Provider/Marcus.py +3 -3
- webscout/Provider/OLLAMA.py +8 -8
- webscout/Provider/Openai.py +7 -3
- webscout/Provider/PI.py +1 -1
- webscout/Provider/Perplexitylabs.py +1 -1
- webscout/Provider/Phind.py +1 -1
- webscout/Provider/PizzaGPT.py +1 -1
- webscout/Provider/QwenLM.py +4 -7
- webscout/Provider/TTI/FreeAIPlayground/async_freeaiplayground.py +3 -1
- webscout/Provider/TTI/FreeAIPlayground/sync_freeaiplayground.py +3 -3
- webscout/Provider/TTI/ImgSys/__init__.py +23 -0
- webscout/Provider/TTI/ImgSys/async_imgsys.py +202 -0
- webscout/Provider/TTI/ImgSys/sync_imgsys.py +195 -0
- webscout/Provider/TTI/__init__.py +3 -1
- webscout/Provider/TTI/artbit/async_artbit.py +1 -1
- webscout/Provider/TTI/artbit/sync_artbit.py +1 -1
- webscout/Provider/TTI/huggingface/async_huggingface.py +1 -1
- webscout/Provider/TTI/huggingface/sync_huggingface.py +1 -1
- webscout/Provider/TTI/piclumen/__init__.py +22 -22
- webscout/Provider/TTI/piclumen/sync_piclumen.py +232 -232
- webscout/Provider/TTI/pixelmuse/__init__.py +4 -0
- webscout/Provider/TTI/pixelmuse/async_pixelmuse.py +249 -0
- webscout/Provider/TTI/pixelmuse/sync_pixelmuse.py +182 -0
- webscout/Provider/TTI/talkai/sync_talkai.py +1 -1
- webscout/Provider/TTS/utils.py +1 -1
- webscout/Provider/TeachAnything.py +1 -1
- webscout/Provider/TextPollinationsAI.py +4 -4
- webscout/Provider/TwoAI.py +1 -2
- webscout/Provider/Venice.py +4 -2
- webscout/Provider/VercelAI.py +234 -0
- webscout/Provider/WebSim.py +3 -2
- webscout/Provider/WiseCat.py +10 -12
- webscout/Provider/Youchat.py +1 -1
- webscout/Provider/__init__.py +10 -0
- webscout/Provider/ai4chat.py +1 -1
- webscout/Provider/aimathgpt.py +2 -6
- webscout/Provider/akashgpt.py +1 -1
- webscout/Provider/askmyai.py +4 -4
- webscout/Provider/asksteve.py +203 -0
- webscout/Provider/bagoodex.py +2 -2
- webscout/Provider/cerebras.py +1 -1
- webscout/Provider/chatglm.py +4 -4
- webscout/Provider/cleeai.py +1 -0
- webscout/Provider/copilot.py +427 -415
- webscout/Provider/elmo.py +1 -1
- webscout/Provider/flowith.py +1 -1
- webscout/Provider/freeaichat.py +57 -31
- webscout/Provider/gaurish.py +3 -5
- webscout/Provider/geminiprorealtime.py +1 -1
- webscout/Provider/granite.py +4 -4
- webscout/Provider/hermes.py +5 -5
- webscout/Provider/julius.py +1 -1
- webscout/Provider/koala.py +1 -1
- webscout/Provider/lepton.py +1 -1
- webscout/Provider/llama3mitril.py +4 -4
- webscout/Provider/llamatutor.py +1 -1
- webscout/Provider/llmchat.py +3 -3
- webscout/Provider/meta.py +1 -1
- webscout/Provider/multichat.py +10 -10
- webscout/Provider/promptrefine.py +1 -1
- webscout/Provider/searchchat.py +293 -0
- webscout/Provider/sonus.py +2 -2
- webscout/Provider/talkai.py +2 -2
- webscout/Provider/turboseek.py +1 -1
- webscout/Provider/tutorai.py +1 -1
- webscout/Provider/typegpt.py +5 -42
- webscout/Provider/uncovr.py +4 -2
- webscout/Provider/x0gpt.py +1 -1
- webscout/__init__.py +36 -36
- webscout/cli.py +293 -332
- webscout/tempid.py +11 -11
- webscout/utils.py +2 -2
- webscout/version.py +1 -1
- webscout/webscout_search.py +1282 -1223
- webscout/webscout_search_async.py +813 -692
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/METADATA +50 -29
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/RECORD +121 -110
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/LICENSE.md +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/WHEEL +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/entry_points.txt +0 -0
- {webscout-7.7.dist-info → webscout-7.8.dist-info}/top_level.txt +0 -0
webscout/Provider/ElectronHub.py
CHANGED
|
@@ -8,7 +8,7 @@ from webscout.AIutel import Conversation
|
|
|
8
8
|
from webscout.AIutel import AwesomePrompts, sanitize_stream
|
|
9
9
|
from webscout.AIbase import Provider, AsyncProvider
|
|
10
10
|
from webscout import exceptions
|
|
11
|
-
from webscout import LitAgent
|
|
11
|
+
from webscout.litagent import LitAgent
|
|
12
12
|
|
|
13
13
|
class ElectronHub(Provider):
|
|
14
14
|
"""
|
|
@@ -16,6 +16,79 @@ class ElectronHub(Provider):
|
|
|
16
16
|
"""
|
|
17
17
|
|
|
18
18
|
AVAILABLE_MODELS = [
|
|
19
|
+
# DeepSeek models
|
|
20
|
+
"deepseek-coder-6.7b-base-awq",
|
|
21
|
+
"deepseek-coder-6.7b-instruct-awq",
|
|
22
|
+
"deepseek-math-7b-instruct",
|
|
23
|
+
"deepseek-r1-distill-qwen-32b",
|
|
24
|
+
|
|
25
|
+
# DiscoLM models
|
|
26
|
+
"discolm-german-7b-v1-awq",
|
|
27
|
+
|
|
28
|
+
# Falcon models
|
|
29
|
+
"falcon-7b-instruct",
|
|
30
|
+
|
|
31
|
+
# Gemma models
|
|
32
|
+
"gemma-7b-it",
|
|
33
|
+
|
|
34
|
+
# Hermes models
|
|
35
|
+
"hermes-2-pro-mistral-7b",
|
|
36
|
+
|
|
37
|
+
# Llama models
|
|
38
|
+
"llama-2-13b-chat-awq",
|
|
39
|
+
"llama-2-7b-chat-fp16",
|
|
40
|
+
"llama-2-7b-chat-int8",
|
|
41
|
+
"llama-3-8b-instruct",
|
|
42
|
+
"llama-3-8b-instruct-awq",
|
|
43
|
+
"llama-3.1-8b-instruct",
|
|
44
|
+
"llama-3.1-8b-instruct-awq",
|
|
45
|
+
"llama-3.1-8b-instruct-fp8",
|
|
46
|
+
"llama-3.2-11b-vision-instruct",
|
|
47
|
+
"llama-3.2-1b-instruct",
|
|
48
|
+
"llama-3.2-3b-instruct",
|
|
49
|
+
"llama-3.3-70b-instruct-fp8-fast",
|
|
50
|
+
"llama-guard-3-8b",
|
|
51
|
+
"llamaguard-7b-awq",
|
|
52
|
+
"meta-llama-3-8b-instruct",
|
|
53
|
+
|
|
54
|
+
# Mistral models
|
|
55
|
+
"mistral-7b-instruct-v0.1",
|
|
56
|
+
"mistral-7b-instruct-v0.1-awq",
|
|
57
|
+
"mistral-7b-instruct-v0.2",
|
|
58
|
+
|
|
59
|
+
# Neural Chat models
|
|
60
|
+
"neural-chat-7b-v3-1-awq",
|
|
61
|
+
|
|
62
|
+
# OpenChat models
|
|
63
|
+
"openchat-3.5-0106",
|
|
64
|
+
|
|
65
|
+
# OpenHermes models
|
|
66
|
+
"openhermes-2.5-mistral-7b-awq",
|
|
67
|
+
|
|
68
|
+
# Phi models
|
|
69
|
+
"phi-2",
|
|
70
|
+
|
|
71
|
+
# Qwen models
|
|
72
|
+
"qwen1.5-0.5b-chat",
|
|
73
|
+
"qwen1.5-1.8b-chat",
|
|
74
|
+
"qwen1.5-14b-chat-awq",
|
|
75
|
+
"qwen1.5-7b-chat-awq",
|
|
76
|
+
|
|
77
|
+
# SQLCoder models
|
|
78
|
+
"sqlcoder-7b-2",
|
|
79
|
+
|
|
80
|
+
# Starling models
|
|
81
|
+
"starling-lm-7b-beta",
|
|
82
|
+
|
|
83
|
+
# TinyLlama models
|
|
84
|
+
"tinyllama-1.1b-chat-v1.0",
|
|
85
|
+
|
|
86
|
+
# UNA models
|
|
87
|
+
"una-cybertron-7b-v2-bf16",
|
|
88
|
+
|
|
89
|
+
# Zephyr models
|
|
90
|
+
"zephyr-7b-beta-awq",
|
|
91
|
+
|
|
19
92
|
# OpenAI GPT models
|
|
20
93
|
"gpt-3.5-turbo",
|
|
21
94
|
"gpt-3.5-turbo-16k",
|
|
@@ -30,8 +103,12 @@ class ElectronHub(Provider):
|
|
|
30
103
|
"gpt-4o-2024-05-13",
|
|
31
104
|
"gpt-4o-2024-08-06",
|
|
32
105
|
"gpt-4o-2024-11-20",
|
|
106
|
+
"gpt-4o-search-preview",
|
|
107
|
+
"gpt-4o-search-preview-2025-03-11",
|
|
33
108
|
"gpt-4o-mini",
|
|
34
109
|
"gpt-4o-mini-2024-07-18",
|
|
110
|
+
"gpt-4o-mini-search-preview",
|
|
111
|
+
"gpt-4o-mini-search-preview-2025-03-11",
|
|
35
112
|
"chatgpt-4o-latest",
|
|
36
113
|
"gpt-4.5-preview",
|
|
37
114
|
"gpt-4.5-preview-2025-02-27",
|
|
@@ -48,17 +125,24 @@ class ElectronHub(Provider):
|
|
|
48
125
|
# Anthropic Claude models
|
|
49
126
|
"claude-2",
|
|
50
127
|
"claude-2.1",
|
|
128
|
+
"claude-3-haiku-20240307",
|
|
129
|
+
"claude-3-5-haiku-20241022",
|
|
51
130
|
"claude-3-opus-20240229",
|
|
52
131
|
"claude-3-sonnet-20240229",
|
|
53
|
-
"claude-3-haiku-20240307",
|
|
54
132
|
"claude-3-5-sonnet-20240620",
|
|
55
133
|
"claude-3-5-sonnet-20241022",
|
|
56
|
-
"claude-3-5-haiku-20241022",
|
|
57
134
|
"claude-3-7-sonnet-20250219",
|
|
58
135
|
"claude-3-7-sonnet-20250219-thinking",
|
|
136
|
+
"claude-3-opus-20240229:safe",
|
|
137
|
+
"claude-3-sonnet-20240229:safe",
|
|
138
|
+
"claude-3-5-sonnet-20240620:safe",
|
|
139
|
+
"claude-3-5-sonnet-20241022:safe",
|
|
140
|
+
"claude-3-7-sonnet-20250219:safe",
|
|
141
|
+
"claude-3-7-sonnet-20250219-thinking:safe",
|
|
59
142
|
|
|
60
143
|
# Google Gemini models
|
|
61
144
|
"gemini-1.0-pro",
|
|
145
|
+
"gemini-1.0-pro-vision",
|
|
62
146
|
"gemini-1.5-pro",
|
|
63
147
|
"gemini-1.5-pro-latest",
|
|
64
148
|
"gemini-1.5-flash-8b",
|
|
@@ -76,6 +160,7 @@ class ElectronHub(Provider):
|
|
|
76
160
|
"gemini-2.0-flash-lite-preview-02-05",
|
|
77
161
|
"gemini-2.0-flash-lite-001",
|
|
78
162
|
"gemini-2.0-pro-exp-02-05",
|
|
163
|
+
"gemini-2.5-pro-exp-03-25",
|
|
79
164
|
|
|
80
165
|
# Google PaLM models
|
|
81
166
|
"palm-2-chat-bison",
|
|
@@ -84,31 +169,34 @@ class ElectronHub(Provider):
|
|
|
84
169
|
"palm-2-codechat-bison-32k",
|
|
85
170
|
|
|
86
171
|
# Meta Llama models
|
|
87
|
-
"llama-2-13b-chat",
|
|
88
172
|
"llama-2-70b-chat",
|
|
89
|
-
"llama-guard-3-8b",
|
|
90
|
-
"code-llama-34b-instruct",
|
|
91
|
-
"llama-3-8b",
|
|
92
173
|
"llama-3-70b",
|
|
93
|
-
"llama-3.1-8b",
|
|
94
174
|
"llama-3.1-70b",
|
|
95
175
|
"llama-3.1-405b",
|
|
96
|
-
"llama-3.2-1b",
|
|
97
|
-
"llama-3.2-3b",
|
|
98
|
-
"llama-3.2-11b",
|
|
99
176
|
"llama-3.2-90b",
|
|
100
|
-
"llama-3.3-70b-instruct",
|
|
101
177
|
"llama-3.1-nemotron-70b-instruct",
|
|
102
|
-
"llama-3.1-tulu-3-8b",
|
|
103
178
|
"llama-3.1-tulu-3-70b",
|
|
104
179
|
"llama-3.1-tulu-3-405b",
|
|
180
|
+
"llama-3.1-sonar-small-128k-online",
|
|
181
|
+
"llama-3.1-sonar-large-128k-online",
|
|
182
|
+
"llama-3.1-sonar-huge-128k-online",
|
|
183
|
+
"llama-3.1-sonar-small-128k-chat",
|
|
184
|
+
"llama-3.1-sonar-large-128k-chat",
|
|
185
|
+
"llama-3.1-swallow-70b-instruct-v0.3",
|
|
186
|
+
"llama-3.1-8b-lexi-uncensored-v2",
|
|
187
|
+
"llama-3.1-lumimaid-8b",
|
|
188
|
+
"llama-3.1-lumimaid-70b",
|
|
189
|
+
"llama3-openbiollm-70b",
|
|
190
|
+
|
|
191
|
+
# Mixtral models
|
|
192
|
+
"mixtral-8x7b",
|
|
193
|
+
"mixtral-8x22b",
|
|
105
194
|
|
|
106
|
-
# Mistral models
|
|
107
|
-
"mistral-7b-instruct",
|
|
108
195
|
"mistral-tiny-latest",
|
|
109
196
|
"mistral-tiny",
|
|
110
197
|
"mistral-tiny-2312",
|
|
111
198
|
"mistral-tiny-2407",
|
|
199
|
+
"mistral-small-3.1-24b-instruct",
|
|
112
200
|
"mistral-small-24b-instruct-2501",
|
|
113
201
|
"mistral-small-latest",
|
|
114
202
|
"mistral-small",
|
|
@@ -122,89 +210,39 @@ class ElectronHub(Provider):
|
|
|
122
210
|
"mistral-large-2411",
|
|
123
211
|
"mistral-large-2407",
|
|
124
212
|
"mistral-large-2402",
|
|
125
|
-
|
|
126
|
-
# Mixtral models
|
|
127
|
-
"mixtral-8x7b",
|
|
128
|
-
"mixtral-8x22b",
|
|
129
|
-
|
|
130
|
-
# DeepSeek models
|
|
131
|
-
"deepseek-r1",
|
|
132
|
-
"deepseek-r1-nitro",
|
|
133
|
-
"deepseek-r1-distill-llama-8b",
|
|
134
|
-
"deepseek-r1-distill-llama-70b",
|
|
135
|
-
"deepseek-r1-distill-qwen-1.5b",
|
|
136
|
-
"deepseek-r1-distill-qwen-7b",
|
|
137
|
-
"deepseek-r1-distill-qwen-14b",
|
|
138
|
-
"deepseek-r1-distill-qwen-32b",
|
|
139
|
-
"deepseek-v3",
|
|
140
|
-
"deepseek-coder",
|
|
141
|
-
"deepseek-v2.5",
|
|
142
|
-
"deepseek-vl2",
|
|
143
|
-
"deepseek-llm-67b-chat",
|
|
144
|
-
"deepseek-math-7b-instruct",
|
|
145
|
-
"deepseek-coder-6.7b-base-awq",
|
|
146
|
-
"deepseek-coder-6.7b-instruct-awq",
|
|
147
|
-
|
|
148
|
-
# Qwen models
|
|
149
|
-
"qwen-1.5-0.5b-chat",
|
|
150
|
-
"qwen-1.5-1.8b-chat",
|
|
151
|
-
"qwen-1.5-14b-chat-awq",
|
|
152
|
-
"qwen-1.5-7b-chat-awq",
|
|
153
|
-
"qwen-2-7b-instruct",
|
|
154
|
-
"qwen-2-72b-instruct",
|
|
155
|
-
"qwen-2-vl-7b-instruct",
|
|
156
|
-
"qwen-2-vl-72b-instruct",
|
|
157
|
-
"qwen-2.5-7b-instruct",
|
|
158
|
-
"qwen-2.5-32b-instruct",
|
|
159
|
-
"qwen-2.5-72b-instruct",
|
|
160
|
-
"qwen-2.5-coder-32b-instruct",
|
|
161
|
-
"qwq-32b-preview",
|
|
162
|
-
"qvq-72b-preview",
|
|
163
|
-
"qwen-vl-plus",
|
|
164
|
-
"qwen2.5-vl-72b-instruct",
|
|
165
|
-
"qwen-turbo",
|
|
166
|
-
"qwen-plus",
|
|
167
|
-
"qwen-max",
|
|
168
|
-
|
|
169
|
-
# Microsoft models
|
|
170
|
-
"phi-4",
|
|
171
|
-
"phi-3.5-mini-128k-instruct",
|
|
172
|
-
"phi-3-medium-128k-instruct",
|
|
173
|
-
"phi-3-mini-128k-instruct",
|
|
174
|
-
"phi-2",
|
|
175
|
-
|
|
176
|
-
# Gemma models
|
|
177
|
-
"gemma-7b-it",
|
|
178
|
-
"gemma-2-9b-it",
|
|
179
|
-
"gemma-2-27b-it",
|
|
180
|
-
|
|
181
|
-
# Various other models
|
|
182
|
-
"nemotron-4-340b",
|
|
183
|
-
"pixtral-large-2411",
|
|
184
|
-
"pixtral-12b",
|
|
185
213
|
"open-mistral-nemo",
|
|
186
214
|
"open-mistral-nemo-2407",
|
|
187
215
|
"open-mixtral-8x22b-2404",
|
|
188
216
|
"open-mixtral-8x7b",
|
|
217
|
+
|
|
218
|
+
# Codestral models
|
|
189
219
|
"codestral-mamba",
|
|
190
220
|
"codestral-latest",
|
|
191
221
|
"codestral-2405",
|
|
192
222
|
"codestral-2412",
|
|
193
223
|
"codestral-2501",
|
|
194
|
-
|
|
224
|
+
|
|
225
|
+
# Ministral models
|
|
195
226
|
"ministral-3b",
|
|
196
227
|
"ministral-3b-2410",
|
|
197
228
|
"ministral-8b",
|
|
198
229
|
"ministral-8b-2410",
|
|
230
|
+
|
|
231
|
+
# Mistral Saba models
|
|
199
232
|
"mistral-saba-latest",
|
|
200
233
|
"mistral-saba-2502",
|
|
234
|
+
|
|
235
|
+
# F1 models
|
|
201
236
|
"f1-mini-preview",
|
|
202
237
|
"f1-preview",
|
|
238
|
+
|
|
239
|
+
# Dolphin models
|
|
203
240
|
"dolphin-mixtral-8x7b",
|
|
204
241
|
"dolphin-mixtral-8x22b",
|
|
205
242
|
"dolphin3.0-mistral-24b",
|
|
206
243
|
"dolphin3.0-r1-mistral-24b",
|
|
207
|
-
|
|
244
|
+
|
|
245
|
+
# Command models
|
|
208
246
|
"command",
|
|
209
247
|
"command-light",
|
|
210
248
|
"command-nightly",
|
|
@@ -216,10 +254,18 @@ class ElectronHub(Provider):
|
|
|
216
254
|
"command-r-plus-04-2024",
|
|
217
255
|
"command-r-plus-08-2024",
|
|
218
256
|
"command-r7b-12-2024",
|
|
257
|
+
"command-a-03-2025",
|
|
258
|
+
|
|
259
|
+
# Other AI models
|
|
260
|
+
"nemotron-4-340b",
|
|
261
|
+
"pixtral-large-2411",
|
|
262
|
+
"pixtral-12b",
|
|
263
|
+
"dbrx-instruct",
|
|
219
264
|
"c4ai-aya-expanse-8b",
|
|
220
265
|
"c4ai-aya-expanse-32b",
|
|
221
266
|
"reka-flash",
|
|
222
267
|
"reka-core",
|
|
268
|
+
"reka-flash-3",
|
|
223
269
|
"grok-2",
|
|
224
270
|
"grok-2-mini",
|
|
225
271
|
"grok-beta",
|
|
@@ -234,19 +280,45 @@ class ElectronHub(Provider):
|
|
|
234
280
|
"sonar-reasoning",
|
|
235
281
|
"sonar-pro",
|
|
236
282
|
"sonar",
|
|
237
|
-
"
|
|
238
|
-
"
|
|
239
|
-
"
|
|
240
|
-
"
|
|
241
|
-
"
|
|
283
|
+
"phi-4",
|
|
284
|
+
"phi-4-multimodal-instruct",
|
|
285
|
+
"phi-3.5-mini-128k-instruct",
|
|
286
|
+
"phi-3-medium-128k-instruct",
|
|
287
|
+
"phi-3-mini-128k-instruct",
|
|
242
288
|
"wizardlm-2-7b",
|
|
243
289
|
"wizardlm-2-8x22b",
|
|
244
290
|
"minimax-01",
|
|
245
291
|
"jamba-1.5-large",
|
|
246
292
|
"jamba-1.5-mini",
|
|
293
|
+
"jamba-1.6-large",
|
|
294
|
+
"jamba-1.6-mini",
|
|
247
295
|
"jamba-instruct",
|
|
248
|
-
"openchat-3.5-7b",
|
|
249
296
|
"openchat-3.6-8b",
|
|
297
|
+
|
|
298
|
+
# Qwen models
|
|
299
|
+
"qwen-1.5-0.5b-chat",
|
|
300
|
+
"qwen-1.5-1.8b-chat",
|
|
301
|
+
"qwen-1.5-14b-chat-awq",
|
|
302
|
+
"qwen-1.5-7b-chat-awq",
|
|
303
|
+
"qwen-2-7b-instruct",
|
|
304
|
+
"qwen-2-72b-instruct",
|
|
305
|
+
"qwen-2-vl-7b-instruct",
|
|
306
|
+
"qwen-2-vl-72b-instruct",
|
|
307
|
+
"qwen-2.5-7b-instruct",
|
|
308
|
+
"qwen-2.5-32b-instruct",
|
|
309
|
+
"qwen-2.5-72b-instruct",
|
|
310
|
+
"qwen-2.5-coder-32b-instruct",
|
|
311
|
+
"qwq-32b-preview",
|
|
312
|
+
"qwq-32b",
|
|
313
|
+
"qwen-vl-plus",
|
|
314
|
+
"qwen2.5-vl-3b-instruct",
|
|
315
|
+
"qwen2.5-vl-7b-instruct",
|
|
316
|
+
"qwen2.5-vl-72b-instruct",
|
|
317
|
+
"qwen-turbo",
|
|
318
|
+
"qwen-plus",
|
|
319
|
+
"qwen-max",
|
|
320
|
+
|
|
321
|
+
# Other models
|
|
250
322
|
"aion-1.0",
|
|
251
323
|
"aion-1.0-mini",
|
|
252
324
|
"aion-rp-llama-3.1-8b",
|
|
@@ -258,9 +330,7 @@ class ElectronHub(Provider):
|
|
|
258
330
|
"mytho-max-l2-13b",
|
|
259
331
|
"deephermes-3-llama-3-8b-preview",
|
|
260
332
|
"nous-hermes-llama2-13b",
|
|
261
|
-
"hermes-3-llama-3.1-8b",
|
|
262
333
|
"hermes-3-llama-3.1-405b",
|
|
263
|
-
"hermes-2-pro-llama-3-8b",
|
|
264
334
|
"nous-hermes-2-mixtral-8x7b-dpo",
|
|
265
335
|
|
|
266
336
|
# Chinese models
|
|
@@ -268,6 +338,7 @@ class ElectronHub(Provider):
|
|
|
268
338
|
"doubao-lite-32k",
|
|
269
339
|
"doubao-pro-4k",
|
|
270
340
|
"doubao-pro-32k",
|
|
341
|
+
"ui-tars-72b-dpo",
|
|
271
342
|
"ernie-lite-8k",
|
|
272
343
|
"ernie-tiny-8k",
|
|
273
344
|
"ernie-speed-8k",
|
|
@@ -288,9 +359,10 @@ class ElectronHub(Provider):
|
|
|
288
359
|
"yi-large-rag",
|
|
289
360
|
"yi-medium",
|
|
290
361
|
"yi-34b-chat-200k",
|
|
362
|
+
"moonlight-16b-a3b-instruct",
|
|
291
363
|
"spark-desk-v1.5",
|
|
292
364
|
|
|
293
|
-
#
|
|
365
|
+
# Additional models
|
|
294
366
|
"step-2-16k-exp-202412",
|
|
295
367
|
"granite-3.1-2b-instruct",
|
|
296
368
|
"granite-3.1-8b-instruct",
|
|
@@ -299,42 +371,52 @@ class ElectronHub(Provider):
|
|
|
299
371
|
"unslopnemo-12b",
|
|
300
372
|
"rocinante-12b-v1.1",
|
|
301
373
|
"rocinante-12b-v1",
|
|
374
|
+
"olympiccoder-7b",
|
|
375
|
+
"olympiccoder-32b",
|
|
376
|
+
"anubis-pro-105b-v1",
|
|
377
|
+
"fallen-llama-3.3-r1-70b-v1",
|
|
378
|
+
"skyfall-36b-v2",
|
|
379
|
+
"wayfarer-large-70b-llama-3.3",
|
|
380
|
+
"qwq-32b-snowdrop-v0",
|
|
381
|
+
"qwq-32b-abliterated",
|
|
302
382
|
"sky-t1-32b-preview",
|
|
383
|
+
"tiny-r1-32b-preview",
|
|
303
384
|
"lfm-3b",
|
|
304
385
|
"lfm-7b",
|
|
305
386
|
"lfm-40b",
|
|
306
|
-
"rogue-rose-103b-v0.2",
|
|
307
387
|
"eva-llama-3.33-70b-v0.0",
|
|
308
388
|
"eva-llama-3.33-70b-v0.1",
|
|
309
389
|
"eva-qwen2.5-72b",
|
|
310
390
|
"eva-qwen2.5-32b-v0.2",
|
|
311
391
|
"sorcererlm-8x22b",
|
|
312
392
|
"mythalion-13b",
|
|
313
|
-
"zephyr-7b-beta",
|
|
314
|
-
"zephyr-7b-alpha",
|
|
315
393
|
"toppy-m-7b",
|
|
316
394
|
"openhermes-2.5-mistral-7b",
|
|
317
395
|
"l3-lunaris-8b",
|
|
318
|
-
"llama-3.1-lumimaid-8b",
|
|
319
|
-
"llama-3.1-lumimaid-70b",
|
|
320
|
-
"llama-3-lumimaid-8b",
|
|
321
|
-
"llama-3-lumimaid-70b",
|
|
322
|
-
"llama3-openbiollm-70b",
|
|
323
396
|
"l3.1-70b-hanami-x1",
|
|
397
|
+
"lumimaid-magnum-v4-12b",
|
|
324
398
|
"magnum-v4-72b",
|
|
399
|
+
"magnum-v4-12b",
|
|
400
|
+
"magnum-v3-34b",
|
|
401
|
+
"magnum-v2.5-12b-kto",
|
|
325
402
|
"magnum-v2-72b",
|
|
403
|
+
"magnum-v2-32b",
|
|
404
|
+
"magnum-v2-12b",
|
|
326
405
|
"magnum-72b",
|
|
327
406
|
"mini-magnum-12b-v1.1",
|
|
328
407
|
"remm-slerp-l2-13b",
|
|
408
|
+
"patricide-12b-unslop-mell",
|
|
329
409
|
"midnight-rose-70b",
|
|
330
|
-
"
|
|
410
|
+
"airoboros-l2-13b-gpt4-m2.0",
|
|
331
411
|
"airoboros-l2-70b",
|
|
332
412
|
"xwin-lm-70b",
|
|
333
413
|
"noromaid-20b",
|
|
334
414
|
"violet-twilight-v0.2",
|
|
335
415
|
"saiga-nemo-12b",
|
|
336
416
|
"l3-8b-stheno-v3.2",
|
|
337
|
-
"
|
|
417
|
+
"l3.3-electra-r1-70b",
|
|
418
|
+
"l3.3-cu-mai-r1-70b",
|
|
419
|
+
"l3.3-mokume-gane-r1-70b-v1.1",
|
|
338
420
|
"l3.3-70b-euryale-v2.3",
|
|
339
421
|
"l3.3-ms-evayale-70b",
|
|
340
422
|
"70b-l3.3-cirrus-x1",
|