janito 2.17.1__py3-none-any.whl → 2.19.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- janito/cli/chat_mode/session.py +32 -14
- janito/cli/cli_commands/list_models.py +16 -2
- janito/cli/cli_commands/list_providers.py +6 -1
- janito/cli/cli_commands/model_utils.py +88 -68
- janito/cli/cli_commands/ping_providers.py +55 -0
- janito/cli/main_cli.py +18 -0
- janito/llm/model.py +1 -0
- janito/providers/alibaba/model_info.py +57 -0
- janito/providers/alibaba/provider.py +1 -1
- janito/providers/openai/model_info.py +33 -0
- janito/providers/openai/provider.py +1 -1
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/METADATA +1 -1
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/RECORD +17 -16
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/WHEEL +0 -0
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/entry_points.txt +0 -0
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/licenses/LICENSE +0 -0
- {janito-2.17.1.dist-info → janito-2.19.0.dist-info}/top_level.txt +0 -0
janito/cli/chat_mode/session.py
CHANGED
@@ -296,21 +296,39 @@ class ChatSession:
|
|
296
296
|
else "Unknown"
|
297
297
|
)
|
298
298
|
|
299
|
-
# Get backend hostname if available
|
300
299
|
backend_hostname = "Unknown"
|
301
|
-
|
302
|
-
|
303
|
-
|
304
|
-
)
|
305
|
-
|
306
|
-
|
307
|
-
|
308
|
-
|
309
|
-
|
310
|
-
|
311
|
-
|
312
|
-
|
313
|
-
|
300
|
+
candidates = []
|
301
|
+
drv = getattr(self.agent, "driver", None)
|
302
|
+
if drv is not None:
|
303
|
+
cfg = getattr(drv, "config", None)
|
304
|
+
if cfg is not None:
|
305
|
+
b = getattr(cfg, "base_url", None)
|
306
|
+
if b:
|
307
|
+
candidates.append(b)
|
308
|
+
direct_base = getattr(drv, "base_url", None)
|
309
|
+
if direct_base:
|
310
|
+
candidates.append(direct_base)
|
311
|
+
cfg2 = getattr(self.agent, "config", None)
|
312
|
+
if cfg2 is not None:
|
313
|
+
b2 = getattr(cfg2, "base_url", None)
|
314
|
+
if b2:
|
315
|
+
candidates.append(b2)
|
316
|
+
top_base = getattr(self.agent, "base_url", None)
|
317
|
+
if top_base:
|
318
|
+
candidates.append(top_base)
|
319
|
+
from urllib.parse import urlparse
|
320
|
+
for candidate in candidates:
|
321
|
+
try:
|
322
|
+
if not candidate:
|
323
|
+
continue
|
324
|
+
parsed = urlparse(str(candidate))
|
325
|
+
host = parsed.netloc or parsed.path
|
326
|
+
if host:
|
327
|
+
backend_hostname = host
|
328
|
+
break
|
329
|
+
except Exception:
|
330
|
+
backend_hostname = str(candidate)
|
331
|
+
break
|
314
332
|
|
315
333
|
self.console.print(
|
316
334
|
Rule(
|
@@ -27,9 +27,23 @@ def handle_list_models(args, provider_instance):
|
|
27
27
|
if models and isinstance(models[0], dict):
|
28
28
|
_print_models_table(models, provider_name)
|
29
29
|
else:
|
30
|
-
|
30
|
+
# Fallback for simple string model lists
|
31
|
+
from rich.table import Table
|
32
|
+
from janito.cli.console import shared_console
|
33
|
+
|
34
|
+
table = Table(title=f"Supported models for provider '{provider_name}'")
|
35
|
+
table.add_column("Model Name", style="cyan")
|
36
|
+
|
31
37
|
for m in models:
|
32
|
-
|
38
|
+
table.add_row(str(m))
|
39
|
+
|
40
|
+
import sys
|
41
|
+
if sys.stdout.isatty():
|
42
|
+
shared_console.print(table)
|
43
|
+
else:
|
44
|
+
print(f"Supported models for provider '{provider_name}':")
|
45
|
+
for m in models:
|
46
|
+
print(f"- {m}")
|
33
47
|
except Exception as e:
|
34
48
|
print(f"Error listing models for provider '{provider_name}': {e}")
|
35
49
|
return
|
@@ -3,8 +3,13 @@ CLI Command: List supported LLM providers
|
|
3
3
|
"""
|
4
4
|
|
5
5
|
from janito.provider_registry import list_providers
|
6
|
+
from janito.cli.cli_commands.ping_providers import handle_ping_providers
|
6
7
|
|
7
8
|
|
8
9
|
def handle_list_providers(args=None):
|
9
|
-
|
10
|
+
# Check if ping flag is set
|
11
|
+
if args and getattr(args, 'ping', False):
|
12
|
+
handle_ping_providers(args)
|
13
|
+
else:
|
14
|
+
list_providers()
|
10
15
|
return
|
@@ -5,59 +5,94 @@ Utilities for model-related CLI output
|
|
5
5
|
|
6
6
|
def _print_models_table(models, provider_name):
|
7
7
|
from rich.table import Table
|
8
|
-
from
|
8
|
+
from janito.cli.console import shared_console
|
9
9
|
|
10
|
-
headers = [
|
11
|
-
"name",
|
12
|
-
"open",
|
13
|
-
"context",
|
14
|
-
"max_input",
|
15
|
-
"max_cot",
|
16
|
-
"max_response",
|
17
|
-
"thinking_supported",
|
18
|
-
"driver",
|
19
|
-
]
|
20
|
-
display_headers = [
|
21
|
-
"Model Name",
|
22
|
-
"Vendor",
|
23
|
-
"context",
|
24
|
-
"max_input",
|
25
|
-
"max_cot",
|
26
|
-
"max_response",
|
27
|
-
"Thinking",
|
28
|
-
"Driver",
|
29
|
-
]
|
30
10
|
table = Table(title=f"Supported models for provider '{provider_name}'")
|
31
|
-
|
32
|
-
|
11
|
+
table.add_column("Model Name", style="cyan")
|
12
|
+
table.add_column("Vendor", style="yellow", justify="center")
|
13
|
+
table.add_column("Context", style="magenta", justify="center")
|
14
|
+
table.add_column("Max Input", style="green", justify="center")
|
15
|
+
table.add_column("CoT", style="blue", justify="center")
|
16
|
+
table.add_column("Max Response", style="red", justify="center")
|
17
|
+
table.add_column("Thinking", style="bright_black", justify="center")
|
18
|
+
table.add_column("Driver", style="white")
|
19
|
+
|
20
|
+
# Get default model for this provider
|
21
|
+
from janito.providers.registry import LLMProviderRegistry
|
22
|
+
try:
|
23
|
+
provider_class = LLMProviderRegistry.get(provider_name)
|
24
|
+
default_model = getattr(provider_class, "DEFAULT_MODEL", None)
|
25
|
+
except:
|
26
|
+
default_model = None
|
27
|
+
|
33
28
|
for m in models:
|
34
|
-
|
35
|
-
|
36
|
-
|
37
|
-
|
29
|
+
name = str(m.get("name", ""))
|
30
|
+
|
31
|
+
# Highlight default model with different color
|
32
|
+
if name == default_model:
|
33
|
+
name = f"[bold green]⭐ {name}[/bold green]"
|
34
|
+
|
35
|
+
vendor = "Open" if m.get("open") is True or m.get("open") == "Open" else "Locked"
|
36
|
+
|
37
|
+
context = _format_context(m.get("context", ""))
|
38
|
+
max_input = _format_k(m.get("max_input", ""))
|
39
|
+
max_cot = _format_k(m.get("max_cot", ""))
|
40
|
+
max_response = _format_k(m.get("max_response", ""))
|
41
|
+
|
42
|
+
# Determine thinking indicators
|
43
|
+
thinking_supported = m.get("thinking_supported") is True or m.get("thinking_supported") == "True"
|
44
|
+
cot_value = m.get("max_cot", "")
|
45
|
+
|
46
|
+
thinking_icon = "📖" if thinking_supported and m.get("thinking", False) else ""
|
47
|
+
# Only show CoT value if it's a valid number and thinking is supported
|
48
|
+
cot_display = ""
|
49
|
+
if thinking_supported and cot_value and str(cot_value).lower() != "n/a":
|
50
|
+
cot_display = _format_k(cot_value)
|
51
|
+
|
52
|
+
driver = _format_driver(m.get("driver", ""))
|
53
|
+
|
54
|
+
table.add_row(name, vendor, context, max_input, cot_display, max_response, thinking_icon, driver)
|
38
55
|
|
56
|
+
import sys
|
39
57
|
if sys.stdout.isatty():
|
40
|
-
|
41
|
-
|
42
|
-
Console().print(table)
|
58
|
+
shared_console.print(table)
|
43
59
|
else:
|
44
60
|
# ASCII-friendly fallback table when output is redirected
|
45
61
|
print(f"Supported models for provider '{provider_name}'")
|
46
|
-
|
47
|
-
|
62
|
+
print("Model Name | Vendor | Context | Max Input | CoT | Max Response | Thinking | Driver")
|
63
|
+
|
64
|
+
# Get default model for fallback
|
65
|
+
from janito.providers.registry import LLMProviderRegistry
|
66
|
+
try:
|
67
|
+
provider_class = LLMProviderRegistry.get(provider_name)
|
68
|
+
default_model = getattr(provider_class, "DEFAULT_MODEL", None)
|
69
|
+
except:
|
70
|
+
default_model = None
|
71
|
+
|
48
72
|
for m in models:
|
49
|
-
|
50
|
-
|
51
|
-
|
52
|
-
|
53
|
-
|
54
|
-
|
55
|
-
|
56
|
-
|
57
|
-
|
73
|
+
name = str(m.get("name", ""))
|
74
|
+
if name == default_model:
|
75
|
+
name = f"⭐ {name} (default)"
|
76
|
+
|
77
|
+
vendor = "Open" if m.get("open") is True or m.get("open") == "Open" else "Locked"
|
78
|
+
context = _format_context(m.get("context", ""))
|
79
|
+
max_input = _format_k(m.get("max_input", ""))
|
80
|
+
max_cot = _format_k(m.get("max_cot", ""))
|
81
|
+
max_response = _format_k(m.get("max_response", ""))
|
82
|
+
thinking_supported = m.get("thinking_supported") is True or m.get("thinking_supported") == "True"
|
83
|
+
cot_value = m.get("max_cot", "")
|
84
|
+
|
85
|
+
thinking = "Y" if thinking_supported and m.get("thinking", False) else ""
|
86
|
+
cot_display = ""
|
87
|
+
if thinking_supported and cot_value and str(cot_value).lower() != "n/a":
|
88
|
+
cot_display = _format_k(cot_value)
|
89
|
+
|
90
|
+
driver = _format_driver(m.get("driver", ""))
|
91
|
+
print(f"{name} | {vendor} | {context} | {max_input} | {cot_display} | {max_response} | {thinking} | {driver}")
|
58
92
|
|
59
93
|
|
60
94
|
def _format_k(val):
|
95
|
+
"""Format numeric values with k suffix for thousands."""
|
61
96
|
try:
|
62
97
|
n = int(val)
|
63
98
|
if n >= 1000:
|
@@ -67,31 +102,16 @@ def _format_k(val):
|
|
67
102
|
return str(val)
|
68
103
|
|
69
104
|
|
70
|
-
def
|
71
|
-
|
72
|
-
|
73
|
-
|
74
|
-
|
75
|
-
|
105
|
+
def _format_context(val):
|
106
|
+
"""Format context field which might be a single value or range."""
|
107
|
+
if isinstance(val, (list, tuple)) and len(val) == 2:
|
108
|
+
return f"{_format_k(val[0])} / {_format_k(val[1])}"
|
109
|
+
return _format_k(val)
|
110
|
+
|
76
111
|
|
77
|
-
|
78
|
-
|
79
|
-
|
80
|
-
|
81
|
-
|
82
|
-
|
83
|
-
and isinstance(v, (list, tuple))
|
84
|
-
and len(v) == 2
|
85
|
-
):
|
86
|
-
row.append(f"{_format_k(v[0])} / {_format_k(v[1])}")
|
87
|
-
else:
|
88
|
-
row.append(_format_k(v))
|
89
|
-
elif h == "open":
|
90
|
-
row.append("Open" if v is True or v == "Open" else "Locked")
|
91
|
-
elif h == "thinking_supported":
|
92
|
-
row.append("📖" if v is True or v == "True" else "")
|
93
|
-
elif h == "driver":
|
94
|
-
row.append(format_driver(v))
|
95
|
-
else:
|
96
|
-
row.append(str(v))
|
97
|
-
return row
|
112
|
+
def _format_driver(val):
|
113
|
+
"""Format driver name by removing ModelDriver suffix."""
|
114
|
+
if isinstance(val, (list, tuple)):
|
115
|
+
return ", ".join(val)
|
116
|
+
val_str = str(val)
|
117
|
+
return val_str.removesuffix("ModelDriver").strip()
|
@@ -0,0 +1,55 @@
|
|
1
|
+
from janito.provider_registry import list_providers
|
2
|
+
from janito.providers.registry import LLMProviderRegistry
|
3
|
+
from janito.cli.console import shared_console
|
4
|
+
from rich.table import Table
|
5
|
+
from rich.progress import Progress, SpinnerColumn, TextColumn
|
6
|
+
import asyncio
|
7
|
+
import time
|
8
|
+
|
9
|
+
|
10
|
+
def handle_ping_providers(args):
|
11
|
+
"""Ping/test connectivity for all providers."""
|
12
|
+
try:
|
13
|
+
# Get all providers
|
14
|
+
providers = list_providers()
|
15
|
+
|
16
|
+
# Create table for results
|
17
|
+
table = Table(title="Provider Connectivity Test")
|
18
|
+
table.add_column("Provider", style="cyan")
|
19
|
+
table.add_column("Status", style="magenta")
|
20
|
+
table.add_column("Time", style="green")
|
21
|
+
table.add_column("Details", style="yellow")
|
22
|
+
|
23
|
+
# Test each provider
|
24
|
+
for provider_name in providers:
|
25
|
+
start_time = time.time()
|
26
|
+
try:
|
27
|
+
# Get provider class
|
28
|
+
provider_class = LLMProviderRegistry.get(provider_name)
|
29
|
+
provider = provider_class()
|
30
|
+
|
31
|
+
# Test the provider (simplified - just check if we can instantiate and get models)
|
32
|
+
models = provider.get_models()
|
33
|
+
if models:
|
34
|
+
status = "✓ Connected"
|
35
|
+
details = f"{len(models)} models available"
|
36
|
+
else:
|
37
|
+
status = "⚠ No models"
|
38
|
+
details = "Provider reachable but no models returned"
|
39
|
+
|
40
|
+
except Exception as e:
|
41
|
+
status = "✗ Failed"
|
42
|
+
details = str(e)
|
43
|
+
|
44
|
+
end_time = time.time()
|
45
|
+
elapsed = f"{(end_time - start_time)*1000:.0f}ms"
|
46
|
+
|
47
|
+
table.add_row(provider_name, status, elapsed, details)
|
48
|
+
|
49
|
+
# Print results
|
50
|
+
shared_console.print(table)
|
51
|
+
|
52
|
+
except Exception as e:
|
53
|
+
print(f"Error testing provider connectivity: {e}")
|
54
|
+
|
55
|
+
return
|
janito/cli/main_cli.py
CHANGED
@@ -124,6 +124,10 @@ definition = [
|
|
124
124
|
["--list-providers"],
|
125
125
|
{"action": "store_true", "help": "List supported LLM providers"},
|
126
126
|
),
|
127
|
+
(
|
128
|
+
["--ping"],
|
129
|
+
{"action": "store_true", "help": "Ping/test connectivity for all providers (use with --list-providers)"},
|
130
|
+
),
|
127
131
|
(
|
128
132
|
["--list-drivers"],
|
129
133
|
{
|
@@ -238,6 +242,19 @@ GETTER_KEYS = [
|
|
238
242
|
"list_drivers",
|
239
243
|
"region_info",
|
240
244
|
"list_providers_region",
|
245
|
+
"ping",
|
246
|
+
]
|
247
|
+
GETTER_KEYS = [
|
248
|
+
"show_config",
|
249
|
+
"list_providers",
|
250
|
+
"list_profiles",
|
251
|
+
"list_models",
|
252
|
+
"list_tools",
|
253
|
+
"list_config",
|
254
|
+
"list_drivers",
|
255
|
+
"region_info",
|
256
|
+
"list_providers_region",
|
257
|
+
"ping",
|
241
258
|
]
|
242
259
|
|
243
260
|
|
@@ -361,6 +378,7 @@ class JanitoCLI:
|
|
361
378
|
or self.args.show_config
|
362
379
|
or self.args.list_config
|
363
380
|
or self.args.list_drivers
|
381
|
+
or self.args.ping
|
364
382
|
):
|
365
383
|
self._maybe_print_verbose_provider_model()
|
366
384
|
handle_getter(self.args)
|
janito/llm/model.py
CHANGED
@@ -7,6 +7,9 @@ MODEL_SPECS = {
|
|
7
7
|
max_response=8192,
|
8
8
|
category="Alibaba Qwen Turbo Model (OpenAI-compatible)",
|
9
9
|
driver="OpenAIModelDriver",
|
10
|
+
thinking_supported=True,
|
11
|
+
thinking=False,
|
12
|
+
max_cot=8192,
|
10
13
|
),
|
11
14
|
"qwen-plus": LLMModelInfo(
|
12
15
|
name="qwen-plus",
|
@@ -14,6 +17,9 @@ MODEL_SPECS = {
|
|
14
17
|
max_response=8192,
|
15
18
|
category="Alibaba Qwen Plus Model (OpenAI-compatible)",
|
16
19
|
driver="OpenAIModelDriver",
|
20
|
+
thinking_supported=True,
|
21
|
+
thinking=False,
|
22
|
+
max_cot=8192,
|
17
23
|
),
|
18
24
|
"qwen-max": LLMModelInfo(
|
19
25
|
name="qwen-max",
|
@@ -21,6 +27,9 @@ MODEL_SPECS = {
|
|
21
27
|
max_response=8192,
|
22
28
|
category="Alibaba Qwen Max Model (OpenAI-compatible)",
|
23
29
|
driver="OpenAIModelDriver",
|
30
|
+
thinking_supported=True,
|
31
|
+
thinking=False,
|
32
|
+
max_cot=8192,
|
24
33
|
),
|
25
34
|
|
26
35
|
"qwen3-coder-plus": LLMModelInfo(
|
@@ -29,6 +38,9 @@ MODEL_SPECS = {
|
|
29
38
|
max_response=65536,
|
30
39
|
category="Alibaba Qwen3 Coder Plus Model (OpenAI-compatible)",
|
31
40
|
driver="OpenAIModelDriver",
|
41
|
+
thinking_supported=True,
|
42
|
+
thinking=False,
|
43
|
+
max_cot=65536,
|
32
44
|
),
|
33
45
|
"qwen3-coder-480b-a35b-instruct": LLMModelInfo(
|
34
46
|
name="qwen3-coder-480b-a35b-instruct",
|
@@ -36,5 +48,50 @@ MODEL_SPECS = {
|
|
36
48
|
max_response=65536,
|
37
49
|
category="Alibaba Qwen3 Coder 480B A35B Instruct Model (OpenAI-compatible)",
|
38
50
|
driver="OpenAIModelDriver",
|
51
|
+
thinking_supported=True,
|
52
|
+
thinking=False,
|
53
|
+
max_cot=65536,
|
54
|
+
),
|
55
|
+
|
56
|
+
# Qwen3 1M context models (July 2025 update)
|
57
|
+
"qwen3-235b-a22b-thinking-2507": LLMModelInfo(
|
58
|
+
name="qwen3-235b-a22b-thinking-2507",
|
59
|
+
context=131072, # Supports up to 1M with special config
|
60
|
+
max_response=32768,
|
61
|
+
category="Alibaba Qwen3 235B A22B Thinking Model (OpenAI-compatible)",
|
62
|
+
driver="OpenAIModelDriver",
|
63
|
+
thinking=True,
|
64
|
+
thinking_supported=True,
|
65
|
+
max_cot=32768,
|
66
|
+
),
|
67
|
+
"qwen3-235b-a22b-instruct-2507": LLMModelInfo(
|
68
|
+
name="qwen3-235b-a22b-instruct-2507",
|
69
|
+
context=129024, # Supports up to 1M with special config
|
70
|
+
max_response=32768,
|
71
|
+
category="Alibaba Qwen3 235B A22B Instruct Model (OpenAI-compatible)",
|
72
|
+
driver="OpenAIModelDriver",
|
73
|
+
thinking_supported=True,
|
74
|
+
thinking=False,
|
75
|
+
max_cot=32768,
|
76
|
+
),
|
77
|
+
"qwen3-30b-a3b-thinking-2507": LLMModelInfo(
|
78
|
+
name="qwen3-30b-a3b-thinking-2507",
|
79
|
+
context=126976, # Supports up to 1M with special config
|
80
|
+
max_response=32768,
|
81
|
+
category="Alibaba Qwen3 30B A3B Thinking Model (OpenAI-compatible)",
|
82
|
+
driver="OpenAIModelDriver",
|
83
|
+
thinking=True,
|
84
|
+
thinking_supported=True,
|
85
|
+
max_cot=32768,
|
86
|
+
),
|
87
|
+
"qwen3-30b-a3b-instruct-2507": LLMModelInfo(
|
88
|
+
name="qwen3-30b-a3b-instruct-2507",
|
89
|
+
context=129024, # Supports up to 1M with special config
|
90
|
+
max_response=32768,
|
91
|
+
category="Alibaba Qwen3 30B A3B Instruct Model (OpenAI-compatible)",
|
92
|
+
driver="OpenAIModelDriver",
|
93
|
+
thinking_supported=True,
|
94
|
+
thinking=False,
|
95
|
+
max_cot=32768,
|
39
96
|
),
|
40
97
|
}
|
@@ -17,7 +17,7 @@ class AlibabaProvider(LLMProvider):
|
|
17
17
|
NAME = "alibaba"
|
18
18
|
MAINTAINER = "João Pinto <janito@ikignosis.org>"
|
19
19
|
MODEL_SPECS = MODEL_SPECS
|
20
|
-
DEFAULT_MODEL = "qwen3-
|
20
|
+
DEFAULT_MODEL = "qwen3-235b-a22b-instruct-2507" # 129k context, general-purpose model
|
21
21
|
|
22
22
|
def __init__(
|
23
23
|
self, auth_manager: LLMAuthManager = None, config: LLMDriverConfig = None
|
@@ -111,6 +111,39 @@ MODEL_SPECS = {
|
|
111
111
|
open="openai",
|
112
112
|
driver="OpenAIModelDriver",
|
113
113
|
),
|
114
|
+
"gpt-5": LLMModelInfo(
|
115
|
+
name="gpt-5",
|
116
|
+
context=200000,
|
117
|
+
max_input=100000,
|
118
|
+
max_cot="N/A",
|
119
|
+
max_response=100000,
|
120
|
+
thinking_supported=True,
|
121
|
+
default_temp=1.0,
|
122
|
+
open="openai",
|
123
|
+
driver="OpenAIModelDriver",
|
124
|
+
),
|
125
|
+
"gpt-5-mini": LLMModelInfo(
|
126
|
+
name="gpt-5-mini",
|
127
|
+
context=200000,
|
128
|
+
max_input=100000,
|
129
|
+
max_cot="N/A",
|
130
|
+
max_response=100000,
|
131
|
+
thinking_supported=True,
|
132
|
+
default_temp=1.0,
|
133
|
+
open="openai",
|
134
|
+
driver="OpenAIModelDriver",
|
135
|
+
),
|
136
|
+
"gpt-5-nano": LLMModelInfo(
|
137
|
+
name="gpt-5-nano",
|
138
|
+
context=200000,
|
139
|
+
max_input=100000,
|
140
|
+
max_cot="N/A",
|
141
|
+
max_response=100000,
|
142
|
+
thinking_supported=True,
|
143
|
+
default_temp=1.0,
|
144
|
+
open="openai",
|
145
|
+
driver="OpenAIModelDriver",
|
146
|
+
),
|
114
147
|
# duplicated gpt-4-turbo with minimal properties for distinction
|
115
148
|
"gpt-4-turbo-alt": LLMModelInfo(
|
116
149
|
name="gpt-4-turbo",
|
@@ -17,7 +17,7 @@ class OpenAIProvider(LLMProvider):
|
|
17
17
|
NAME = "openai"
|
18
18
|
MAINTAINER = "João Pinto <janito@ikignosis.org>"
|
19
19
|
MODEL_SPECS = MODEL_SPECS
|
20
|
-
DEFAULT_MODEL = "gpt-
|
20
|
+
DEFAULT_MODEL = "gpt-5" # Options: gpt-4.1, gpt-4o, o3-mini, o4-mini, gpt-5, gpt-5-nano
|
21
21
|
|
22
22
|
def __init__(
|
23
23
|
self, auth_manager: LLMAuthManager = None, config: LLMDriverConfig = None
|
@@ -28,7 +28,7 @@ janito/cli/__init__.py,sha256=xaPDOrWphBbCR63Xpcx_yfpXSJIlCaaICc4j2qpWqrM,194
|
|
28
28
|
janito/cli/config.py,sha256=HkZ14701HzIqrvaNyDcDhGlVHfpX_uHlLp2rHmhRm_k,872
|
29
29
|
janito/cli/console.py,sha256=gJolqzWL7jEPLxeuH-CwBDRFpXt976KdZOEAB2tdBDs,64
|
30
30
|
janito/cli/main.py,sha256=s5odou0txf8pzTf1ADk2yV7T5m8B6cejJ81e7iu776U,312
|
31
|
-
janito/cli/main_cli.py,sha256=
|
31
|
+
janito/cli/main_cli.py,sha256=aLkWYT3qG4AmGGx7RXkLJACgVhurhpwFrz6_0rwpQCQ,15703
|
32
32
|
janito/cli/prompt_core.py,sha256=F68J4Xl6jZMYFN4oBBYZFj15Jp-HTYoLub4bw2XpNRU,11648
|
33
33
|
janito/cli/prompt_handler.py,sha256=SnPTlL64noeAMGlI08VBDD5IDD8jlVMIYA4-fS8zVLg,215
|
34
34
|
janito/cli/prompt_setup.py,sha256=1SSvvgS568-3BO_4Sw9A-QF_iLWiIXsNHT0JVqaLwkU,2120
|
@@ -39,7 +39,7 @@ janito/cli/chat_mode/bindings.py,sha256=odjc5_-YW1t2FRhBUNRNoBMoQIg5sMz3ktV7xG0A
|
|
39
39
|
janito/cli/chat_mode/chat_entry.py,sha256=RFdPd23jsA2DMHRacpjAdwI_1dFBaWrtnwyQEgb2fHA,475
|
40
40
|
janito/cli/chat_mode/prompt_style.py,sha256=vsqQ9xxmrYjj1pWuVe9CayQf39fo2EIXrkKPkflSVn4,805
|
41
41
|
janito/cli/chat_mode/script_runner.py,sha256=wOwEn4bgmjqHqjTqtfyaSOnRPsGf4ZVW-YAWhEeqxXU,6507
|
42
|
-
janito/cli/chat_mode/session.py,sha256=
|
42
|
+
janito/cli/chat_mode/session.py,sha256=gY87m6PlC_fNfSTdaJ5T-iq2i3Xc6JRb6rz9NraIUc4,15572
|
43
43
|
janito/cli/chat_mode/session_profile_select.py,sha256=CJ2g3VbPGWfBNrNkYYX57oIJZJ-hIZBNGB-zcdjC9vk,5379
|
44
44
|
janito/cli/chat_mode/toolbar.py,sha256=bJ9jPaTInp2gB3yjSVJp8mpNEFiOslzNhVaiqpXJhKc,3025
|
45
45
|
janito/cli/chat_mode/shell/autocomplete.py,sha256=lE68MaVaodbA2VfUM0_YLqQVLBJAE_BJsd5cMtwuD-g,793
|
@@ -74,13 +74,14 @@ janito/cli/chat_mode/shell/session/history.py,sha256=tYav6GgjAZkvWhlI_rfG6OArNqW
|
|
74
74
|
janito/cli/chat_mode/shell/session/manager.py,sha256=MwD9reHsRaly0CyRB-S1JJ0wPKz2g8Xdj2VvlU35Hgc,1001
|
75
75
|
janito/cli/cli_commands/list_config.py,sha256=oiQEGaGPjwjG-PrOcakpNMbbqISTsBEs7rkGH3ceQsI,1179
|
76
76
|
janito/cli/cli_commands/list_drivers.py,sha256=r2ENykUcvf_9XYp6LHd3RvLXGXyVUA6oe_Pr0dyv92I,5124
|
77
|
-
janito/cli/cli_commands/list_models.py,sha256=
|
77
|
+
janito/cli/cli_commands/list_models.py,sha256=DAXH9CT3Q4krJ_NlrXseCGsc4WECYeJSGX6z0ks8_uA,1661
|
78
78
|
janito/cli/cli_commands/list_profiles.py,sha256=9-HV2EbtP2AdubbMoakjbu7Oq4Ss9UDyO7Eb6CC52wI,2681
|
79
|
-
janito/cli/cli_commands/list_providers.py,sha256=
|
79
|
+
janito/cli/cli_commands/list_providers.py,sha256=3ywm1Ohv7yVqV1E9hB-3Jz8BwzhyCScKxffq6iDI4nA,391
|
80
80
|
janito/cli/cli_commands/list_providers_region.py,sha256=qrMj_gtgEMty8UH0P_O5SgWCVJ9ZKxGUp_GdsE4_EH4,2548
|
81
81
|
janito/cli/cli_commands/list_tools.py,sha256=JFRdlhPeA3BzhJ2PkjIt3u137IJoNc-vYSjUuPlaOXw,3593
|
82
82
|
janito/cli/cli_commands/model_selection.py,sha256=ANWtwC5glZkGMdaNtARDbEG3QmuBUcDLVxzzC5jeBNo,1643
|
83
|
-
janito/cli/cli_commands/model_utils.py,sha256=
|
83
|
+
janito/cli/cli_commands/model_utils.py,sha256=PO64PW-TIlWyPY8CzYnI0y5Zp6ukV_NjaSj8CEXflV0,4889
|
84
|
+
janito/cli/cli_commands/ping_providers.py,sha256=U3iTETPXDEGr3dqlJqNTPvpkhKo2KfXe_bN2_LGJtx0,2015
|
84
85
|
janito/cli/cli_commands/set_api_key.py,sha256=ZItSuB0HO14UsbyXXCgTKAXS-EUCHfCkntzg3WAAtK0,1048
|
85
86
|
janito/cli/cli_commands/show_config.py,sha256=eYMcuvU-d7mvvuctbQacZFERqcKHEnxaRRjasyj-_lE,2004
|
86
87
|
janito/cli/cli_commands/show_system_prompt.py,sha256=9ZJGW7lIGJ9LX2JZiWVEm4AbaD0qEQO7LF89jPgk52I,5232
|
@@ -120,14 +121,14 @@ janito/llm/driver_config.py,sha256=OW0ae49EfgKDqaThuDjZBiaN78voNzwiZ6szERMqJos,1
|
|
120
121
|
janito/llm/driver_config_builder.py,sha256=BvWGx7vaBR5NyvPY1XNAP3lAgo1uf-T25CSsIo2kkCU,1401
|
121
122
|
janito/llm/driver_input.py,sha256=Zq7IO4KdQPUraeIo6XoOaRy1IdQAyYY15RQw4JU30uA,389
|
122
123
|
janito/llm/message_parts.py,sha256=QY_0kDjaxdoErDgKPRPv1dNkkYJuXIBmHWNLiOEKAH4,1365
|
123
|
-
janito/llm/model.py,sha256=
|
124
|
+
janito/llm/model.py,sha256=EioBkdgn8hJ0iQaKN-0KbXlsrk3YKmwR9IbvoEbdVTE,1159
|
124
125
|
janito/llm/provider.py,sha256=3FbhQPrWBSEoIdIi-5DWIh0DD_CM570EFf1NcuGyGko,7961
|
125
126
|
janito/providers/__init__.py,sha256=wlb8-dfnWRsZclFTT5cbwYqSWFpppChrAxZEgpdBgng,450
|
126
127
|
janito/providers/dashscope.bak.zip,sha256=BwXxRmZreEivvRtmqbr5BR62IFVlNjAf4y6DrF2BVJo,5998
|
127
128
|
janito/providers/registry.py,sha256=Ygwv9eVrTXOKhv0EKxSWQXO5WMHvajWE2Q_Lc3p7dKo,730
|
128
129
|
janito/providers/alibaba/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
129
|
-
janito/providers/alibaba/model_info.py,sha256=
|
130
|
-
janito/providers/alibaba/provider.py,sha256=
|
130
|
+
janito/providers/alibaba/model_info.py,sha256=G8TC-SaH96R2EnvVLWb7TjLXAmskeVylDzPho97dSTM,3326
|
131
|
+
janito/providers/alibaba/provider.py,sha256=L7oK_TeJn8p_ZaFlls5eb9YF03VrFkR_PLdopVvZQ7w,4192
|
131
132
|
janito/providers/anthropic/model_info.py,sha256=m6pBh0Ia8_xC1KZ7ke_4HeHIFw7nWjnYVItnRpkCSWc,1206
|
132
133
|
janito/providers/anthropic/provider.py,sha256=aGynBxCFc7oTyvGNDUkbutJCKurC_9J4AkReC2LTPYo,3023
|
133
134
|
janito/providers/azure_openai/model_info.py,sha256=TMSqEpQROIIYUGAyulYJ5xGhj7CbLoaKL_JXeLbXaG0,689
|
@@ -145,8 +146,8 @@ janito/providers/moonshotai/__init__.py,sha256=nThTAtynq4O2Iidm95daKOCKXi5APRJYt
|
|
145
146
|
janito/providers/moonshotai/model_info.py,sha256=MpPAB3tZVvZ8V7tZsiJpk5SReHjVcnwwbp63aUebx9Y,718
|
146
147
|
janito/providers/moonshotai/provider.py,sha256=Lynw2cAF0IA9QPIRN4oY6Yr0cEwyVggpaQQrmkcgHuE,3874
|
147
148
|
janito/providers/openai/__init__.py,sha256=f0m16-sIqScjL9Mp4A0CQBZx6H3PTEy0cnE08jeaB5U,38
|
148
|
-
janito/providers/openai/model_info.py,sha256=
|
149
|
-
janito/providers/openai/provider.py,sha256=
|
149
|
+
janito/providers/openai/model_info.py,sha256=VTkq3xcx2vk0tXlFVHQxKeFzl-DL1T1J2elVOEwCdHI,4265
|
150
|
+
janito/providers/openai/provider.py,sha256=UyG__9yen6MAcCS3SbKzKIj1UIauidYhDBOKn-cptn0,4743
|
150
151
|
janito/providers/openai/schema_generator.py,sha256=hTqeLcPTR8jeKn5DUUpo7b-EZ-V-g1WwXiX7MbHnFzE,2234
|
151
152
|
janito/providers/zai/__init__.py,sha256=qtIr9_QBFaXG8xB6cRDGhS7se6ir11CWseI9azLMRBo,24
|
152
153
|
janito/providers/zai/model_info.py,sha256=ldwD8enpxXv1G-YsDw4YJn31YsVueQ4vj5HgoYvnPxo,1183
|
@@ -217,9 +218,9 @@ janito/tools/adapters/local/validate_file_syntax/ps1_validator.py,sha256=TeIkPt0
|
|
217
218
|
janito/tools/adapters/local/validate_file_syntax/python_validator.py,sha256=BfCO_K18qy92m-2ZVvHsbEU5e11OPo1pO9Vz4G4616E,130
|
218
219
|
janito/tools/adapters/local/validate_file_syntax/xml_validator.py,sha256=AijlsP_PgNuC8ZbGsC5vOTt3Jur76otQzkd_7qR0QFY,284
|
219
220
|
janito/tools/adapters/local/validate_file_syntax/yaml_validator.py,sha256=TgyI0HRL6ug_gBcWEm5TGJJuA4E34ZXcIzMpAbv3oJs,155
|
220
|
-
janito-2.
|
221
|
-
janito-2.
|
222
|
-
janito-2.
|
223
|
-
janito-2.
|
224
|
-
janito-2.
|
225
|
-
janito-2.
|
221
|
+
janito-2.19.0.dist-info/licenses/LICENSE,sha256=GSAKapQH5ZIGWlpQTA7v5YrfECyaxaohUb1vJX-qepw,1090
|
222
|
+
janito-2.19.0.dist-info/METADATA,sha256=KgBACQLGdnw7inirQR1Snxe86lMSNnUdhXDkUpJN-U0,16365
|
223
|
+
janito-2.19.0.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
224
|
+
janito-2.19.0.dist-info/entry_points.txt,sha256=wIo5zZxbmu4fC-ZMrsKD0T0vq7IqkOOLYhrqRGypkx4,48
|
225
|
+
janito-2.19.0.dist-info/top_level.txt,sha256=m0NaVCq0-ivxbazE2-ND0EA9Hmuijj_OGkmCbnBcCig,7
|
226
|
+
janito-2.19.0.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|