llms-py 2.0.17__tar.gz → 2.0.18__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llms_py-2.0.17/llms_py.egg-info → llms_py-2.0.18}/PKG-INFO +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/main.cpython-314.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/llms.json +19 -21
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/main.py +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/ai.mjs +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18/llms_py.egg-info}/PKG-INFO +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18}/pyproject.toml +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18}/setup.py +1 -1
- {llms_py-2.0.17 → llms_py-2.0.18}/LICENSE +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/MANIFEST.in +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/README.md +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__init__.py +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__main__.py +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/__init__.cpython-312.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/__init__.cpython-313.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/__init__.cpython-314.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/__main__.cpython-312.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/__main__.cpython-314.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/llms.cpython-312.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/main.cpython-312.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/__pycache__/main.cpython-313.pyc +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/index.html +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Analytics.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/App.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Avatar.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Brand.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/ChatPrompt.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Main.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/ModelSelector.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/ProviderIcon.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/ProviderStatus.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Recents.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/SettingsDialog.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Sidebar.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/SignIn.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/SystemPromptEditor.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/SystemPromptSelector.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/Welcome.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/app.css +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/fav.svg +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/chart.js +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/charts.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/color.js +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/highlight.min.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/idb.min.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/marked.min.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/servicestack-client.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/servicestack-vue.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/vue-router.min.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/vue.min.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/lib/vue.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/markdown.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/tailwind.input.css +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/threadStore.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/typography.css +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui/utils.mjs +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms/ui.json +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/SOURCES.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/dependency_links.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/entry_points.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/not-zip-safe +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/requires.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/llms_py.egg-info/top_level.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/requirements.txt +0 -0
- {llms_py-2.0.17 → llms_py-2.0.18}/setup.cfg +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: llms-py
|
|
3
|
-
Version: 2.0.
|
|
3
|
+
Version: 2.0.18
|
|
4
4
|
Summary: A lightweight CLI tool and OpenAI-compatible server for querying multiple Large Language Model (LLM) providers
|
|
5
5
|
Home-page: https://github.com/ServiceStack/llms
|
|
6
6
|
Author: ServiceStack
|
|
Binary file
|
|
@@ -93,10 +93,6 @@
|
|
|
93
93
|
}
|
|
94
94
|
],
|
|
95
95
|
"max_completion_tokens": 16,
|
|
96
|
-
"reasoning": {
|
|
97
|
-
"max_tokens": 16,
|
|
98
|
-
"reasoning_effort": "low"
|
|
99
|
-
},
|
|
100
96
|
"stream": false
|
|
101
97
|
}
|
|
102
98
|
},
|
|
@@ -147,7 +143,6 @@
|
|
|
147
143
|
"api_key": "$GROQ_API_KEY",
|
|
148
144
|
"models": {
|
|
149
145
|
"allam-2-7b": "allam-2-7b",
|
|
150
|
-
"gemma2:9b": "gemma2-9b-it",
|
|
151
146
|
"compound": "groq/compound",
|
|
152
147
|
"compound-mini": "groq/compound-mini",
|
|
153
148
|
"llama3.1:8b": "llama-3.1-8b-instant",
|
|
@@ -599,7 +594,6 @@
|
|
|
599
594
|
"base_url": "https://api.anthropic.com",
|
|
600
595
|
"api_key": "$ANTHROPIC_API_KEY",
|
|
601
596
|
"models": {
|
|
602
|
-
"claude-opus-4-1": "claude-opus-4-1",
|
|
603
597
|
"claude-sonnet-4-5": "claude-sonnet-4-5",
|
|
604
598
|
"claude-sonnet-4-0": "claude-sonnet-4-0",
|
|
605
599
|
"claude-3-7-sonnet": "claude-3-7-sonnet-latest",
|
|
@@ -640,6 +634,25 @@
|
|
|
640
634
|
"input": "0.00000025",
|
|
641
635
|
"output": "0.00000125"
|
|
642
636
|
}
|
|
637
|
+
},
|
|
638
|
+
"check": {
|
|
639
|
+
"messages": [
|
|
640
|
+
{
|
|
641
|
+
"role": "user",
|
|
642
|
+
"content": [
|
|
643
|
+
{
|
|
644
|
+
"type": "text",
|
|
645
|
+
"text": "1+1="
|
|
646
|
+
}
|
|
647
|
+
]
|
|
648
|
+
}
|
|
649
|
+
],
|
|
650
|
+
"max_completion_tokens": 512,
|
|
651
|
+
"reasoning": {
|
|
652
|
+
"max_tokens": 128,
|
|
653
|
+
"reasoning_effort": "low"
|
|
654
|
+
},
|
|
655
|
+
"stream": false
|
|
643
656
|
}
|
|
644
657
|
},
|
|
645
658
|
"openai": {
|
|
@@ -735,21 +748,6 @@
|
|
|
735
748
|
"input": "0.000003",
|
|
736
749
|
"output": "0.000006"
|
|
737
750
|
}
|
|
738
|
-
},
|
|
739
|
-
"check": {
|
|
740
|
-
"messages": [
|
|
741
|
-
{
|
|
742
|
-
"role": "user",
|
|
743
|
-
"content": [
|
|
744
|
-
{
|
|
745
|
-
"type": "text",
|
|
746
|
-
"text": "1+1="
|
|
747
|
-
}
|
|
748
|
-
]
|
|
749
|
-
}
|
|
750
|
-
],
|
|
751
|
-
"max_completion_tokens": 16,
|
|
752
|
-
"stream": false
|
|
753
751
|
}
|
|
754
752
|
},
|
|
755
753
|
"grok": {
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: llms-py
|
|
3
|
-
Version: 2.0.
|
|
3
|
+
Version: 2.0.18
|
|
4
4
|
Summary: A lightweight CLI tool and OpenAI-compatible server for querying multiple Large Language Model (LLM) providers
|
|
5
5
|
Home-page: https://github.com/ServiceStack/llms
|
|
6
6
|
Author: ServiceStack
|
|
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|
|
4
4
|
|
|
5
5
|
[project]
|
|
6
6
|
name = "llms-py"
|
|
7
|
-
version = "2.0.
|
|
7
|
+
version = "2.0.18"
|
|
8
8
|
description = "A lightweight CLI tool and OpenAI-compatible server for querying multiple Large Language Model (LLM) providers"
|
|
9
9
|
readme = "README.md"
|
|
10
10
|
license = "BSD-3-Clause"
|
|
@@ -16,7 +16,7 @@ with open(os.path.join(this_directory, "requirements.txt"), encoding="utf-8") as
|
|
|
16
16
|
|
|
17
17
|
setup(
|
|
18
18
|
name="llms-py",
|
|
19
|
-
version="2.0.
|
|
19
|
+
version="2.0.18",
|
|
20
20
|
author="ServiceStack",
|
|
21
21
|
author_email="team@servicestack.net",
|
|
22
22
|
description="A lightweight CLI tool and OpenAI-compatible server for querying multiple Large Language Model (LLM) providers",
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|