lollms-client 1.1.2__tar.gz → 1.3.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of lollms-client might be problematic. Click here for more details.

Files changed (78) hide show
  1. {lollms_client-1.1.2/src/lollms_client.egg-info → lollms_client-1.3.0}/PKG-INFO +1 -1
  2. {lollms_client-1.1.2 → lollms_client-1.3.0}/pyproject.toml +2 -0
  3. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/__init__.py +1 -1
  4. lollms_client-1.3.0/src/lollms_client/assets/models_ctx_sizes.json +382 -0
  5. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/lollms/__init__.py +2 -2
  6. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/ollama/__init__.py +56 -0
  7. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/openai/__init__.py +3 -3
  8. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_core.py +285 -131
  9. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_discussion.py +419 -147
  10. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_tti_binding.py +32 -82
  11. lollms_client-1.3.0/src/lollms_client/tti_bindings/diffusers/__init__.py +819 -0
  12. lollms_client-1.3.0/src/lollms_client/tti_bindings/openai/__init__.py +124 -0
  13. {lollms_client-1.1.2 → lollms_client-1.3.0/src/lollms_client.egg-info}/PKG-INFO +1 -1
  14. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client.egg-info/SOURCES.txt +2 -1
  15. lollms_client-1.1.2/src/lollms_client/tti_bindings/dalle/__init__.py +0 -454
  16. lollms_client-1.1.2/src/lollms_client/tti_bindings/diffusers/__init__.py +0 -656
  17. {lollms_client-1.1.2 → lollms_client-1.3.0}/LICENSE +0 -0
  18. {lollms_client-1.1.2 → lollms_client-1.3.0}/README.md +0 -0
  19. {lollms_client-1.1.2 → lollms_client-1.3.0}/setup.cfg +0 -0
  20. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/__init__.py +0 -0
  21. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/azure_openai/__init__.py +0 -0
  22. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/claude/__init__.py +0 -0
  23. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/gemini/__init__.py +0 -0
  24. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/grok/__init__.py +0 -0
  25. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/groq/__init__.py +0 -0
  26. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/hugging_face_inference_api/__init__.py +0 -0
  27. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/litellm/__init__.py +0 -0
  28. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/llamacpp/__init__.py +0 -0
  29. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/lollms_webui/__init__.py +0 -0
  30. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/mistral/__init__.py +0 -0
  31. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/open_router/__init__.py +0 -0
  32. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/openllm/__init__.py +0 -0
  33. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/pythonllamacpp/__init__.py +0 -0
  34. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/tensor_rt/__init__.py +0 -0
  35. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/transformers/__init__.py +0 -0
  36. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/llm_bindings/vllm/__init__.py +0 -0
  37. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_config.py +0 -0
  38. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_js_analyzer.py +0 -0
  39. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_llm_binding.py +0 -0
  40. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_mcp_binding.py +0 -0
  41. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_mcp_security.py +0 -0
  42. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_personality.py +0 -0
  43. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_python_analyzer.py +0 -0
  44. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_stt_binding.py +0 -0
  45. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_ttm_binding.py +0 -0
  46. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_tts_binding.py +0 -0
  47. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_ttv_binding.py +0 -0
  48. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_types.py +0 -0
  49. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/lollms_utilities.py +0 -0
  50. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/local_mcp/__init__.py +0 -0
  51. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/file_writer/file_writer.py +0 -0
  52. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/generate_image_from_prompt/generate_image_from_prompt.py +0 -0
  53. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/internet_search/internet_search.py +0 -0
  54. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/python_interpreter/python_interpreter.py +0 -0
  55. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/remote_mcp/__init__.py +0 -0
  56. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/mcp_bindings/standard_mcp/__init__.py +0 -0
  57. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/stt_bindings/__init__.py +0 -0
  58. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/stt_bindings/lollms/__init__.py +0 -0
  59. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/stt_bindings/whisper/__init__.py +0 -0
  60. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/stt_bindings/whispercpp/__init__.py +0 -0
  61. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tti_bindings/__init__.py +0 -0
  62. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tti_bindings/gemini/__init__.py +0 -0
  63. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tti_bindings/lollms/__init__.py +0 -0
  64. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttm_bindings/__init__.py +0 -0
  65. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttm_bindings/audiocraft/__init__.py +0 -0
  66. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttm_bindings/bark/__init__.py +0 -0
  67. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttm_bindings/lollms/__init__.py +0 -0
  68. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tts_bindings/__init__.py +0 -0
  69. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tts_bindings/bark/__init__.py +0 -0
  70. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tts_bindings/lollms/__init__.py +0 -0
  71. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tts_bindings/piper_tts/__init__.py +0 -0
  72. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/tts_bindings/xtts/__init__.py +0 -0
  73. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttv_bindings/__init__.py +0 -0
  74. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client/ttv_bindings/lollms/__init__.py +0 -0
  75. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client.egg-info/dependency_links.txt +0 -0
  76. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client.egg-info/requires.txt +0 -0
  77. {lollms_client-1.1.2 → lollms_client-1.3.0}/src/lollms_client.egg-info/top_level.txt +0 -0
  78. {lollms_client-1.1.2 → lollms_client-1.3.0}/test/test_lollms_discussion.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: lollms_client
3
- Version: 1.1.2
3
+ Version: 1.3.0
4
4
  Summary: A client library for LoLLMs generate endpoint
5
5
  Author-email: ParisNeo <parisneoai@gmail.com>
6
6
  License: Apache Software License
@@ -46,3 +46,5 @@ where = ["src"]
46
46
 
47
47
  [tool.setuptools.dynamic]
48
48
  version = {attr = "lollms_client.__version__"}
49
+ [tool.setuptools.package-data]
50
+ lollms_client = ["assets/*.json"]
@@ -8,7 +8,7 @@ from lollms_client.lollms_utilities import PromptReshaper # Keep general utiliti
8
8
  from lollms_client.lollms_mcp_binding import LollmsMCPBinding, LollmsMCPBindingManager
9
9
  from lollms_client.lollms_llm_binding import LollmsLLMBindingManager
10
10
 
11
- __version__ = "1.1.2" # Updated version
11
+ __version__ = "1.3.0" # Updated version
12
12
 
13
13
  # Optionally, you could define __all__ if you want to be explicit about exports
14
14
  __all__ = [
@@ -0,0 +1,382 @@
1
+ {
2
+ "agentica-org/deepcoder-14b-preview": 8192,
3
+ "agentica-org/deepcoder-14b-preview:free": 8192,
4
+ "ai21/jamba-large-1.7": 256000,
5
+ "ai21/jamba-mini-1.7": 256000,
6
+ "aion-labs/aion-1.0": 8192,
7
+ "aion-labs/aion-1.0-mini": 8192,
8
+ "aion-labs/aion-rp-llama-3.1-8b": 131072,
9
+ "alfredpros/codellama-7b-instruct-solidity": 16384,
10
+ "alpindale/goliath-120b": 4096,
11
+ "amazon/nova-lite-v1": 32768,
12
+ "amazon/nova-micro-v1": 32768,
13
+ "amazon/nova-pro-v1": 32768,
14
+ "anthracite-org/magnum-v2-72b": 131072,
15
+ "anthracite-org/magnum-v4-72b": 131072,
16
+ "claude-3-haiku": 200000,
17
+ "claude-3-haiku:beta": 200000,
18
+ "claude-3-opus": 200000,
19
+ "claude-3-opus:beta": 200000,
20
+ "claude-3.5-haiku": 200000,
21
+ "claude-3.5-haiku-20241022": 200000,
22
+ "claude-3.5-haiku:beta": 200000,
23
+ "claude-3.5-sonnet": 200000,
24
+ "claude-3.5-sonnet-20240620": 200000,
25
+ "claude-3.5-sonnet-20240620:beta": 200000,
26
+ "claude-3.5-sonnet:beta": 200000,
27
+ "claude-3.7-sonnet": 200000,
28
+ "claude-3.7-sonnet:beta": 200000,
29
+ "claude-3.7-sonnet:thinking": 200000,
30
+ "claude-3-5-haiku": 200000,
31
+ "claude-3-5-haiku-20241022": 200000,
32
+ "claude-3-5-haiku:beta": 200000,
33
+ "claude-3-5-sonnet": 200000,
34
+ "claude-3-5-sonnet-20240620": 200000,
35
+ "claude-3-5-sonnet-20240620:beta": 200000,
36
+ "claude-3-5-sonnet:beta": 200000,
37
+ "claude-3-7-sonnet": 200000,
38
+ "claude-3-7-sonnet:beta": 200000,
39
+ "claude-3-7-sonnet:thinking": 200000,
40
+ "claude-opus-4": 200000,
41
+ "claude-opus-4.1": 200000,
42
+ "claude-sonnet-4": 200000,
43
+ "arcee-ai/coder-large": 32768,
44
+ "arcee-ai/maestro-reasoning": 32768,
45
+ "arcee-ai/spotlight": 32768,
46
+ "arcee-ai/virtuoso-large": 32768,
47
+ "arliai/qwq-32b-arliai-rpr-v1": 8192,
48
+ "arliai/qwq-32b-arliai-rpr-v1:free": 8192,
49
+ "baidu/ernie-4.5-300b-a47b": 128000,
50
+ "bytedance/ui-tars-1.5-7b": 8192,
51
+ "cognitivecomputations/dolphin-mistral-24b-venice-edition:free": 32768,
52
+ "cognitivecomputations/dolphin-mixtral-8x22b": 65536,
53
+ "cognitivecomputations/dolphin3.0-mistral-24b": 32768,
54
+ "cognitivecomputations/dolphin3.0-mistral-24b:free": 32768,
55
+ "cognitivecomputations/dolphin3.0-r1-mistral-24b": 32768,
56
+ "cognitivecomputations/dolphin3.0-r1-mistral-24b:free": 32768,
57
+ "cohere/command": 8192,
58
+ "cohere/command-a": 8192,
59
+ "cohere/command-r": 128000,
60
+ "cohere/command-r-03-2024": 128000,
61
+ "cohere/command-r-08-2024": 128000,
62
+ "cohere/command-r-plus": 128000,
63
+ "cohere/command-r-plus-04-2024": 128000,
64
+ "cohere/command-r-plus-08-2024": 128000,
65
+ "cohere/command-r7b-12-2024": 128000,
66
+ "deepseek/deepseek-chat": 32768,
67
+ "deepseek/deepseek-chat-v3-0324": 32768,
68
+ "deepseek/deepseek-chat-v3-0324:free": 32768,
69
+ "deepseek/deepseek-prover-v2": 131072,
70
+ "deepseek/deepseek-r1": 32768,
71
+ "deepseek/deepseek-r1-0528": 32768,
72
+ "deepseek/deepseek-r1-0528-qwen3-8b": 32768,
73
+ "deepseek/deepseek-r1-0528-qwen3-8b:free": 32768,
74
+ "deepseek/deepseek-r1-0528:free": 32768,
75
+ "deepseek/deepseek-r1-distill-llama-70b": 131072,
76
+ "deepseek/deepseek-r1-distill-llama-70b:free": 131072,
77
+ "deepseek/deepseek-r1-distill-llama-8b": 131072,
78
+ "deepseek/deepseek-r1-distill-qwen-1.5b": 32768,
79
+ "deepseek/deepseek-r1-distill-qwen-14b": 32768,
80
+ "deepseek/deepseek-r1-distill-qwen-14b:free": 32768,
81
+ "deepseek/deepseek-r1-distill-qwen-32b": 32768,
82
+ "deepseek/deepseek-r1-distill-qwen-7b": 32768,
83
+ "deepseek/deepseek-r1:free": 32768,
84
+ "deepseek/deepseek-v3-base": 32768,
85
+ "eleutherai/llemma_7b": 8192,
86
+ "featherless/qwerky-72b:free": 8192,
87
+ "google/gemini-2.0-flash-001": 1000000,
88
+ "google/gemini-2.0-flash-exp:free": 1000000,
89
+ "google/gemini-2.0-flash-lite-001": 1000000,
90
+ "google/gemini-2.5-flash": 1000000,
91
+ "google/gemini-2.5-flash-lite": 1000000,
92
+ "google/gemini-2.5-flash-lite-preview-06-17": 1000000,
93
+ "google/gemini-2.5-pro": 2000000,
94
+ "google/gemini-2.5-pro-exp-03-25": 2000000,
95
+ "google/gemini-2.5-pro-preview": 2000000,
96
+ "google/gemini-2.5-pro-preview-05-06": 2000000,
97
+ "google/gemini-flash-1.5": 1000000,
98
+ "google/gemini-flash-1.5-8b": 1000000,
99
+ "google/gemini-pro-1.5": 2000000,
100
+ "google/gemma-2-27b-it": 8192,
101
+ "google/gemma-2-9b-it": 8192,
102
+ "google/gemma-2-9b-it:free": 8192,
103
+ "google/gemma-3-12b-it": 131072,
104
+ "google/gemma-3-12b-it:free": 131072,
105
+ "google/gemma-3-27b-it": 131072,
106
+ "google/gemma-3-27b-it:free": 131072,
107
+ "google/gemma-3-4b-it": 131072,
108
+ "google/gemma-3-4b-it:free": 131072,
109
+ "google/gemma-3n-e2b-it:free": 131072,
110
+ "google/gemma-3n-e4b-it": 131072,
111
+ "google/gemma-3n-e4b-it:free": 131072,
112
+ "gryphe/mythomax-l2-13b": 4096,
113
+ "inception/mercury": 32768,
114
+ "inception/mercury-coder": 32768,
115
+ "infermatic/mn-inferor-12b": 8192,
116
+ "inflection/inflection-3-pi": 128000,
117
+ "inflection/inflection-3-productivity": 128000,
118
+ "liquid/lfm-3b": 8192,
119
+ "liquid/lfm-40b": 8192,
120
+ "liquid/lfm-7b": 8192,
121
+ "mancer/weaver": 8192,
122
+ "meta-llama/llama-3-70b-instruct": 8192,
123
+ "meta-llama/llama-3-8b-instruct": 8192,
124
+ "meta-llama/llama-3.1-405b": 131072,
125
+ "meta-llama/llama-3.1-405b-instruct": 131072,
126
+ "meta-llama/llama-3.1-405b-instruct:free": 131072,
127
+ "meta-llama/llama-3.1-70b-instruct": 131072,
128
+ "meta-llama/llama-3.1-8b-instruct": 131072,
129
+ "meta-llama/llama-3.2-11b-vision-instruct": 131072,
130
+ "meta-llama/llama-3.2-11b-vision-instruct:free": 131072,
131
+ "meta-llama/llama-3.2-1b-instruct": 131072,
132
+ "meta-llama/llama-3.2-3b-instruct": 131072,
133
+ "meta-llama/llama-3.2-3b-instruct:free": 131072,
134
+ "meta-llama/llama-3.2-90b-vision-instruct": 131072,
135
+ "meta-llama/llama-3.3-70b-instruct": 131072,
136
+ "meta-llama/llama-3.3-70b-instruct:free": 131072,
137
+ "meta-llama/llama-4-maverick": 131072,
138
+ "meta-llama/llama-4-scout": 131072,
139
+ "meta-llama/llama-guard-2-8b": 8192,
140
+ "meta-llama/llama-guard-3-8b": 131072,
141
+ "meta-llama/llama-guard-4-12b": 131072,
142
+ "microsoft/mai-ds-r1": 32768,
143
+ "microsoft/mai-ds-r1:free": 32768,
144
+ "microsoft/phi-3-medium-128k-instruct": 131072,
145
+ "microsoft/phi-3-mini-128k-instruct": 131072,
146
+ "microsoft/phi-3.5-mini-128k-instruct": 131072,
147
+ "microsoft/phi-4": 131072,
148
+ "microsoft/phi-4-multimodal-instruct": 131072,
149
+ "microsoft/phi-4-reasoning-plus": 131072,
150
+ "microsoft/wizardlm-2-8x22b": 65536,
151
+ "minimax/minimax-01": 200000,
152
+ "minimax/minimax-m1": 200000,
153
+ "mistralai/codestral-2501": 32768,
154
+ "mistralai/codestral-2508": 32768,
155
+ "mistralai/devstral-medium": 32768,
156
+ "mistralai/devstral-small": 32768,
157
+ "mistralai/devstral-small-2505": 32768,
158
+ "mistralai/devstral-small-2505:free": 32768,
159
+ "mistralai/magistral-medium-2506": 32768,
160
+ "mistralai/magistral-medium-2506:thinking": 32768,
161
+ "mistralai/magistral-small-2506": 32768,
162
+ "mistralai/ministral-3b": 32768,
163
+ "mistralai/ministral-8b": 32768,
164
+ "mistralai/mistral-7b-instruct": 32768,
165
+ "mistralai/mistral-7b-instruct-v0.1": 8192,
166
+ "mistralai/mistral-7b-instruct-v0.2": 32768,
167
+ "mistralai/mistral-7b-instruct-v0.3": 32768,
168
+ "mistralai/mistral-7b-instruct:free": 32768,
169
+ "mistralai/mistral-large": 32768,
170
+ "mistralai/mistral-large-2407": 128000,
171
+ "mistralai/mistral-large-2411": 128000,
172
+ "mistralai/mistral-medium-3": 32768,
173
+ "mistralai/mistral-nemo": 128000,
174
+ "mistralai/mistral-nemo:free": 128000,
175
+ "mistralai/mistral-saba": 32768,
176
+ "mistralai/mistral-small": 32768,
177
+ "mistralai/mistral-small-24b-instruct-2501": 32768,
178
+ "mistralai/mistral-small-24b-instruct-2501:free": 32768,
179
+ "mistralai/mistral-small-3.1-24b-instruct": 32768,
180
+ "mistralai/mistral-small-3.1-24b-instruct:free": 32768,
181
+ "mistralai/mistral-small-3.2-24b-instruct": 32768,
182
+ "mistralai/mistral-small-3.2-24b-instruct:free": 32768,
183
+ "mistralai/mistral-tiny": 32768,
184
+ "mistralai/mixtral-8x22b-instruct": 65536,
185
+ "mistralai/mixtral-8x7b-instruct": 32768,
186
+ "mistralai/pixtral-12b": 128000,
187
+ "mistralai/pixtral-large-2411": 128000,
188
+ "moonshotai/kimi-dev-72b:free": 200000,
189
+ "moonshotai/kimi-k2": 200000,
190
+ "moonshotai/kimi-k2:free": 200000,
191
+ "moonshotai/kimi-vl-a3b-thinking": 200000,
192
+ "moonshotai/kimi-vl-a3b-thinking:free": 200000,
193
+ "morph/morph-v3-fast": 8192,
194
+ "morph/morph-v3-large": 8192,
195
+ "neversleep/llama-3-lumimaid-70b": 8192,
196
+ "neversleep/llama-3.1-lumimaid-8b": 131072,
197
+ "neversleep/noromaid-20b": 32768,
198
+ "nousresearch/deephermes-3-llama-3-8b-preview:free": 8192,
199
+ "nousresearch/deephermes-3-mistral-24b-preview": 32768,
200
+ "nousresearch/hermes-2-pro-llama-3-8b": 8192,
201
+ "nousresearch/hermes-3-llama-3.1-405b": 131072,
202
+ "nousresearch/hermes-3-llama-3.1-70b": 131072,
203
+ "nousresearch/nous-hermes-2-mixtral-8x7b-dpo": 32768,
204
+ "nvidia/llama-3.1-nemotron-70b-instruct": 131072,
205
+ "nvidia/llama-3.1-nemotron-ultra-253b-v1": 131072,
206
+ "nvidia/llama-3.1-nemotron-ultra-253b-v1:free": 131072,
207
+ "nvidia/llama-3.3-nemotron-super-49b-v1": 131072,
208
+ "openai/chatgpt-4o-latest": 128000,
209
+ "openai/codex-mini": 2048,
210
+ "openai/gpt-3.5-turbo": 4096,
211
+ "openai/gpt-3.5-turbo-0613": 4096,
212
+ "openai/gpt-3.5-turbo-16k": 16384,
213
+ "openai/gpt-3.5-turbo-instruct": 4096,
214
+ "openai/gpt-4": 8192,
215
+ "openai/gpt-4-0314": 8192,
216
+ "openai/gpt-4-1106-preview": 128000,
217
+ "openai/gpt-4-turbo": 128000,
218
+ "openai/gpt-4-turbo-preview": 128000,
219
+ "openai/gpt-4.1": 128000,
220
+ "openai/gpt-4.1-mini": 128000,
221
+ "openai/gpt-4.1-nano": 128000,
222
+ "openai/gpt-4o": 128000,
223
+ "openai/gpt-4o-2024-05-13": 128000,
224
+ "openai/gpt-4o-2024-08-06": 128000,
225
+ "openai/gpt-4o-2024-11-20": 128000,
226
+ "openai/gpt-4o-mini": 128000,
227
+ "openai/gpt-4o-mini-2024-07-18": 128000,
228
+ "openai/gpt-4o-mini-search-preview": 128000,
229
+ "openai/gpt-4o-search-preview": 128000,
230
+ "openai/gpt-4o:extended": 128000,
231
+ "openai/gpt-5": 200000,
232
+ "openai/gpt-5-chat": 200000,
233
+ "openai/gpt-5-mini": 200000,
234
+ "openai/gpt-5-nano": 200000,
235
+ "openai/gpt-oss-120b": 128000,
236
+ "openai/gpt-oss-20b": 128000,
237
+ "openai/gpt-oss-20b:free": 128000,
238
+ "openai/o1": 128000,
239
+ "openai/o1-mini": 128000,
240
+ "openai/o1-mini-2024-09-12": 128000,
241
+ "openai/o1-pro": 128000,
242
+ "openai/o3": 200000,
243
+ "openai/o3-mini": 200000,
244
+ "openai/o3-mini-high": 200000,
245
+ "openai/o3-pro": 200000,
246
+ "openai/o4-mini": 128000,
247
+ "openai/o4-mini-high": 128000,
248
+ "opengvlab/internvl3-14b": 8192,
249
+ "openrouter/auto": 8192,
250
+ "perplexity/r1-1776": 32768,
251
+ "perplexity/sonar": 32768,
252
+ "perplexity/sonar-deep-research": 32768,
253
+ "perplexity/sonar-pro": 32768,
254
+ "perplexity/sonar-reasoning": 32768,
255
+ "perplexity/sonar-reasoning-pro": 32768,
256
+ "pygmalionai/mythalion-13b": 4096,
257
+ "qwen/qwen-2-72b-instruct": 32768,
258
+ "qwen/qwen-2.5-72b-instruct": 131072,
259
+ "qwen/qwen-2.5-72b-instruct:free": 131072,
260
+ "qwen/qwen-2.5-7b-instruct": 131072,
261
+ "qwen/qwen-2.5-coder-32b-instruct": 131072,
262
+ "qwen/qwen-2.5-coder-32b-instruct:free": 131072,
263
+ "qwen/qwen-2.5-vl-7b-instruct": 131072,
264
+ "qwen/qwen-max": 32768,
265
+ "qwen/qwen-plus": 32768,
266
+ "qwen/qwen-turbo": 8192,
267
+ "qwen/qwen-vl-max": 32768,
268
+ "qwen/qwen-vl-plus": 32768,
269
+ "qwen/qwen2.5-vl-32b-instruct": 131072,
270
+ "qwen/qwen2.5-vl-32b-instruct:free": 131072,
271
+ "qwen/qwen2.5-vl-72b-instruct": 131072,
272
+ "qwen/qwen2.5-vl-72b-instruct:free": 131072,
273
+ "qwen/qwen3-14b": 32768,
274
+ "qwen/qwen3-14b:free": 32768,
275
+ "qwen/qwen3-235b-a22b": 32768,
276
+ "qwen/qwen3-235b-a22b-2507": 32768,
277
+ "qwen/qwen3-235b-a22b-thinking-2507": 32768,
278
+ "qwen/qwen3-235b-a22b:free": 32768,
279
+ "qwen/qwen3-30b-a3b": 32768,
280
+ "qwen/qwen3-30b-a3b-instruct-2507": 32768,
281
+ "qwen/qwen3-30b-a3b:free": 32768,
282
+ "qwen/qwen3-32b": 32768,
283
+ "qwen/qwen3-4b:free": 32768,
284
+ "qwen/qwen3-8b": 32768,
285
+ "qwen/qwen3-8b:free": 32768,
286
+ "qwen/qwen3-coder": 32768,
287
+ "qwen/qwen3-coder:free": 32768,
288
+ "qwen/qwq-32b": 32768,
289
+ "qwen/qwq-32b-preview": 32768,
290
+ "qwen/qwq-32b:free": 32768,
291
+ "raifle/sorcererlm-8x22b": 65536,
292
+ "rekaai/reka-flash-3:free": 128000,
293
+ "sao10k/l3-euryale-70b": 8192,
294
+ "sao10k/l3-lunaris-8b": 8192,
295
+ "sao10k/l3.1-euryale-70b": 131072,
296
+ "sao10k/l3.3-euryale-70b": 131072,
297
+ "sarvamai/sarvam-m:free": 8192,
298
+ "scb10x/llama3.1-typhoon2-70b-instruct": 131072,
299
+ "shisa-ai/shisa-v2-llama3.3-70b": 131072,
300
+ "shisa-ai/shisa-v2-llama3.3-70b:free": 131072,
301
+ "sophosympatheia/midnight-rose-70b": 4096,
302
+ "switchpoint/router": 8192,
303
+ "tencent/hunyuan-a13b-instruct": 8192,
304
+ "tencent/hunyuan-a13b-instruct:free": 8192,
305
+ "thedrummer/anubis-70b-v1.1": 8192,
306
+ "thedrummer/anubis-pro-105b-v1": 8192,
307
+ "thedrummer/rocinante-12b": 8192,
308
+ "thedrummer/skyfall-36b-v2": 8192,
309
+ "thedrummer/unslopnemo-12b": 128000,
310
+ "thedrummer/valkyrie-49b-v1": 8192,
311
+ "thudm/glm-4-32b": 2000000,
312
+ "thudm/glm-4.1v-9b-thinking": 2000000,
313
+ "thudm/glm-z1-32b:free": 2000000,
314
+ "tngtech/deepseek-r1t-chimera": 32768,
315
+ "tngtech/deepseek-r1t-chimera:free": 32768,
316
+ "tngtech/deepseek-r1t2-chimera:free": 32768,
317
+ "undi95/remm-slerp-l2-13b": 4096,
318
+ "x-ai/grok-2-1212": 128000,
319
+ "x-ai/grok-2-vision-1212": 128000,
320
+ "x-ai/grok-3": 128000,
321
+ "x-ai/grok-3-beta": 128000,
322
+ "x-ai/grok-3-mini": 128000,
323
+ "x-ai/grok-3-mini-beta": 128000,
324
+ "x-ai/grok-4": 128000,
325
+ "x-ai/grok-vision-beta": 128000,
326
+ "z-ai/glm-4-32b": 2000000,
327
+ "z-ai/glm-4.5": 2000000,
328
+ "z-ai/glm-4.5-air": 2000000,
329
+ "z-ai/glm-4.5-air:free": 2000000,
330
+ "llama3.1": 131072,
331
+ "llama3.2": 131072,
332
+ "llama3.3": 131072,
333
+ "llama3": 8192,
334
+ "llama2": 4096,
335
+ "mixtral8x22b": 65536,
336
+ "mixtral": 32768,
337
+ "mistral": 32768,
338
+ "gemma3": 131072,
339
+ "gemma2": 8192,
340
+ "gemma": 8192,
341
+ "phi3": 131072,
342
+ "phi2": 2048,
343
+ "phi": 2048,
344
+ "qwen2.5": 131072,
345
+ "qwen2": 32768,
346
+ "qwen": 8192,
347
+ "codellama": 16384,
348
+ "codegemma": 8192,
349
+ "deepseek-coder-v2": 131072,
350
+ "deepseek-coder": 16384,
351
+ "deepseek-v2": 131072,
352
+ "deepseek-llm": 4096,
353
+ "yi1.5": 32768,
354
+ "yi": 4096,
355
+ "command-r": 131072,
356
+ "wizardlm2": 32768,
357
+ "wizardlm": 16384,
358
+ "zephyr": 65536,
359
+ "vicuna": 2048,
360
+ "falcon": 2048,
361
+ "starcoder": 8192,
362
+ "stablelm": 4096,
363
+ "orca2": 4096,
364
+ "orca": 4096,
365
+ "dolphin": 32768,
366
+ "openhermes": 8192,
367
+ "gpt-oss": 128000,
368
+ "gpt-3.5-turbo": 4096,
369
+ "gpt-4": 8192,
370
+ "grok-2": 128000,
371
+ "grok-2-1212": 128000,
372
+ "grok-2-vision-1212": 128000,
373
+ "grok-3": 128000,
374
+ "grok-3-fast": 128000,
375
+ "grok-3-beta": 128000,
376
+ "grok-3-mini": 128000,
377
+ "grok-3-mini-beta": 128000,
378
+ "grok-3-mini-fast": 128000,
379
+ "grok-4-0709": 128000,
380
+ "grok-4": 128000,
381
+ "grok-vision-beta": 128000
382
+ }
@@ -9,7 +9,7 @@ from lollms_client.lollms_discussion import LollmsDiscussion
9
9
  from typing import Optional, Callable, List, Union
10
10
  from ascii_colors import ASCIIColors, trace_exception
11
11
  from typing import List, Dict
12
-
12
+ import httpx
13
13
  import pipmaster as pm
14
14
 
15
15
  pm.ensure_packages(["openai","tiktoken"])
@@ -49,7 +49,7 @@ class LollmsBinding(LollmsLLMBinding):
49
49
 
50
50
  if not self.service_key:
51
51
  self.service_key = os.getenv("LOLLMS_API_KEY", self.service_key)
52
- self.client = openai.OpenAI(api_key=self.service_key, base_url=None if self.host_address is None else self.host_address if len(self.host_address)>0 else None)
52
+ self.client = openai.OpenAI(api_key=self.service_key, base_url=None if self.host_address is None else self.host_address if len(self.host_address)>0 else None, http_client=httpx.Client(verify=self.verify_ssl_certificate))
53
53
  self.completion_format = ELF_COMPLETION_FORMAT.Chat
54
54
 
55
55
  def lollms_listMountedPersonalities(self, host_address:str|None=None):
@@ -680,7 +680,63 @@ class OllamaBinding(LollmsLLMBinding):
680
680
 
681
681
  ASCIIColors.warning(f"Context size not found for model '{model_name}'")
682
682
  return None
683
+
684
+ def ps(self):
685
+ """
686
+ Lists running models in a standardized, flat format.
687
+
688
+ This method corresponds to the /api/ps endpoint in the Ollama API. It retrieves
689
+ the models currently loaded into memory and transforms the data into a simplified,
690
+ flat list of dictionaries.
691
+
692
+ Returns:
693
+ list[dict]: A list of dictionaries, each representing a running model with a standardized set of keys.
694
+ Returns an empty list if the client is not initialized or if an error occurs.
683
695
 
696
+ Example of a returned model dictionary:
697
+ {
698
+ "model_name": "gemma3:12b",
699
+ "size": 13861175232,
700
+ "vram_size": 10961479680,
701
+ "parameters_size": "12.2B",
702
+ "quantization_level": "Q4_K_M",
703
+ "context_size": 32000,
704
+ "parent_model": "",
705
+ "expires_at": "2025-08-20T22:28:18.6708784+02:00"
706
+ }
707
+ """
708
+ if not self.ollama_client:
709
+ ASCIIColors.warning("Ollama client not initialized. Cannot list running models.")
710
+ return []
711
+
712
+ try:
713
+ running_models_response = self.ollama_client.ps()
714
+
715
+ models_list = running_models_response.get('models', [])
716
+ standardized_models = []
717
+
718
+ for model_data in models_list:
719
+ details = model_data.get('details', {})
720
+
721
+ flat_model_info = {
722
+ "model_name": model_data.get("name"),
723
+ "size": model_data.get("size"),
724
+ "vram_size": model_data.get("size_vram"),
725
+ "expires_at": model_data.get("expires_at"),
726
+ "parameters_size": details.get("parameter_size"),
727
+ "quantization_level": details.get("quantization_level"),
728
+ "parent_model": details.get("parent_model"),
729
+ # Add context_size if it exists in the details
730
+ "context_size": details.get("context_length")
731
+ }
732
+ standardized_models.append(flat_model_info)
733
+
734
+ return standardized_models
735
+
736
+ except Exception as e:
737
+ ASCIIColors.error(f"Failed to list running models from Ollama at {self.host_address}: {e}")
738
+ return []
739
+
684
740
  if __name__ == '__main__':
685
741
  global full_streamed_text
686
742
  # Example Usage (requires an Ollama server running)
@@ -10,7 +10,7 @@ from typing import Optional, Callable, List, Union
10
10
  from ascii_colors import ASCIIColors, trace_exception
11
11
  from typing import List, Dict
12
12
  import math
13
-
13
+ import httpx
14
14
  import pipmaster as pm
15
15
 
16
16
  pm.ensure_packages(["openai","tiktoken"])
@@ -47,7 +47,7 @@ class OpenAIBinding(LollmsLLMBinding):
47
47
 
48
48
  if not self.service_key:
49
49
  self.service_key = os.getenv("OPENAI_API_KEY", self.service_key)
50
- self.client = openai.OpenAI(api_key=self.service_key, base_url=None if self.host_address is None else self.host_address if len(self.host_address)>0 else None)
50
+ self.client = openai.OpenAI(api_key=self.service_key, base_url=None if self.host_address is None else self.host_address if len(self.host_address)>0 else None, http_client=httpx.Client(verify=self.verify_ssl_certificate))
51
51
  self.completion_format = ELF_COMPLETION_FORMAT.Chat
52
52
 
53
53
  def _build_openai_params(self, messages: list, **kwargs) -> dict:
@@ -668,4 +668,4 @@ class OpenAIBinding(LollmsLLMBinding):
668
668
  """
669
669
  self.model = model_name
670
670
  self.model_name = model_name
671
- return True
671
+ return True