lollms-client 1.3.8__tar.gz → 1.4.0__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of lollms-client might be problematic. Click here for more details.
- {lollms_client-1.3.8/src/lollms_client.egg-info → lollms_client-1.4.0}/PKG-INFO +4 -4
- {lollms_client-1.3.8 → lollms_client-1.4.0}/README.md +3 -3
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/__init__.py +1 -1
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_core.py +2200 -110
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_discussion.py +227 -75
- {lollms_client-1.3.8 → lollms_client-1.4.0/src/lollms_client.egg-info}/PKG-INFO +4 -4
- {lollms_client-1.3.8 → lollms_client-1.4.0}/LICENSE +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/pyproject.toml +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/setup.cfg +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/assets/models_ctx_sizes.json +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/azure_openai/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/claude/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/gemini/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/grok/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/groq/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/hugging_face_inference_api/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/litellm/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/llamacpp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/lollms_webui/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/mistral/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/ollama/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/open_router/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/openai/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/openllm/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/pythonllamacpp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/tensor_rt/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/transformers/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/llm_bindings/vllm/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_agentic.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_config.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_js_analyzer.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_llm_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_mcp_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_mcp_security.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_personality.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_python_analyzer.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_stt_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_tti_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_ttm_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_tts_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_ttv_binding.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_types.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/lollms_utilities.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/local_mcp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/file_writer/file_writer.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/generate_image_from_prompt/generate_image_from_prompt.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/internet_search/internet_search.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/local_mcp/default_tools/python_interpreter/python_interpreter.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/remote_mcp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/mcp_bindings/standard_mcp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/stt_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/stt_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/stt_bindings/whisper/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/stt_bindings/whispercpp/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tti_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tti_bindings/diffusers/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tti_bindings/gemini/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tti_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tti_bindings/openai/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttm_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttm_bindings/audiocraft/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttm_bindings/bark/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttm_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/bark/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/bark/server/install_bark.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/bark/server/main.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/piper_tts/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/piper_tts/server/install_piper.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/piper_tts/server/main.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/piper_tts/server/setup_voices.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/xtts/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/xtts/server/main.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/tts_bindings/xtts/server/setup_voices.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttv_bindings/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client/ttv_bindings/lollms/__init__.py +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client.egg-info/SOURCES.txt +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client.egg-info/dependency_links.txt +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client.egg-info/requires.txt +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/src/lollms_client.egg-info/top_level.txt +0 -0
- {lollms_client-1.3.8 → lollms_client-1.4.0}/test/test_lollms_discussion.py +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: lollms_client
|
|
3
|
-
Version: 1.
|
|
3
|
+
Version: 1.4.0
|
|
4
4
|
Summary: A client library for LoLLMs generate endpoint
|
|
5
5
|
Author-email: ParisNeo <parisneoai@gmail.com>
|
|
6
6
|
License: Apache Software License
|
|
@@ -56,7 +56,7 @@ Whether you're connecting to a remote LoLLMs server, an Ollama instance, the Ope
|
|
|
56
56
|
* 📝 **Advanced Structured Content Generation:** Reliably generate structured JSON output from natural language prompts using the `generate_structured_content` helper method, enforcing a specific schema.
|
|
57
57
|
* 💬 **Advanced Discussion Management:** Robustly manage conversation histories with `LollmsDiscussion`, featuring branching, context exporting, and automatic pruning.
|
|
58
58
|
* 🧠 **Persistent Memory & Data Zones:** `LollmsDiscussion` now supports multiple, distinct data zones (`user_data_zone`, `discussion_data_zone`, `personality_data_zone`) and a long-term `memory` field. This allows for sophisticated context layering and state management, enabling agents to learn and remember over time.
|
|
59
|
-
* ✍️ **
|
|
59
|
+
* ✍️ **Structured Memorization:** The `memorize()` method analyzes a conversation to extract its essence (e.g., a problem and its solution), creating a structured "memory" with a title and content. These memories are stored and can be explicitly loaded into the AI's context, providing a more robust and manageable long-term memory system.
|
|
60
60
|
* 📊 **Detailed Context Analysis:** The `get_context_status()` method provides a rich, detailed breakdown of the prompt context, showing the content and token count for each individual component (system prompt, data zones, message history).
|
|
61
61
|
* ⚙️ **Standardized Configuration Management:** A unified dictionary-based system (`llm_binding_config`) to configure any binding in a consistent manner.
|
|
62
62
|
* 🧩 **Extensible:** Designed to easily incorporate new LLM backends and modality services, including custom MCP toolsets.
|
|
@@ -384,7 +384,7 @@ with tempfile.TemporaryDirectory() as tmpdir:
|
|
|
384
384
|
# The 'breakdown' shows the individual zones that were combined
|
|
385
385
|
for name, content in sys_ctx.get('breakdown', {}).items():
|
|
386
386
|
# For brevity, show only first line of content
|
|
387
|
-
print(f" -> Contains '{name}': {content.split(os.linesep)
|
|
387
|
+
print(f" -> Contains '{name}': {content.split(os.linesep)}...")
|
|
388
388
|
|
|
389
389
|
# Print the message history details
|
|
390
390
|
if 'message_history' in status['zones']:
|
|
@@ -424,7 +424,7 @@ with tempfile.TemporaryDirectory() as tmpdir:
|
|
|
424
424
|
if name == 'memory':
|
|
425
425
|
ASCIIColors.yellow(f" -> Full '{name}' content:\n{content}")
|
|
426
426
|
else:
|
|
427
|
-
print(f" -> Contains '{name}': {content.split(os.linesep)
|
|
427
|
+
print(f" -> Contains '{name}': {content.split(os.linesep)}...")
|
|
428
428
|
print("------------------------------------------")
|
|
429
429
|
|
|
430
430
|
```
|
|
@@ -25,7 +25,7 @@ Whether you're connecting to a remote LoLLMs server, an Ollama instance, the Ope
|
|
|
25
25
|
* 📝 **Advanced Structured Content Generation:** Reliably generate structured JSON output from natural language prompts using the `generate_structured_content` helper method, enforcing a specific schema.
|
|
26
26
|
* 💬 **Advanced Discussion Management:** Robustly manage conversation histories with `LollmsDiscussion`, featuring branching, context exporting, and automatic pruning.
|
|
27
27
|
* 🧠 **Persistent Memory & Data Zones:** `LollmsDiscussion` now supports multiple, distinct data zones (`user_data_zone`, `discussion_data_zone`, `personality_data_zone`) and a long-term `memory` field. This allows for sophisticated context layering and state management, enabling agents to learn and remember over time.
|
|
28
|
-
* ✍️ **
|
|
28
|
+
* ✍️ **Structured Memorization:** The `memorize()` method analyzes a conversation to extract its essence (e.g., a problem and its solution), creating a structured "memory" with a title and content. These memories are stored and can be explicitly loaded into the AI's context, providing a more robust and manageable long-term memory system.
|
|
29
29
|
* 📊 **Detailed Context Analysis:** The `get_context_status()` method provides a rich, detailed breakdown of the prompt context, showing the content and token count for each individual component (system prompt, data zones, message history).
|
|
30
30
|
* ⚙️ **Standardized Configuration Management:** A unified dictionary-based system (`llm_binding_config`) to configure any binding in a consistent manner.
|
|
31
31
|
* 🧩 **Extensible:** Designed to easily incorporate new LLM backends and modality services, including custom MCP toolsets.
|
|
@@ -353,7 +353,7 @@ with tempfile.TemporaryDirectory() as tmpdir:
|
|
|
353
353
|
# The 'breakdown' shows the individual zones that were combined
|
|
354
354
|
for name, content in sys_ctx.get('breakdown', {}).items():
|
|
355
355
|
# For brevity, show only first line of content
|
|
356
|
-
print(f" -> Contains '{name}': {content.split(os.linesep)
|
|
356
|
+
print(f" -> Contains '{name}': {content.split(os.linesep)}...")
|
|
357
357
|
|
|
358
358
|
# Print the message history details
|
|
359
359
|
if 'message_history' in status['zones']:
|
|
@@ -393,7 +393,7 @@ with tempfile.TemporaryDirectory() as tmpdir:
|
|
|
393
393
|
if name == 'memory':
|
|
394
394
|
ASCIIColors.yellow(f" -> Full '{name}' content:\n{content}")
|
|
395
395
|
else:
|
|
396
|
-
print(f" -> Contains '{name}': {content.split(os.linesep)
|
|
396
|
+
print(f" -> Contains '{name}': {content.split(os.linesep)}...")
|
|
397
397
|
print("------------------------------------------")
|
|
398
398
|
|
|
399
399
|
```
|
|
@@ -8,7 +8,7 @@ from lollms_client.lollms_utilities import PromptReshaper # Keep general utiliti
|
|
|
8
8
|
from lollms_client.lollms_mcp_binding import LollmsMCPBinding, LollmsMCPBindingManager
|
|
9
9
|
from lollms_client.lollms_llm_binding import LollmsLLMBindingManager
|
|
10
10
|
|
|
11
|
-
__version__ = "1.
|
|
11
|
+
__version__ = "1.4.0" # Updated version
|
|
12
12
|
|
|
13
13
|
# Optionally, you could define __all__ if you want to be explicit about exports
|
|
14
14
|
__all__ = [
|