lollms-client 0.24.0__py3-none-any.whl → 0.24.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of lollms-client might be problematic. Click here for more details.
- lollms_client/__init__.py +1 -1
- lollms_client/lollms_core.py +6 -1
- {lollms_client-0.24.0.dist-info → lollms_client-0.24.1.dist-info}/METADATA +1 -1
- {lollms_client-0.24.0.dist-info → lollms_client-0.24.1.dist-info}/RECORD +7 -7
- {lollms_client-0.24.0.dist-info → lollms_client-0.24.1.dist-info}/WHEEL +0 -0
- {lollms_client-0.24.0.dist-info → lollms_client-0.24.1.dist-info}/licenses/LICENSE +0 -0
- {lollms_client-0.24.0.dist-info → lollms_client-0.24.1.dist-info}/top_level.txt +0 -0
lollms_client/__init__.py
CHANGED
|
@@ -8,7 +8,7 @@ from lollms_client.lollms_utilities import PromptReshaper # Keep general utiliti
|
|
|
8
8
|
from lollms_client.lollms_mcp_binding import LollmsMCPBinding, LollmsMCPBindingManager
|
|
9
9
|
|
|
10
10
|
|
|
11
|
-
__version__ = "0.24.
|
|
11
|
+
__version__ = "0.24.1" # Updated version
|
|
12
12
|
|
|
13
13
|
# Optionally, you could define __all__ if you want to be explicit about exports
|
|
14
14
|
__all__ = [
|
lollms_client/lollms_core.py
CHANGED
|
@@ -1606,16 +1606,19 @@ Provide your response as a single JSON object inside a JSON markdown tag. Use th
|
|
|
1606
1606
|
tool_params = action.get("tool_params", {})
|
|
1607
1607
|
except (json.JSONDecodeError, TypeError) as e:
|
|
1608
1608
|
current_scratchpad += f"\n\n### Step {i+1} Failure\n- **Error:** Failed to generate a valid JSON action: {e}"
|
|
1609
|
+
log_step(f"\n\n### Step {i+1} Failure\n- **Error:** Failed to generate a valid JSON action: {e}", "scratchpad", is_start=False)
|
|
1609
1610
|
if reasoning_step_id:
|
|
1610
1611
|
log_step(f"Reasoning Step {i+1}/{max_reasoning_steps}", "reasoning_step", metadata={"id": reasoning_step_id, "error": str(e)}, is_start=False)
|
|
1611
1612
|
break
|
|
1612
1613
|
|
|
1613
1614
|
current_scratchpad += f"\n\n### Step {i+1}: Thought\n{thought}"
|
|
1615
|
+
log_step(f"\n\n### Step {i+1}: Thought\n{thought}", "scratchpad", is_start=False)
|
|
1614
1616
|
if streaming_callback:
|
|
1615
1617
|
streaming_callback(thought, MSG_TYPE.MSG_TYPE_INFO, {"type": "thought"})
|
|
1616
1618
|
|
|
1617
1619
|
if not tool_name:
|
|
1618
1620
|
current_scratchpad += f"\n\n### Step {i+1} Failure\n- **Error:** Did not specify a tool name."
|
|
1621
|
+
log_step(f"\n\n### Step {i+1} Failure\n- **Error:** Did not specify a tool name.", "scratchpad", is_start=False)
|
|
1619
1622
|
if reasoning_step_id:
|
|
1620
1623
|
log_step(f"Reasoning Step {i+1}/{max_reasoning_steps}", "reasoning_step", metadata={"id": reasoning_step_id}, is_start=False)
|
|
1621
1624
|
break
|
|
@@ -1623,12 +1626,14 @@ Provide your response as a single JSON object inside a JSON markdown tag. Use th
|
|
|
1623
1626
|
if tool_name == "request_clarification":
|
|
1624
1627
|
clarification_question = action.get("clarification_question", "Could you please provide more details?")
|
|
1625
1628
|
current_scratchpad += f"\n\n### Step {i+1}: Action\n- **Action:** Decided to request clarification.\n- **Question:** {clarification_question}"
|
|
1629
|
+
log_step(f"\n\n### Step {i+1}: Action\n- **Action:** Decided to request clarification.\n- **Question:** {clarification_question}", "scratchpad", is_start=False)
|
|
1626
1630
|
if reasoning_step_id:
|
|
1627
1631
|
log_step(f"Reasoning Step {i+1}/{max_reasoning_steps}", "reasoning_step", metadata={"id": reasoning_step_id}, is_start=False)
|
|
1628
1632
|
return {"final_answer": clarification_question, "final_scratchpad": current_scratchpad, "tool_calls": tool_calls_this_turn, "sources": sources_this_turn, "clarification_required": True, "error": None}
|
|
1629
1633
|
|
|
1630
1634
|
if tool_name == "final_answer":
|
|
1631
1635
|
current_scratchpad += f"\n\n### Step {i+1}: Action\n- **Action:** Decided to formulate the final answer."
|
|
1636
|
+
log_step(f"\n\n### Step {i+1}: Action\n- **Action:** Decided to formulate the final answer.", "scratchpad", is_start=False)
|
|
1632
1637
|
if reasoning_step_id:
|
|
1633
1638
|
log_step(f"Reasoning Step {i+1}/{max_reasoning_steps}", "reasoning_step", metadata={"id": reasoning_step_id}, is_start=False)
|
|
1634
1639
|
break
|
|
@@ -1681,7 +1686,7 @@ Provide your response as a single JSON object inside a JSON markdown tag. Use th
|
|
|
1681
1686
|
|
|
1682
1687
|
tool_calls_this_turn.append({"name": tool_name, "params": tool_params, "result": tool_result})
|
|
1683
1688
|
current_scratchpad += f"\n\n### Step {i+1}: Observation\n- **Action:** Called `{tool_name}`\n- **Result:**\n{observation_text}"
|
|
1684
|
-
log_step("{
|
|
1689
|
+
log_step(f"### Step {i+1}: Observation\n- **Action:** Called `{tool_name}`\n", "scratchpad", is_start=False)
|
|
1685
1690
|
|
|
1686
1691
|
if reasoning_step_id:
|
|
1687
1692
|
log_step(f"Reasoning Step {i+1}/{max_reasoning_steps}", "reasoning_step", metadata={"id": reasoning_step_id}, is_start=False)
|
|
@@ -26,9 +26,9 @@ examples/mcp_examples/openai_mcp.py,sha256=7IEnPGPXZgYZyiES_VaUbQ6viQjenpcUxGiHE
|
|
|
26
26
|
examples/mcp_examples/run_remote_mcp_example_v2.py,sha256=bbNn93NO_lKcFzfIsdvJJijGx2ePFTYfknofqZxMuRM,14626
|
|
27
27
|
examples/mcp_examples/run_standard_mcp_example.py,sha256=GSZpaACPf3mDPsjA8esBQVUsIi7owI39ca5avsmvCxA,9419
|
|
28
28
|
examples/test_local_models/local_chat.py,sha256=slakja2zaHOEAUsn2tn_VmI4kLx6luLBrPqAeaNsix8,456
|
|
29
|
-
lollms_client/__init__.py,sha256=
|
|
29
|
+
lollms_client/__init__.py,sha256=hphTI6chtuepiPoUBzI6cYGTu07c4BTD8JsniICS2qo,1047
|
|
30
30
|
lollms_client/lollms_config.py,sha256=goEseDwDxYJf3WkYJ4IrLXwg3Tfw73CXV2Avg45M_hE,21876
|
|
31
|
-
lollms_client/lollms_core.py,sha256=
|
|
31
|
+
lollms_client/lollms_core.py,sha256=odJ97aTFkc_tXAfnbyO5C2MstJC4fPqDR0Zusw-vcW0,152646
|
|
32
32
|
lollms_client/lollms_discussion.py,sha256=e0lfbkj3Bh-1YbC1JncjO0Dvkx9maFXprAwLLiXxSFc,44578
|
|
33
33
|
lollms_client/lollms_js_analyzer.py,sha256=01zUvuO2F_lnUe_0NLxe1MF5aHE1hO8RZi48mNPv-aw,8361
|
|
34
34
|
lollms_client/lollms_llm_binding.py,sha256=Kpzhs5Jx8eAlaaUacYnKV7qIq2wbME5lOEtKSfJKbpg,12161
|
|
@@ -79,8 +79,8 @@ lollms_client/tts_bindings/piper_tts/__init__.py,sha256=0IEWG4zH3_sOkSb9WbZzkeV5
|
|
|
79
79
|
lollms_client/tts_bindings/xtts/__init__.py,sha256=FgcdUH06X6ZR806WQe5ixaYx0QoxtAcOgYo87a2qxYc,18266
|
|
80
80
|
lollms_client/ttv_bindings/__init__.py,sha256=UZ8o2izQOJLQgtZ1D1cXoNST7rzqW22rL2Vufc7ddRc,3141
|
|
81
81
|
lollms_client/ttv_bindings/lollms/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
82
|
-
lollms_client-0.24.
|
|
83
|
-
lollms_client-0.24.
|
|
84
|
-
lollms_client-0.24.
|
|
85
|
-
lollms_client-0.24.
|
|
86
|
-
lollms_client-0.24.
|
|
82
|
+
lollms_client-0.24.1.dist-info/licenses/LICENSE,sha256=HrhfyXIkWY2tGFK11kg7vPCqhgh5DcxleloqdhrpyMY,11558
|
|
83
|
+
lollms_client-0.24.1.dist-info/METADATA,sha256=RMCh7EHn4YiKXnGjDssZdehDM-Wza5OYdl7iWEB9FuQ,13401
|
|
84
|
+
lollms_client-0.24.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
85
|
+
lollms_client-0.24.1.dist-info/top_level.txt,sha256=NI_W8S4OYZvJjb0QWMZMSIpOrYzpqwPGYaklhyWKH2w,23
|
|
86
|
+
lollms_client-0.24.1.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|