@1mancompany/onemancompany 0.7.47 → 0.7.48

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -26,11 +26,11 @@ const STATUS_STYLES = {
26
26
  blocked: { color: '#f97316', label: '\u2298 Blocked' },
27
27
  };
28
28
 
29
- const CARD_W = 240;
29
+ const CARD_W = 260;
30
30
  const CARD_MIN_H = 72; // minimum card height (no description)
31
31
  const DESC_LINE_H = 14; // height per description line
32
- const DESC_MAX_LINES = 6; // cap description lines
33
- const DESC_CHARS_PER_LINE = 34;
32
+ const DESC_MAX_LINES = 12; // cap description lines (was 6 — caused overflow/truncation)
33
+ const DESC_CHARS_PER_LINE = 38;
34
34
  const CHILDREN_PAGE_SIZE = 5;
35
35
 
36
36
  /* ─────────────────── Word-wrap helper (shared with old code) ────── */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@1mancompany/onemancompany",
3
- "version": "0.7.47",
3
+ "version": "0.7.48",
4
4
  "description": "The AI Operating System for One-Person Companies",
5
5
  "bin": {
6
6
  "onemancompany": "bin/cli.js"
package/pyproject.toml CHANGED
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "onemancompany"
3
- version = "0.7.47"
3
+ version = "0.7.48"
4
4
  description = "A one-man company simulation with pixel art visualization and LangChain AI agents"
5
5
  requires-python = ">=3.12"
6
6
  dependencies = [
@@ -641,6 +641,7 @@ class BaseAgentRunner:
641
641
  content = last_msg.content or ""
642
642
  if isinstance(content, str):
643
643
  on_log("llm_input", f"[{type(last_msg).__name__}] {content}")
644
+ logger.debug("[LLM INPUT] employee={}: {}", self.employee_id, content[:3000])
644
645
  elif kind == "on_chat_model_end":
645
646
  output = data.get("output", None)
646
647
  if output:
@@ -671,6 +672,7 @@ class BaseAgentRunner:
671
672
  if text.strip():
672
673
  final_content = text # track last AI output
673
674
  on_log("llm_output", text)
675
+ logger.debug("[LLM OUTPUT] employee={}: {}", self.employee_id, text[:3000])
674
676
  tool_calls = getattr(output, "tool_calls", None)
675
677
  if tool_calls:
676
678
  last_tool_calls = []
@@ -685,11 +687,13 @@ class BaseAgentRunner:
685
687
  "tool_args": args_dict,
686
688
  "content": f"{name}({args})",
687
689
  })
690
+ logger.debug("[TOOL CALL] employee={}: {}({})", self.employee_id, name, args[:1000])
688
691
  elif kind == "on_tool_end":
689
692
  output = data.get("output", "")
690
693
  name = event.get("name", "tool")
691
694
  result_str = str(output)
692
695
  last_tool_results.append(f"{name} → {result_str}")
696
+ logger.debug("[TOOL RESULT] employee={}: {} → {}", self.employee_id, name, result_str[:2000])
693
697
  on_log("tool_result", {
694
698
  "tool_name": name,
695
699
  "tool_result": result_str,