owlmind 0.1.5.1__tar.gz → 0.1.7__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (23) hide show
  1. {owlmind-0.1.5.1/src/owlmind.egg-info → owlmind-0.1.7}/PKG-INFO +24 -2
  2. {owlmind-0.1.5.1 → owlmind-0.1.7}/README.md +23 -1
  3. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/__init__.py +1 -1
  4. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/models/ollama.py +1 -2
  5. {owlmind-0.1.5.1 → owlmind-0.1.7/src/owlmind.egg-info}/PKG-INFO +24 -2
  6. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind.egg-info/SOURCES.txt +0 -1
  7. owlmind-0.1.5.1/src/owlmind/graphk/pipeline.py +0 -94
  8. {owlmind-0.1.5.1 → owlmind-0.1.7}/LICENSE.txt +0 -0
  9. {owlmind-0.1.5.1 → owlmind-0.1.7}/pyproject.toml +0 -0
  10. {owlmind-0.1.5.1 → owlmind-0.1.7}/setup.cfg +0 -0
  11. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/cli.py +0 -0
  12. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/core/__init__.py +0 -0
  13. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/core/component.py +0 -0
  14. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/graphk/__init__.py +0 -0
  15. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/graphk/node.py +0 -0
  16. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind/models/__init__.py +0 -0
  17. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind.egg-info/dependency_links.txt +0 -0
  18. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind.egg-info/entry_points.txt +0 -0
  19. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind.egg-info/requires.txt +0 -0
  20. {owlmind-0.1.5.1 → owlmind-0.1.7}/src/owlmind.egg-info/top_level.txt +0 -0
  21. {owlmind-0.1.5.1 → owlmind-0.1.7}/tests/test_core_1.py +0 -0
  22. {owlmind-0.1.5.1 → owlmind-0.1.7}/tests/test_models_1.py +0 -0
  23. {owlmind-0.1.5.1 → owlmind-0.1.7}/tests/test_pipeline_1.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: owlmind
3
- Version: 0.1.5.1
3
+ Version: 0.1.7
4
4
  Summary: Experimentation environment and pedagogical sandbox for studying generative intelligence systems.
5
5
  Author-email: Fernando Koch <your-email@example.com>
6
6
  License: MIT
@@ -73,6 +73,10 @@ Verify if your model provider is online.
73
73
  owlmind ping
74
74
  ```
75
75
 
76
+ It should return:
77
+ ```bash
78
+ Status: ONLINE (Host: http://localhost:11434)
79
+ ```
76
80
 
77
81
  #### Information
78
82
  View your environment information.
@@ -81,11 +85,29 @@ View your environment information.
81
85
  owlmind info
82
86
  ```
83
87
 
88
+ It should return something like:
89
+
90
+ ```bash
91
+ $ owlmind info
92
+ ----------------------------------------
93
+ Status : online
94
+ Host : http://localhost:11434
95
+ Model : llama3
96
+ ----------------------------------------
97
+ Available Models: 5
98
+ - gpt-oss:latest
99
+ - gemma:latest
100
+ - tinyllama:latest
101
+ - llama3:latest
102
+ - llama3.2:latest
103
+ ----------------------------------------
104
+ ```
105
+
84
106
  #### Generation with Parameters
85
107
  Run inference with full control over sampling parameters.
86
108
 
87
109
  ```bash
88
- owlmind query "How do AI-driven organizations scale?" --temp 1.2 --ctx-size 4096
110
+ owlmind query "How do AI-driven organizations scale?" -p temperature=1.2, ctx-size=4096
89
111
  ```
90
112
 
91
113
  Other parameters:
@@ -60,6 +60,10 @@ Verify if your model provider is online.
60
60
  owlmind ping
61
61
  ```
62
62
 
63
+ It should return:
64
+ ```bash
65
+ Status: ONLINE (Host: http://localhost:11434)
66
+ ```
63
67
 
64
68
  #### Information
65
69
  View your environment information.
@@ -68,11 +72,29 @@ View your environment information.
68
72
  owlmind info
69
73
  ```
70
74
 
75
+ It should return something like:
76
+
77
+ ```bash
78
+ $ owlmind info
79
+ ----------------------------------------
80
+ Status : online
81
+ Host : http://localhost:11434
82
+ Model : llama3
83
+ ----------------------------------------
84
+ Available Models: 5
85
+ - gpt-oss:latest
86
+ - gemma:latest
87
+ - tinyllama:latest
88
+ - llama3:latest
89
+ - llama3.2:latest
90
+ ----------------------------------------
91
+ ```
92
+
71
93
  #### Generation with Parameters
72
94
  Run inference with full control over sampling parameters.
73
95
 
74
96
  ```bash
75
- owlmind query "How do AI-driven organizations scale?" --temp 1.2 --ctx-size 4096
97
+ owlmind query "How do AI-driven organizations scale?" -p temperature=1.2, ctx-size=4096
76
98
  ```
77
99
 
78
100
  Other parameters:
@@ -2,4 +2,4 @@
2
2
  OwlMind Framework - experimentation environment for Generative Intelligence Systems.
3
3
  """
4
4
 
5
- __version__ = "0.1.5.1"
5
+ __version__ = "0.1.7"
@@ -25,7 +25,6 @@ class Ollama(Component):
25
25
 
26
26
  DEFAULT_SERVER = "http://localhost:11434"
27
27
  DEFAULT_MODEL = "llama3"
28
- DEFAULT_TIMEOUT = 10
29
28
  DEFAULT_LOG_LEVEL = logging.INFO
30
29
 
31
30
  OLLAMA_PARAMS = {
@@ -54,7 +53,7 @@ class Ollama(Component):
54
53
  if not hasattr(self, param): setattr(self, param, default_value)
55
54
 
56
55
  # Initialize the client using the framework internal
57
- self._client_ = ollama.Client(host=self.url, timeout=self.DEFAULT_TIMEOUT)
56
+ self._client_ = ollama.Client(host=self.url)
58
57
 
59
58
  # Obfuscate the client-related keys just in case
60
59
  self.obfuscate(['_client_', '_models_cache_'])
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: owlmind
3
- Version: 0.1.5.1
3
+ Version: 0.1.7
4
4
  Summary: Experimentation environment and pedagogical sandbox for studying generative intelligence systems.
5
5
  Author-email: Fernando Koch <your-email@example.com>
6
6
  License: MIT
@@ -73,6 +73,10 @@ Verify if your model provider is online.
73
73
  owlmind ping
74
74
  ```
75
75
 
76
+ It should return:
77
+ ```bash
78
+ Status: ONLINE (Host: http://localhost:11434)
79
+ ```
76
80
 
77
81
  #### Information
78
82
  View your environment information.
@@ -81,11 +85,29 @@ View your environment information.
81
85
  owlmind info
82
86
  ```
83
87
 
88
+ It should return something like:
89
+
90
+ ```bash
91
+ $ owlmind info
92
+ ----------------------------------------
93
+ Status : online
94
+ Host : http://localhost:11434
95
+ Model : llama3
96
+ ----------------------------------------
97
+ Available Models: 5
98
+ - gpt-oss:latest
99
+ - gemma:latest
100
+ - tinyllama:latest
101
+ - llama3:latest
102
+ - llama3.2:latest
103
+ ----------------------------------------
104
+ ```
105
+
84
106
  #### Generation with Parameters
85
107
  Run inference with full control over sampling parameters.
86
108
 
87
109
  ```bash
88
- owlmind query "How do AI-driven organizations scale?" --temp 1.2 --ctx-size 4096
110
+ owlmind query "How do AI-driven organizations scale?" -p temperature=1.2, ctx-size=4096
89
111
  ```
90
112
 
91
113
  Other parameters:
@@ -13,7 +13,6 @@ src/owlmind/core/__init__.py
13
13
  src/owlmind/core/component.py
14
14
  src/owlmind/graphk/__init__.py
15
15
  src/owlmind/graphk/node.py
16
- src/owlmind/graphk/pipeline.py
17
16
  src/owlmind/models/__init__.py
18
17
  src/owlmind/models/ollama.py
19
18
  tests/test_core_1.py
@@ -1,94 +0,0 @@
1
- ##
2
- ## GraphK - Framework for Graph programming.
3
- ## pipeline.py — data structure for Processing Lines.
4
- ##
5
- # Copyright (c) 2025, Dr. Fernando Koch
6
- # http://github.com/kochf1/graphk
7
- #
8
- # Disclosure:
9
- # This code was developed through 'vibe coding'. Certain components
10
- # required manual implementation, and human-in-the-loop review and refinement
11
- # were applied throughout the project.
12
- #
13
-
14
-
15
- from abc import ABC, abstractmethod
16
- from typing import Union, Callable, List, Any, Iterator, Iterable
17
- from .node import Node
18
-
19
- class Pipeline(Node):
20
- """
21
- Orchestrator for a sequence of Nodes.
22
- Compatible with GraphK Node and Gate structures.
23
- """
24
- # Execution Policies
25
- SKIP_ON_FAIL = 0
26
- STOP_ON_FAIL = 1
27
-
28
- def __init__(self,
29
- nodes: Iterable[Node] = None,
30
- _session: dict = None,
31
- fail_policy: int = STOP_ON_FAIL,
32
- **kwargs):
33
- """
34
- Initialize the pipeline as a Node containing other Nodes.
35
- """
36
- super().__init__(_session=_session, **kwargs)
37
- self.nodes = list(nodes) if nodes else []
38
- self.fail_policy = fail_policy
39
- self.obfuscate(['nodes', 'fail_policy'])
40
-
41
- # --- Node Access (Pythonic Interface) ---
42
-
43
- def __len__(self) -> int: return len(self.nodes)
44
-
45
- def __getitem__(self, index): return self.nodes[index]
46
-
47
- def add(self, node: Node):
48
- self.nodes.append(node)
49
- return self
50
-
51
- # --- Execution Logic ---
52
-
53
- def ping(self) -> bool:
54
- """Pipeline is healthy if all internal nodes are healthy."""
55
- return all(n.ping() for n in self.nodes)
56
-
57
- def info(self) -> dict:
58
- return {
59
- "type": "Pipeline",
60
- "node_count": len(self.nodes),
61
- "nodes": [type(n).__name__ for n in self.nodes]
62
- }
63
-
64
- def step(self) -> Iterator[Any]:
65
- """
66
- Executes the pipeline sequence.
67
- Passes output of one node as input to the next.
68
- """
69
- current_input = getattr(self, 'input', None)
70
-
71
- for node in self.nodes:
72
- # 1. Check Condition Gate (if assigned to the node)
73
- if hasattr(node, 'can_proceed') and not node.can_proceed(current_input):
74
- if self.fail_policy == self.STOP_ON_FAIL: break
75
- continue
76
-
77
- # 2. Set input and execute
78
- node.input = current_input
79
-
80
- # 3. Collect output (assuming step() returns an iterator)
81
- node_output = None
82
- for chunk in node.step():
83
- node_output = chunk # Capture last chunk as the state
84
- yield chunk
85
-
86
- # 4. Check Validation Gate
87
- if hasattr(node, 'is_valid') and not node.is_valid(node_output):
88
- if self.fail_policy == self.STOP_ON_FAIL: break
89
-
90
- # 5. Chain output to next input
91
- current_input = node_output
92
-
93
- self._output = current_input
94
-
File without changes
File without changes
File without changes
File without changes
File without changes