not-again-ai 0.10.2__tar.gz → 0.10.3__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/PKG-INFO +14 -13
  2. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/README.md +2 -1
  3. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/pyproject.toml +12 -12
  4. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/tokens.py +2 -0
  5. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/LICENSE +0 -0
  6. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/__init__.py +0 -0
  7. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/base/__init__.py +0 -0
  8. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/base/file_system.py +0 -0
  9. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/base/parallel.py +0 -0
  10. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/__init__.py +0 -0
  11. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/__init__.py +0 -0
  12. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/chat_completion.py +0 -0
  13. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/context_management.py +0 -0
  14. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/embeddings.py +0 -0
  15. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/openai_client.py +0 -0
  16. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/llm/openai_api/prompts.py +0 -0
  17. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/__init__.py +0 -0
  18. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/chat_completion.py +0 -0
  19. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/huggingface/__init__.py +0 -0
  20. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/huggingface/chat_completion.py +0 -0
  21. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/huggingface/helpers.py +0 -0
  22. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/__init__.py +0 -0
  23. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/chat_completion.py +0 -0
  24. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/model_mapping.py +0 -0
  25. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/ollama_client.py +0 -0
  26. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/service.py +0 -0
  27. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/ollama/tokens.py +0 -0
  28. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/prompts.py +0 -0
  29. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/local_llm/tokens.py +0 -0
  30. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/py.typed +0 -0
  31. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/statistics/__init__.py +0 -0
  32. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/statistics/dependence.py +0 -0
  33. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/__init__.py +0 -0
  34. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/barplots.py +0 -0
  35. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/distributions.py +0 -0
  36. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/scatterplot.py +0 -0
  37. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/time_series.py +0 -0
  38. {not_again_ai-0.10.2 → not_again_ai-0.10.3}/src/not_again_ai/viz/utils.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: not-again-ai
3
- Version: 0.10.2
3
+ Version: 0.10.3
4
4
  Summary: Designed to once and for all collect all the little things that come up over and over again in AI projects and put them in one place.
5
5
  Home-page: https://github.com/DaveCoDev/not-again-ai
6
6
  License: MIT
@@ -21,18 +21,18 @@ Provides-Extra: llm
21
21
  Provides-Extra: local-llm
22
22
  Provides-Extra: statistics
23
23
  Provides-Extra: viz
24
- Requires-Dist: jinja2 (==3.1.4) ; extra == "local-llm"
24
+ Requires-Dist: jinja2 (>=3.1,<4.0) ; extra == "local-llm"
25
25
  Requires-Dist: loguru (==0.7.2)
26
- Requires-Dist: numpy (==1.26.4) ; extra == "statistics" or extra == "viz"
27
- Requires-Dist: ollama (==0.2.1) ; extra == "local-llm"
28
- Requires-Dist: openai (==1.35.10) ; extra == "llm"
29
- Requires-Dist: pandas (==2.2.2) ; extra == "viz"
30
- Requires-Dist: python-liquid (==1.12.1) ; extra == "llm"
31
- Requires-Dist: scikit-learn (==1.5.1) ; extra == "statistics"
32
- Requires-Dist: scipy (==1.14.0) ; extra == "statistics"
33
- Requires-Dist: seaborn (==0.13.2) ; extra == "viz"
34
- Requires-Dist: tiktoken (==0.7.0) ; extra == "llm"
35
- Requires-Dist: transformers (==4.42.3) ; extra == "local-llm"
26
+ Requires-Dist: numpy (>=1.26,<2.0) ; extra == "statistics" or extra == "viz"
27
+ Requires-Dist: ollama (>=0.2,<0.3) ; extra == "local-llm"
28
+ Requires-Dist: openai (>=1.36,<2.0) ; extra == "llm"
29
+ Requires-Dist: pandas (>=2.2,<3.0) ; extra == "viz"
30
+ Requires-Dist: python-liquid (>=1.12,<2.0) ; extra == "llm"
31
+ Requires-Dist: scikit-learn (>=1.5,<2.0) ; extra == "statistics"
32
+ Requires-Dist: scipy (>=1.14,<2.0) ; extra == "statistics"
33
+ Requires-Dist: seaborn (>=0.13,<0.14) ; extra == "viz"
34
+ Requires-Dist: tiktoken (>=0.7,<0.8) ; extra == "llm"
35
+ Requires-Dist: transformers (>=4.42,<5.0) ; extra == "local-llm"
36
36
  Project-URL: Documentation, https://github.com/DaveCoDev/not-again-ai
37
37
  Project-URL: Repository, https://github.com/DaveCoDev/not-again-ai
38
38
  Description-Content-Type: text/markdown
@@ -86,7 +86,8 @@ The package is split into subpackages, so you can install only the parts you nee
86
86
  ...
87
87
  Environment="OLLAMA_HOST=0.0.0.0"
88
88
  ```
89
- 1. HuggingFace transformers and other requirements are hardware dependent so for providers other than Ollama, this only installs some generic dependencies. Check the [notebooks](notebooks/local_llm/) for more details on what is available and how to install it.
89
+ 1. It is recommended to always have the latest version of Ollama. To update Ollama check the [docs](https://github.com/ollama/ollama/blob/main/docs/). The command for Linux is: `curl -fsSL https://ollama.com/install.sh | sh`
90
+ 2. HuggingFace transformers and other requirements are hardware dependent so for providers other than Ollama, this only installs some generic dependencies. Check the [notebooks](notebooks/local_llm/) for more details on what is available and how to install it.
90
91
  * **Statistics**: `pip install not_again_ai[statistics]`
91
92
  * **Visualization**: `pip install not_again_ai[viz]`
92
93
 
@@ -47,7 +47,8 @@ The package is split into subpackages, so you can install only the parts you nee
47
47
  ...
48
48
  Environment="OLLAMA_HOST=0.0.0.0"
49
49
  ```
50
- 1. HuggingFace transformers and other requirements are hardware dependent so for providers other than Ollama, this only installs some generic dependencies. Check the [notebooks](notebooks/local_llm/) for more details on what is available and how to install it.
50
+ 1. It is recommended to always have the latest version of Ollama. To update Ollama check the [docs](https://github.com/ollama/ollama/blob/main/docs/). The command for Linux is: `curl -fsSL https://ollama.com/install.sh | sh`
51
+ 2. HuggingFace transformers and other requirements are hardware dependent so for providers other than Ollama, this only installs some generic dependencies. Check the [notebooks](notebooks/local_llm/) for more details on what is available and how to install it.
51
52
  * **Statistics**: `pip install not_again_ai[statistics]`
52
53
  * **Visualization**: `pip install not_again_ai[viz]`
53
54
 
@@ -1,6 +1,6 @@
1
1
  [tool.poetry]
2
2
  name = "not-again-ai"
3
- version = "0.10.2"
3
+ version = "0.10.3"
4
4
  description = "Designed to once and for all collect all the little things that come up over and over again in AI projects and put them in one place."
5
5
  authors = ["DaveCoDev <dave.co.dev@gmail.com>"]
6
6
  license = "MIT"
@@ -29,17 +29,17 @@ python = "^3.11 || ^3.12"
29
29
  loguru = { version = "==0.7.2" }
30
30
 
31
31
  # Optional dependencies are defined here, and groupings are defined below.
32
- jinja2 = { version = "==3.1.4", optional = true }
33
- numpy = { version = "==1.26.4", optional = true }
34
- ollama = { version = "==0.2.1", optional = true }
35
- openai = { version = "==1.35.10", optional = true }
36
- pandas = { version = "==2.2.2", optional = true }
37
- python-liquid = { version = "==1.12.1", optional = true }
38
- scipy = { version = "==1.14.0", optional = true }
39
- scikit-learn = { version = "==1.5.1", optional = true }
40
- seaborn = { version = "==0.13.2", optional = true }
41
- tiktoken = { version = "==0.7.0", optional = true }
42
- transformers = { version = "==4.42.3", optional = true }
32
+ jinja2 = { version = "^3.1", optional = true }
33
+ numpy = { version = "^1.26", optional = true }
34
+ ollama = { version = "^0.2", optional = true }
35
+ openai = { version = "^1.36", optional = true }
36
+ pandas = { version = "^2.2", optional = true }
37
+ python-liquid = { version = "^1.12", optional = true }
38
+ scipy = { version = "^1.14", optional = true }
39
+ scikit-learn = { version = "^1.5", optional = true }
40
+ seaborn = { version = "^0.13", optional = true }
41
+ tiktoken = { version = "^0.7", optional = true }
42
+ transformers = { version = "^4.42", optional = true }
43
43
 
44
44
  [tool.poetry.extras]
45
45
  llm = ["openai", "python-liquid", "tiktoken"]
@@ -86,6 +86,8 @@ def num_tokens_from_messages(
86
86
  "gpt-4-turbo-2024-04-09",
87
87
  "gpt-4o",
88
88
  "gpt-4o-2024-05-13",
89
+ "gpt-4o-mini",
90
+ "gpt-4o-mini-2024-07-18",
89
91
  }:
90
92
  tokens_per_message = 3 # every message follows <|start|>{role/name}\n{content}<|end|>\n
91
93
  tokens_per_name = 1 # if there's a name, the role is omitted
File without changes