git-llm-tool 0.1.15__tar.gz → 0.1.16__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/PKG-INFO +1 -1
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/config.py +7 -7
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/pyproject.toml +1 -1
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/LICENSE +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/README.md +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/__init__.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/__main__.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/cli.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/commands/__init__.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/commands/changelog_cmd.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/commands/commit_cmd.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/__init__.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/diff_optimizer.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/exceptions.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/git_helper.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/jira_helper.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/rate_limiter.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/smart_chunker.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/core/token_counter.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/__init__.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/anthropic_langchain.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/azure_openai_langchain.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/base.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/factory.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/gemini_langchain.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/langchain_base.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/ollama_langchain.py +0 -0
- {git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/openai_langchain.py +0 -0
|
@@ -21,8 +21,8 @@ class LlmConfig:
|
|
|
21
21
|
chunking_threshold: int = 12000 # Token threshold to trigger chunking + parallel processing
|
|
22
22
|
|
|
23
23
|
# Ollama configuration for hybrid processing
|
|
24
|
-
use_ollama_for_chunks: bool =
|
|
25
|
-
ollama_model: str = "
|
|
24
|
+
use_ollama_for_chunks: bool = True # Use Ollama for chunk processing (map phase)
|
|
25
|
+
ollama_model: str = "phi3:mini" # Ollama model for chunk processing
|
|
26
26
|
ollama_base_url: str = "http://localhost:11434" # Ollama API base URL
|
|
27
27
|
|
|
28
28
|
# Parallel processing configuration
|
|
@@ -30,14 +30,14 @@ class LlmConfig:
|
|
|
30
30
|
ollama_max_parallel_chunks: int = 16 # Maximum concurrent chunks for Ollama (local)
|
|
31
31
|
|
|
32
32
|
# Chunking configuration
|
|
33
|
-
chunk_size: int =
|
|
33
|
+
chunk_size: int = 2048 # Maximum chunk size in tokens
|
|
34
34
|
chunk_overlap: int = 150 # Overlap between chunks in tokens
|
|
35
35
|
|
|
36
36
|
# Internal constants (not user-configurable)
|
|
37
|
-
_chunk_size: int =
|
|
37
|
+
_chunk_size: int = 2048 # Maximum chunk size in characters
|
|
38
38
|
_chunk_overlap: int = 300 # Overlap between chunks to maintain context
|
|
39
39
|
_max_parallel_chunks: int = 4 # Maximum number of chunks to process in parallel (remote APIs)
|
|
40
|
-
_ollama_max_parallel_chunks: int =
|
|
40
|
+
_ollama_max_parallel_chunks: int = 4 # Maximum number of chunks to process in parallel (Ollama local)
|
|
41
41
|
_chunk_processing_timeout: float = 120.0 # Timeout for each chunk processing (seconds)
|
|
42
42
|
_max_retries: int = 5 # Maximum number of retries
|
|
43
43
|
_initial_delay: float = 1.0 # Initial retry delay in seconds
|
|
@@ -188,8 +188,8 @@ class ConfigLoader:
|
|
|
188
188
|
# Processing settings
|
|
189
189
|
chunking_threshold=llm_data.get("chunking_threshold", 12000),
|
|
190
190
|
# Ollama settings
|
|
191
|
-
use_ollama_for_chunks=llm_data.get("use_ollama_for_chunks",
|
|
192
|
-
ollama_model=llm_data.get("ollama_model", "
|
|
191
|
+
use_ollama_for_chunks=llm_data.get("use_ollama_for_chunks", True),
|
|
192
|
+
ollama_model=llm_data.get("ollama_model", "phi3:mini"),
|
|
193
193
|
ollama_base_url=llm_data.get("ollama_base_url", "http://localhost:11434"),
|
|
194
194
|
# Parallel processing settings
|
|
195
195
|
max_parallel_chunks=llm_data.get("max_parallel_chunks", 4),
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{git_llm_tool-0.1.15 → git_llm_tool-0.1.16}/git_llm_tool/providers/azure_openai_langchain.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|