ollamadiffuser 1.1.6__py3-none-any.whl → 1.2.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: ollamadiffuser
3
- Version: 1.1.6
3
+ Version: 1.2.1
4
4
  Summary: 🎨 Local AI Image Generation with Ollama-style CLI for Stable Diffusion, FLUX.1, and LoRA support
5
5
  Home-page: https://github.com/ollamadiffuser/ollamadiffuser
6
6
  Author: OllamaDiffuser Team
@@ -52,15 +52,16 @@ Requires-Dist: psutil>=5.9.0
52
52
  Requires-Dist: jinja2>=3.0.0
53
53
  Requires-Dist: peft>=0.10.0
54
54
  Requires-Dist: numpy>=1.21.0
55
+ Requires-Dist: controlnet-aux>=0.0.7
56
+ Requires-Dist: opencv-python>=4.8.0
57
+ Requires-Dist: stable-diffusion-cpp-python>=0.1.0
58
+ Requires-Dist: gguf>=0.1.0
55
59
  Provides-Extra: dev
56
60
  Requires-Dist: pytest>=7.0.0; extra == "dev"
57
61
  Requires-Dist: pytest-asyncio>=0.21.0; extra == "dev"
58
62
  Requires-Dist: black>=23.0.0; extra == "dev"
59
63
  Requires-Dist: isort>=5.12.0; extra == "dev"
60
64
  Requires-Dist: flake8>=6.0.0; extra == "dev"
61
- Provides-Extra: full
62
- Requires-Dist: controlnet-aux>=0.0.7; extra == "full"
63
- Requires-Dist: opencv-python>=4.8.0; extra == "full"
64
65
  Dynamic: author
65
66
  Dynamic: home-page
66
67
  Dynamic: license-file
@@ -81,11 +82,69 @@ Dynamic: requires-python
81
82
 
82
83
  ---
83
84
 
85
+ ## 🔑 Hugging Face Authentication
86
+
87
+ **Do you need a Hugging Face token?** It depends on which models you want to use!
88
+
89
+ ### 🟢 Models that DON'T require a token:
90
+ - **FLUX.1-schnell** - Apache 2.0 license, ready to use ✅
91
+ - **Stable Diffusion 1.5** - Basic model, no authentication needed ✅
92
+ - **Most ControlNet models** - Generally public access ✅
93
+
94
+ ### 🟡 Models that DO require a token:
95
+ - **FLUX.1-dev** - Requires HF token and license agreement ⚠️
96
+ - **Stable Diffusion 3.5** - Requires HF token and license agreement ⚠️
97
+ - **Some premium LoRAs** - Gated models from Hugging Face ⚠️
98
+
99
+ ### 🚀 Quick Setup
100
+
101
+ **For basic usage** (no token needed):
102
+ ```bash
103
+ # These work immediately without any setup:
104
+ ollamadiffuser pull flux.1-schnell
105
+ ollamadiffuser pull stable-diffusion-1.5
106
+ ```
107
+
108
+ **For advanced models** (token required):
109
+ ```bash
110
+ # 1. Set your token
111
+ export HF_TOKEN=your_token_here
112
+
113
+ # 2. Now you can access gated models
114
+ ollamadiffuser pull flux.1-dev
115
+ ollamadiffuser pull stable-diffusion-3.5-medium
116
+ ```
117
+
118
+ ### 🔧 How to get a Hugging Face token:
119
+
120
+ 1. **Create account**: Visit [huggingface.co](https://huggingface.co) and sign up
121
+ 2. **Generate token**: Go to Settings → Access Tokens → Create new token
122
+ 3. **Accept licenses**: Visit the model pages and accept license agreements:
123
+ - [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev)
124
+ - [Stable Diffusion 3.5](https://huggingface.co/stabilityai/stable-diffusion-3.5-medium)
125
+ 4. **Set environment variable**:
126
+ ```bash
127
+ # Temporary (current session)
128
+ export HF_TOKEN=your_token_here
129
+
130
+ # Permanent (add to ~/.bashrc or ~/.zshrc)
131
+ echo 'export HF_TOKEN=your_token_here' >> ~/.bashrc
132
+ ```
133
+
134
+ ### 💡 Pro Tips:
135
+ - **Start simple**: Begin with FLUX.1-schnell (no token required, commercial use OK)
136
+ - **Token scope**: Use "read" permissions for downloading models
137
+ - **Privacy**: Your token stays local - never shared with OllamaDiffuser servers
138
+ - **Troubleshooting**: If downloads fail, verify your token and model access permissions
139
+
140
+ ---
141
+
84
142
  ## ✨ Features
85
143
 
86
144
  - **🚀 Fast Startup**: Instant application launch with lazy loading architecture
87
145
  - **🎛️ ControlNet Support**: Precise image generation control with 10+ control types
88
146
  - **🔄 LoRA Integration**: Dynamic LoRA loading and management
147
+ - **📦 GGUF Support**: Memory-efficient quantized models (3GB VRAM minimum!)
89
148
  - **🌐 Multiple Interfaces**: CLI, Python API, Web UI, and REST API
90
149
  - **📦 Model Management**: Easy installation and switching between models
91
150
  - **⚡ Performance Optimized**: Memory-efficient with GPU acceleration
@@ -109,6 +168,18 @@ curl -X POST http://localhost:8000/api/generate \
109
168
  --output image.png
110
169
  ```
111
170
 
171
+ ### GGUF Quick Start (Low VRAM)
172
+ ```bash
173
+ # For systems with limited VRAM (3GB+)
174
+ pip install ollamadiffuser stable-diffusion-cpp-python gguf
175
+
176
+ # Download memory-efficient GGUF model
177
+ ollamadiffuser pull flux.1-dev-gguf-q4ks
178
+
179
+ # Generate with reduced memory usage
180
+ ollamadiffuser run flux.1-dev-gguf-q4ks
181
+ ```
182
+
112
183
  ### Option 2: Development Installation
113
184
  ```bash
114
185
  # Clone the repository
@@ -162,12 +233,26 @@ curl -X POST http://localhost:8000/api/generate/controlnet \
162
233
 
163
234
  Choose from a variety of state-of-the-art image generation models:
164
235
 
165
- | Model | License | Quality | Speed | Commercial Use |
166
- |-------|---------|---------|-------|----------------|
167
- | **FLUX.1-schnell** | Apache 2.0 | High | **4 steps** (12x faster) | ✅ Commercial OK |
168
- | **FLUX.1-dev** | Non-commercial | High | 50 steps | ❌ Non-commercial |
169
- | **Stable Diffusion 3.5** | CreativeML | Medium | 28 steps | ⚠️ Check License |
170
- | **Stable Diffusion 1.5** | CreativeML | Fast | Lightweight | ⚠️ Check License |
236
+ | Model | License | Quality | Speed | Commercial Use | VRAM |
237
+ |-------|---------|---------|-------|----------------|------|
238
+ | **FLUX.1-schnell** | Apache 2.0 | High | **4 steps** (12x faster) | ✅ Commercial OK | 20GB+ |
239
+ | **FLUX.1-dev** | Non-commercial | High | 50 steps | ❌ Non-commercial | 20GB+ |
240
+ | **FLUX.1-dev-gguf** | Non-commercial | High | 4 steps | Non-commercial | **3-16GB** |
241
+ | **Stable Diffusion 3.5** | CreativeML | Medium | 28 steps | ⚠️ Check License | 12GB+ |
242
+ | **Stable Diffusion 1.5** | CreativeML | Fast | Lightweight | ⚠️ Check License | 6GB+ |
243
+
244
+ ### 💾 GGUF Models - Reduced Memory Requirements
245
+
246
+ **NEW**: GGUF quantized models enable running FLUX.1-dev on budget hardware!
247
+
248
+ | GGUF Variant | VRAM | Quality | Best For |
249
+ |--------------|------|---------|----------|
250
+ | `flux.1-dev-gguf-q4ks` | 6GB | ⭐⭐⭐⭐ | **Recommended** - RTX 3060/4060 |
251
+ | `flux.1-dev-gguf-q3ks` | 4GB | ⭐⭐⭐ | Mobile GPUs, GTX 1660 Ti |
252
+ | `flux.1-dev-gguf-q2k` | 3GB | ⭐⭐ | Entry-level hardware |
253
+ | `flux.1-dev-gguf-q6k` | 10GB | ⭐⭐⭐⭐⭐ | RTX 3080/4070+ |
254
+
255
+ 📖 **[Complete GGUF Guide](GGUF_GUIDE.md)** - Hardware recommendations, installation, and optimization tips
171
256
 
172
257
  ### Why Choose FLUX.1-schnell?
173
258
  - **Apache 2.0 license** - Perfect for commercial use
@@ -305,6 +390,11 @@ else:
305
390
  - **Stable Diffusion 3**: Latest architecture
306
391
  - **FLUX.1**: State-of-the-art quality
307
392
 
393
+ ### GGUF Quantized Models
394
+ - **FLUX.1-dev GGUF**: 7 quantization levels (3GB-16GB VRAM)
395
+ - **Memory Efficient**: Run high-quality models on budget hardware
396
+ - **Same API**: Works seamlessly with existing commands
397
+
308
398
  ### ControlNet Models
309
399
  - **SD 1.5 ControlNet**: 4 control types (canny, depth, openpose, scribble)
310
400
  - **SDXL ControlNet**: 2 control types (canny, depth)
@@ -338,6 +428,19 @@ image = engine.generate_image(
338
428
  )
339
429
  ```
340
430
 
431
+ ### GGUF Model Usage
432
+ ```bash
433
+ # Check GGUF support
434
+ ollamadiffuser registry check-gguf
435
+
436
+ # Download GGUF model for your hardware
437
+ ollamadiffuser pull flux.1-dev-gguf-q4ks # 6GB VRAM
438
+ ollamadiffuser pull flux.1-dev-gguf-q3ks # 4GB VRAM
439
+
440
+ # Use with optimized settings
441
+ ollamadiffuser run flux.1-dev-gguf-q4ks
442
+ ```
443
+
341
444
  ### Batch Processing
342
445
  ```python
343
446
  from ollamadiffuser.core.utils.controlnet_preprocessors import controlnet_preprocessor
@@ -375,6 +478,7 @@ with open("control.jpg", "rb") as f:
375
478
 
376
479
  ## 📚 Documentation & Guides
377
480
 
481
+ - **[GGUF Models Guide](GGUF_GUIDE.md)**: Complete guide to memory-efficient GGUF models
378
482
  - **[ControlNet Guide](CONTROLNET_GUIDE.md)**: Comprehensive ControlNet usage and examples
379
483
  - **[Website Documentation](https://www.ollamadiffuser.com/)**: Complete tutorials and guides
380
484
 
@@ -386,10 +490,17 @@ with open("control.jpg", "rb") as f:
386
490
  - **Python**: 3.8+
387
491
 
388
492
  ### Recommended Hardware
493
+
494
+ #### For Regular Models
389
495
  - **GPU**: 8GB+ VRAM (NVIDIA/AMD)
390
496
  - **RAM**: 16GB+ system RAM
391
497
  - **Storage**: SSD with 50GB+ free space
392
498
 
499
+ #### For GGUF Models (Memory Efficient)
500
+ - **GPU**: 3GB+ VRAM (or CPU only)
501
+ - **RAM**: 8GB+ system RAM (16GB+ for CPU inference)
502
+ - **Storage**: SSD with 20GB+ free space
503
+
393
504
  ### Supported Platforms
394
505
  - **CUDA**: NVIDIA GPUs (recommended)
395
506
  - **MPS**: Apple Silicon (M1/M2/M3)
@@ -420,6 +531,18 @@ pip install "ollamadiffuser[full]"
420
531
  pip install 'ollamadiffuser[full]'
421
532
  ```
422
533
 
534
+ #### GGUF Support Issues
535
+ ```bash
536
+ # Install GGUF dependencies
537
+ pip install stable-diffusion-cpp-python gguf
538
+
539
+ # Check GGUF support
540
+ ollamadiffuser registry check-gguf
541
+
542
+ # See full GGUF troubleshooting guide
543
+ # Read GGUF_GUIDE.md for detailed troubleshooting
544
+ ```
545
+
423
546
  #### Complete Dependency Check
424
547
  ```bash
425
548
  # Run comprehensive system diagnostics
@@ -474,6 +597,10 @@ curl -X POST http://localhost:8000/api/controlnet/initialize
474
597
 
475
598
  #### Memory Issues
476
599
  ```bash
600
+ # Use GGUF models for lower memory usage
601
+ ollamadiffuser pull flux.1-dev-gguf-q4ks # 6GB VRAM
602
+ ollamadiffuser pull flux.1-dev-gguf-q3ks # 4GB VRAM
603
+
477
604
  # Use smaller image sizes via API
478
605
  curl -X POST http://localhost:8000/api/generate \
479
606
  -H "Content-Type: application/json" \
@@ -492,12 +619,18 @@ curl -X POST http://localhost:8000/api/generate \
492
619
  # If you encounter OpenCV issues on Apple Silicon
493
620
  pip uninstall opencv-python
494
621
  pip install opencv-python-headless>=4.8.0
622
+
623
+ # For GGUF Metal acceleration
624
+ CMAKE_ARGS="-DSD_METAL=ON" pip install stable-diffusion-cpp-python
495
625
  ```
496
626
 
497
627
  #### Windows
498
628
  ```bash
499
629
  # If you encounter build errors
500
630
  pip install --only-binary=all opencv-python>=4.8.0
631
+
632
+ # For GGUF CUDA acceleration
633
+ CMAKE_ARGS="-DSD_CUDA=ON" pip install stable-diffusion-cpp-python
501
634
  ```
502
635
 
503
636
  #### Linux
@@ -540,6 +673,8 @@ This project is licensed under the MIT License - see the [LICENSE](LICENSE) file
540
673
  ## 🙏 Acknowledgments
541
674
 
542
675
  - **Stability AI**: For Stable Diffusion models
676
+ - **Black Forest Labs**: For FLUX.1 models
677
+ - **city96**: For FLUX.1-dev GGUF quantizations
543
678
  - **Hugging Face**: For model hosting and diffusers library
544
679
  - **ControlNet Team**: For ControlNet architecture
545
680
  - **Community**: For feedback and contributions
@@ -1,20 +1,23 @@
1
- ollamadiffuser/__init__.py,sha256=DWqHxjTy1ekkwx7FhVOWIannwUuINWT2aPNKNOEcC7o,1127
1
+ ollamadiffuser/__init__.py,sha256=n9dmXqrOp56VjdptVANGVeHea_UiuGHqP1d-b6dyq6g,1127
2
2
  ollamadiffuser/__main__.py,sha256=tNWMvEHq4ddtKLp7DrhIoOdnFw3F8RNrETC_u5xpkFI,141
3
3
  ollamadiffuser/api/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
- ollamadiffuser/api/server.py,sha256=4-3gT8W1404bxvJ7y9htvKbd2yxrrbtAUvT7shOlJss,17679
4
+ ollamadiffuser/api/server.py,sha256=kc-Up50zmLfSM4f4InYR4Btkl35lyEN6UFcxlvxICSQ,18237
5
5
  ollamadiffuser/cli/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
6
6
  ollamadiffuser/cli/commands.py,sha256=Pe0vyfGiffwd10QlVxBCTtNnMqHi8nJ3oNn_k8nAi5k,8903
7
- ollamadiffuser/cli/main.py,sha256=Iea_jVRu_Z3CViIcRTZM_SQJ7jtx91ZNaRvilxXqfqk,41908
7
+ ollamadiffuser/cli/main.py,sha256=qj0VKTOjw_gox2dPVtbU-9kCo25TqshyVyqW8qsv4Pk,56081
8
8
  ollamadiffuser/core/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
9
9
  ollamadiffuser/core/config/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
10
+ ollamadiffuser/core/config/model_registry.py,sha256=o408Qk-UQJ4NBM-AcyWAj4EhRUzrvl6zdOkKk1-ATHg,31238
10
11
  ollamadiffuser/core/config/settings.py,sha256=VhI1vLGmOAQ7-XtyHrT5KoMpcGeGt-Mij-9NxX_ZKsI,4881
11
12
  ollamadiffuser/core/inference/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
12
- ollamadiffuser/core/inference/engine.py,sha256=ky76lAjWexlrgmHSZZILa3FPQP7xx2WQ0_DdW34M9Xk,57808
13
+ ollamadiffuser/core/inference/engine.py,sha256=-EVcH4NyRVBQ-puvI_Az3KOZJYG9b-ySzsvoBTY3mlY,73962
13
14
  ollamadiffuser/core/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
- ollamadiffuser/core/models/manager.py,sha256=vO1Az_aO5lZKMgSyK_6j2wT5nzPMowZgKhcH2mQVLkI,24139
15
+ ollamadiffuser/core/models/gguf_loader.py,sha256=ocfl3_MDVXC9nSjW8YJdz4kX1Q-Qe2ltu6w4fbqhxVY,35724
16
+ ollamadiffuser/core/models/manager.py,sha256=rTEAameGih3wPcVG_Y-4k_brBeEqEoBjoI7fjggNtiY,16799
17
+ ollamadiffuser/core/models/registry.py,sha256=YPx3xcHnCHogyowi9fQ6oXZg7_jz0fM5bDyyg-BgSFY,15125
15
18
  ollamadiffuser/core/utils/__init__.py,sha256=ZdXZWX1hfDnnV6OmRD6UStNljDJIQ892da2CtC-zdDw,31
16
19
  ollamadiffuser/core/utils/controlnet_preprocessors.py,sha256=v21X_Bk-a4gKbUZUKoeP2W8TSGlv-ST8IYNsn3NrZ2c,15446
17
- ollamadiffuser/core/utils/download_utils.py,sha256=DvCt-cjH6WSBJniJT112b4a9AUzlwOYhQtPuEfISmtM,20961
20
+ ollamadiffuser/core/utils/download_utils.py,sha256=dqHf7loKSqDHzgQEV-p1kDwoHxwlgurLPX8J96vD5fA,22421
18
21
  ollamadiffuser/core/utils/lora_manager.py,sha256=SrZydPSGJqCS_Vek35bEdG2Q51qCOLZmPvnNzUjjIN0,14328
19
22
  ollamadiffuser/ui/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
20
23
  ollamadiffuser/ui/web.py,sha256=hK41sF9uVR_14lc4AlTXaGjw5m7aiJXpVp1cSOOmkT8,27715
@@ -34,9 +37,9 @@ ollamadiffuser/ui/samples/scribble/face_sketch.png,sha256=MVVYy_aS48xoS_RnIDzLUa
34
37
  ollamadiffuser/ui/samples/scribble/tree_sketch.png,sha256=3P-NGgW25xRwreDxiBYKcDhd2oHZAwKSkjNVM5oPTWY,3017
35
38
  ollamadiffuser/ui/templates/index.html,sha256=qTQVFxiTbeZ90O-iNqWC_4pYP6yyIs2z6U69VJPqAB4,38176
36
39
  ollamadiffuser/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
37
- ollamadiffuser-1.1.6.dist-info/licenses/LICENSE,sha256=cnGL9l2P510Uk3TCnv62kot6vAfdSawhOZh7Y-oYoIE,1071
38
- ollamadiffuser-1.1.6.dist-info/METADATA,sha256=bDoFxbnLvkp40zA1PIkGDXBvznaS7FQ2PEKPimPLfJI,17070
39
- ollamadiffuser-1.1.6.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
40
- ollamadiffuser-1.1.6.dist-info/entry_points.txt,sha256=tHXXO3N0GSnIobDe_eSOLfHPjjVFjeTg2Fd-APoD6sY,64
41
- ollamadiffuser-1.1.6.dist-info/top_level.txt,sha256=97wOGgTCxDE765Nr_o7B4Kwr_M_jy8fCCeQ81sMKlC4,15
42
- ollamadiffuser-1.1.6.dist-info/RECORD,,
40
+ ollamadiffuser-1.2.1.dist-info/licenses/LICENSE,sha256=cnGL9l2P510Uk3TCnv62kot6vAfdSawhOZh7Y-oYoIE,1071
41
+ ollamadiffuser-1.2.1.dist-info/METADATA,sha256=FoUPXS8xMBHOSfMgQcv2E1ITGhPU2IwF_nDU7BBHvQs,21756
42
+ ollamadiffuser-1.2.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
43
+ ollamadiffuser-1.2.1.dist-info/entry_points.txt,sha256=tHXXO3N0GSnIobDe_eSOLfHPjjVFjeTg2Fd-APoD6sY,64
44
+ ollamadiffuser-1.2.1.dist-info/top_level.txt,sha256=97wOGgTCxDE765Nr_o7B4Kwr_M_jy8fCCeQ81sMKlC4,15
45
+ ollamadiffuser-1.2.1.dist-info/RECORD,,