ollamadiffuser 1.1.6__py3-none-any.whl → 1.2.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ollamadiffuser/__init__.py +1 -1
- ollamadiffuser/api/server.py +10 -2
- ollamadiffuser/cli/main.py +366 -28
- ollamadiffuser/core/config/model_registry.py +757 -0
- ollamadiffuser/core/inference/engine.py +334 -4
- ollamadiffuser/core/models/gguf_loader.py +771 -0
- ollamadiffuser/core/models/manager.py +139 -312
- ollamadiffuser/core/models/registry.py +384 -0
- ollamadiffuser/core/utils/download_utils.py +35 -2
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/METADATA +145 -10
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/RECORD +15 -12
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/WHEEL +0 -0
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/entry_points.txt +0 -0
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/licenses/LICENSE +0 -0
- {ollamadiffuser-1.1.6.dist-info → ollamadiffuser-1.2.1.dist-info}/top_level.txt +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: ollamadiffuser
|
|
3
|
-
Version: 1.1
|
|
3
|
+
Version: 1.2.1
|
|
4
4
|
Summary: 🎨 Local AI Image Generation with Ollama-style CLI for Stable Diffusion, FLUX.1, and LoRA support
|
|
5
5
|
Home-page: https://github.com/ollamadiffuser/ollamadiffuser
|
|
6
6
|
Author: OllamaDiffuser Team
|
|
@@ -52,15 +52,16 @@ Requires-Dist: psutil>=5.9.0
|
|
|
52
52
|
Requires-Dist: jinja2>=3.0.0
|
|
53
53
|
Requires-Dist: peft>=0.10.0
|
|
54
54
|
Requires-Dist: numpy>=1.21.0
|
|
55
|
+
Requires-Dist: controlnet-aux>=0.0.7
|
|
56
|
+
Requires-Dist: opencv-python>=4.8.0
|
|
57
|
+
Requires-Dist: stable-diffusion-cpp-python>=0.1.0
|
|
58
|
+
Requires-Dist: gguf>=0.1.0
|
|
55
59
|
Provides-Extra: dev
|
|
56
60
|
Requires-Dist: pytest>=7.0.0; extra == "dev"
|
|
57
61
|
Requires-Dist: pytest-asyncio>=0.21.0; extra == "dev"
|
|
58
62
|
Requires-Dist: black>=23.0.0; extra == "dev"
|
|
59
63
|
Requires-Dist: isort>=5.12.0; extra == "dev"
|
|
60
64
|
Requires-Dist: flake8>=6.0.0; extra == "dev"
|
|
61
|
-
Provides-Extra: full
|
|
62
|
-
Requires-Dist: controlnet-aux>=0.0.7; extra == "full"
|
|
63
|
-
Requires-Dist: opencv-python>=4.8.0; extra == "full"
|
|
64
65
|
Dynamic: author
|
|
65
66
|
Dynamic: home-page
|
|
66
67
|
Dynamic: license-file
|
|
@@ -81,11 +82,69 @@ Dynamic: requires-python
|
|
|
81
82
|
|
|
82
83
|
---
|
|
83
84
|
|
|
85
|
+
## 🔑 Hugging Face Authentication
|
|
86
|
+
|
|
87
|
+
**Do you need a Hugging Face token?** It depends on which models you want to use!
|
|
88
|
+
|
|
89
|
+
### 🟢 Models that DON'T require a token:
|
|
90
|
+
- **FLUX.1-schnell** - Apache 2.0 license, ready to use ✅
|
|
91
|
+
- **Stable Diffusion 1.5** - Basic model, no authentication needed ✅
|
|
92
|
+
- **Most ControlNet models** - Generally public access ✅
|
|
93
|
+
|
|
94
|
+
### 🟡 Models that DO require a token:
|
|
95
|
+
- **FLUX.1-dev** - Requires HF token and license agreement ⚠️
|
|
96
|
+
- **Stable Diffusion 3.5** - Requires HF token and license agreement ⚠️
|
|
97
|
+
- **Some premium LoRAs** - Gated models from Hugging Face ⚠️
|
|
98
|
+
|
|
99
|
+
### 🚀 Quick Setup
|
|
100
|
+
|
|
101
|
+
**For basic usage** (no token needed):
|
|
102
|
+
```bash
|
|
103
|
+
# These work immediately without any setup:
|
|
104
|
+
ollamadiffuser pull flux.1-schnell
|
|
105
|
+
ollamadiffuser pull stable-diffusion-1.5
|
|
106
|
+
```
|
|
107
|
+
|
|
108
|
+
**For advanced models** (token required):
|
|
109
|
+
```bash
|
|
110
|
+
# 1. Set your token
|
|
111
|
+
export HF_TOKEN=your_token_here
|
|
112
|
+
|
|
113
|
+
# 2. Now you can access gated models
|
|
114
|
+
ollamadiffuser pull flux.1-dev
|
|
115
|
+
ollamadiffuser pull stable-diffusion-3.5-medium
|
|
116
|
+
```
|
|
117
|
+
|
|
118
|
+
### 🔧 How to get a Hugging Face token:
|
|
119
|
+
|
|
120
|
+
1. **Create account**: Visit [huggingface.co](https://huggingface.co) and sign up
|
|
121
|
+
2. **Generate token**: Go to Settings → Access Tokens → Create new token
|
|
122
|
+
3. **Accept licenses**: Visit the model pages and accept license agreements:
|
|
123
|
+
- [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev)
|
|
124
|
+
- [Stable Diffusion 3.5](https://huggingface.co/stabilityai/stable-diffusion-3.5-medium)
|
|
125
|
+
4. **Set environment variable**:
|
|
126
|
+
```bash
|
|
127
|
+
# Temporary (current session)
|
|
128
|
+
export HF_TOKEN=your_token_here
|
|
129
|
+
|
|
130
|
+
# Permanent (add to ~/.bashrc or ~/.zshrc)
|
|
131
|
+
echo 'export HF_TOKEN=your_token_here' >> ~/.bashrc
|
|
132
|
+
```
|
|
133
|
+
|
|
134
|
+
### 💡 Pro Tips:
|
|
135
|
+
- **Start simple**: Begin with FLUX.1-schnell (no token required, commercial use OK)
|
|
136
|
+
- **Token scope**: Use "read" permissions for downloading models
|
|
137
|
+
- **Privacy**: Your token stays local - never shared with OllamaDiffuser servers
|
|
138
|
+
- **Troubleshooting**: If downloads fail, verify your token and model access permissions
|
|
139
|
+
|
|
140
|
+
---
|
|
141
|
+
|
|
84
142
|
## ✨ Features
|
|
85
143
|
|
|
86
144
|
- **🚀 Fast Startup**: Instant application launch with lazy loading architecture
|
|
87
145
|
- **🎛️ ControlNet Support**: Precise image generation control with 10+ control types
|
|
88
146
|
- **🔄 LoRA Integration**: Dynamic LoRA loading and management
|
|
147
|
+
- **📦 GGUF Support**: Memory-efficient quantized models (3GB VRAM minimum!)
|
|
89
148
|
- **🌐 Multiple Interfaces**: CLI, Python API, Web UI, and REST API
|
|
90
149
|
- **📦 Model Management**: Easy installation and switching between models
|
|
91
150
|
- **⚡ Performance Optimized**: Memory-efficient with GPU acceleration
|
|
@@ -109,6 +168,18 @@ curl -X POST http://localhost:8000/api/generate \
|
|
|
109
168
|
--output image.png
|
|
110
169
|
```
|
|
111
170
|
|
|
171
|
+
### GGUF Quick Start (Low VRAM)
|
|
172
|
+
```bash
|
|
173
|
+
# For systems with limited VRAM (3GB+)
|
|
174
|
+
pip install ollamadiffuser stable-diffusion-cpp-python gguf
|
|
175
|
+
|
|
176
|
+
# Download memory-efficient GGUF model
|
|
177
|
+
ollamadiffuser pull flux.1-dev-gguf-q4ks
|
|
178
|
+
|
|
179
|
+
# Generate with reduced memory usage
|
|
180
|
+
ollamadiffuser run flux.1-dev-gguf-q4ks
|
|
181
|
+
```
|
|
182
|
+
|
|
112
183
|
### Option 2: Development Installation
|
|
113
184
|
```bash
|
|
114
185
|
# Clone the repository
|
|
@@ -162,12 +233,26 @@ curl -X POST http://localhost:8000/api/generate/controlnet \
|
|
|
162
233
|
|
|
163
234
|
Choose from a variety of state-of-the-art image generation models:
|
|
164
235
|
|
|
165
|
-
| Model | License | Quality | Speed | Commercial Use |
|
|
166
|
-
|
|
167
|
-
| **FLUX.1-schnell** | Apache 2.0 | High | **4 steps** (12x faster) | ✅ Commercial OK |
|
|
168
|
-
| **FLUX.1-dev** | Non-commercial | High | 50 steps | ❌ Non-commercial |
|
|
169
|
-
| **
|
|
170
|
-
| **Stable Diffusion
|
|
236
|
+
| Model | License | Quality | Speed | Commercial Use | VRAM |
|
|
237
|
+
|-------|---------|---------|-------|----------------|------|
|
|
238
|
+
| **FLUX.1-schnell** | Apache 2.0 | High | **4 steps** (12x faster) | ✅ Commercial OK | 20GB+ |
|
|
239
|
+
| **FLUX.1-dev** | Non-commercial | High | 50 steps | ❌ Non-commercial | 20GB+ |
|
|
240
|
+
| **FLUX.1-dev-gguf** | Non-commercial | High | 4 steps | ❌ Non-commercial | **3-16GB** |
|
|
241
|
+
| **Stable Diffusion 3.5** | CreativeML | Medium | 28 steps | ⚠️ Check License | 12GB+ |
|
|
242
|
+
| **Stable Diffusion 1.5** | CreativeML | Fast | Lightweight | ⚠️ Check License | 6GB+ |
|
|
243
|
+
|
|
244
|
+
### 💾 GGUF Models - Reduced Memory Requirements
|
|
245
|
+
|
|
246
|
+
**NEW**: GGUF quantized models enable running FLUX.1-dev on budget hardware!
|
|
247
|
+
|
|
248
|
+
| GGUF Variant | VRAM | Quality | Best For |
|
|
249
|
+
|--------------|------|---------|----------|
|
|
250
|
+
| `flux.1-dev-gguf-q4ks` | 6GB | ⭐⭐⭐⭐ | **Recommended** - RTX 3060/4060 |
|
|
251
|
+
| `flux.1-dev-gguf-q3ks` | 4GB | ⭐⭐⭐ | Mobile GPUs, GTX 1660 Ti |
|
|
252
|
+
| `flux.1-dev-gguf-q2k` | 3GB | ⭐⭐ | Entry-level hardware |
|
|
253
|
+
| `flux.1-dev-gguf-q6k` | 10GB | ⭐⭐⭐⭐⭐ | RTX 3080/4070+ |
|
|
254
|
+
|
|
255
|
+
📖 **[Complete GGUF Guide](GGUF_GUIDE.md)** - Hardware recommendations, installation, and optimization tips
|
|
171
256
|
|
|
172
257
|
### Why Choose FLUX.1-schnell?
|
|
173
258
|
- **Apache 2.0 license** - Perfect for commercial use
|
|
@@ -305,6 +390,11 @@ else:
|
|
|
305
390
|
- **Stable Diffusion 3**: Latest architecture
|
|
306
391
|
- **FLUX.1**: State-of-the-art quality
|
|
307
392
|
|
|
393
|
+
### GGUF Quantized Models
|
|
394
|
+
- **FLUX.1-dev GGUF**: 7 quantization levels (3GB-16GB VRAM)
|
|
395
|
+
- **Memory Efficient**: Run high-quality models on budget hardware
|
|
396
|
+
- **Same API**: Works seamlessly with existing commands
|
|
397
|
+
|
|
308
398
|
### ControlNet Models
|
|
309
399
|
- **SD 1.5 ControlNet**: 4 control types (canny, depth, openpose, scribble)
|
|
310
400
|
- **SDXL ControlNet**: 2 control types (canny, depth)
|
|
@@ -338,6 +428,19 @@ image = engine.generate_image(
|
|
|
338
428
|
)
|
|
339
429
|
```
|
|
340
430
|
|
|
431
|
+
### GGUF Model Usage
|
|
432
|
+
```bash
|
|
433
|
+
# Check GGUF support
|
|
434
|
+
ollamadiffuser registry check-gguf
|
|
435
|
+
|
|
436
|
+
# Download GGUF model for your hardware
|
|
437
|
+
ollamadiffuser pull flux.1-dev-gguf-q4ks # 6GB VRAM
|
|
438
|
+
ollamadiffuser pull flux.1-dev-gguf-q3ks # 4GB VRAM
|
|
439
|
+
|
|
440
|
+
# Use with optimized settings
|
|
441
|
+
ollamadiffuser run flux.1-dev-gguf-q4ks
|
|
442
|
+
```
|
|
443
|
+
|
|
341
444
|
### Batch Processing
|
|
342
445
|
```python
|
|
343
446
|
from ollamadiffuser.core.utils.controlnet_preprocessors import controlnet_preprocessor
|
|
@@ -375,6 +478,7 @@ with open("control.jpg", "rb") as f:
|
|
|
375
478
|
|
|
376
479
|
## 📚 Documentation & Guides
|
|
377
480
|
|
|
481
|
+
- **[GGUF Models Guide](GGUF_GUIDE.md)**: Complete guide to memory-efficient GGUF models
|
|
378
482
|
- **[ControlNet Guide](CONTROLNET_GUIDE.md)**: Comprehensive ControlNet usage and examples
|
|
379
483
|
- **[Website Documentation](https://www.ollamadiffuser.com/)**: Complete tutorials and guides
|
|
380
484
|
|
|
@@ -386,10 +490,17 @@ with open("control.jpg", "rb") as f:
|
|
|
386
490
|
- **Python**: 3.8+
|
|
387
491
|
|
|
388
492
|
### Recommended Hardware
|
|
493
|
+
|
|
494
|
+
#### For Regular Models
|
|
389
495
|
- **GPU**: 8GB+ VRAM (NVIDIA/AMD)
|
|
390
496
|
- **RAM**: 16GB+ system RAM
|
|
391
497
|
- **Storage**: SSD with 50GB+ free space
|
|
392
498
|
|
|
499
|
+
#### For GGUF Models (Memory Efficient)
|
|
500
|
+
- **GPU**: 3GB+ VRAM (or CPU only)
|
|
501
|
+
- **RAM**: 8GB+ system RAM (16GB+ for CPU inference)
|
|
502
|
+
- **Storage**: SSD with 20GB+ free space
|
|
503
|
+
|
|
393
504
|
### Supported Platforms
|
|
394
505
|
- **CUDA**: NVIDIA GPUs (recommended)
|
|
395
506
|
- **MPS**: Apple Silicon (M1/M2/M3)
|
|
@@ -420,6 +531,18 @@ pip install "ollamadiffuser[full]"
|
|
|
420
531
|
pip install 'ollamadiffuser[full]'
|
|
421
532
|
```
|
|
422
533
|
|
|
534
|
+
#### GGUF Support Issues
|
|
535
|
+
```bash
|
|
536
|
+
# Install GGUF dependencies
|
|
537
|
+
pip install stable-diffusion-cpp-python gguf
|
|
538
|
+
|
|
539
|
+
# Check GGUF support
|
|
540
|
+
ollamadiffuser registry check-gguf
|
|
541
|
+
|
|
542
|
+
# See full GGUF troubleshooting guide
|
|
543
|
+
# Read GGUF_GUIDE.md for detailed troubleshooting
|
|
544
|
+
```
|
|
545
|
+
|
|
423
546
|
#### Complete Dependency Check
|
|
424
547
|
```bash
|
|
425
548
|
# Run comprehensive system diagnostics
|
|
@@ -474,6 +597,10 @@ curl -X POST http://localhost:8000/api/controlnet/initialize
|
|
|
474
597
|
|
|
475
598
|
#### Memory Issues
|
|
476
599
|
```bash
|
|
600
|
+
# Use GGUF models for lower memory usage
|
|
601
|
+
ollamadiffuser pull flux.1-dev-gguf-q4ks # 6GB VRAM
|
|
602
|
+
ollamadiffuser pull flux.1-dev-gguf-q3ks # 4GB VRAM
|
|
603
|
+
|
|
477
604
|
# Use smaller image sizes via API
|
|
478
605
|
curl -X POST http://localhost:8000/api/generate \
|
|
479
606
|
-H "Content-Type: application/json" \
|
|
@@ -492,12 +619,18 @@ curl -X POST http://localhost:8000/api/generate \
|
|
|
492
619
|
# If you encounter OpenCV issues on Apple Silicon
|
|
493
620
|
pip uninstall opencv-python
|
|
494
621
|
pip install opencv-python-headless>=4.8.0
|
|
622
|
+
|
|
623
|
+
# For GGUF Metal acceleration
|
|
624
|
+
CMAKE_ARGS="-DSD_METAL=ON" pip install stable-diffusion-cpp-python
|
|
495
625
|
```
|
|
496
626
|
|
|
497
627
|
#### Windows
|
|
498
628
|
```bash
|
|
499
629
|
# If you encounter build errors
|
|
500
630
|
pip install --only-binary=all opencv-python>=4.8.0
|
|
631
|
+
|
|
632
|
+
# For GGUF CUDA acceleration
|
|
633
|
+
CMAKE_ARGS="-DSD_CUDA=ON" pip install stable-diffusion-cpp-python
|
|
501
634
|
```
|
|
502
635
|
|
|
503
636
|
#### Linux
|
|
@@ -540,6 +673,8 @@ This project is licensed under the MIT License - see the [LICENSE](LICENSE) file
|
|
|
540
673
|
## 🙏 Acknowledgments
|
|
541
674
|
|
|
542
675
|
- **Stability AI**: For Stable Diffusion models
|
|
676
|
+
- **Black Forest Labs**: For FLUX.1 models
|
|
677
|
+
- **city96**: For FLUX.1-dev GGUF quantizations
|
|
543
678
|
- **Hugging Face**: For model hosting and diffusers library
|
|
544
679
|
- **ControlNet Team**: For ControlNet architecture
|
|
545
680
|
- **Community**: For feedback and contributions
|
|
@@ -1,20 +1,23 @@
|
|
|
1
|
-
ollamadiffuser/__init__.py,sha256=
|
|
1
|
+
ollamadiffuser/__init__.py,sha256=n9dmXqrOp56VjdptVANGVeHea_UiuGHqP1d-b6dyq6g,1127
|
|
2
2
|
ollamadiffuser/__main__.py,sha256=tNWMvEHq4ddtKLp7DrhIoOdnFw3F8RNrETC_u5xpkFI,141
|
|
3
3
|
ollamadiffuser/api/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
|
-
ollamadiffuser/api/server.py,sha256=
|
|
4
|
+
ollamadiffuser/api/server.py,sha256=kc-Up50zmLfSM4f4InYR4Btkl35lyEN6UFcxlvxICSQ,18237
|
|
5
5
|
ollamadiffuser/cli/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
6
6
|
ollamadiffuser/cli/commands.py,sha256=Pe0vyfGiffwd10QlVxBCTtNnMqHi8nJ3oNn_k8nAi5k,8903
|
|
7
|
-
ollamadiffuser/cli/main.py,sha256=
|
|
7
|
+
ollamadiffuser/cli/main.py,sha256=qj0VKTOjw_gox2dPVtbU-9kCo25TqshyVyqW8qsv4Pk,56081
|
|
8
8
|
ollamadiffuser/core/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
9
9
|
ollamadiffuser/core/config/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
10
|
+
ollamadiffuser/core/config/model_registry.py,sha256=o408Qk-UQJ4NBM-AcyWAj4EhRUzrvl6zdOkKk1-ATHg,31238
|
|
10
11
|
ollamadiffuser/core/config/settings.py,sha256=VhI1vLGmOAQ7-XtyHrT5KoMpcGeGt-Mij-9NxX_ZKsI,4881
|
|
11
12
|
ollamadiffuser/core/inference/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
12
|
-
ollamadiffuser/core/inference/engine.py,sha256
|
|
13
|
+
ollamadiffuser/core/inference/engine.py,sha256=-EVcH4NyRVBQ-puvI_Az3KOZJYG9b-ySzsvoBTY3mlY,73962
|
|
13
14
|
ollamadiffuser/core/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
14
|
-
ollamadiffuser/core/models/
|
|
15
|
+
ollamadiffuser/core/models/gguf_loader.py,sha256=ocfl3_MDVXC9nSjW8YJdz4kX1Q-Qe2ltu6w4fbqhxVY,35724
|
|
16
|
+
ollamadiffuser/core/models/manager.py,sha256=rTEAameGih3wPcVG_Y-4k_brBeEqEoBjoI7fjggNtiY,16799
|
|
17
|
+
ollamadiffuser/core/models/registry.py,sha256=YPx3xcHnCHogyowi9fQ6oXZg7_jz0fM5bDyyg-BgSFY,15125
|
|
15
18
|
ollamadiffuser/core/utils/__init__.py,sha256=ZdXZWX1hfDnnV6OmRD6UStNljDJIQ892da2CtC-zdDw,31
|
|
16
19
|
ollamadiffuser/core/utils/controlnet_preprocessors.py,sha256=v21X_Bk-a4gKbUZUKoeP2W8TSGlv-ST8IYNsn3NrZ2c,15446
|
|
17
|
-
ollamadiffuser/core/utils/download_utils.py,sha256=
|
|
20
|
+
ollamadiffuser/core/utils/download_utils.py,sha256=dqHf7loKSqDHzgQEV-p1kDwoHxwlgurLPX8J96vD5fA,22421
|
|
18
21
|
ollamadiffuser/core/utils/lora_manager.py,sha256=SrZydPSGJqCS_Vek35bEdG2Q51qCOLZmPvnNzUjjIN0,14328
|
|
19
22
|
ollamadiffuser/ui/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
20
23
|
ollamadiffuser/ui/web.py,sha256=hK41sF9uVR_14lc4AlTXaGjw5m7aiJXpVp1cSOOmkT8,27715
|
|
@@ -34,9 +37,9 @@ ollamadiffuser/ui/samples/scribble/face_sketch.png,sha256=MVVYy_aS48xoS_RnIDzLUa
|
|
|
34
37
|
ollamadiffuser/ui/samples/scribble/tree_sketch.png,sha256=3P-NGgW25xRwreDxiBYKcDhd2oHZAwKSkjNVM5oPTWY,3017
|
|
35
38
|
ollamadiffuser/ui/templates/index.html,sha256=qTQVFxiTbeZ90O-iNqWC_4pYP6yyIs2z6U69VJPqAB4,38176
|
|
36
39
|
ollamadiffuser/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
37
|
-
ollamadiffuser-1.1.
|
|
38
|
-
ollamadiffuser-1.1.
|
|
39
|
-
ollamadiffuser-1.1.
|
|
40
|
-
ollamadiffuser-1.1.
|
|
41
|
-
ollamadiffuser-1.1.
|
|
42
|
-
ollamadiffuser-1.1.
|
|
40
|
+
ollamadiffuser-1.2.1.dist-info/licenses/LICENSE,sha256=cnGL9l2P510Uk3TCnv62kot6vAfdSawhOZh7Y-oYoIE,1071
|
|
41
|
+
ollamadiffuser-1.2.1.dist-info/METADATA,sha256=FoUPXS8xMBHOSfMgQcv2E1ITGhPU2IwF_nDU7BBHvQs,21756
|
|
42
|
+
ollamadiffuser-1.2.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
43
|
+
ollamadiffuser-1.2.1.dist-info/entry_points.txt,sha256=tHXXO3N0GSnIobDe_eSOLfHPjjVFjeTg2Fd-APoD6sY,64
|
|
44
|
+
ollamadiffuser-1.2.1.dist-info/top_level.txt,sha256=97wOGgTCxDE765Nr_o7B4Kwr_M_jy8fCCeQ81sMKlC4,15
|
|
45
|
+
ollamadiffuser-1.2.1.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|