wangkanai commited on
Commit
bcb4296
·
verified ·
1 Parent(s): 7fc2a75

Add files using upload-large-folder tool

Browse files
README.md ADDED
@@ -0,0 +1,248 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- README Version: v1.0 -->
2
+
3
+ ---
4
+ license: openrail++
5
+ library_name: diffusers
6
+ pipeline_tag: text-to-image
7
+ tags:
8
+ - text-to-image
9
+ - stable-diffusion
10
+ - sdxl
11
+ - sdxl-base
12
+ - sdxl-turbo
13
+ - fp16
14
+ - image-generation
15
+ base_model: stabilityai/stable-diffusion-xl-base-1.0
16
+ ---
17
+
18
+ # Stable Diffusion XL FP16 Model Repository
19
+
20
+ Local repository containing Stable Diffusion XL (SDXL) checkpoint models in FP16 precision for high-quality text-to-image generation.
21
+
22
+ ## Model Description
23
+
24
+ This repository contains two SDXL checkpoint models optimized for different use cases:
25
+
26
+ - **SDXL Base**: Full-featured SDXL 1.0 base model for high-quality image generation with standard inference steps
27
+ - **SDXL Turbo**: Fast inference variant optimized for fewer steps (1-4 steps) while maintaining quality
28
+
29
+ Both models use FP16 (16-bit floating point) precision, providing a balance between quality and VRAM efficiency.
30
+
31
+ ## Repository Contents
32
+
33
+ ```
34
+ E:\huggingface\sdxl-fp16\
35
+ ├── checkpoints/
36
+ │ └── sdxl/
37
+ │ ├── sdxl-base.safetensors (6.94 GB)
38
+ │ └── sdxl-turbo.safetensors (13.88 GB)
39
+ ├── diffusion_models/
40
+ │ └── sdxl/ (empty - reserved)
41
+ └── loras/
42
+ └── sdxl/ (empty - reserved)
43
+ ```
44
+
45
+ **Total Repository Size**: ~20.82 GB
46
+
47
+ ### Model Files
48
+
49
+ | File | Size | Description |
50
+ |------|------|-------------|
51
+ | `sdxl-base.safetensors` | 6.94 GB | SDXL 1.0 base checkpoint (FP16) |
52
+ | `sdxl-turbo.safetensors` | 13.88 GB | SDXL Turbo checkpoint (FP16) |
53
+
54
+ ## Hardware Requirements
55
+
56
+ ### SDXL Base
57
+ - **VRAM**: 8GB minimum, 12GB+ recommended
58
+ - **Disk Space**: 7GB for model file
59
+ - **System RAM**: 16GB+ recommended
60
+ - **GPU**: NVIDIA GPU with CUDA support
61
+
62
+ ### SDXL Turbo
63
+ - **VRAM**: 12GB minimum, 16GB+ recommended
64
+ - **Disk Space**: 14GB for model file
65
+ - **System RAM**: 16GB+ recommended
66
+ - **GPU**: NVIDIA GPU with CUDA support
67
+
68
+ ## Usage Examples
69
+
70
+ ### SDXL Base (Standard Quality)
71
+
72
+ ```python
73
+ from diffusers import DiffusionPipeline
74
+ import torch
75
+
76
+ # Load SDXL base model from local path
77
+ pipe = DiffusionPipeline.from_single_file(
78
+ "E:/huggingface/sdxl-fp16/checkpoints/sdxl/sdxl-base.safetensors",
79
+ torch_dtype=torch.float16
80
+ )
81
+
82
+ pipe.to("cuda")
83
+
84
+ # Generate image with standard settings
85
+ image = pipe(
86
+ prompt="a beautiful mountain landscape at sunset, photorealistic, highly detailed",
87
+ negative_prompt="blurry, low quality, distorted",
88
+ num_inference_steps=50,
89
+ guidance_scale=7.5,
90
+ width=1024,
91
+ height=1024
92
+ ).images[0]
93
+
94
+ image.save("output.png")
95
+ ```
96
+
97
+ ### SDXL Turbo (Fast Generation)
98
+
99
+ ```python
100
+ from diffusers import DiffusionPipeline
101
+ import torch
102
+
103
+ # Load SDXL Turbo for fast inference
104
+ pipe = DiffusionPipeline.from_single_file(
105
+ "E:/huggingface/sdxl-fp16/checkpoints/sdxl/sdxl-turbo.safetensors",
106
+ torch_dtype=torch.float16
107
+ )
108
+
109
+ pipe.to("cuda")
110
+
111
+ # Generate with minimal steps (1-4 steps)
112
+ image = pipe(
113
+ prompt="a futuristic cityscape at night, neon lights, cyberpunk",
114
+ num_inference_steps=4, # Turbo optimized for 1-4 steps
115
+ guidance_scale=0.0, # Turbo works best with guidance_scale=0
116
+ width=1024,
117
+ height=1024
118
+ ).images[0]
119
+
120
+ image.save("turbo_output.png")
121
+ ```
122
+
123
+ ### Memory Optimization
124
+
125
+ ```python
126
+ import torch
127
+ from diffusers import DiffusionPipeline
128
+
129
+ # Enable memory-efficient attention
130
+ pipe = DiffusionPipeline.from_single_file(
131
+ "E:/huggingface/sdxl-fp16/checkpoints/sdxl/sdxl-base.safetensors",
132
+ torch_dtype=torch.float16
133
+ )
134
+
135
+ # Apply optimizations
136
+ pipe.enable_attention_slicing()
137
+ pipe.enable_vae_slicing()
138
+ pipe.to("cuda")
139
+
140
+ # Generate with optimized memory usage
141
+ image = pipe(
142
+ prompt="your prompt here",
143
+ num_inference_steps=30
144
+ ).images[0]
145
+ ```
146
+
147
+ ## Model Specifications
148
+
149
+ ### SDXL Base
150
+ - **Architecture**: Latent Diffusion Model with UNet
151
+ - **Parameters**: ~2.6B (UNet backbone)
152
+ - **Precision**: FP16 (16-bit floating point)
153
+ - **Format**: SafeTensors (secure, efficient)
154
+ - **Resolution**: 1024x1024 native, supports 512-2048px
155
+ - **Text Encoders**: Dual CLIP (OpenCLIP ViT-bigG, OpenAI CLIP ViT-L)
156
+ - **Inference Steps**: 30-50 recommended
157
+
158
+ ### SDXL Turbo
159
+ - **Architecture**: Adversarial Diffusion Distillation (ADD)
160
+ - **Parameters**: Similar to base with distillation optimizations
161
+ - **Precision**: FP16 (16-bit floating point)
162
+ - **Format**: SafeTensors
163
+ - **Resolution**: 1024x1024 native
164
+ - **Inference Steps**: 1-4 steps (optimized)
165
+ - **Guidance Scale**: 0.0 recommended (classifier-free guidance disabled)
166
+
167
+ ## Performance Tips
168
+
169
+ ### Speed Optimization
170
+ - **SDXL Turbo**: Use 1-4 steps with `guidance_scale=0.0` for fastest generation
171
+ - **Attention Slicing**: Enable with `pipe.enable_attention_slicing()` for memory efficiency
172
+ - **VAE Slicing**: Enable with `pipe.enable_vae_slicing()` to reduce VRAM usage
173
+ - **Lower Resolutions**: Use 768x768 or 512x512 for faster generation
174
+ - **Batch Processing**: Process multiple prompts together when VRAM allows
175
+
176
+ ### Quality Optimization
177
+ - **SDXL Base**: Use 40-50 steps for highest quality
178
+ - **Guidance Scale**: 7.0-9.0 for base model (higher = more prompt adherence)
179
+ - **Negative Prompts**: Use detailed negative prompts to avoid unwanted elements
180
+ - **Resolution**: 1024x1024 is the native resolution for best results
181
+ - **Aspect Ratios**: Multiples of 64 recommended (1024x768, 768x1024, etc.)
182
+
183
+ ### VRAM Management
184
+ - **8GB VRAM**: Use attention slicing, VAE slicing, lower batch sizes
185
+ - **12GB VRAM**: Standard settings with optimizations
186
+ - **16GB+ VRAM**: Can handle higher resolutions and batch sizes
187
+
188
+ ## Changelog
189
+
190
+ ### v1.0 (2025-10-13)
191
+ - Initial repository documentation
192
+ - Added SDXL Base checkpoint (6.94 GB)
193
+ - Added SDXL Turbo checkpoint (13.88 GB)
194
+ - Organized directory structure for checkpoints, diffusion models, and LoRAs
195
+
196
+ ## License
197
+
198
+ **License**: CreativeML Open RAIL++-M License
199
+
200
+ Stable Diffusion XL models are released under the [CreativeML Open RAIL++-M license](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/blob/main/LICENSE.md), which permits commercial use with the following key terms:
201
+
202
+ - ✅ Commercial use permitted
203
+ - ✅ Modification and redistribution allowed
204
+ - ⚠️ Use restrictions apply (see full license)
205
+ - ⚠️ Must include license and attribution
206
+
207
+ **Key Restrictions**: Cannot be used for illegal activities, generating harmful content, or violating privacy rights. See full license for complete terms.
208
+
209
+ ## Citation
210
+
211
+ If you use these models in your research or applications, please cite:
212
+
213
+ ```bibtex
214
+ @misc{podell2023sdxl,
215
+ title={SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis},
216
+ author={Dustin Podell and Zion English and Kyle Lacey and Andreas Blattmann and Tim Dockhorn and Jonas Müller and Joe Penna and Robin Rombach},
217
+ year={2023},
218
+ eprint={2307.01952},
219
+ archivePrefix={arXiv},
220
+ primaryClass={cs.CV}
221
+ }
222
+
223
+ @inproceedings{sauer2023adversarial,
224
+ title={Adversarial Diffusion Distillation},
225
+ author={Sauer, Axel and Lorenz, Dominik and Blattmann, Andreas and Rombach, Robin},
226
+ booktitle={arXiv preprint arXiv:2311.17042},
227
+ year={2023}
228
+ }
229
+ ```
230
+
231
+ ## Official Resources
232
+
233
+ - [SDXL Base Model](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0)
234
+ - [SDXL Turbo Model](https://huggingface.co/stabilityai/sdxl-turbo)
235
+ - [SDXL Documentation](https://huggingface.co/docs/diffusers/using-diffusers/sdxl)
236
+ - [Diffusers Library](https://github.com/huggingface/diffusers)
237
+ - [SDXL Paper](https://arxiv.org/abs/2307.01952)
238
+ - [SDXL Turbo Paper](https://arxiv.org/abs/2311.17042)
239
+
240
+ ## Contact & Support
241
+
242
+ - **Issues**: Report issues with models or documentation on [Hugging Face Discussions](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/discussions)
243
+ - **Community**: Join [Hugging Face Discord](https://discord.gg/hugging-face) for community support
244
+ - **Repository**: This is a local storage repository - for upstream issues, see official model pages
245
+
246
+ ---
247
+
248
+ **Repository maintained locally** | Last updated: 2025-10-13
checkpoints/sdxl/sdxl-base.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31e35c80fc4829d14f90153f4c74cd59c90b779f6afe05a74cd6120b893f7e5b
3
+ size 6938078334
checkpoints/sdxl/sdxl-turbo.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e58e3704b4c0831bf848e0507c9b5ff2cd8d007b8d0719dba3874156f631050
3
+ size 13875761366