Appendum
Descripion
Some extra useful information that is not part of the course, but could be useful in some way.
FP32 Version of Stable Diffusion 1.5 Pruned EMAOnly
Section | Video Links |
---|---|
Using ComfyUI with a 10 Series Nvidia GPU | ![]() |
If using a 10 Series NVidia GPU, modern generative AI will not be an enjoyable experience.
Many of the earlier lessons in this course use the Stable Diffusion 1.5 Pruned EMAOnly FP16 model.
If you have a 10 series Nvidia card, then you will be very limited in choices, since many AI models are released in FP16 format.
There is a FP32 version of SD1.5 Pruned EMAOnly that you can try instead.
It is twice as large to download and load into memory than the FP16 version, but may be faster to use if your 10 Series has enough VRAM.
SD1.5 Pruned EMAOnly Version | filesize | Link |
---|---|---|
FP16 | 2.13 GB | Files and versions (huggingface) |
FP32 | 4.27 GB | Files and versions (huggingface) |
Why is FP16 not advisable on 10 series GPUs
10 Series GPUs don't have native FP16 acceleration, so they need to emulate FP16 and this will cause extra resource management on the GPU.
10 Series don't have tensor cores, which are specialized hardware units designed to accelerate FP16 operations.
On 10-series GPUs, FP16 operations fall back to standard CUDA cores, which are optimized for FP32.
Which 10 Series GPUs can I use
GPU | VRAM | Notes |
---|---|---|
GTX 1050 / Ti | 2–4 GB | Might crash with FP32 due to VRAM limits. |
GTX 1060 3GB | 3 GB | Might crash with FP32 due to VRAM limits. |
GTX 1060 6GB | 6 GB | May just barely run FP32 SD1.5; FP16 helps memory, not speed. |
GTX 1070 | 8 GB | Can run FP32 model more comfortably. FP16 model saves memory but doesn’t improve speed. |
GTX 1080 | 8 GB | Can run FP32 model more comfortably. FP16 model saves memory but doesn’t improve speed. |
GTX 1080 Ti | 11 GB | Runs FP32 fine; FP16 likely slightly slower due to conversion overhead. |
Useful Links
v1-5-pruned-emaonly-fp16.safetensors (huggingface)
v1-5-pruned-emaonly.safetensors (FP32 Version, huggingface)
What’s the Difference Between Single-, Double-, Multi- and Mixed-Precision Computing? (Nvidia Blog)