Skip to content

Appendum

Descripion

Some extra useful information that is not part of the course, but could be useful in some way.

FP32 Version of Stable Diffusion 1.5 Pruned EMAOnly

Section Video Links
Using ComfyUI with a 10 Series Nvidia GPU Using ComfyUI with a 10 Series Nvidia GPU

If using a 10 Series NVidia GPU, modern generative AI will not be an enjoyable experience.

Many of the earlier lessons in this course use the Stable Diffusion 1.5 Pruned EMAOnly FP16 model.

If you have a 10 series Nvidia card, then you will be very limited in choices, since many AI models are released in FP16 format.

There is a FP32 version of SD1.5 Pruned EMAOnly that you can try instead.

It is twice as large to download and load into memory than the FP16 version, but may be faster to use if your 10 Series has enough VRAM.

SD1.5 Pruned EMAOnly Version filesize Link
FP16 2.13 GB Files and versions (huggingface)
FP32 4.27 GB Files and versions (huggingface)

Why is FP16 not advisable on 10 series GPUs

10 Series GPUs don't have native FP16 acceleration, so they need to emulate FP16 and this will cause extra resource management on the GPU.

10 Series don't have tensor cores, which are specialized hardware units designed to accelerate FP16 operations.

On 10-series GPUs, FP16 operations fall back to standard CUDA cores, which are optimized for FP32.

Which 10 Series GPUs can I use

GPU VRAM Notes
GTX 1050 / Ti 2–4 GB Might crash with FP32 due to VRAM limits.
GTX 1060 3GB 3 GB Might crash with FP32 due to VRAM limits.
GTX 1060 6GB 6 GB May just barely run FP32 SD1.5; FP16 helps memory, not speed.
GTX 1070 8 GB Can run FP32 model more comfortably. FP16 model saves memory but doesn’t improve speed.
GTX 1080 8 GB Can run FP32 model more comfortably. FP16 model saves memory but doesn’t improve speed.
GTX 1080 Ti 11 GB Runs FP32 fine; FP16 likely slightly slower due to conversion overhead.

v1-5-pruned-emaonly-fp16.safetensors (huggingface)

v1-5-pruned-emaonly.safetensors (FP32 Version, huggingface)

What’s the Difference Between Single-, Double-, Multi- and Mixed-Precision Computing? (Nvidia Blog)