ComfyUI-GGUF is a project that provides GGUF quantization support for native ComfyUI models. It allows model files to be stored in GGUF format, which is promoted by llama.cpp. Although standard UNET models (conv2d) are not suitable for quantization, transformer/DiT models like flux appear to be minimally affected by quantization. This enables them to run on low-end GPUs with lower bits per weight variable rate.