This is an automated archive made by the Lemmit Bot.

The original was posted on /r/stablediffusion by /u/lazyspock on 2024-09-21 23:28:34+00:00.


I’ve been using (and struggling with) the original Flux Dev FP16 since launch. It works in my 3060 12Gb, but it takes 2 to 3 minutes to generate an image without any LoRa, makes the computer unusable for anything else during generation, and it’s even worse with bigger LoRas (when it needs to reload the model for each generation). But it was the price to pay for being able to use Flux Dev, right?

Wrong. After this marvelous post from u/Lory1998 (thanks again!), I’ve decided to test Flux-Dev-Q5_1.gguf as he suggested and, man, what a difference! Now I can generate images considerably faster even with two LoRas, as the model fits entirely in my VRAM. There’s no reload of the model ever, as long as I don’t change the checkpoint, and even the LoRas are loaded in an instant. Also, I can use my computer for other non-intensive things like YouTube, Reddit, ect, while generating the image without Windows almost choking and without making the generation slower. And the best part is that there are no discernible quality differences in the generated images,

So, if you’re also a 12Gb VRAM person, try it. It’s worth it.