I've only had this happen when I messed something up in my training settings. Keep in mind that a very slow learning rate can cause this as well. That said, maybe try training with kohya_ss, it's much simpler and less resource-heavy than the training extension in Auto1111.
Stable Diffusion
Discuss matters related to our favourite AI Art generation technology
Also see
- Stable Diffusion Art (See its sidebar for more GenAI Art comms)
- !aihorde@lemmy.dbzer0.com
Other communities
Depending on how the dependencies (i.e. xformers) are versioned you can make a new clone of A1111 and checkout a commit from ~8 months ago to see if it works again. Of course I would also recommend trying a fresh install.
Of course I would also recommend trying a fresh install.
Way ahead of you there. I've reinstalled the current version four or five times at this point.
make a new clone of A1111 and checkout a commit from ~8 months ago
This is a good idea. I've tried two different old versions from old commit hashes so far and both have crashed with other problems. It seems like (lol) both versions of A1111 put their venv in the same place, so the old versions are barfing on some dependencies with version numbers that are too high and they ALSO broke my current version by downgrading some other dependencies (easy fix, just wipe it out and reinstall it again). I'm trying to debug this, because I COULD see a world where I have an old version of A1111 training on one card while the NEW version generates on the other.
I don't know much about training, but maybe these can help.
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
Check out the readme and the advanced parameters. Iirc there's literally a checkbox for like half vae or something that says "check this if you're getting NaN errors"