• Josh Watzman's avatar
    Reduce peak memory usage when changing models · b50ff4f4
    Josh Watzman authored
    A few tweaks to reduce peak memory usage, the biggest being that if we
    aren't using the checkpoint cache, we shouldn't duplicate the model
    state dict just to immediately throw it away.
    
    On my machine with 16GB of RAM, this change means I can typically change
    models, whereas before it would typically OOM.
    b50ff4f4
Name
Last commit
Last update
.github Loading commit data...
embeddings Loading commit data...
extensions Loading commit data...
javascript Loading commit data...
localizations Loading commit data...
models Loading commit data...
modules Loading commit data...
scripts Loading commit data...
textual_inversion_templates Loading commit data...
.gitignore Loading commit data...
.pylintrc Loading commit data...
CODEOWNERS Loading commit data...
README.md Loading commit data...
artists.csv Loading commit data...
environment-wsl2.yaml Loading commit data...
launch.py Loading commit data...
requirements.txt Loading commit data...
requirements_versions.txt Loading commit data...
screenshot.png Loading commit data...
script.js Loading commit data...
style.css Loading commit data...
txt2img_Screenshot.png Loading commit data...
webui-user.bat Loading commit data...
webui-user.sh Loading commit data...
webui.bat Loading commit data...
webui.py Loading commit data...
webui.sh Loading commit data...