Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
S
stable-diffusion-webui
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Administrator
stable-diffusion-webui
Commits
9553a7e0
Unverified
Commit
9553a7e0
authored
Oct 29, 2022
by
AUTOMATIC1111
Committed by
GitHub
Oct 29, 2022
Browse files
Options
Browse Files
Download
Plain Diff
Merge pull request #3818 from jwatzman/master
Reduce peak memory usage when changing models
parents
28e6d4a5
b50ff4f4
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
7 additions
and
4 deletions
+7
-4
sd_models.py
modules/sd_models.py
+7
-4
No files found.
modules/sd_models.py
View file @
9553a7e0
...
...
@@ -173,7 +173,9 @@ def load_model_weights(model, checkpoint_info):
print
(
f
"Global Step: {pl_sd['global_step']}"
)
sd
=
get_state_dict_from_checkpoint
(
pl_sd
)
missing
,
extra
=
model
.
load_state_dict
(
sd
,
strict
=
False
)
del
pl_sd
model
.
load_state_dict
(
sd
,
strict
=
False
)
del
sd
if
shared
.
cmd_opts
.
opt_channelslast
:
model
.
to
(
memory_format
=
torch
.
channels_last
)
...
...
@@ -197,9 +199,10 @@ def load_model_weights(model, checkpoint_info):
model
.
first_stage_model
.
to
(
devices
.
dtype_vae
)
checkpoints_loaded
[
checkpoint_info
]
=
model
.
state_dict
()
.
copy
()
while
len
(
checkpoints_loaded
)
>
shared
.
opts
.
sd_checkpoint_cache
:
checkpoints_loaded
.
popitem
(
last
=
False
)
# LRU
if
shared
.
opts
.
sd_checkpoint_cache
>
0
:
checkpoints_loaded
[
checkpoint_info
]
=
model
.
state_dict
()
.
copy
()
while
len
(
checkpoints_loaded
)
>
shared
.
opts
.
sd_checkpoint_cache
:
checkpoints_loaded
.
popitem
(
last
=
False
)
# LRU
else
:
print
(
f
"Loading weights [{sd_model_hash}] from cache"
)
checkpoints_loaded
.
move_to_end
(
checkpoint_info
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment