Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
S
stable-diffusion-webui
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Administrator
stable-diffusion-webui
Commits
89d8ecff
Commit
89d8ecff
authored
Nov 22, 2022
by
flamelaw
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
small fixes
parent
5b57f61b
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
4 additions
and
4 deletions
+4
-4
hypernetwork.py
modules/hypernetworks/hypernetwork.py
+3
-3
textual_inversion.py
modules/textual_inversion/textual_inversion.py
+1
-1
No files found.
modules/hypernetworks/hypernetwork.py
View file @
89d8ecff
...
@@ -435,8 +435,8 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step,
...
@@ -435,8 +435,8 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step,
optimizer_name
=
hypernetwork
.
optimizer_name
optimizer_name
=
hypernetwork
.
optimizer_name
else
:
else
:
print
(
f
"Optimizer type {hypernetwork.optimizer_name} is not defined!"
)
print
(
f
"Optimizer type {hypernetwork.optimizer_name} is not defined!"
)
optimizer
=
torch
.
optim
.
AdamW
(
params
=
weights
,
lr
=
scheduler
.
learn_rate
)
optimizer
=
torch
.
optim
.
AdamW
(
params
=
weights
,
lr
=
scheduler
.
learn_rate
)
optimizer_name
=
'AdamW'
optimizer_name
=
'AdamW'
if
hypernetwork
.
optimizer_state_dict
:
# This line must be changed if Optimizer type can be different from saved optimizer.
if
hypernetwork
.
optimizer_state_dict
:
# This line must be changed if Optimizer type can be different from saved optimizer.
try
:
try
:
...
@@ -582,7 +582,7 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step,
...
@@ -582,7 +582,7 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, gradient_step,
shared
.
state
.
textinfo
=
f
"""
shared
.
state
.
textinfo
=
f
"""
<p>
<p>
Loss: {loss_step:.7f}<br/>
Loss: {loss_step:.7f}<br/>
Step: {
hypernetwork.step
}<br/>
Step: {
steps_done
}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last saved hypernetwork: {html.escape(last_saved_file)}<br/>
Last saved hypernetwork: {html.escape(last_saved_file)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
...
...
modules/textual_inversion/textual_inversion.py
View file @
89d8ecff
...
@@ -436,7 +436,7 @@ def train_embedding(embedding_name, learn_rate, batch_size, gradient_step, data_
...
@@ -436,7 +436,7 @@ def train_embedding(embedding_name, learn_rate, batch_size, gradient_step, data_
shared
.
state
.
textinfo
=
f
"""
shared
.
state
.
textinfo
=
f
"""
<p>
<p>
Loss: {loss_step:.7f}<br/>
Loss: {loss_step:.7f}<br/>
Step: {
embedding.step
}<br/>
Step: {
steps_done
}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last prompt: {html.escape(batch.cond_text[0])}<br/>
Last saved embedding: {html.escape(last_saved_file)}<br/>
Last saved embedding: {html.escape(last_saved_file)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
Last saved image: {html.escape(last_saved_image)}<br/>
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment