Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
S
stable-diffusion-webui
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Administrator
stable-diffusion-webui
Commits
cbb857b6
Commit
cbb857b6
authored
Oct 26, 2022
by
AUTOMATIC
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
enable creating embedding with --medvram
parent
ee73341f
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
0 deletions
+3
-0
textual_inversion.py
modules/textual_inversion/textual_inversion.py
+3
-0
No files found.
modules/textual_inversion/textual_inversion.py
View file @
cbb857b6
...
@@ -157,6 +157,9 @@ def create_embedding(name, num_vectors_per_token, overwrite_old, init_text='*'):
...
@@ -157,6 +157,9 @@ def create_embedding(name, num_vectors_per_token, overwrite_old, init_text='*'):
cond_model
=
shared
.
sd_model
.
cond_stage_model
cond_model
=
shared
.
sd_model
.
cond_stage_model
embedding_layer
=
cond_model
.
wrapped
.
transformer
.
text_model
.
embeddings
embedding_layer
=
cond_model
.
wrapped
.
transformer
.
text_model
.
embeddings
with
devices
.
autocast
():
cond_model
([
""
])
# will send cond model to GPU if lowvram/medvram is active
ids
=
cond_model
.
tokenizer
(
init_text
,
max_length
=
num_vectors_per_token
,
return_tensors
=
"pt"
,
add_special_tokens
=
False
)[
"input_ids"
]
ids
=
cond_model
.
tokenizer
(
init_text
,
max_length
=
num_vectors_per_token
,
return_tensors
=
"pt"
,
add_special_tokens
=
False
)[
"input_ids"
]
embedded
=
embedding_layer
.
token_embedding
.
wrapped
(
ids
.
to
(
devices
.
device
))
.
squeeze
(
0
)
embedded
=
embedding_layer
.
token_embedding
.
wrapped
(
ids
.
to
(
devices
.
device
))
.
squeeze
(
0
)
vec
=
torch
.
zeros
((
num_vectors_per_token
,
embedded
.
shape
[
1
]),
device
=
devices
.
device
)
vec
=
torch
.
zeros
((
num_vectors_per_token
,
embedded
.
shape
[
1
]),
device
=
devices
.
device
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment