Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
S
stable-diffusion-webui
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Administrator
stable-diffusion-webui
Commits
1740c335
Commit
1740c335
authored
Jan 07, 2023
by
AUTOMATIC
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
more comments
parent
08066676
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
16 additions
and
5 deletions
+16
-5
sd_hijack_clip.py
modules/sd_hijack_clip.py
+16
-5
No files found.
modules/sd_hijack_clip.py
View file @
1740c335
...
@@ -3,7 +3,7 @@ from collections import namedtuple
...
@@ -3,7 +3,7 @@ from collections import namedtuple
import
torch
import
torch
from
modules
import
prompt_parser
,
devices
from
modules
import
prompt_parser
,
devices
,
sd_hijack
from
modules.shared
import
opts
from
modules.shared
import
opts
...
@@ -22,14 +22,24 @@ class PromptChunk:
...
@@ -22,14 +22,24 @@ class PromptChunk:
PromptChunkFix
=
namedtuple
(
'PromptChunkFix'
,
[
'offset'
,
'embedding'
])
PromptChunkFix
=
namedtuple
(
'PromptChunkFix'
,
[
'offset'
,
'embedding'
])
"""This is a marker showing that textual inversion embedding's vectors have to placed at offset in the prompt chunk"""
"""An object of this type is a marker showing that textual inversion embedding's vectors have to placed at offset in the prompt
chunk. Thos objects are found in PromptChunk.fixes and, are placed into FrozenCLIPEmbedderWithCustomWordsBase.hijack.fixes, and finally
are applied by sd_hijack.EmbeddingsWithFixes's forward function."""
class
FrozenCLIPEmbedderWithCustomWordsBase
(
torch
.
nn
.
Module
):
class
FrozenCLIPEmbedderWithCustomWordsBase
(
torch
.
nn
.
Module
):
"""A pytorch module that is a wrapper for FrozenCLIPEmbedder module. it enhances FrozenCLIPEmbedder, making it possible to
have unlimited prompt length and assign weights to tokens in prompt.
"""
def
__init__
(
self
,
wrapped
,
hijack
):
def
__init__
(
self
,
wrapped
,
hijack
):
super
()
.
__init__
()
super
()
.
__init__
()
self
.
wrapped
=
wrapped
self
.
wrapped
=
wrapped
self
.
hijack
=
hijack
"""Original FrozenCLIPEmbedder module; can also be FrozenOpenCLIPEmbedder or xlmr.BertSeriesModelWithTransformation,
depending on model."""
self
.
hijack
:
sd_hijack
.
StableDiffusionModelHijack
=
hijack
self
.
chunk_length
=
75
self
.
chunk_length
=
75
def
empty_chunk
(
self
):
def
empty_chunk
(
self
):
...
@@ -55,7 +65,8 @@ class FrozenCLIPEmbedderWithCustomWordsBase(torch.nn.Module):
...
@@ -55,7 +65,8 @@ class FrozenCLIPEmbedderWithCustomWordsBase(torch.nn.Module):
converts a batch of token ids (in python lists) into a single tensor with numeric respresentation of those tokens;
converts a batch of token ids (in python lists) into a single tensor with numeric respresentation of those tokens;
All python lists with tokens are assumed to have same length, usually 77.
All python lists with tokens are assumed to have same length, usually 77.
if input is a list with B elements and each element has T tokens, expected output shape is (B, T, C), where C depends on
if input is a list with B elements and each element has T tokens, expected output shape is (B, T, C), where C depends on
model - can be 768 and 1024
model - can be 768 and 1024.
Among other things, this call will read self.hijack.fixes, apply it to its inputs, and clear it (setting it to None).
"""
"""
raise
NotImplementedError
raise
NotImplementedError
...
@@ -113,7 +124,7 @@ class FrozenCLIPEmbedderWithCustomWordsBase(torch.nn.Module):
...
@@ -113,7 +124,7 @@ class FrozenCLIPEmbedderWithCustomWordsBase(torch.nn.Module):
last_comma
=
len
(
chunk
.
tokens
)
last_comma
=
len
(
chunk
.
tokens
)
# this is when we are at the end of alloted 75 tokens for the current chunk, and the current token is not a comma. opts.comma_padding_backtrack
# this is when we are at the end of alloted 75 tokens for the current chunk, and the current token is not a comma. opts.comma_padding_backtrack
# is a setting that specifies that i
s there is a comma nearby, the text after
comma should be moved out of this chunk and into the next.
# is a setting that specifies that i
f there is a comma nearby, the text after the
comma should be moved out of this chunk and into the next.
elif
opts
.
comma_padding_backtrack
!=
0
and
len
(
chunk
.
tokens
)
==
self
.
chunk_length
and
last_comma
!=
-
1
and
len
(
chunk
.
tokens
)
-
last_comma
<=
opts
.
comma_padding_backtrack
:
elif
opts
.
comma_padding_backtrack
!=
0
and
len
(
chunk
.
tokens
)
==
self
.
chunk_length
and
last_comma
!=
-
1
and
len
(
chunk
.
tokens
)
-
last_comma
<=
opts
.
comma_padding_backtrack
:
break_location
=
last_comma
+
1
break_location
=
last_comma
+
1
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment