Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
S
stable-diffusion-webui
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Administrator
stable-diffusion-webui
Commits
29eff4a1
Unverified
Commit
29eff4a1
authored
Nov 08, 2022
by
pepe10-gpu
Committed by
GitHub
Nov 08, 2022
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
terrible hack
parent
cd6c55c1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
9 additions
and
2 deletions
+9
-2
devices.py
modules/devices.py
+9
-2
No files found.
modules/devices.py
View file @
29eff4a1
...
@@ -39,8 +39,15 @@ def torch_gc():
...
@@ -39,8 +39,15 @@ def torch_gc():
def
enable_tf32
():
def
enable_tf32
():
if
torch
.
cuda
.
is_available
():
if
torch
.
cuda
.
is_available
():
torch
.
backends
.
cudnn
.
benchmark
=
True
#TODO: make this better; find a way to check if it is a turing card
torch
.
backends
.
cudnn
.
enabled
=
True
turing
=
[
"1630"
,
"1650"
,
"1660"
,
"Quadro RTX 3000"
,
"Quadro RTX 4000"
,
"Quadro RTX 4000"
,
"Quadro RTX 5000"
,
"Quadro RTX 5000"
,
"Quadro RTX 6000"
,
"Quadro RTX 6000"
,
"Quadro RTX 8000"
,
"Quadro RTX T400"
,
"Quadro RTX T400"
,
"Quadro RTX T600"
,
"Quadro RTX T1000"
,
"Quadro RTX T1000"
,
"2060"
,
"2070"
,
"2080"
,
"Titan RTX"
,
"Tesla T4"
,
"MX450"
,
"MX550"
]
for
devid
in
range
(
0
,
torch
.
cuda
.
device_count
()):
for
i
in
turing
:
if
i
in
torch
.
cuda
.
get_device_name
(
devid
):
shd
=
True
if
shd
:
torch
.
backends
.
cudnn
.
benchmark
=
True
torch
.
backends
.
cudnn
.
enabled
=
True
torch
.
backends
.
cuda
.
matmul
.
allow_tf32
=
True
torch
.
backends
.
cuda
.
matmul
.
allow_tf32
=
True
torch
.
backends
.
cudnn
.
allow_tf32
=
True
torch
.
backends
.
cudnn
.
allow_tf32
=
True
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment