Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
OpenFold
Commits
aebda3d8
"runtime/git@developer.sourcefind.cn:OpenDAS/dynamo.git" did not exist on "8588e33a464d9f82d6ad93a433590a3bc3ff92de"
Commit
aebda3d8
authored
Nov 13, 2023
by
Christina Floristean
Browse files
Config fixes for when using ds kernel
parent
f1563999
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
12 additions
and
1 deletion
+12
-1
openfold/config.py
openfold/config.py
+12
-1
No files found.
openfold/config.py
View file @
aebda3d8
...
@@ -28,6 +28,7 @@ def enforce_config_constraints(config):
...
@@ -28,6 +28,7 @@ def enforce_config_constraints(config):
(
(
"globals.use_lma"
,
"globals.use_lma"
,
"globals.use_flash"
,
"globals.use_flash"
,
"globals.use_deepspeed_evo_attention"
),
),
]
]
...
@@ -38,9 +39,18 @@ def enforce_config_constraints(config):
...
@@ -38,9 +39,18 @@ def enforce_config_constraints(config):
raise
ValueError
(
f
"Only one of
{
s1
}
and
{
s2
}
may be set at a time"
)
raise
ValueError
(
f
"Only one of
{
s1
}
and
{
s2
}
may be set at a time"
)
fa_is_installed
=
importlib
.
util
.
find_spec
(
"flash_attn"
)
is
not
None
fa_is_installed
=
importlib
.
util
.
find_spec
(
"flash_attn"
)
is
not
None
if
(
config
.
globals
.
use_flash
and
not
fa_is_installed
)
:
if
config
.
globals
.
use_flash
and
not
fa_is_installed
:
raise
ValueError
(
"use_flash requires that FlashAttention is installed"
)
raise
ValueError
(
"use_flash requires that FlashAttention is installed"
)
deepspeed_is_installed
=
importlib
.
util
.
find_spec
(
"deepspeed"
)
is
not
None
ds4s_is_installed
=
deepspeed_is_installed
and
importlib
.
util
.
find_spec
(
"deepspeed.ops.deepspeed4science"
)
is
not
None
if
config
.
globals
.
use_deepspeed_evo_attention
and
not
ds4s_is_installed
:
raise
ValueError
(
"use_deepspeed_evo_attention requires that DeepSpeed be installed "
"and that the deepspeed.ops.deepspeed4science package exists"
)
if
(
if
(
config
.
globals
.
offload_inference
and
config
.
globals
.
offload_inference
and
not
config
.
model
.
template
.
average_templates
not
config
.
model
.
template
.
average_templates
...
@@ -193,6 +203,7 @@ def model_config(
...
@@ -193,6 +203,7 @@ def model_config(
c
.
globals
.
offload_inference
=
True
c
.
globals
.
offload_inference
=
True
c
.
globals
.
use_lma
=
True
c
.
globals
.
use_lma
=
True
c
.
globals
.
use_flash
=
False
c
.
globals
.
use_flash
=
False
c
.
globals
.
use_deepspeed_evo_attention
=
False
c
.
model
.
template
.
offload_inference
=
True
c
.
model
.
template
.
offload_inference
=
True
c
.
model
.
template
.
template_pair_stack
.
tune_chunk_size
=
False
c
.
model
.
template
.
template_pair_stack
.
tune_chunk_size
=
False
c
.
model
.
extra_msa
.
extra_msa_stack
.
tune_chunk_size
=
False
c
.
model
.
extra_msa
.
extra_msa_stack
.
tune_chunk_size
=
False
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment