Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
OpenFold
Commits
89dee905
Unverified
Commit
89dee905
authored
Jun 22, 2022
by
Gustaf Ahdritz
Committed by
GitHub
Jun 22, 2022
Browse files
Merge pull request #117 from CyrusBiotechnology/run-multiple-models
Use multiple models for inference
parents
a48860cb
a2ab7ab7
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
169 additions
and
148 deletions
+169
-148
run_pretrained_openfold.py
run_pretrained_openfold.py
+169
-148
tests/test_data/sample_feats.pickle.gz
tests/test_data/sample_feats.pickle.gz
+0
-0
No files found.
run_pretrained_openfold.py
View file @
89dee905
...
@@ -19,6 +19,7 @@ import gc
...
@@ -19,6 +19,7 @@ import gc
import
logging
import
logging
import
numpy
as
np
import
numpy
as
np
import
os
import
os
from
copy
import
deepcopy
import
pickle
import
pickle
from
pytorch_lightning.utilities.deepspeed
import
(
from
pytorch_lightning.utilities.deepspeed
import
(
...
@@ -161,69 +162,125 @@ def prep_output(out, batch, feature_dict, feature_processor, args):
...
@@ -161,69 +162,125 @@ def prep_output(out, batch, feature_dict, feature_processor, args):
return
unrelaxed_protein
return
unrelaxed_protein
def
main
(
args
):
def
generate_batch
(
fasta_file
,
fasta_dir
,
alignment_dir
,
data_processor
,
feature_processor
,
prediction_dir
):
# Create the output directory
with
open
(
os
.
path
.
join
(
fasta_dir
,
fasta_file
),
"r"
)
as
fp
:
os
.
makedirs
(
args
.
output_dir
,
exist_ok
=
True
)
data
=
fp
.
read
(
)
# Prep the model
lines
=
[
config
=
model_config
(
args
.
config_preset
)
l
.
replace
(
'
\n
'
,
''
)
for
prot
in
data
.
split
(
'>'
)
for
l
in
prot
.
strip
().
split
(
'
\n
'
,
1
)
][
1
:]
tags
,
seqs
=
lines
[::
2
],
lines
[
1
::
2
]
tags
=
[
t
.
split
()[
0
]
for
t
in
tags
]
# assert len(tags) == len(set(tags)), "All FASTA tags must be unique"
tag
=
'-'
.
join
(
tags
)
output_name
=
f
'
{
tag
}
_
{
args
.
config_preset
}
'
if
args
.
output_postfix
is
not
None
:
output_name
=
f
'
{
output_name
}
_
{
args
.
output_postfix
}
'
# Save the unrelaxed PDB.
unrelaxed_output_path
=
os
.
path
.
join
(
prediction_dir
,
f
'
{
output_name
}
_unrelaxed.pdb'
)
if
os
.
path
.
exists
(
unrelaxed_output_path
):
return
precompute_alignments
(
tags
,
seqs
,
alignment_dir
,
args
)
tmp_fasta_path
=
os
.
path
.
join
(
args
.
output_dir
,
f
"tmp_
{
os
.
getpid
()
}
.fasta"
)
if
len
(
seqs
)
==
1
:
seq
=
seqs
[
0
]
with
open
(
tmp_fasta_path
,
"w"
)
as
fp
:
fp
.
write
(
f
">
{
tag
}
\n
{
seq
}
"
)
logger
.
info
(
f
"Using config preset
{
args
.
config_preset
}
..."
)
local_alignment_dir
=
os
.
path
.
join
(
alignment_dir
,
tag
)
feature_dict
=
data_processor
.
process_fasta
(
fasta_path
=
tmp_fasta_path
,
alignment_dir
=
local_alignment_dir
)
else
:
with
open
(
tmp_fasta_path
,
"w"
)
as
fp
:
fp
.
write
(
'
\n
'
.
join
([
f
">
{
tag
}
\n
{
seq
}
"
for
tag
,
seq
in
zip
(
tags
,
seqs
)])
)
feature_dict
=
data_processor
.
process_multiseq_fasta
(
fasta_path
=
tmp_fasta_path
,
super_alignment_dir
=
alignment_dir
,
)
# Remove temporary FASTA file
os
.
remove
(
tmp_fasta_path
)
processed_feature_dict
=
feature_processor
.
process_features
(
feature_dict
,
mode
=
'predict'
,
)
return
processed_feature_dict
,
tag
,
feature_dict
def
load_models_from_command_line
(
args
,
config
):
# Create the output directory
os
.
makedirs
(
args
.
output_dir
,
exist_ok
=
True
)
if
args
.
jax_param_path
:
for
path
in
args
.
jax_param_path
.
split
(
","
):
model
=
AlphaFold
(
config
)
model
=
AlphaFold
(
config
)
model
=
model
.
eval
()
model
=
model
.
eval
()
if
(
args
.
jax_param_path
):
import_jax_weights_
(
import_jax_weights_
(
model
,
args
.
jax_param_
path
,
version
=
args
.
config_preset
model
,
path
,
version
=
args
.
model_name
)
)
model
=
model
.
to
(
args
.
model_device
)
logger
.
info
(
logger
.
info
(
f
"Successfully loaded JAX parameters at
{
args
.
jax_param_path
}
..."
f
"Successfully loaded JAX parameters at
{
args
.
jax_param_path
}
..."
)
)
elif
(
args
.
openfold_checkpoint_path
):
yield
model
,
None
if
(
os
.
path
.
isdir
(
args
.
openfold_checkpoint_path
)):
if
args
.
openfold_checkpoint_path
:
# A DeepSpeed checkpoint
for
path
in
args
.
openfold_checkpoint_path
.
split
(
","
):
model
=
AlphaFold
(
config
)
model
=
model
.
eval
()
checkpoint_basename
=
os
.
path
.
splitext
(
checkpoint_basename
=
os
.
path
.
splitext
(
os
.
path
.
basename
(
os
.
path
.
basename
(
os
.
path
.
normpath
(
args
.
openfold_checkpoint_
path
)
os
.
path
.
normpath
(
path
)
)
)
)[
0
]
)[
0
]
if
os
.
path
.
isdir
(
path
):
# A DeepSpeed checkpoint
ckpt_path
=
os
.
path
.
join
(
ckpt_path
=
os
.
path
.
join
(
args
.
output_dir
,
args
.
output_dir
,
checkpoint_basename
+
".pt"
,
checkpoint_basename
+
".pt"
,
)
)
if
(
not
os
.
path
.
isfile
(
ckpt_path
)
)
:
if
not
os
.
path
.
isfile
(
ckpt_path
):
convert_zero_checkpoint_to_fp32_state_dict
(
convert_zero_checkpoint_to_fp32_state_dict
(
args
.
openfold_checkpoint_
path
,
path
,
ckpt_path
,
ckpt_path
,
)
)
d
=
torch
.
load
(
ckpt_path
)
d
=
torch
.
load
(
ckpt_path
)
model
.
load_state_dict
(
d
[
"ema"
][
"params"
])
model
.
load_state_dict
(
d
[
"ema"
][
"params"
])
else
:
else
:
# A checkpoint from the public release, which only contains EMA
ckpt_path
=
path
# params
ckpt_path
=
args
.
openfold_checkpoint_path
d
=
torch
.
load
(
ckpt_path
)
d
=
torch
.
load
(
ckpt_path
)
if
(
"ema"
in
d
):
if
(
"ema"
in
d
):
# The public weights have had this done to them already
# The public weights have had this done to them already
d
=
d
[
"ema"
][
"params"
]
d
=
d
[
"ema"
][
"params"
]
model
.
load_state_dict
(
d
)
model
.
load_state_dict
(
d
)
model
=
model
.
to
(
args
.
model_device
)
logger
.
info
(
logger
.
info
(
f
"Loaded OpenFold parameters at
{
args
.
openfold_checkpoint_path
}
..."
f
"Loaded OpenFold parameters at
{
args
.
openfold_checkpoint_path
}
..."
)
)
else
:
yield
model
,
checkpoint_basename
if
not
args
.
jax_param_path
and
not
args
.
openfold_checkpoint_path
:
raise
ValueError
(
raise
ValueError
(
"At least one of jax_param_path or openfold_checkpoint_path must "
"At least one of jax_param_path or openfold_checkpoint_path must "
"be specified."
"be specified."
)
)
model
=
model
.
to
(
args
.
model_device
)
def
main
(
args
):
# Create the output directory
os
.
makedirs
(
args
.
output_dir
,
exist_ok
=
True
)
config
=
model_config
(
args
.
config_preset
)
template_featurizer
=
templates
.
TemplateHitFeaturizer
(
template_featurizer
=
templates
.
TemplateHitFeaturizer
(
mmcif_dir
=
args
.
template_mmcif_dir
,
mmcif_dir
=
args
.
template_mmcif_dir
,
max_template_date
=
args
.
max_template_date
,
max_template_date
=
args
.
max_template_date
,
...
@@ -244,7 +301,7 @@ def main(args):
...
@@ -244,7 +301,7 @@ def main(args):
feature_processor
=
feature_pipeline
.
FeaturePipeline
(
config
.
data
)
feature_processor
=
feature_pipeline
.
FeaturePipeline
(
config
.
data
)
if
not
os
.
path
.
exists
(
output_dir_base
):
if
not
os
.
path
.
exists
(
output_dir_base
):
os
.
makedirs
(
output_dir_base
)
os
.
makedirs
(
output_dir_base
)
if
(
args
.
use_precomputed_alignments
is
None
)
:
if
args
.
use_precomputed_alignments
is
None
:
alignment_dir
=
os
.
path
.
join
(
output_dir_base
,
"alignments"
)
alignment_dir
=
os
.
path
.
join
(
output_dir_base
,
"alignments"
)
else
:
else
:
alignment_dir
=
args
.
use_precomputed_alignments
alignment_dir
=
args
.
use_precomputed_alignments
...
@@ -254,73 +311,39 @@ def main(args):
...
@@ -254,73 +311,39 @@ def main(args):
os
.
makedirs
(
prediction_dir
,
exist_ok
=
True
)
os
.
makedirs
(
prediction_dir
,
exist_ok
=
True
)
for
fasta_file
in
os
.
listdir
(
args
.
fasta_dir
):
for
fasta_file
in
os
.
listdir
(
args
.
fasta_dir
):
# Gather input sequences
with
open
(
os
.
path
.
join
(
args
.
fasta_dir
,
fasta_file
),
"r"
)
as
fp
:
data
=
fp
.
read
()
lines
=
[
batch_data
=
generate_batch
(
l
.
replace
(
'
\n
'
,
''
)
fasta_file
,
for
prot
in
data
.
split
(
'>'
)
for
l
in
prot
.
strip
().
split
(
'
\n
'
,
1
)
args
.
fasta_dir
,
][
1
:]
alignment_dir
,
tags
,
seqs
=
lines
[::
2
],
lines
[
1
::
2
]
data_processor
,
feature_processor
,
tags
=
[
t
.
split
()[
0
]
for
t
in
tags
]
prediction_dir
)
# assert len(tags) == len(set(tags)), "All FASTA tags must be unique"
tag
=
'-'
.
join
(
tags
)
output_name
=
f
'
{
tag
}
_
{
args
.
config_preset
}
'
if
(
args
.
output_postfix
is
not
None
):
output_name
=
f
'
{
output_name
}
_
{
args
.
output_postfix
}
'
# Save the unrelaxed PDB.
unrelaxed_output_path
=
os
.
path
.
join
(
prediction_dir
,
f
'
{
output_name
}
_unrelaxed.pdb'
)
if
(
os
.
path
.
exists
(
unrelaxed_output_path
)):
if
batch_data
is
None
:
# this file has already been processed
continue
continue
precompute_alignments
(
tag
s
,
seqs
,
alignment_dir
,
args
)
batch
,
tag
,
feature_dict
=
batch_data
tmp_fasta_path
=
os
.
path
.
join
(
args
.
output_dir
,
f
"tmp_
{
os
.
getpid
()
}
.fasta"
)
for
model
,
model_version
in
load_models_from_command_line
(
args
,
config
):
if
(
len
(
seqs
)
==
1
):
seq
=
seqs
[
0
]
with
open
(
tmp_fasta_path
,
"w"
)
as
fp
:
fp
.
write
(
f
">
{
tag
}
\n
{
seq
}
"
)
local_alignment_dir
=
os
.
path
.
join
(
alignment_dir
,
tag
)
working_batch
=
deepcopy
(
batch
)
feature_dict
=
data_processor
.
process_fasta
(
out
=
run_model
(
model
,
working_batch
,
tag
,
args
)
fasta_path
=
tmp_fasta_path
,
alignment_dir
=
local_alignment_dir
)
else
:
with
open
(
tmp_fasta_path
,
"w"
)
as
fp
:
fp
.
write
(
'
\n
'
.
join
([
f
">
{
tag
}
\n
{
seq
}
"
for
tag
,
seq
in
zip
(
tags
,
seqs
)])
)
feature_dict
=
data_processor
.
process_multiseq_fasta
(
fasta_path
=
tmp_fasta_path
,
super_alignment_dir
=
alignment_dir
,
)
# Remove temporary FASTA file
os
.
remove
(
tmp_fasta_path
)
processed_feature_dict
=
feature_processor
.
process_features
(
feature_dict
,
mode
=
'predict'
,
)
batch
=
processed_feature_dict
out
=
run_model
(
model
,
batch
,
tag
,
args
)
# Toss out the recycling dimensions --- we don't need them anymore
# Toss out the recycling dimensions --- we don't need them anymore
batch
=
tensor_tree_map
(
lambda
x
:
np
.
array
(
x
[...,
-
1
].
cpu
()),
batch
)
working_
batch
=
tensor_tree_map
(
lambda
x
:
np
.
array
(
x
[...,
-
1
].
cpu
()),
working_
batch
)
out
=
tensor_tree_map
(
lambda
x
:
np
.
array
(
x
.
cpu
()),
out
)
out
=
tensor_tree_map
(
lambda
x
:
np
.
array
(
x
.
cpu
()),
out
)
unrelaxed_protein
=
prep_output
(
unrelaxed_protein
=
prep_output
(
out
,
batch
,
feature_dict
,
feature_processor
,
args
out
,
working_
batch
,
feature_dict
,
feature_processor
,
args
)
)
output_name
=
f
'
{
tag
}
_
{
args
.
config_preset
}
'
output_name
=
f
'
{
tag
}
_
{
args
.
config_preset
}
'
if
(
args
.
output_postfix
is
not
None
):
if
model_version
is
not
None
:
output_name
=
f
'
{
output_name
}
_
{
model_version
}
'
if
args
.
output_postfix
is
not
None
:
output_name
=
f
'
{
output_name
}
_
{
args
.
output_postfix
}
'
output_name
=
f
'
{
output_name
}
_
{
args
.
output_postfix
}
'
# Save the unrelaxed PDB.
# Save the unrelaxed PDB.
...
@@ -331,8 +354,7 @@ def main(args):
...
@@ -331,8 +354,7 @@ def main(args):
fp
.
write
(
protein
.
to_pdb
(
unrelaxed_protein
))
fp
.
write
(
protein
.
to_pdb
(
unrelaxed_protein
))
logger
.
info
(
f
"Output written to
{
unrelaxed_output_path
}
..."
)
logger
.
info
(
f
"Output written to
{
unrelaxed_output_path
}
..."
)
if
not
args
.
skip_relaxation
:
if
(
not
args
.
skip_relaxation
):
amber_relaxer
=
relax
.
AmberRelaxation
(
amber_relaxer
=
relax
.
AmberRelaxation
(
use_gpu
=
(
args
.
model_device
!=
"cpu"
),
use_gpu
=
(
args
.
model_device
!=
"cpu"
),
**
config
.
relax
,
**
config
.
relax
,
...
@@ -342,7 +364,7 @@ def main(args):
...
@@ -342,7 +364,7 @@ def main(args):
logger
.
info
(
f
"Running relaxation on
{
unrelaxed_output_path
}
..."
)
logger
.
info
(
f
"Running relaxation on
{
unrelaxed_output_path
}
..."
)
t
=
time
.
perf_counter
()
t
=
time
.
perf_counter
()
visible_devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
,
default
=
""
)
visible_devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
,
default
=
""
)
if
(
"cuda"
in
args
.
model_device
)
:
if
"cuda"
in
args
.
model_device
:
device_no
=
args
.
model_device
.
split
(
":"
)[
-
1
]
device_no
=
args
.
model_device
.
split
(
":"
)[
-
1
]
os
.
environ
[
"CUDA_VISIBLE_DEVICES"
]
=
device_no
os
.
environ
[
"CUDA_VISIBLE_DEVICES"
]
=
device_no
relaxed_pdb_str
,
_
,
_
=
amber_relaxer
.
process
(
prot
=
unrelaxed_protein
)
relaxed_pdb_str
,
_
,
_
=
amber_relaxer
.
process
(
prot
=
unrelaxed_protein
)
...
@@ -355,10 +377,9 @@ def main(args):
...
@@ -355,10 +377,9 @@ def main(args):
)
)
with
open
(
relaxed_output_path
,
'w'
)
as
fp
:
with
open
(
relaxed_output_path
,
'w'
)
as
fp
:
fp
.
write
(
relaxed_pdb_str
)
fp
.
write
(
relaxed_pdb_str
)
logger
.
info
(
f
"Relaxed output written to
{
relaxed_output_path
}
..."
)
logger
.
info
(
f
"Relaxed output written to
{
relaxed_output_path
}
..."
)
if
(
args
.
save_outputs
)
:
if
args
.
save_outputs
:
output_dict_path
=
os
.
path
.
join
(
output_dict_path
=
os
.
path
.
join
(
args
.
output_dir
,
f
'
{
output_name
}
_output_dict.pkl'
args
.
output_dir
,
f
'
{
output_name
}
_output_dict.pkl'
)
)
...
...
tests/test_data/sample_feats.pickle.gz
deleted
100755 → 0
View file @
a48860cb
File deleted
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment