Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
OpenFold
Commits
5ff5177b
Commit
5ff5177b
authored
Mar 27, 2024
by
Jennifer
Browse files
more logging changes
parent
0c3435cc
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
12 additions
and
38 deletions
+12
-38
train_openfold.py
train_openfold.py
+12
-38
No files found.
train_openfold.py
View file @
5ff5177b
...
@@ -410,9 +410,7 @@ def main(args):
...
@@ -410,9 +410,7 @@ def main(args):
wdb_logger
.
experiment
.
save
(
f
"
{
freeze_path
}
"
)
wdb_logger
.
experiment
.
save
(
f
"
{
freeze_path
}
"
)
# Raw dump of all args from pl.Trainer constructor
# Raw dump of all args from pl.Trainer constructor
trainer_kws
=
set
([
trainer_kws
=
[
'num_nodes'
,
'precision'
,
'max_epochs'
,
'log_every_n_steps'
,
'flush_logs_ever_n_steps'
,
'num_sanity_val_steps'
]
'accelerator'
,
'strategy'
,
'devices'
,
'num_nodes'
,
'precision'
,
'logger'
,
'callbacks'
,
'fast_dev_run'
,
'max_epochs'
,
'min_epochs'
,
'max_steps'
,
'min_steps'
,
'max_tim'
,
'limit_train_batches'
,
'limit_val_batches'
,
'limit_test_batches'
,
'limit_predict_batches'
,
'overfit_batches'
,
'val_check_interval'
,
'check_val_every_n_epoch'
,
'num_sanity_val_steps'
,
'log_every_n_steps'
,
'enable_checkpointing'
,
'enable_progress_bar'
,
'enable_model_summary'
,
'accumulate_grad_batches'
,
'gradient_clip_val'
,
'gradient_clip_algorithm'
,
'deterministic'
,
'benchmark'
,
'inference_mode'
,
'use_distributed_sampler'
,
'profiler'
,
'detect_anomaly'
,
'barebones'
,
'plugins'
,
'sync_batchnorm'
,
'reload_dataloaders_every_n_epochs'
,
'default_root_dir'
,
])
trainer_args
=
{
k
:
v
for
k
,
v
in
vars
(
args
).
items
()
if
k
in
trainer_kws
}
trainer_args
=
{
k
:
v
for
k
,
v
in
vars
(
args
).
items
()
if
k
in
trainer_kws
}
trainer_args
.
update
({
trainer_args
.
update
({
'default_root_dir'
:
args
.
output_dir
,
'default_root_dir'
:
args
.
output_dir
,
...
@@ -630,54 +628,30 @@ if __name__ == "__main__":
...
@@ -630,54 +628,30 @@ if __name__ == "__main__":
parser
.
add_argument
(
parser
.
add_argument
(
"--experiment_config_json"
,
default
=
""
,
help
=
"Path to a json file with custom config values to overwrite config setting"
,
"--experiment_config_json"
,
default
=
""
,
help
=
"Path to a json file with custom config values to overwrite config setting"
,
)
)
# Trainer additional arguments
# Ideally we'd want something like config.add_trainer_args()
parser
.
add_argument
(
parser
.
add_argument
(
"--num_nodes"
,
type
=
int
,
default
=
1
,
"--gpus"
,
type
=
int
,
default
=
1
,
help
=
'For determining optimal strategy and effective batch size.'
)
parser
.
add_argument
(
"--gpus"
,
type
=
int
,
default
=
1
,
)
parser
.
add_argument
(
"--num_workers"
,
type
=
int
,
default
=
4
,
# interaction with num_data_workers?
)
)
parser
.
add_argument
(
"--precision"
,
type
=
str
,
default
=
None
,
trainer_group
=
parser
.
add_argument_group
(
'PyTorch Lightning Trainer Args'
)
trainer_group
.
add_argument
(
"--num_nodes"
,
type
=
int
,
default
=
1
,
)
)
parser
.
add_argument
(
trainer_group
.
add_argument
(
"--re
place_sampler_ddp"
,
type
=
bool_type
,
default
=
True
,
"--
p
re
cision"
,
type
=
str
,
default
=
'bf16'
,
help
=
'Sets precision, lower precision improves runtime performance.'
)
)
parser
.
add_argument
(
trainer_group
.
add_argument
(
"--max_epochs"
,
type
=
int
,
default
=
1
,
"--max_epochs"
,
type
=
int
,
default
=
1
,
)
)
parser
.
add_argument
(
trainer_group
.
add_argument
(
"--log_every_n_steps"
,
type
=
int
,
default
=
25
,
"--log_every_n_steps"
,
type
=
int
,
default
=
25
,
)
)
parser
.
add_argument
(
trainer_group
.
add_argument
(
"--flush_logs_every_n_steps"
,
type
=
int
,
default
=
5
,
"--flush_logs_every_n_steps"
,
type
=
int
,
default
=
5
,
)
)
parser
.
add_argument
(
trainer_group
.
add_argument
(
"--num_sanity_val_steps"
,
type
=
int
,
default
=
0
,
"--num_sanity_val_steps"
,
type
=
int
,
default
=
0
,
)
)
# parser = pl.Trainer.add_argparse_args(parser)
#
# # Disable the initial validation pass
# parser.set_defaults(
# num_sanity_val_steps=0,
# )
# # Remove some buggy/redundant arguments introduced by the Trainer
# remove_arguments(
# parser,
# [
# "--accelerator",
# "--resume_from_checkpoint",
# "--reload_dataloaders_every_epoch",
# "--reload_dataloaders_every_n_epochs",
# ]
# )
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
if
(
args
.
seed
is
None
and
if
(
args
.
seed
is
None
and
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment