Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
0bf5e500
Commit
0bf5e500
authored
Nov 28, 2022
by
Tri Dao
Browse files
Release training code
parent
9bc63d1e
Changes
139
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
134 additions
and
0 deletions
+134
-0
training/configs/optimizer/adamw-apex-zero.yaml
training/configs/optimizer/adamw-apex-zero.yaml
+7
-0
training/configs/optimizer/adamw-apex.yaml
training/configs/optimizer/adamw-apex.yaml
+3
-0
training/configs/optimizer/adamw-zero.yaml
training/configs/optimizer/adamw-zero.yaml
+7
-0
training/configs/optimizer/adamw.yaml
training/configs/optimizer/adamw.yaml
+2
-0
training/configs/optimizer/fusedlamb-ds.yaml
training/configs/optimizer/fusedlamb-ds.yaml
+2
-0
training/configs/optimizer/fusedlamb.yaml
training/configs/optimizer/fusedlamb.yaml
+2
-0
training/configs/optimizer/sgd.yaml
training/configs/optimizer/sgd.yaml
+2
-0
training/configs/scheduler/cosine-warmup-timm.yaml
training/configs/scheduler/cosine-warmup-timm.yaml
+2
-0
training/configs/scheduler/cosine-warmup.yaml
training/configs/scheduler/cosine-warmup.yaml
+2
-0
training/configs/scheduler/invsqrt.yaml
training/configs/scheduler/invsqrt.yaml
+3
-0
training/configs/scheduler/linear-warmup.yaml
training/configs/scheduler/linear-warmup.yaml
+2
-0
training/configs/scheduler/multi-step.yaml
training/configs/scheduler/multi-step.yaml
+2
-0
training/configs/scheduler/plateau.yaml
training/configs/scheduler/plateau.yaml
+9
-0
training/configs/scheduler/poly-warmup.yaml
training/configs/scheduler/poly-warmup.yaml
+2
-0
training/configs/scheduler/step.yaml
training/configs/scheduler/step.yaml
+3
-0
training/configs/task/sequence-model.yaml
training/configs/task/sequence-model.yaml
+1
-0
training/configs/trainer/all_params.yaml
training/configs/trainer/all_params.yaml
+49
-0
training/configs/trainer/ddp.yaml
training/configs/trainer/ddp.yaml
+6
-0
training/configs/trainer/debug.yaml
training/configs/trainer/debug.yaml
+21
-0
training/configs/trainer/default.yaml
training/configs/trainer/default.yaml
+7
-0
No files found.
training/configs/optimizer/adamw-apex-zero.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
torch.distributed.optim.ZeroRedundancyOptimizer
_recursive_
:
True
optimizer_class
:
_target_
:
apex.optimizers.FusedAdam
_partial_
:
True
adam_w_mode
:
True
training/configs/optimizer/adamw-apex.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
apex.optimizers.FusedAdam
adam_w_mode
:
True
training/configs/optimizer/adamw-zero.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
torch.distributed.optim.ZeroRedundancyOptimizer
_recursive_
:
True
optimizer_class
:
_target_
:
torch.optim.__getattribute__
_args_
:
-
"
AdamW"
training/configs/optimizer/adamw.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
torch.optim.AdamW
training/configs/optimizer/fusedlamb-ds.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
deepspeed.ops.lamb.FusedLamb
training/configs/optimizer/fusedlamb.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
apex.optimizers.FusedLAMB
training/configs/optimizer/sgd.yaml
0 → 100644
View file @
0bf5e500
# @package train.optimizer
_target_
:
torch.optim.SGD
training/configs/scheduler/cosine-warmup-timm.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
src.optim.timm_lr_scheduler.TimmCosineLRScheduler
training/configs/scheduler/cosine-warmup.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
transformers.get_cosine_schedule_with_warmup
training/configs/scheduler/invsqrt.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
src.optim.lr_scheduler.InvSqrt
num_warmup_steps
:
???
training/configs/scheduler/linear-warmup.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
transformers.get_linear_schedule_with_warmup
training/configs/scheduler/multi-step.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
torch.optim.lr_scheduler.MultiStepLR
training/configs/scheduler/plateau.yaml
0 → 100644
View file @
0bf5e500
# @package _global_
train
:
scheduler_interval
:
epoch
scheduler_monitor
:
???
scheduler
:
_target_
:
torch.optim.lr_scheduler.ReduceLROnPlateau
factor
:
0.2
# Decay factor when ReduceLROnPlateau is used
patience
:
20
min_lr
:
0.0
# Minimum learning rate during annealing
training/configs/scheduler/poly-warmup.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
transformers.get_polynomial_decay_schedule_with_warmup
training/configs/scheduler/step.yaml
0 → 100644
View file @
0bf5e500
# @package train.scheduler
_target_
:
torch.optim.lr_scheduler.StepLR
step_size
:
???
training/configs/task/sequence-model.yaml
0 → 100644
View file @
0bf5e500
_target_
:
src.tasks.seq.SequenceModel
training/configs/trainer/all_params.yaml
0 → 100644
View file @
0bf5e500
_target_
:
pytorch_lightning.Trainer
# default values for all trainer parameters
checkpoint_callback
:
True
default_root_dir
:
null
gradient_clip_val
:
0.0
process_position
:
0
num_nodes
:
1
num_processes
:
1
gpus
:
null
auto_select_gpus
:
False
tpu_cores
:
null
log_gpu_memory
:
null
overfit_batches
:
0.0
track_grad_norm
:
-1
check_val_every_n_epoch
:
1
fast_dev_run
:
False
accumulate_grad_batches
:
1
max_epochs
:
1
min_epochs
:
1
max_steps
:
null
min_steps
:
null
limit_train_batches
:
1.0
limit_val_batches
:
1.0
limit_test_batches
:
1.0
val_check_interval
:
1.0
flush_logs_every_n_steps
:
100
log_every_n_steps
:
50
accelerator
:
null
sync_batchnorm
:
False
precision
:
32
weights_summary
:
"
top"
weights_save_path
:
null
num_sanity_val_steps
:
2
truncated_bptt_steps
:
null
resume_from_checkpoint
:
null
profiler
:
null
benchmark
:
False
deterministic
:
False
reload_dataloaders_every_epoch
:
False
auto_lr_find
:
False
replace_sampler_ddp
:
True
terminate_on_nan
:
False
auto_scale_batch_size
:
False
prepare_data_per_node
:
True
plugins
:
null
amp_backend
:
"
native"
amp_level
:
"
O2"
move_metrics_to_cpu
:
False
training/configs/trainer/ddp.yaml
0 → 100644
View file @
0bf5e500
defaults
:
-
default.yaml
accelerator
:
gpu
devices
:
4
strategy
:
ddp
training/configs/trainer/debug.yaml
0 → 100644
View file @
0bf5e500
defaults
:
-
default.yaml
gpus
:
0
min_epochs
:
1
max_epochs
:
2
# prints
weights_summary
:
"
full"
profiler
:
null
# debugs
fast_dev_run
:
true
num_sanity_val_steps
:
2
overfit_batches
:
0
limit_train_batches
:
1.0
limit_val_batches
:
1.0
limit_test_batches
:
1.0
track_grad_norm
:
-1
terminate_on_nan
:
true
training/configs/trainer/default.yaml
0 → 100644
View file @
0bf5e500
_target_
:
pytorch_lightning.Trainer
# set `gpu` to train on GPU, null to train on CPU only
accelerator
:
null
min_epochs
:
1
max_epochs
:
1000
Prev
1
2
3
4
5
6
7
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment