Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
huaerkl
fairseq-data2vec_pytorch
Commits
72f5785f
Commit
72f5785f
authored
Aug 15, 2023
by
huaerkl
Browse files
v1.0
parents
Pipeline
#505
canceled with stages
Changes
508
Pipelines
1
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
914 additions
and
0 deletions
+914
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_1.yaml
...data2vec/config/audio/pretraining/run_config/slurm_1.yaml
+37
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_1_aws.yaml
...2vec/config/audio/pretraining/run_config/slurm_1_aws.yaml
+36
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_2.yaml
...data2vec/config/audio/pretraining/run_config/slurm_2.yaml
+37
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_2_aws.yaml
...2vec/config/audio/pretraining/run_config/slurm_2_aws.yaml
+37
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_3.yaml
...data2vec/config/audio/pretraining/run_config/slurm_3.yaml
+36
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_4.yaml
...data2vec/config/audio/pretraining/run_config/slurm_4.yaml
+36
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_4_aws.yaml
...2vec/config/audio/pretraining/run_config/slurm_4_aws.yaml
+37
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_6_aws.yaml
...2vec/config/audio/pretraining/run_config/slurm_6_aws.yaml
+36
-0
examples/data2vec/config/audio/pretraining/run_config/slurm_8_aws.yaml
...2vec/config/audio/pretraining/run_config/slurm_8_aws.yaml
+36
-0
examples/data2vec/config/text/pretraining/base.yaml
examples/data2vec/config/text/pretraining/base.yaml
+77
-0
examples/data2vec/config/text/pretraining/run_config/local.yaml
...es/data2vec/config/text/pretraining/run_config/local.yaml
+15
-0
examples/data2vec/config/text/pretraining/run_config/slurm_1_aws.yaml
...a2vec/config/text/pretraining/run_config/slurm_1_aws.yaml
+37
-0
examples/data2vec/config/text/pretraining/run_config/slurm_2.yaml
.../data2vec/config/text/pretraining/run_config/slurm_2.yaml
+37
-0
examples/data2vec/config/text/pretraining/run_config/slurm_2_aws.yaml
...a2vec/config/text/pretraining/run_config/slurm_2_aws.yaml
+37
-0
examples/data2vec/config/text/pretraining/run_config/slurm_3.yaml
.../data2vec/config/text/pretraining/run_config/slurm_3.yaml
+36
-0
examples/data2vec/config/text/pretraining/run_config/slurm_4.yaml
.../data2vec/config/text/pretraining/run_config/slurm_4.yaml
+36
-0
examples/data2vec/config/text/pretraining/run_config/slurm_4_aws.yaml
...a2vec/config/text/pretraining/run_config/slurm_4_aws.yaml
+41
-0
examples/data2vec/config/text/pretraining/run_config/slurm_8_aws.yaml
...a2vec/config/text/pretraining/run_config/slurm_8_aws.yaml
+41
-0
examples/data2vec/config/v2/base_audio_only_task.yaml
examples/data2vec/config/v2/base_audio_only_task.yaml
+113
-0
examples/data2vec/config/v2/base_images_only_task.yaml
examples/data2vec/config/v2/base_images_only_task.yaml
+116
-0
No files found.
Too many changes to show.
To preserve performance only
508 of 508+
files are displayed.
Plain diff
Email patch
examples/data2vec/config/audio/pretraining/run_config/slurm_1.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
80
gpus_per_node
:
8
tasks_per_node
:
1
mem_gb
:
450
nodes
:
1
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_1_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
80
gpus_per_node
:
8
tasks_per_node
:
1
mem_gb
:
0
nodes
:
1
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec,learnlab,learnfair
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_2.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
450
nodes
:
2
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_2_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
task.post_save_script
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
2
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec,learnlab,learnfair
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_3.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
80
gpus_per_node
:
8
tasks_per_node
:
1
mem_gb
:
450
nodes
:
3
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_4.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
450
nodes
:
4
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_4_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
task.post_save_script
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
4
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec,learnlab,learnfair
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_6_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
6
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec,learnlab,learnfair
max_num_timeout
:
30
examples/data2vec/config/audio/pretraining/run_config/slurm_8_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
8
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec,learnlab,learnfair
max_num_timeout
:
30
examples/data2vec/config/text/pretraining/base.yaml
0 → 100644
View file @
72f5785f
# @package _group_
common
:
fp16
:
true
log_format
:
json
log_interval
:
200
tensorboard_logdir
:
tb
checkpoint
:
no_epoch_checkpoints
:
true
save_interval_updates
:
50000
keep_interval_updates
:
1
distributed_training
:
distributed_world_size
:
16
ddp_backend
:
legacy_ddp
task
:
_name
:
masked_lm
data
:
???
sample_break_mode
:
complete_doc
tokens_per_sample
:
512
include_target_tokens
:
true
random_token_prob
:
0
leave_unmasked_prob
:
0
mask_prob
:
0.35
mask_multiple_length
:
4
criterion
:
model
dataset
:
max_tokens
:
8192
ignore_unused_valid_subsets
:
true
skip_invalid_size_inputs_valid_test
:
true
optimizer
:
_name
:
adam
weight_decay
:
0.01
adam_betas
:
(0.9,0.98)
adam_eps
:
1e-06
lr_scheduler
:
_name
:
cosine
warmup_updates
:
10000
optimization
:
clip_norm
:
5
lr
:
[
0.0002
]
max_update
:
1000000
update_freq
:
[
1
]
model
:
_name
:
data2vec_text
head_layers
:
2
average_top_k_layers
:
10
layer_norm_target_layer
:
true
loss_scale
:
1
ema_decay
:
0.999
ema_end_decay
:
0.9999
ema_anneal_end_step
:
300000
loss_beta
:
4
ema_transformer_layers_only
:
true
transformer
:
dropout
:
0.1
attention_dropout
:
0.1
layernorm_embedding
:
true
activation_fn
:
gelu
no_scale_embedding
:
true
max_source_positions
:
512
encoder
:
embed_dim
:
768
ffn_embed_dim
:
3072
layers
:
12
attention_heads
:
12
normalize_before
:
false
learned_pos
:
true
layerdrop
:
0
examples/data2vec/config/text/pretraining/run_config/local.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
sweep
:
dir
:
${env:PWD}/tmp_dbg/${now:%H-%M-%S}
distributed_training
:
distributed_world_size
:
1
nprocs_per_node
:
1
distributed_port
:
-1
common
:
log_interval
:
1
dataset
:
num_workers
:
0
examples/data2vec/config/text/pretraining/run_config/slurm_1_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
_'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}/submitit
timeout_min
:
4320
cpus_per_task
:
80
gpus_per_node
:
8
tasks_per_node
:
1
mem_gb
:
0
nodes
:
1
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec
max_num_timeout
:
30
exclude
:
a100-st-p4d24xlarge-471
examples/data2vec/config/text/pretraining/run_config/slurm_2.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
80
gpus_per_node
:
8
tasks_per_node
:
1
mem_gb
:
450
nodes
:
2
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/text/pretraining/run_config/slurm_2_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
_'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}/submitit
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
2
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec
max_num_timeout
:
30
exclude
:
a100-st-p4d24xlarge-471
examples/data2vec/config/text/pretraining/run_config/slurm_3.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
450
nodes
:
3
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/text/pretraining/run_config/slurm_4.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
:'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
sweep
:
dir
:
/checkpoint/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
450
nodes
:
4
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
devlab,learnlab,learnfair,scavenge
constraint
:
volta32gb,ib4
max_num_timeout
:
30
examples/data2vec/config/text/pretraining/run_config/slurm_4_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
_'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}/submitit
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
4
name
:
${env:PREFIX}_${hydra.job.config_name}
partition
:
wav2vec
max_num_timeout
:
30
exclude
:
a100-st-p4d24xlarge-471
distributed_training
:
distributed_world_size
:
32
ddp_backend
:
legacy_ddp
examples/data2vec/config/text/pretraining/run_config/slurm_8_aws.yaml
0 → 100644
View file @
72f5785f
# @package _global_
hydra
:
job
:
config
:
override_dirname
:
kv_sep
:
'
_'
item_sep
:
'
/'
exclude_keys
:
-
run_config
-
distributed_training.distributed_port
-
distributed_training.distributed_world_size
-
model.pretrained_model_path
-
model.target_network_path
-
next_script
-
task.cache_in_scratch
-
task.data
-
checkpoint.save_interval_updates
-
checkpoint.keep_interval_updates
-
checkpoint.save_on_overflow
-
common.log_interval
-
common.user_dir
sweep
:
dir
:
/fsx-wav2vec/${env:USER}/${env:PREFIX}/${hydra.job.config_name}_${hydra.launcher.gpus_per_node}/${hydra.job.override_dirname}
subdir
:
'
'
launcher
:
submitit_folder
:
${hydra.sweep.dir}/submitit
timeout_min
:
4320
cpus_per_task
:
10
gpus_per_node
:
8
tasks_per_node
:
8
mem_gb
:
0
nodes
:
8
name
:
pt
partition
:
wav2vec
max_num_timeout
:
30
exclude
:
a100-st-p4d24xlarge-471
distributed_training
:
distributed_world_size
:
64
ddp_backend
:
legacy_ddp
examples/data2vec/config/v2/base_audio_only_task.yaml
0 → 100644
View file @
72f5785f
# @package _group_
common
:
fp16
:
true
log_format
:
json
log_interval
:
200
tensorboard_logdir
:
tb
min_loss_scale
:
1e-6
fp16_no_flatten_grads
:
false
user_dir
:
${env:PWD}/examples/data2vec
checkpoint
:
save_interval
:
1
save_interval_updates
:
25000
keep_interval_updates
:
1
no_epoch_checkpoints
:
true
task
:
_name
:
audio_pretraining
data
:
/private/home/abaevski/data/librispeech/full
max_sample_size
:
320000
min_sample_size
:
32000
normalize
:
true
precompute_mask_config
:
{}
dataset
:
num_workers
:
6
max_tokens
:
1000000
skip_invalid_size_inputs_valid_test
:
true
validate_interval
:
5
required_batch_size_multiple
:
1
disable_validation
:
true
distributed_training
:
distributed_world_size
:
8
ddp_backend
:
legacy_ddp
criterion
:
_name
:
model
log_keys
:
-
ema_decay
-
target_var
-
pred_var
-
model_norm
-
ema_norm
-
masked_pct
optimization
:
max_update
:
400000
lr
:
[
0.00075
]
debug_param_names
:
true
optimizer
:
_name
:
adam
adam_betas
:
[
0.9
,
0.98
]
adam_eps
:
1e-06
weight_decay
:
0.01
lr_scheduler
:
_name
:
cosine
warmup_updates
:
8000
model
:
_name
:
data2vec_multi
loss_beta
:
0
loss_scale
:
null
depth
:
12
embed_dim
:
768
clone_batch
:
8
ema_decay
:
0.999
ema_end_decay
:
0.99999
ema_anneal_end_step
:
75000
ema_encoder_only
:
false
average_top_k_layers
:
8
instance_norm_target_layer
:
true
layer_norm_target_layer
:
false
layer_norm_targets
:
false
layerdrop
:
0.05
norm_eps
:
1e-5
supported_modality
:
AUDIO
modalities
:
audio
:
feature_encoder_spec
:
'
[(512,
10,
5)]
+
[(512,
3,
2)]
*
4
+
[(512,2,2)]
+
[(512,2,2)]'
conv_pos_depth
:
5
conv_pos_width
:
95
conv_pos_groups
:
16
prenet_depth
:
0
mask_prob
:
0.5
mask_prob_adjust
:
0.05
inverse_mask
:
false
mask_length
:
5
mask_noise_std
:
0.01
mask_dropout
:
0
add_masks
:
false
ema_local_encoder
:
false
use_alibi_encoder
:
true
prenet_layerdrop
:
0.05
prenet_dropout
:
0.1
learned_alibi_scale
:
true
learned_alibi_scale_per_head
:
true
decoder
:
input_dropout
:
0.1
decoder_dim
:
384
decoder_groups
:
16
decoder_kernel
:
7
decoder_layers
:
4
examples/data2vec/config/v2/base_images_only_task.yaml
0 → 100644
View file @
72f5785f
# @package _group_
common
:
fp16
:
true
log_format
:
json
log_interval
:
200
tensorboard_logdir
:
tb
min_loss_scale
:
1e-6
fp16_no_flatten_grads
:
true
user_dir
:
${env:PWD}/examples/data2vec
checkpoint
:
save_interval
:
5
save_interval_updates
:
25000
keep_interval_updates
:
1
no_epoch_checkpoints
:
true
task
:
_name
:
mae_image_pretraining
data
:
/datasets01/imagenet_full_size/061417/
rebuild_batches
:
true
local_cache_path
:
/scratch/cache_abaevski/imagenet
key
:
source
precompute_mask_config
:
{}
dataset
:
num_workers
:
10
batch_size
:
8
skip_invalid_size_inputs_valid_test
:
true
required_batch_size_multiple
:
1
disable_validation
:
true
distributed_training
:
distributed_world_size
:
4
ddp_backend
:
c10d
criterion
:
_name
:
model
log_keys
:
-
ema_decay
-
target_var
-
pred_var
-
model_norm
-
ema_norm
-
masked_pct
optimization
:
max_update
:
375300
lr
:
[
0.001
]
debug_param_names
:
true
clip_norm
:
4
optimizer
:
_name
:
composite
dynamic_groups
:
true
groups
:
default
:
lr_float
:
1e-3
optimizer
:
_name
:
adam
adam_betas
:
[
0.9
,
0.95
]
weight_decay
:
0.05
lr_scheduler
:
_name
:
cosine
warmup_updates
:
50040
lr_scheduler
:
pass_through
model
:
_name
:
data2vec_multi
ema_decay
:
0.9998
ema_end_decay
:
0.99999
ema_anneal_end_step
:
100000
instance_norm_target_layer
:
true
layer_norm_target_layer
:
false
layer_norm_targets
:
true
end_of_block_targets
:
false
depth
:
10
average_top_k_layers
:
10
clone_batch
:
16
norm_eps
:
1e-6
min_target_var
:
0
min_pred_var
:
0
encoder_dropout
:
0
post_mlp_drop
:
0
attention_dropout
:
0
activation_dropout
:
0
supported_modality
:
IMAGE
cls_loss
:
0.01
ema_encoder_only
:
false
modalities
:
image
:
inverse_mask
:
true
mask_prob
:
0.8
mask_prob_adjust
:
0.07
mask_length
:
3
mask_noise_std
:
0.01
prenet_depth
:
2
ema_local_encoder
:
true
num_extra_tokens
:
1
init_extra_token_zero
:
false
use_alibi_encoder
:
false
decoder
:
decoder_dim
:
768
decoder_groups
:
16
decoder_kernel
:
3
decoder_layers
:
6
input_dropout
:
0
Prev
1
…
9
10
11
12
13
14
15
16
17
…
26
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment