Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
flash-attention
Commits
c2407dec
Commit
c2407dec
authored
Dec 21, 2022
by
Tri Dao
Browse files
Fix typo in config: train.gpu -> train.gpu_mem
parent
13cdceb3
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
4 additions
and
4 deletions
+4
-4
training/configs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary-8k.yaml
...gs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary-8k.yaml
+1
-1
training/configs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary.yaml
...nfigs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary.yaml
+1
-1
training/configs/experiment/pile/gpt3-2.7B-flash-rotary-8k.yaml
...ng/configs/experiment/pile/gpt3-2.7B-flash-rotary-8k.yaml
+1
-1
training/configs/experiment/pile/gpt3-2.7B-flash-rotary.yaml
training/configs/experiment/pile/gpt3-2.7B-flash-rotary.yaml
+1
-1
No files found.
training/configs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary-8k.yaml
View file @
c2407dec
...
...
@@ -11,7 +11,7 @@ model:
mlp_checkpoint_lvl
:
0
datamodule
:
batch_size
:
${eval:"1 if ${train.gpu_mem} < 24 else (2 if ${train.gpu_mem} < 40 else (4 if ${train.gpu} < 80 else 8))"}
batch_size
:
${eval:"1 if ${train.gpu_mem} < 24 else (2 if ${train.gpu_mem} < 40 else (4 if ${train.gpu
_mem
} < 80 else 8))"}
train
:
optimizer
:
...
...
training/configs/experiment/pile/gpt3-2.7B-flash-hdim128-rotary.yaml
View file @
c2407dec
...
...
@@ -11,7 +11,7 @@ model:
mlp_checkpoint_lvl
:
0
datamodule
:
batch_size
:
${eval:"4 if ${train.gpu_mem} < 24 else (8 if ${train.gpu_mem} < 40 else (16 if ${train.gpu} < 80 else 32))"}
batch_size
:
${eval:"4 if ${train.gpu_mem} < 24 else (8 if ${train.gpu_mem} < 40 else (16 if ${train.gpu
_mem
} < 80 else 32))"}
train
:
optimizer
:
...
...
training/configs/experiment/pile/gpt3-2.7B-flash-rotary-8k.yaml
View file @
c2407dec
...
...
@@ -11,7 +11,7 @@ model:
mlp_checkpoint_lvl
:
0
datamodule
:
batch_size
:
${eval:"1 if ${train.gpu_mem} < 24 else (2 if ${train.gpu_mem} < 40 else (4 if ${train.gpu} < 80 else 8))"}
batch_size
:
${eval:"1 if ${train.gpu_mem} < 24 else (2 if ${train.gpu_mem} < 40 else (4 if ${train.gpu
_mem
} < 80 else 8))"}
train
:
optimizer
:
...
...
training/configs/experiment/pile/gpt3-2.7B-flash-rotary.yaml
View file @
c2407dec
...
...
@@ -11,7 +11,7 @@ model:
mlp_checkpoint_lvl
:
0
datamodule
:
batch_size
:
${eval:"4 if ${train.gpu_mem} < 24 else (8 if ${train.gpu_mem} < 40 else (16 if ${train.gpu} < 80 else 32))"}
batch_size
:
${eval:"4 if ${train.gpu_mem} < 24 else (8 if ${train.gpu_mem} < 40 else (16 if ${train.gpu
_mem
} < 80 else 32))"}
train
:
optimizer
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment