Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ColossalAI
Commits
df66741f
Unverified
Commit
df66741f
authored
Sep 21, 2023
by
Baizhou Zhang
Committed by
GitHub
Sep 21, 2023
Browse files
[bug] fix get_default_parser in examples (#4764)
parent
c0a03370
Changes
14
Hide whitespace changes
Inline
Side-by-side
Showing
14 changed files
with
36 additions
and
26 deletions
+36
-26
colossalai/legacy/__init__.py
colossalai/legacy/__init__.py
+9
-1
examples/community/roberta/pretraining/arguments.py
examples/community/roberta/pretraining/arguments.py
+2
-2
examples/images/vit/args.py
examples/images/vit/args.py
+3
-3
examples/images/vit/run_benchmark.sh
examples/images/vit/run_benchmark.sh
+3
-3
examples/images/vit/run_demo.sh
examples/images/vit/run_demo.sh
+2
-2
examples/images/vit/test_ci.sh
examples/images/vit/test_ci.sh
+2
-2
examples/language/gpt/titans/train_gpt.py
examples/language/gpt/titans/train_gpt.py
+2
-1
examples/language/opt/args.py
examples/language/opt/args.py
+3
-3
examples/language/opt/run_benchmark.sh
examples/language/opt/run_benchmark.sh
+2
-2
examples/language/opt/run_demo.sh
examples/language/opt/run_demo.sh
+2
-2
examples/language/opt/test_ci.sh
examples/language/opt/test_ci.sh
+2
-2
examples/language/palm/run.sh
examples/language/palm/run.sh
+1
-1
examples/language/palm/test_ci.sh
examples/language/palm/test_ci.sh
+1
-1
examples/language/palm/train.py
examples/language/palm/train.py
+2
-1
No files found.
colossalai/legacy/__init__.py
View file @
df66741f
from
.initialize
import
initialize
,
launch
,
launch_from_openmpi
,
launch_from_slurm
,
launch_from_torch
from
.initialize
import
(
get_default_parser
,
initialize
,
launch
,
launch_from_openmpi
,
launch_from_slurm
,
launch_from_torch
,
)
__all__
=
[
__all__
=
[
"launch"
,
"launch"
,
...
@@ -6,4 +13,5 @@ __all__ = [
...
@@ -6,4 +13,5 @@ __all__ = [
"launch_from_slurm"
,
"launch_from_slurm"
,
"launch_from_torch"
,
"launch_from_torch"
,
"initialize"
,
"initialize"
,
"get_default_parser"
,
]
]
examples/community/roberta/pretraining/arguments.py
View file @
df66741f
import
colossalai
import
argparse
__all__
=
[
"parse_args"
]
__all__
=
[
"parse_args"
]
def
parse_args
():
def
parse_args
():
parser
=
colossalai
.
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--distplan"
,
"--distplan"
,
...
...
examples/images/vit/args.py
View file @
df66741f
from
colossalai
import
get_default_
parse
r
import
arg
parse
def
parse_demo_args
():
def
parse_demo_args
():
parser
=
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--model_name_or_path"
,
"--model_name_or_path"
,
type
=
str
,
type
=
str
,
...
@@ -52,7 +52,7 @@ def parse_demo_args():
...
@@ -52,7 +52,7 @@ def parse_demo_args():
def
parse_benchmark_args
():
def
parse_benchmark_args
():
parser
=
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--model_name_or_path"
,
"--model_name_or_path"
,
...
...
examples/images/vit/run_benchmark.sh
View file @
df66741f
...
@@ -11,9 +11,9 @@ for PLUGIN in "torch_ddp" "torch_ddp_fp16" "low_level_zero" "gemini" "hybrid_par
...
@@ -11,9 +11,9 @@ for PLUGIN in "torch_ddp" "torch_ddp_fp16" "low_level_zero" "gemini" "hybrid_par
do
do
MODEL_PATH
=
"google/vit-base-patch16-224"
MODEL_PATH
=
"google/vit-base-patch16-224"
torch
run
\
colossalai
run
\
--
standalone
\
--
nproc_per_node
${
GPUNUM
}
\
--
nproc_per_node
4
\
--
master_port
29505
\
vit_benchmark.py
\
vit_benchmark.py
\
--model_name_or_path
${
MODEL_PATH
}
\
--model_name_or_path
${
MODEL_PATH
}
\
--mem_cap
${
MEMCAP
}
\
--mem_cap
${
MEMCAP
}
\
...
...
examples/images/vit/run_demo.sh
View file @
df66741f
...
@@ -35,9 +35,9 @@ WEIGHT_DECAY=0.05
...
@@ -35,9 +35,9 @@ WEIGHT_DECAY=0.05
WARMUP_RATIO
=
0.3
WARMUP_RATIO
=
0.3
# run the script for demo
# run the script for demo
torchrun
\
colossalai run
\
--standalone
\
--nproc_per_node
${
GPUNUM
}
\
--nproc_per_node
${
GPUNUM
}
\
--master_port
29505
\
vit_train_demo.py
\
vit_train_demo.py
\
--model_name_or_path
${
MODEL
}
\
--model_name_or_path
${
MODEL
}
\
--output_path
${
OUTPUT_PATH
}
\
--output_path
${
OUTPUT_PATH
}
\
...
...
examples/images/vit/test_ci.sh
View file @
df66741f
...
@@ -5,9 +5,9 @@ BS=8
...
@@ -5,9 +5,9 @@ BS=8
for
PLUGIN
in
"torch_ddp"
"torch_ddp_fp16"
"low_level_zero"
"gemini"
"hybrid_parallel"
for
PLUGIN
in
"torch_ddp"
"torch_ddp_fp16"
"low_level_zero"
"gemini"
"hybrid_parallel"
do
do
torchrun
\
colossalai run
\
--standalone
\
--nproc_per_node
4
\
--nproc_per_node
4
\
--master_port
29505
\
vit_benchmark.py
\
vit_benchmark.py
\
--model_name_or_path
"google/vit-base-patch16-224"
\
--model_name_or_path
"google/vit-base-patch16-224"
\
--plugin
${
PLUGIN
}
\
--plugin
${
PLUGIN
}
\
...
...
examples/language/gpt/titans/train_gpt.py
View file @
df66741f
import
argparse
import
contextlib
import
contextlib
import
os
import
os
...
@@ -29,7 +30,7 @@ VOCAB_SIZE = 50257
...
@@ -29,7 +30,7 @@ VOCAB_SIZE = 50257
def
main
():
def
main
():
parser
=
colossalai
.
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
"--from_torch"
,
default
=
False
,
action
=
"store_true"
)
parser
.
add_argument
(
"--from_torch"
,
default
=
False
,
action
=
"store_true"
)
parser
.
add_argument
(
"--use_dummy_dataset"
,
default
=
False
,
action
=
"store_true"
)
parser
.
add_argument
(
"--use_dummy_dataset"
,
default
=
False
,
action
=
"store_true"
)
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
...
...
examples/language/opt/args.py
View file @
df66741f
from
colossalai
import
get_default_
parse
r
import
arg
parse
def
parse_demo_args
():
def
parse_demo_args
():
parser
=
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--model_name_or_path"
,
"--model_name_or_path"
,
type
=
str
,
type
=
str
,
...
@@ -39,7 +39,7 @@ def parse_demo_args():
...
@@ -39,7 +39,7 @@ def parse_demo_args():
def
parse_benchmark_args
():
def
parse_benchmark_args
():
parser
=
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--model_name_or_path"
,
"--model_name_or_path"
,
type
=
str
,
type
=
str
,
...
...
examples/language/opt/run_benchmark.sh
View file @
df66741f
...
@@ -16,9 +16,9 @@ for GPUNUM in 1 4
...
@@ -16,9 +16,9 @@ for GPUNUM in 1 4
do
do
MODLE_PATH
=
"facebook/opt-
${
MODEL
}
"
MODLE_PATH
=
"facebook/opt-
${
MODEL
}
"
torchrun
\
colossalai run
\
--standalone
\
--nproc_per_node
${
GPUNUM
}
\
--nproc_per_node
${
GPUNUM
}
\
--master_port
29505
\
opt_benchmark.py
\
opt_benchmark.py
\
--model_name_or_path
${
MODLE_PATH
}
\
--model_name_or_path
${
MODLE_PATH
}
\
--mem_cap
${
MEMCAP
}
\
--mem_cap
${
MEMCAP
}
\
...
...
examples/language/opt/run_demo.sh
View file @
df66741f
...
@@ -30,9 +30,9 @@ WEIGHT_DECAY=0.01
...
@@ -30,9 +30,9 @@ WEIGHT_DECAY=0.01
WARMUP_RATIO
=
0.1
WARMUP_RATIO
=
0.1
# run the script for demo
# run the script for demo
torchrun
\
colossalai run
\
--standalone
\
--nproc_per_node
${
GPUNUM
}
\
--nproc_per_node
${
GPUNUM
}
\
--master_port
29505
\
opt_train_demo.py
\
opt_train_demo.py
\
--model_name_or_path
${
MODEL
}
\
--model_name_or_path
${
MODEL
}
\
--output_path
${
OUTPUT_PATH
}
\
--output_path
${
OUTPUT_PATH
}
\
...
...
examples/language/opt/test_ci.sh
View file @
df66741f
...
@@ -7,9 +7,9 @@ do
...
@@ -7,9 +7,9 @@ do
for
GPUNUM
in
1 4
for
GPUNUM
in
1 4
do
do
torchrun
\
colossalai run
\
--standalone
\
--nproc_per_node
${
GPUNUM
}
\
--nproc_per_node
${
GPUNUM
}
\
--master_port
29505
\
opt_benchmark.py
\
opt_benchmark.py
\
--model_name_or_path
"facebook/opt-125m"
\
--model_name_or_path
"facebook/opt-125m"
\
--plugin
${
PLUGIN
}
\
--plugin
${
PLUGIN
}
\
...
...
examples/language/palm/run.sh
View file @
df66741f
...
@@ -8,6 +8,6 @@ export PLACEMENT='cpu'
...
@@ -8,6 +8,6 @@ export PLACEMENT='cpu'
export
USE_SHARD_INIT
=
False
export
USE_SHARD_INIT
=
False
export
BATCH_SIZE
=
1
export
BATCH_SIZE
=
1
env
OMP_NUM_THREADS
=
12
torchrun
--standalone
--nproc_per_node
=
${
GPUNUM
}
--master_port
2950
1
train.py
\
env
OMP_NUM_THREADS
=
12
colossalai run
--nproc_per_node
${
GPUNUM
}
--master_port
2950
5
train.py
\
--dummy_data
=
True
--tp_degree
=
${
TPDEGREE
}
--batch_size
=
${
BATCH_SIZE
}
--plugin
=
'gemini'
\
--dummy_data
=
True
--tp_degree
=
${
TPDEGREE
}
--batch_size
=
${
BATCH_SIZE
}
--plugin
=
'gemini'
\
--placement
${
PLACEMENT
}
--shardinit
${
USE_SHARD_INIT
}
--distplan
${
DISTPAN
}
2>&1 |
tee
run.log
--placement
${
PLACEMENT
}
--shardinit
${
USE_SHARD_INIT
}
--distplan
${
DISTPAN
}
2>&1 |
tee
run.log
examples/language/palm/test_ci.sh
View file @
df66741f
...
@@ -4,6 +4,6 @@ for BATCH_SIZE in 2
...
@@ -4,6 +4,6 @@ for BATCH_SIZE in 2
do
do
for
GPUNUM
in
1 4
for
GPUNUM
in
1 4
do
do
env
OMP_NUM_THREADS
=
12
torchrun
--standalone
--nproc_per_node
=
${
GPUNUM
}
--
standalone
train.py
--dummy_data
=
True
--batch_size
=
${
BATCH_SIZE
}
--plugin
=
'gemini'
2>&1 |
tee
run.log
env
OMP_NUM_THREADS
=
12
colossalai run
--nproc_per_node
${
GPUNUM
}
--
master_port
29505
train.py
--dummy_data
=
True
--batch_size
=
${
BATCH_SIZE
}
--plugin
=
'gemini'
2>&1 |
tee
run.log
done
done
done
done
examples/language/palm/train.py
View file @
df66741f
import
argparse
import
gzip
import
gzip
from
contextlib
import
nullcontext
from
contextlib
import
nullcontext
from
functools
import
partial
from
functools
import
partial
...
@@ -33,7 +34,7 @@ SEQ_LEN = 1024
...
@@ -33,7 +34,7 @@ SEQ_LEN = 1024
def
parse_args
():
def
parse_args
():
parser
=
colossalai
.
get_default_p
arser
()
parser
=
argparse
.
ArgumentP
arser
()
parser
.
add_argument
(
parser
.
add_argument
(
"--distplan"
,
"--distplan"
,
type
=
str
,
type
=
str
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment