Unverified Commit 3f57f46a authored by Robert Clark's avatar Robert Clark
Browse files

Add seq-length argument to distribued BERT example



The --seq-length argument was missing from the example distributed BERT
pre-training script which throws an error while attempting to parse the
arguments as the code expects all arguments to be specified and have a
non-None value.
Signed-Off-By: default avatarRobert Clark <roclark@nvidia.com>
parent 90e0a0dd
...@@ -23,6 +23,7 @@ python -m torch.distributed.launch $DISTRIBUTED_ARGS \ ...@@ -23,6 +23,7 @@ python -m torch.distributed.launch $DISTRIBUTED_ARGS \
--num-attention-heads 16 \ --num-attention-heads 16 \
--micro-batch-size 2 \ --micro-batch-size 2 \
--global-batch-size 16 \ --global-batch-size 16 \
--seq-length 512 \
--max-position-embeddings 512 \ --max-position-embeddings 512 \
--train-iters 1000000 \ --train-iters 1000000 \
--save $CHECKPOINT_PATH \ --save $CHECKPOINT_PATH \
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment