Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
transformers
Commits
1306b7d3
Unverified
Commit
1306b7d3
authored
Apr 12, 2023
by
Stas Bekman
Committed by
GitHub
Apr 12, 2023
Browse files
[tests] switch to torchrun (#22712)
parent
d87ef00c
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
4 additions
and
4 deletions
+4
-4
tests/extended/test_trainer_ext.py
tests/extended/test_trainer_ext.py
+1
-1
tests/trainer/test_trainer_distributed.py
tests/trainer/test_trainer_distributed.py
+3
-3
No files found.
tests/extended/test_trainer_ext.py
View file @
1306b7d3
...
...
@@ -366,7 +366,7 @@ class TestTrainerExt(TestCasePlus):
n_gpus_to_use
=
get_gpu_count
()
master_port
=
get_torch_dist_unique_port
()
distributed_args
=
f
"""
-m torch.distributed.
launch
-m torch.distributed.
run
--nproc_per_node=
{
n_gpus_to_use
}
--master_port=
{
master_port
}
{
self
.
examples_dir_str
}
/pytorch/translation/run_translation.py
...
...
tests/trainer/test_trainer_distributed.py
View file @
1306b7d3
...
...
@@ -67,7 +67,7 @@ class TestTrainerDistributedNeuronCore(TestCasePlus):
@
require_torch_neuroncore
def
test_trainer
(
self
):
distributed_args
=
f
"""
-m torch.distributed.
launch
-m torch.distributed.
run
--nproc_per_node=2
--master_port=
{
get_torch_dist_unique_port
()
}
{
self
.
test_file_dir
}
/test_trainer_distributed.py
...
...
@@ -83,7 +83,7 @@ class TestTrainerDistributed(TestCasePlus):
@
require_torch_multi_gpu
def
test_trainer
(
self
):
distributed_args
=
f
"""
-m torch.distributed.
launch
-m torch.distributed.
run
--nproc_per_node=
{
torch
.
cuda
.
device_count
()
}
--master_port=
{
get_torch_dist_unique_port
()
}
{
self
.
test_file_dir
}
/test_trainer_distributed.py
...
...
@@ -98,7 +98,7 @@ class TestTrainerDistributed(TestCasePlus):
if
__name__
==
"__main__"
:
# The script below is meant to be run under torch.distributed, on a machine with multiple GPUs:
#
# PYTHONPATH="src" python -m torch.distributed.
launch
--nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py
# PYTHONPATH="src" python -m torch.distributed.
run
--nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py
parser
=
HfArgumentParser
((
TrainingArguments
,))
training_args
=
parser
.
parse_args_into_dataclasses
()[
0
]
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment