Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
6defe188
Commit
6defe188
authored
Oct 05, 2022
by
Eric Harper
Committed by
Jared Casper
Oct 05, 2022
Browse files
Setter for pipeline parallel split rank, remove print
parent
fb8c09eb
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
6 additions
and
7 deletions
+6
-7
megatron/core/parallel_state.py
megatron/core/parallel_state.py
+6
-0
megatron/core/tensor_parallel/random.py
megatron/core/tensor_parallel/random.py
+0
-7
No files found.
megatron/core/parallel_state.py
View file @
6defe188
...
@@ -296,6 +296,12 @@ def set_pipeline_model_parallel_rank(rank):
...
@@ -296,6 +296,12 @@ def set_pipeline_model_parallel_rank(rank):
_MPU_PIPELINE_MODEL_PARALLEL_RANK
=
rank
_MPU_PIPELINE_MODEL_PARALLEL_RANK
=
rank
def
set_pipeline_model_parallel_split_rank
(
rank
):
"""Set pipeline model parallel split rank."""
global
_MPU_PIPELINE_MODEL_PARALLEL_SPLIT_RANK
_MPU_PIPELINE_MODEL_PARALLEL_SPLIT_RANK
=
rank
def
get_tensor_model_parallel_rank
():
def
get_tensor_model_parallel_rank
():
"""Return my rank for the tensor model parallel group."""
"""Return my rank for the tensor model parallel group."""
global
_MPU_TENSOR_MODEL_PARALLEL_RANK
global
_MPU_TENSOR_MODEL_PARALLEL_RANK
...
...
megatron/core/tensor_parallel/random.py
View file @
6defe188
...
@@ -163,13 +163,6 @@ def model_parallel_cuda_manual_seed(seed):
...
@@ -163,13 +163,6 @@ def model_parallel_cuda_manual_seed(seed):
# Data parallel gets the original seed.
# Data parallel gets the original seed.
data_parallel_seed
=
seed
data_parallel_seed
=
seed
if
torch
.
distributed
.
get_rank
()
==
0
:
print
(
'> initializing model parallel cuda seeds on global rank {}, '
'model parallel rank {}, and data parallel rank {} with '
'model parallel seed: {} and data parallel seed: {}'
.
format
(
torch
.
distributed
.
get_rank
(),
get_tensor_model_parallel_rank
(),
get_data_parallel_rank
(),
tensor_model_parallel_seed
,
data_parallel_seed
),
flush
=
True
)
_CUDA_RNG_STATE_TRACKER
.
reset
()
_CUDA_RNG_STATE_TRACKER
.
reset
()
# Set the default state.
# Set the default state.
torch
.
cuda
.
manual_seed
(
data_parallel_seed
)
torch
.
cuda
.
manual_seed
(
data_parallel_seed
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment