Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
Megatron-LM
Commits
7ac342b7
Commit
7ac342b7
authored
Mar 09, 2022
by
Lawrence McAfee
Browse files
removed old (pre-gbuf-padding) get_model_grad_buffer_dp_views().
parent
0528bf77
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
0 additions
and
31 deletions
+0
-31
megatron/optimizer/distrib_optimizer.py
megatron/optimizer/distrib_optimizer.py
+0
-31
No files found.
megatron/optimizer/distrib_optimizer.py
View file @
7ac342b7
...
...
@@ -309,37 +309,6 @@ class DistributedOptimizer(MixedPrecisionOptimizer):
_zero_grad_group_helper
(
model_params
,
set_to_none
=
False
)
# _zero_grad_group_helper(params, set_to_none = False)
# def get_model_grad_buffer_dp_views(self):
# # >>>
# # ** only contiguous grad buffer supported, for now [ TEMPORARY ] **
# args = get_args()
# assert args.use_contiguous_buffers_in_local_ddp
# # <<<
# # Grad buffer views.
# gbuf_view_items = []
# for model_index, model in enumerate(self.models):
# for dtype, gbuf_shard in self.model_gbuf_shards[model_index].items():
# world_shards = gbuf_shard["world_all"]
# gbuf = model._grad_buffers[dtype].data
# gbuf_views = [ gbuf[s.start:s.end] for s in world_shards ]
# gbuf_view_items.append((model_index, dtype, gbuf_views))
# # pax(0, {
# # "world_shards" : world_shards,
# # "gbuf_views" : gbuf_views,
# # })
# pax(0, {
# "gbuf_view_items" : gbuf_view_items,
# **{
# "views / %d" % i : item[2]
# for i, item in enumerate(gbuf_view_items)
# },
# })
# return gbuf_view_items
def
get_model_grad_buffer_dp_views
(
self
):
data_parallel_world_size
=
mpu
.
get_data_parallel_world_size
()
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment