Commit 0f6bfdeb authored by Sandeep Subramanian's avatar Sandeep Subramanian Committed by Jimmy Zhang
Browse files

Fix output layer init func when untie embeddings is True

parent a9f6569d
...@@ -416,7 +416,7 @@ class TransformerLanguageModel(MegatronModule): ...@@ -416,7 +416,7 @@ class TransformerLanguageModel(MegatronModule):
args.hidden_size, args.hidden_size,
args.padded_vocab_size, args.padded_vocab_size,
bias=False, # Setting bias to False always to keep it consistent with embedding tying that also does not have a bias. bias=False, # Setting bias to False always to keep it consistent with embedding tying that also does not have a bias.
init_method=output_layer_init_method) init_method=self.init_method)
self._output_layer_key = 'output_layer' self._output_layer_key = 'output_layer'
def set_input_tensor(self, input_tensor): def set_input_tensor(self, input_tensor):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment