Commit e1f1aa06 authored by Mostofa Patwary's avatar Mostofa Patwary
Browse files

removed self_attention_pos_emb

parent 5b7736ca
......@@ -741,7 +741,6 @@ class ParallelTransformerLayer(MegatronModule):
# Layer norm at the beginning of the transformer layer.
layernorm_output = self.input_layernorm(hidden_states)
# Self attention.
self_attention_pos_emb = None
attention_output, attention_bias = \
self.self_attention(
layernorm_output,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment