"megatron/git@developer.sourcefind.cn:OpenDAS/megatron-lm.git" did not exist on "6728a780dc76ff9684217033d9127f7b1186230b"
Commit befa593d authored by xinliupitt's avatar xinliupitt
Browse files

indent

parent 7e4837c2
...@@ -433,23 +433,23 @@ class TransformerEncoder(tf.keras.layers.Layer): ...@@ -433,23 +433,23 @@ class TransformerEncoder(tf.keras.layers.Layer):
"num_layers": "num_layers":
self._num_layers, self._num_layers,
"num_attention_heads": "num_attention_heads":
self._num_attention_heads, self._num_attention_heads,
"intermediate_size": "intermediate_size":
self._intermediate_size, self._intermediate_size,
"activation": "activation":
self._activation, self._activation,
"dropout_rate": "dropout_rate":
self._dropout_rate, self._dropout_rate,
"attention_dropout_rate": "attention_dropout_rate":
self._attention_dropout_rate, self._attention_dropout_rate,
"use_bias": "use_bias":
self._use_bias, self._use_bias,
"norm_first": "norm_first":
self._norm_first, self._norm_first,
"norm_epsilon": "norm_epsilon":
self._norm_epsilon, self._norm_epsilon,
"intermediate_dropout": "intermediate_dropout":
self._intermediate_dropout self._intermediate_dropout
} }
def call(self, def call(self,
...@@ -548,23 +548,23 @@ class TransformerDecoder(tf.keras.layers.Layer): ...@@ -548,23 +548,23 @@ class TransformerDecoder(tf.keras.layers.Layer):
"num_layers": "num_layers":
self._num_layers, self._num_layers,
"num_attention_heads": "num_attention_heads":
self._num_attention_heads, self._num_attention_heads,
"intermediate_size": "intermediate_size":
self._intermediate_size, self._intermediate_size,
"activation": "activation":
self._activation, self._activation,
"dropout_rate": "dropout_rate":
self._dropout_rate, self._dropout_rate,
"attention_dropout_rate": "attention_dropout_rate":
self._attention_dropout_rate, self._attention_dropout_rate,
"use_bias": "use_bias":
self._use_bias, self._use_bias,
"norm_first": "norm_first":
self._norm_first, self._norm_first,
"norm_epsilon": "norm_epsilon":
self._norm_epsilon, self._norm_epsilon,
"intermediate_dropout": "intermediate_dropout":
self._intermediate_dropout self._intermediate_dropout
} }
def call(self, def call(self,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment