Unverified Commit f8521900 authored by drbh's avatar drbh Committed by GitHub
Browse files

fix: prefer hidden_activation over hidden_act in gemma2 (#2381)

parent 2ca59806
...@@ -265,7 +265,7 @@ class FlashGemma2Attention(torch.nn.Module): ...@@ -265,7 +265,7 @@ class FlashGemma2Attention(torch.nn.Module):
class Gemma2MLP(nn.Module): class Gemma2MLP(nn.Module):
def __init__(self, prefix, config, weights): def __init__(self, prefix, config, weights):
super().__init__() super().__init__()
act = config.hidden_act act = config.hidden_activation
self.act = ( self.act = (
ACT2FN[act] ACT2FN[act]
if "gelu" not in act if "gelu" not in act
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment