Support `num_attention_heads` != `num_key_value_heads` in Flax Llama Implementation (#29557)
* fix tinyllama flax modelling * rename vars to minimize changes * move * formatting * remove unused var
Showing
Please register or sign in to comment