- 22 Feb, 2022 1 commit
-
-
Funtowicz Morgan authored
* Add GeLU10 (clipped version of GeLU) to transformers to improve quantization performances. * Add unittests. * Import tensorflow after `is_tf_available` check. * Fix tensorflow wrong function `tf.tensor` to `tf.constant` * style. * use `tf.math.max` * Fix tf tests. * style. * style style style style style style * style style style style style style * Address @sgugger comments. * Fix wrong operator for raising ValueError for ClippedGELUActivation.
-
- 14 Jan, 2022 1 commit
-
-
Joao Gante authored
-
- 07 Dec, 2020 1 commit
-
-
Sylvain Gugger authored
* Add copyright everywhere missing * Style
-
- 30 Oct, 2020 1 commit
-
-
TFUsers authored
* Replace swish with silu * revert nn.silu to nn.swish due to older version * simplify optimized silu conditional and fix format * Update activations.py * Update activations_tf.py * Update modeling_flax_utils.py * Update modeling_openai.py * add swish testcase * add pytorch swish testcase * Add more robust python version check * more formatting fixes Co-authored-by:TFUsers <TFUsers@gmail.com>
-
- 16 Sep, 2020 1 commit
-
-
Julien Plu authored
* Refactoring the activations functions into a common file * Apply style * remove unused import * fix tests * Fix tests.
-