Improvements to apex.mlp (#804)
* update fused bias relu backward kernel * adding support for not require first layer dgrad * fix bug: wrong layer in requires grad * add infrastructure for optional bias and activation, currently only support no bias and no relu * make bias and relu optional separately * add sigmoid activation option
Showing
This diff is collapsed.
Please register or sign in to comment