[PyTorch] Refactor caching of cumulative sequence lengths (#630)
Do not cache sequence lengths based on layer number Signed-off-by:Tim Moon <tmoon@nvidia.com> Co-authored-by:
Kirthi Shankar Sivamani <ksivamani@nvidia.com>
Showing
Please register or sign in to comment