Fix LLaMA tokenization issue (#531)
* Fix tokenization issue in BaseLM.loglikelihood
* Add a regression script
* Use entire non-continuation length as context
---------
Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
Showing
scripts/regression.py
0 → 100644
Please register or sign in to comment