"...git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "e11e26df93f98104d6ef99c9458059fa787b857a"
TF safetensors reduced mem usage (#24404)
* Slight comment cleanup * Reduce peak mem usage when loading TF-format safetensor weights * Tweak the PyTorch loading code to support lazy loading from safetensors * Pass safe_open objects to the PyTorch loading function * Do GPU transposes for speed * One more tweak to reduce peak usage further * One-line hasattr * Fix bug when there's a shape mismatch * Rename state_dict in the loading code to be clearer * Use TF format everywhere for consistency
Showing
Please register or sign in to comment