"git@developer.sourcefind.cn:OpenDAS/bitsandbytes.git" did not exist on "3fd06fb6206f46b6d18fbb8a512da63832dea98b"
Initial FSDP Support for QLoRA Finetuning (#970)
This PR adds initial FSDP support for training QLoRA models. It enables basic FSDP and CPU Offload support, with low memory training via FSDP.sync_module_states option unsupported.
This PR builds off of #840 commit 8278fca and BNB FSDP by @TimDettmers and @Titus-von-Koeller.
An example of using this PR to finetune QLoRA models with FSDP can be found in the demo repo: AnswerDotAi/fsdp_qlora.
* Minimal changes for fp32 4bit storage from BNB commit 8278fca
* Params4bit with selectable storage dtype
* possible fix for double quantizing linear weight & quant storage dtype
* minor fixes in Params4bit for peft tests
* remove redundant
* add float16
* update test
* Remove float16 quant cast as there are fp32, bf16, & fp16 quant kernels
---------
Co-authored-by:
Kerem Turgutlu <keremturgutlu@gmail.com>
Showing
Please register or sign in to comment