-
Notifications
You must be signed in to change notification settings - Fork 429
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FSDP + QDoRA Support #159
Comments
I have done some FSDP to train full parameters mistral 7b maybe its useful for you |
Thanks, good job! I want to finetune Llama-3-70B with 8 A6000 48G, which are not enough for training full parameters. FSDP + QDoRA is the method I have found to be feasible and probably the most effective. |
yes it should work
|
I've tried this command and encountered the issue described in huggingface/peft#1674 Currently, I am following the official example provided in PEFT for further troubleshooting: https://github.com/huggingface/peft/blob/main/examples/sft/run_peft_qlora_fsdp.sh |
FSDP + QDoRA for Zephyr 141b would be really good |
AFAIK, FSDP+QDoRA is not supported feature in HF official releases like transformers, peft, ... |
Hi the team, great work!
QDoRA seems to be better than QLoRA, refer to Efficient finetuning of Llama 3 with FSDP QDoRA
I wonder whether there will be demo / example about FSDP + QDoRA during finetuning?
Thanks!
The text was updated successfully, but these errors were encountered: