-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
there are lots of bugs in TrainStage1 #33
Comments
Thanks for your interest in our work. There might be some small code typos when we push on github, while you could simply fix them for further usage. |
Hello, I am confused about the inconsistencies between the first training stage and the MLLMSD training stage:
This discrepancy in the number of new tokens causes the MLLMSD model's In the first training stage, the LLama checkpoint is loaded, but in the MLLMSD training stage, the LLava checkpoint is loaded, which is puzzling. Why not directly align LLava with CLIP?" |
@Bilibilee Hi, I'm encountering the same issue regarding the token inconsistency between training stages. Could you share how you resolved this, specifically:
Any insights would be greatly appreciated. |
Thank you for your excellent work, but the open-source code indeed has many minor issues, which makes others hesitant to follow your work.
During the TrainStage1 phase, the issues are as follows:
torchrun --nproc_per_node=8 --master_port=20001 fastchat/train/TrainStage1.py
fastchat directory seemingly doesn't exist,it should betrain/TrainStage1.py
.load_LLaVA_ckpt_v1_1
should beload_LLaVA_ckpt_v1_1_7b
.Could you provide Trainstage1 result checkpoint.
The text was updated successfully, but these errors were encountered: