-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Low loss but highly disorded output #1
Comments
@KaiWU5 Hi, I’m not entirely sure what you mean by “disordered” output—do you mean that when you use similar data, the results turn out unexpectedly bad? Or are you saying that after fine-tuning, your model performs worse even on Qwen2.5-VL’s original examples? Either way, here are some of my thoughts—hopefully, they address your concerns. LoRA fine-tuning can sometimes lead to overfitting, where the model becomes too specialized to the fine-tuning dataset and loses its ability to generalize. This often happens when the dataset is too narrow, the LoRA rank is too high (e.g., 64 or 128), or the learning rate is too aggressive. Additionally, training for too many epochs can overwrite the model’s pre-trained knowledge, making performance worse instead of better. To improve stability, consider increasing data diversity, lowering the LoRA rank (e.g., 8 or 16), and reducing the learning rate (e.g., 1e-5 instead of 1e-4). Using early stopping can help prevent overfitting, and mixing some of the original pre-training data with fine-tuning data can help maintain the model’s generalization ability. Hope this helps! Looking forward to hearing your thoughts. |
@sandy1990418 Thanks for your reply. My training is not lora fine-tuning but full parameter tunning for one epoch on large dataset like llavanext. I even simultaneously train Qwen2VL and Qwen2.5VL to compare. Observations:
Thoughts: |
After finetuning on qwen2.5vl , I got a very low train and validation loss around 0.67, but the output is disordered and worse than before.
Could you share some experiences about the training qwen2.5vl series
The text was updated successfully, but these errors were encountered: