-
Notifications
You must be signed in to change notification settings - Fork 59
feat: add CE-U loss #72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Hi, thank you very much! We'll try to get this in soon. |
@molereddy Sure. Fixed! |
Thank you. Can you run your method and populate the relevant tables in |
Also, please link your paper in one of the files. In the future, we plan to update documentation with all links in one place, for now, make your link easily available. |
Sure. I found the result is different from unleaning using the old
https://github.com/locuslab/tofu code base, even when my hyperparameter is the same. Some
guess:
1. https://huggingface.co/locuslab/tofu_ft_llama2-7b and
https://huggingface.co/open-unlearning/tofu_Llama-2-7b-chat-hf_full are not
the same model.
2. Different handling of the last mini-batch of each epoch between this repository and the old https://github.com/locuslab/tofu
…On Wed, Mar 19, 2025 at 11:02 PM Anmol Mekala ***@***.***> wrote:
Thank you. Can you run your method and populate the relevant tables in
docs/results.md (for reproducibility, not comparison)?
—
Reply to this email directly, view it on GitHub
<#72 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAES3ORLVCPO4NZSWEU7H4T2VJKYDAVCNFSM6AAAAABZEIOCDGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDOMZZGI4TIMRTHA>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
[image: molereddy]*molereddy* left a comment (locuslab/open-unlearning#72)
<#72 (comment)>
Thank you. Can you run your method and populate the relevant tables in
docs/results.md (for reproducibility, not comparison)?
—
Reply to this email directly, view it on GitHub
<#72 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAES3ORLVCPO4NZSWEU7H4T2VJKYDAVCNFSM6AAAAABZEIOCDGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDOMZZGI4TIMRTHA>
.
You are receiving this because you authored the thread.Message ID:
***@***.***>
|
Yes, these models are different and with some other standardizations we've done to the code, results are expected to be different. We are planning to create a leaderboard with best hyperparams in the future but are currently unsure how to be fair to hp tuning across methods. |
I have two questions:
|
|
But, 1e-5 learning rate / 10 epochs are way too high for CE-U, always resulting in 0.0 model utility. Do you really want me to add them to Also I found 4e-5 learning rate / 5 epochs can produce reasonable results with https://github.com/locuslab/tofu, but not with this repository. Are there extra gradient clipping or regularization applied in https://github.com/locuslab/tofu? |
I see the issue. Let's not add that to results.md, then. We definitely have to change the way results are reported. For now, save your results with your hyperparams so that they can be added later. |
2e-6 learning rate / 9 epochs can produce reasonable results with this repository but still different from https://github.com/locuslab/tofu . I really need to figure out why. 2e-6 learning rate / 9 epochs against
|
I think the results difference is partially due to different fine-tuned models: I tried to run CE-U with hyperparameter 2e-6 learning rate / 9 epochs against
In short:
|
Any way, the benchmark here and locuslab/tofu are not comparable. They should be considered different benchmark. |
Hi @Atry Thank you for taking the time to add your method to our repository. We've added guidelines on contributing to our repository, including how to share your settings, results, and reproducible scripts. Could you review them and update this PR accordingly? This should help others use and reproduce your work in their studies. Quick summary of the contribution guide:
We've provided a template folder to get you started. Let us know if you have any questions! |
@Atry we agree that the numbers in these versions of TOFU are not directly comparable. But you can tune the hyperparameters again to get a new setting for your best results. You can now report those best results on our leaderboard as @Dornavineeth mentioned, instead of just the default repro parameters as before. This includes changing the number of epochs, early stoppage, etc. as you mentioned. Please merge the latest changes into your branch, make the above-mentioned updates and we'll get this in ASAP! |
Following up here: it would be great if you could integrate the latest changes, document your contributions and add results. I understand that for results you have to re-run experiments and tune things again, as results are not comparable. If that is not possible, make the updates and test the code to report on which models/datasets it is working and we can get it in without the results. |
Sure! Will do. But maybe not next week.
…On Sun, Apr 13, 2025 at 9:21 AM Anmol Mekala ***@***.***> wrote:
Following up here: it would be great if you could integrate the latest
changes, document your contributions and add results.
I understand that for results you have to re-run experiments and tune
things again, as results are not comparable. If that is not possible, make
the updates and test the code to report on which models/datasets it is
working and we can get it in without the results.
—
Reply to this email directly, view it on GitHub
<#72 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAES3OQHSGK54QWJNDK76TT2ZKFIVAVCNFSM6AAAAABZEIOCDGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDQMBQGAYTMOBZGU>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
*molereddy* left a comment (locuslab/open-unlearning#72)
<#72 (comment)>
Following up here: it would be great if you could integrate the latest
changes, document your contributions and add results.
I understand that for results you have to re-run experiments and tune
things again, as results are not comparable. If that is not possible, make
the updates and test the code to report on which models/datasets it is
working and we can get it in without the results.
—
Reply to this email directly, view it on GitHub
<#72 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAES3OQHSGK54QWJNDK76TT2ZKFIVAVCNFSM6AAAAABZEIOCDGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDQMBQGAYTMOBZGU>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
What does this PR do?
Before submitting
documentation guidelines.