Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unusual Model Utility Gap: Gradient Difference vs Ascent (Llama2) #45

Open
jeongjin0 opened this issue Dec 2, 2024 · 1 comment
Open

Comments

@jeongjin0
Copy link

jeongjin0 commented Dec 2, 2024

Hi, I noticed that for Llama2(forget10), gradient difference shows much lower model utility (~0.27) than gradient ascent (~0.63) in the leaderboard. This seems unusual since gradient difference is designed to maintain performance on the retain set while unlearning.

Interestingly, in Phi model results, gradient difference shows higher utility than ascent as expected. Could you help explain this significant performance gap in Llama2 implementation?

Thanks!

@zhilif
Copy link
Collaborator

zhilif commented Dec 19, 2024

Apologize for the confusing. Previously there's a bug in the leaderboard csv generation. We have uploaded a new version. Meanwhile, I will rerun evals again to double check. Thanks for your comments!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants