-
Notifications
You must be signed in to change notification settings - Fork 242
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Poor result of MEMIT over llama3.2 3B #450
Comments
It's strange, I've been working on fixing MEMIT multi-gpu errors recently, and I'll debug this issue as well. |
Thank you for your response. To reduce GPU memory usage, I have adjusted the maximum token sequence length in the layer stat file to 512. I'm not sure if this change has negatively affected the results. Looking forward your response after debug. |
Hi, my download speed for wikipedia/20200501.en is very slow. Do you have a precomputed cache of Wikipedia (with the file extension .npz) so that I can debug and help you identify the problem faster? |
Sure. Upload the file will cost a few hours. |
I upload my npc file in the following link: By the way, I guess you can use hf-mirror to load the wikipedia/20200501.en. |
I'm glad to have received your npz file. We are currently debugging the issue, and we will inform you promptly once we have results. |
Thank you for the wonderful jobs and provide the hyper-meters of MEMIT over llama3.2 3B.
However, my edit accuracy under 200 facts is low that 50%, while the MEMIT over gpt2-xl have 99% accuracy under 2000 facts. Do you have any tips, thank you very much?
The text was updated successfully, but these errors were encountered: