-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Jarvis data #54
Comments
Hi @Nokimann Which property/task did you try to reproduce, and how much difference did you find? |
I got 0.029 MAE (~600 epochs) and 0.002 MAE (~300 epochs) of U0 data for the unnormalized data and the normalized data, respectively. std ~10 of the QM9 data, which means one order high. So, it is pretty reasonable. |
You are right. We didn't multiply the std with corresponding MAEs but we should have. For some properties (with std <1) in QM9 ALIGNN model performance becomes better than reported now but for models such as U0 it becomes worse. We are working on an erratum right now and will update the arXiv preprint as well as the README file soon. The performance on JARVIS-DFT and MP dataset remains intact. We note that if we train for 1000 or so epochs, we can get U0 MAE upto 0.014 eV. For reference, the std for QM9 tasks: Thanks @Nokimann for catching this mistake. |
It seems like this impacts one of the main claims in the paper, but unfortunately there has been no update in the last month. The paper, readme, and arXiv still show the wrong results. Would you have any update on the progress of fixing this? |
@klicperajo We have updated the README file now with the 1000 epoch run and multiplication of MAEs with corresponding standard deviations. On a related point ( usnistgov/jarvis#202 (comment) ), I see using different package datasets such as from PyG or DGL might give you different graphs. Hence, we choose to learn directly from xyz/POSCAR files. Our goal is that after we train a model, a user can feed a POSCAR/xyz file to get predictions using pretrained.py which might be possible but not too easy using PyG/DGL based datasets.
|
That is great to see, thank you! Any progress on arXiv and npj? I was not suggesting to use the PyG or DGL datasets, but rather to provide the non-standardized data (in eV or similar). I have seen this mistake of reporting standardized error instead of real units several times now. We should make sure that the straightforward way of evaluation is the correct one. Otherwise this error will be repeated again. |
@gasteigerjo Author correction is now available at: https://www.nature.com/articles/s41524-022-00913-5 |
Thank you for making the effort to amend these numbers! |
Thank you for your work on the efficient way to predict the ML method for the molecular system.
But, I couldn't reproduce the paper.
I found that Jarvis summarizes QM9 datasets with normalization, and I issued in jarvis.
I tested ALIGNN, but cannot reproduce it for the unnormalized QM9 datasets.
Only the normalized QM9 dataset provided by Jarvis works to reproduce prediction values in paper.
The text was updated successfully, but these errors were encountered: