-
Notifications
You must be signed in to change notification settings - Fork 3.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[GNMT v2/Tensorflow] Loss not decreasing when training custom dataset #603
Comments
Hello! There are plenty of potential reasons. It might be too high learning rate value, a problem with data preprocessing, and many other reasons. I suggest looking for some article explaining, how to deal with such problems. Following article seems fine to me: https://blog.slavv.com/37-reasons-why-your-neural-network-is-not-working-4020854bd607 Many articles begin with reducing the problem to the simplest example. For instance, reducing the dataset to just a few examples and checking if the model is able to overfit. If the simplest example works, other elements can be verified. Can you try to follow this guide (or any other you find helpful)? If you will face a problem in some steps, it may be easier to help, knowing, what already works. |
Thanks for the tips! I have done preprocessing on the data already and tweaked the learning rate value as well as other hyperparameters. There's not much difference though. I haven't changed anything with the core architecture of the model so I assumed it would work with other datasets as well.
|
@abbyDC Have you tried to use some guide as I suggested before? Have it helped you to find any issue? Can you check the preprocessed data, for example, if all original data exists in the created dataset? |
@mwawrzos Yup double checked them and there seems to be no problem with the dataset itself. Also tried doing batch inference after training and got okay results despite the loss being like that |
I just wanted to ask the following to help me train a custom model which allows me to translate <src_lang> to english. I have an issue where the loss ranges from 17-200 for a single epoch. It goes up and down drastically. I'm not sure what else I need to tweak.
Steps I've done:
Questions:
The text was updated successfully, but these errors were encountered: