Replies: 1 comment 1 reply
-
Hello @ghsanti Thank you for pointing the issue out. Can you please set I suspect division by 0 or near-0 numbers are the culprit here. Best Regards, |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I've just read this issue in tf-keras, and ended up testing a simple convnet with mnist to understand the problem.
I suspect the reason may be due to BatchNormalization, but don't know enough to understand why.
Here is a simple image comparing the mean absolute error for
model(X, training=True)
andmodel.predict(X)
which disappears istraining=False
:But why would that make the outputs different, between
model(X, training=False)
(ormodel.predict
), andmodel(X, training=True)
I don't have a colab link, but the snippet I used is:
Is there a simple explanation? I should read the code but maybe someone knows already.
Anyone that could help, please?
Beta Was this translation helpful? Give feedback.
All reactions