The reconstruction loss is very low on training images.
But when I test on new data, outputs look distorted.
The model seems confident but wrong.
It feels like it memorized the dataset.
Why does my autoencoder reconstruct training images well but fails on new ones?
Anushrita GhoshBegginer
This happens because the autoencoder has overfit the training distribution. Instead of learning general representations, it memorized pixel-level details of the training images, which do not generalize.
Autoencoders with too much capacity can easily become identity mappings, especially when trained on small or uniform datasets. In this case, low loss simply means the network copied what it saw.
Reducing model size, adding noise, or using variational autoencoders forces the model to learn meaningful latent representations instead of memorization.
Common mistakes:
Using too large a bottleneck
No noise or regularization
Training on limited data
The practical takeaway is that low reconstruction loss does not mean useful representations.