Training Loss Analysis
![]() |
![]() |
![]() |
---|
- D performs too well (low D loss).
- G can’t keep up (high G loss).
- The generator still struggles to catch up (G loss is higher than D loss).
- Possibly a much better result (G loss is lower than D's).
- However, the trend suggests a role reversal, which is not desirable.
(alternative - decreese D's learning rate)
![]() |
![]() |
---|
The plot flips here, indicating the G started to fail. I think that best performance was likely around 30 iterations (see second plot).
Training the D too often hinders the G’s laearning. Training it too rarely weakens its ability to distinguish. The best results come from balancing the learning pace of both networks.
Symbol | Meaning |
---|---|
🟥 Red line | Discriminator loss |
🟦 Blue line | Generator loss |
D | Discriminator |
G | Generator |
📊 | Average loss per epoch |
Generated numbers by the trained model (0-9):
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
---|
between class 1 and 9: