During training,train_loss suddenly increase meanwhile val_dice decrease to 0 #1623
Unanswered
overwhelmedyy
asked this question in
Q&A
Replies: 1 comment 1 reply
-
Hi @yy4551, |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I was using the jupyter notebook tutorial code " spleen_segmentation_3d_lightning.ipynb",work on a pancrease dataset given by tutor.The changes I made is learning rate from 1e-4 to 8e-4, max_epoch from 600 to 100, and I changed CacheDataset to PersistentDataset. Other than these I believe the training process is keep unchanged.
Below is the code pieces start from setting PersistentDataset
100 epoch is very little and the learning rate is quite large,basically my intention is to check if the code is functioning.
However,during the training process when the Dice Metric pretty much hit the best score,the train_loss suddenly oscillating then going way up:
and the val_metric drops to zero :
and val_loss is like this:
this is reproducible and I' ve seen it a few times,seems like larger the learning rate,earlier the loss increasing (and dice dropping) will happen.But at the first 50 epochs it works pretty well to me.So confusing...
I'm really a beginner and didnt find similar problem raised by others.Anyone please give me a little idea or suggestion,I will be more than grateful.
So many thanks!
Beta Was this translation helpful? Give feedback.
All reactions