Submitted by Apprehensive_Air8919 t3_11dfgfm in deeplearning
Im currently working with the transformer architecture and doing depth estimation. My dataset is 6700 images of dimensions 3x256x256. I've run into a wierd thing. My validation loss suddenly falls alot around epoch 30-40 while my training loss barely does. I cant seem to find out why it is happening. Hope you can help me! I use Adam with lr=0.000001
​
The code for the vision transformer is here.
yannbouteiller t1_ja8cd7n wrote
That is pretty strange indeed. Perhaps this would be a magical effect of dropout ?