Submitted by AutoModerator t3_100mjlp in MachineLearning
emmytau t1_j2r7cbc wrote
Is it problematic that my Bart summarization model's training loss drop below validation loss? I could for example stop the training already after 2 epochs. However, it would be nice to train more epochs but maybe it would just require more data - or do you have any training argument suggestions?
See graph of training- and validation loss https://imgur.com/mF7Frfd
Model here: https://huggingface.co/emmyapi/distilbart-podimo-data-eval-2
Viewing a single comment thread. View all comments