Viewing a single comment thread. View all comments

Dartagnjan OP t1_j108k4y wrote

That is what I have already done. So far, the loss just oscillates but remains high, which leads me to believe that either I am not training in the right way i.e. maybe the difference between the easy and hard training examples is too drastic to bridge. Or my model is just not capable of handing the harder examples.

1

JustOneAvailableName t1_j1096lz wrote

Sounds like you need a higher batch size. What happens on a plateaued model on the hard examples when you take a huge batch size?

2