uristmcderp t1_j8db0gw wrote
Reply to comment by diviludicrum in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
The whole assessing its own success is the bottleneck for most interesting problems. You can't have a feedback loop unless it can accurately evaluate if it's doing better or worse. This isn't a trivial problem either, since humans aren't all that great at using absolute metrics to describe quality, once past a minimum threshold.
ksatriamelayu t1_j8ebpx4 wrote
Do people use things like evolutionary fitness + changing environments to describe those quality? Seems dynamic environment might be the answer?
Oat-is-the-Best t1_j8ef5x0 wrote
How do you calculate your fitness? That has the same problem of a model not being able to assess its own success
Viewing a single comment thread. View all comments