Submitted by _atswi_ t3_11aq4qo in MachineLearning
What's the best way to quantify the uncertainty of a trained LLM? I assume the entropy of the model's final probability distribution is a decent measure. Just wanted to know if the NLP community sticks to this measure, or if there's something more specific to language?
Would really appreciate recent references that may have popped up over the past few months (if any). Also if there are any cool & easy to integrate implementations. Thanks!
activatedgeek t1_j9tq7qo wrote
Came across this recently - Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation