Submitted by Longjumping_Essay498 t3_1003d7w in MachineLearning
currentscurrents t1_j2f996k wrote
Attention maps can be a type of explanation.
It tells you what the model was looking at when it generated a word or identified an image, but it doesn't tell you why it looked at those bits or why it made the decision it did. You can get some useful information by looking at it, but not everything you need to explain the model.
Longjumping_Essay498 OP t1_j2f9r9s wrote
Let say if for some example we dig into these attention maps, and find some perspective of some head for attending words. For an example in gpt some head focus on parts of speech. Will it always reliably do it for all example? What do you think. Can we manually evaluate and categorize the learnings??
IntelArtiGen t1_j2fcirf wrote
You can just say "the network evaluated that it needed to give more attention to these parts to perform the task". You can speculate why but you can't be sure.
currentscurrents t1_j2fduvv wrote
You can get some information this way, but not everything you would want to know. You can try it yourself with BertViz.
The information you do get can be useful though. For example in image processing, you can use the attention map from an object classifier to see where the object is in the image.
Viewing a single comment thread. View all comments