029187 OP t1_iqth0xk wrote
Reply to comment by pia322 in [D] - Why do Attention layers work so well? Don't weights in DNNs already tell the network how much weight/attention to give to a specific input? (High weight = lots of attention, low weight = little attention) by 029187
I'm kinda scared by the idea that we get all the way to strong AI and still don't understand why it works.
Viewing a single comment thread. View all comments