Submitted by abc220022 t3_100y331 in MachineLearning
kdqg t1_j2oo4rl wrote
Reply to comment by lukeiy in [D] What are good ways of incorporating non-sequential context into a transformer model? by abc220022
Also have a look at the slot attention mechanism, which does something similar but arguably more elegantly
Viewing a single comment thread. View all comments