Submitted by 51616 t3_yt6slt in MachineLearning
mgostIH t1_iw4dks1 wrote
Like how a reviewer noted, the "zero shot" part is a bit overclaimed, given that one of the models has to be already trained with these relatives encodings, but the concept of the paper is an interesting phenomenon that points to there being a "true layout" of concepts in latent space that different type of models end up discovering.
lynnharry t1_iwa5pha wrote
From my understanding, the authors meant zero-shot communication (in the title) or stitching (in the text), where two NN components trained in different setups can be stitched together without further finetuning. This is just one useful application of the shared relative representation proposed in the paper.
Viewing a single comment thread. View all comments