Viewing a single comment thread. View all comments

mgostIH t1_iw4dks1 wrote

Like how a reviewer noted, the "zero shot" part is a bit overclaimed, given that one of the models has to be already trained with these relatives encodings, but the concept of the paper is an interesting phenomenon that points to there being a "true layout" of concepts in latent space that different type of models end up discovering.

30

lynnharry t1_iwa5pha wrote

From my understanding, the authors meant zero-shot communication (in the title) or stitching (in the text), where two NN components trained in different setups can be stitched together without further finetuning. This is just one useful application of the shared relative representation proposed in the paper.

1