Submitted by Blutorangensaft t3_10ltyki in MachineLearning
Blutorangensaft OP t1_j5zu9ti wrote
Reply to comment by jackilion in [D] Quantitative measure for smoothness of NLP autoencoder latent space by Blutorangensaft
Thank you for your answer. If a paper on diffusion models pops into your mind that uses this method, feel free to post it.
How would you derive a quantitative evaluation from t-SNE? I thought it's mostly used for visualisation. I'm looking to compute some kind of score from the interpolation.
jackilion t1_j634fkx wrote
What's the point of this score?
Blutorangensaft OP t1_j6356ho wrote
Compare different autoencoders in their ability to create valid language in a continuous space. Later, I want to generate sentences in its latent space by using another neural network, and have them decoded to real sentences by the autoencoder. I want the space to be smooth because the second neural net will naturally be using gradient descent, which involves infinitesimal changes. I believe this network will perform better if the changes that happen actually represent meaningful distances between real sentences.
jackilion t1_j63e6ah wrote
There is no reason to assume your latent space will be smooth by itself. I remember a paper for image generation that had techniques for smoothing out the latent space that can be applied during training:
https://arxiv.org/abs/2106.09016
​
It's about GANs, not autoencoders, but maybe you can find some ideas in there.
Blutorangensaft OP t1_j654qyd wrote
Thank you for the reference, it looks very promising. I've heard of ways to smooth the latent space through Lipschitz regularisation, but then got disappointed again when I read "ah well it's just layer normalisation". So many things in ML come in a different appearance and actually mean the same thing once you implement them.
Viewing a single comment thread. View all comments