1. The Deep Feature Consistent Variational Autoencoder paper proposes using a perceptual loss function to improve over standard VAEs which can generate blurry images. The perceptual loss matches deep feature activations between real and generated images.
2. The model consists of an encoder that outputs mean and variance, a reparameterized decoder, and is trained with a KL divergence loss and perceptual loss on deep feature maps.
3. Experiments show the model generates higher quality images from latent vectors, can manipulate images through vector arithmetic, and the latent space supports facial attribute prediction tasks.
Related topics: