β-variational autoencoders for learning invertible local image descriptors

Abstract

In this paper, we propose an efficient method for learning a local image descriptor and its inversion function using a modified version of a variational autoencoder (VAE) - a β-VAE. We examine different values of β in the loss function of the β-VAE to find an optimal balance between incentivising the similarities between input patches to be preserved in latent space, and ensuring good reconstruction of the patches from their encodings in latent space. Our proposed descriptor demonstrates patch retrieval comparable to the reference autoencoder-based local image descriptor, and also shows improved reconstruction of patches from their encodings.

Publication
IMAGE PROCESSING & COMMUNICATIONS