Self-supervised Bernoulli Autoencoders for Semi-supervised Hashing
Abstract
Semantic hashing is a technique to represent high-dimensional data using similarity-preserving binary codes for efficient indexing and search. Recently, variational autoencoders with Bernoulli latent representations achieved remarkable success in learning such codes in supervised and unsupervised scenarios, outperforming traditional methods thanks to their ability to handle the binary constraints architecturally. In this paper, we propose a novel method for supervision (self-supervised) of variational autoencoders where the model uses its own predictions of the label distribution to implement the pairwise objective function. Also, we investigate the robustness of hashing methods based on variational autoencoders to the lack of supervision, focusing on two semi-supervised approaches currently in use. Our experiments on text and image retrieval tasks show that, as expected, both methods can significantly increase the quality of the hash codes as the number of labelled observations increases, but deteriorates when the amount of labelled samples decreases. In this scenario, the proposed self-supervised approach outperforms the classical approaches and yields similar performance in fully-supervised settings.
Más información
| Título según WOS: | Self-supervised Bernoulli Autoencoders for Semi-supervised Hashing |
| Título según SCOPUS: | Self-supervised Bernoulli Autoencoders for Semi-supervised Hashing |
| Título de la Revista: | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
| Volumen: | 12702 |
| Editorial: | Springer Science and Business Media Deutschland GmbH |
| Fecha de publicación: | 2021 |
| Página final: | 268 |
| Idioma: | English |
| DOI: |
10.1007/978-3-030-93420-0_25 |
| Notas: | ISI, SCOPUS |