Skip to Main content Skip to Navigation
Conference papers

Spreading vectors for similarity search

Abstract : Discretizing multi-dimensional data distributions is a fundamental step of modern indexing methods. State-of-the-art techniques learn parameters of quantizers on training data for optimal performance, thus adapting quantizers to the data. In this work, we propose to reverse this paradigm and adapt the data to the quantizer: we train a neural net which last layer forms a fixed parameter-free quantizer, such as pre-defined points of a hyper-sphere. As a proxy objective, we design and train a neural network that favors uniformity in the spherical latent space, while preserving the neighborhood structure after the mapping. We propose a new regularizer derived from the Kozachenko–Leonenko differential entropy estimator to enforce uniformity and combine it with a locality-aware triplet loss. Experiments show that our end-to-end approach outperforms most learned quantization methods, and is competitive with the state of the art on widely adopted benchmarks. Furthermore, we show that training without the quantization step results in almost no difference in accuracy, but yields a generic catalyzer that can be applied with any subsequent quantizer. The code is available online.
Complete list of metadatas
Contributor : Alexandre Sablayrolles <>
Submitted on : Wednesday, September 4, 2019 - 5:46:01 PM
Last modification on : Thursday, November 19, 2020 - 1:01:31 PM

Links full text


  • HAL Id : hal-02278905, version 1
  • ARXIV : 1806.03198



Alexandre Sablayrolles, Matthijs Douze, Cordelia Schmid, Hervé Jégou. Spreading vectors for similarity search. ICLR 2019 - 7th International Conference on Learning Representations, May 2019, New Orleans, United States. pp.1-13. ⟨hal-02278905⟩



Record views