Ultra-high Dimensional Sparse Representations With Binarization For Efficient Text Retrieval | Awesome Learning to Hash Add your paper to Learning2Hash

Ultra-high Dimensional Sparse Representations With Binarization For Efficient Text Retrieval

Kyoung-Rok Jang, Junmo Kang, Giwon Hong, Sung-Hyon Myaeng, Joohee Park, Taewon Yoon, Heecheol Seo . Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021 – 10 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
EMNLP Efficiency Hybrid ANN Methods Re-Ranking Text Retrieval

The semantic matching capabilities of neural information retrieval can ameliorate synonymy and polysemy problems of symbolic approaches. However, neural models’ dense representations are more suitable for re-ranking, due to their inefficiency. Sparse representations, either in symbolic or latent form, are more efficient with an inverted index. Taking the merits of the sparse and dense representations, we propose an ultra-high dimensional (UHD) representation scheme equipped with directly controllable sparsity. UHD’s large capacity and minimal noise and interference among the dimensions allow for binarized representations, which are highly efficient for storage and search. Also proposed is a bucketing method, where the embeddings from multiple layers of BERT are selected/merged to represent diverse linguistic aspects. We test our models with MS MARCO and TREC CAR, showing that our models outperforms other sparse models

Similar Work