ASMR: Learning Attribute-based Person Search With Adaptive Semantic Margin Regularizer | Awesome Learning to Hash Add your paper to Learning2Hash

ASMR: Learning Attribute-based Person Search With Adaptive Semantic Margin Regularizer

Boseung Jeong, Jicheol Park, Suha Kwak . 2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021 – 23 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Distance Metric Learning Evaluation ICCV

Attribute-based person search is the task of finding person images that are best matched with a set of text attributes given as query. The main challenge of this task is the large modality gap between attributes and images. To reduce the gap, we present a new loss for learning cross-modal embeddings in the context of attribute-based person search. We regard a set of attributes as a category of people sharing the same traits. In a joint embedding space of the two modalities, our loss pulls images close to their person categories for modality alignment. More importantly, it pushes apart a pair of person categories by a margin determined adaptively by their semantic distance, where the distance metric is learned end-to-end so that the loss considers importance of each attribute when relating person categories. Our loss guided by the adaptive semantic margin leads to more discriminative and semantically well-arranged distributions of person images. As a consequence, it enables a simple embedding model to achieve state-of-the-art records on public benchmarks without bells and whistles.

Similar Work