Hit 325
Subject [MMM 2021] Robust Multispectral Pedestrian Detection via Uncertain-Aware Cross-Modal Learning (by Sungjune Park) is accepted in MMM 2021
Date 2020-10-16
Title: Robust Multispectral Pedestrian Detection via Uncertain-Aware Cross-Modal Learning
Authors: Sungjune Park, Jung Uk Kim, Yeon Gyun Kim, Sang-Keun Moon and Yong Man Ro
With the development of deep neural networks, multispectral pedestrian detection has been received a great attention by exploiting complementary properties of multiple modalities (e.g., color-visible and thermal modalities). Previous works usually rely on network prediction scores in combining complementary modal information. However, it is widely known that deep neural networks often show overconfident problem which results in limited performance. In this paper, we propose a novel uncertainty-aware cross-modal learning to alleviate the aforementioned problem in multispectral pedestrian detection. First, we extract
object region uncertainty which represents the reliability of object region features in multiple modalities. Then, we combine each modal object region feature considering object region uncertainty. Second, we guide the classifier of detection framework with soft target labels to be aware of the level of object region uncertainty in multiple modalities. To verify the effectiveness of the proposed methods, we conduct extensive experiments with various detection frameworks on two public datasets (i.e., KAIST Multispectral Pedestrian Dataset and CVC-14).