Weakly supervised object localization via knowledge distillation based on foreground-background contrast

NEUROCOMPUTING(2024)

引用 0|浏览7
暂无评分
摘要
Weakly supervised object localization (WSOL) is a challenging task that aims to localize objects in images using only image -level labels. Despite the widespread use of WSOL methods based on class activation mapping (CAM), such methods do not consider that the network may overly focus on local regions of the most interesting objects during localization, thus neglecting the overall information. To address this issue, we introduces an additional attention branch for convolutional neural networks (CNNs) that utilizes the attention mechanism of multi -layer perceptron (MLP) to enhance the network's learning of global information and supervise the feature learning of CNNs online through knowledge distillation, thereby improving the localization accuracy of WSOL. Specifically, we designs a new loss function using the generated features to combine with contrastive learning, effectively dividing the foreground and background of the image to provide more accurate pseudolabels for subsequent classification and localization tasks. In the experiments, we tested our method on the CUB -200-2011 dataset and compared it with existing methods. The experimental results show that our method achieves good performance in WSOL tasks.
更多
查看译文
关键词
Weakly supervised object localization,Class activation mapping,Multi-layer perceptron,Knowledge distillation,Pseudo-labels
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要