中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Self Correspondence Distillation For End-to-End Weakly-Supervised Semantic Segmentation

文献类型:会议论文

作者Rongtao Xu2,3; Changwei Wang2,3; Jiaxi Sun2,3; Shibiao Xu1; Meng WL(孟维亮)2,3,4; Xiaopeng Zhang2,3
出版日期2023-02
会议日期Feb7-14,2023
会议地点Washington, DC,USA
英文摘要

Efficiently training accurate deep models for weakly supervised
semantic segmentation (WSSS) with image-level labels
is challenging and important. Recently, end-to-end WSSS
methods have become the focus of research due to their high
training efficiency. However, current methods suffer from
insufficient extraction of comprehensive semantic information,
resulting in low-quality pseudo-labels and sub-optimal
solutions for end-to-end WSSS. To this end, we propose a
simple and novel Self Correspondence Distillation (SCD)
method to refine pseudo-labels without introducing external
supervision. Our SCD enables the network to utilize feature
correspondence derived from itself as a distillation target,
which can enhance the network’s feature learning process by
complementing semantic information. In addition, to further
improve the segmentation accuracy, we design a Variationaware
Refine Module to enhance the local consistency of
pseudo-labels by computing pixel-level variation. Finally, we
present an efficient end-to-end Transformer-based framework
(TSCD) via SCD and Variation-aware Refine Module for the
accurate WSSS task. Extensive experiments on the PASCAL
VOC 2012 and MS COCO 2014 datasets demonstrate that our
method significantly outperforms other state-of-the-art methods.
Our code is available at https://github.com/Rongtao-
Xu/RepresentationLearning/tree/main/SCD-AAAI2023.

会议录出版者AAAI
会议录出版地USA
源URL[http://ir.ia.ac.cn/handle/173211/51609]  
专题模式识别国家重点实验室_三维可视计算
多模态人工智能系统全国重点实验室
作者单位1.School of Artificial Intelligence, Beijing University of Posts and Telecommunications
2.NLPR, Institute of Automation, Chinese Academy of Sciences
3.School of Artificial Intelligence, University of Chinese Academy of Sciences
4.Zhejiang Lab
推荐引用方式
GB/T 7714
Rongtao Xu,Changwei Wang,Jiaxi Sun,et al. Self Correspondence Distillation For End-to-End Weakly-Supervised Semantic Segmentation[C]. 见:. Washington, DC,USA. Feb7-14,2023.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。