Open Access
Weakly-Supervised Semantic Segmentation via Self-training
Author(s) -
Hao Cheng,
Chenglin Gu,
Dongwen Zhang
Publication year - 2020
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1487/1/012001
Subject(s) - computer science , segmentation , artificial intelligence , object (grammar) , task (project management) , top down and bottom up design , process (computing) , bridge (graph theory) , image (mathematics) , image segmentation , pattern recognition (psychology) , pixel , machine learning , computer vision , medicine , management , software engineering , economics , operating system
Weakly-supervised semantic segmentation with image tags is a challenging computer vision task. Unlike pixel-level masks, image tags give high level semantic information, without low level appearance information. In this paper, we propose an iteratively self-training framework to bridge this two information, which expand and refine the pseudo-labels with training process going. Initial masks are generated from classification network. In the top-down step, rendered images and its labels as well as spatially weight loss are added to jointly training the model for alleviate the effect of inaccurate object masks. Then in the bottom-up step, an adaptive threshold to the confidence model predictions to keep predicted masks reliable. The top-down and bottom-up steps are conducted iteratively to extract the fine object mask. Experiments on our self-build dataset and GTA5 to CityScapes demonstrate the effectiveness of proposed framework.