Version 2 2021-10-07, 12:12Version 2 2021-10-07, 12:12
Version 1 2021-03-23, 14:46Version 1 2021-03-23, 14:46
conference contribution
posted on 2021-07-09, 00:33authored byD Liang, Z Wei, H Sun, Huiyu Zhou
Training only one deep model for large-scale cross-scenevideo foreground segmentation is challenging due to the off-the-shelf deep learning based segmentor relies on scene-specific structural information. This results in deep mod-els that are scene-biased and evaluations that are scene-influenced.In this paper, we integrate dual modalities(foregrounds’ motion and appearance), and then eliminat-ing features without representativeness of foreground throughattention-module-guided selective-connection structures. It isin an end-to-end training manner and to achieve scene adap-tation in the plug and play style. Experiments indicate theproposed method significantly outperforms the state-of-the-art deep models and background subtraction methods in un-trained scenes – LIMU and LASIESTA. (Codes and datasetwill be available after the anonymous stage.)
History
Author affiliation
School of Informatics
Source
IEEE International Conference on Multimedia and Expo (ICME) 2021
July 5-9, 2021
Version
AM (Accepted Manuscript)
Published in
IEEE International Conference on Multimedia and Expo
Publisher
Institute of Electrical and Electronics Engineers (IEEE)