Authors
Di Hu, Rui Qian, Minyue Jiang, Xiao Tan, Shilei Wen, Errui Ding, Weiyao Lin, Dejing Dou
Publication date
2020
Journal
Advances in Neural Information Processing Systems
Volume
33
Pages
10077-10087
Description
Discriminatively localizing sounding objects in cocktail-party, ie, mixed sound scenes, is commonplace for humans, but still challenging for machines. In this paper, we propose a two-stage learning framework to perform self-supervised class-aware sounding object localization. First, we propose to learn robust object representations by aggregating the candidate sound localization results in the single source scenes. Then, class-aware object localization maps are generated in the cocktail-party scenarios by referring the pre-learned object knowledge, and the sounding objects are accordingly selected by matching audio and visual object category distributions, where the audiovisual consistency is viewed as the self-supervised signal. Experimental results in both realistic and synthesized cocktail-party videos demonstrate that our model is superior in filtering out silent objects and pointing out the location of sounding objects of different classes. Code is available at https://github. com/DTaoo/Discriminative-Sounding-Objects-Localization.
Total citations
202120222023202420444531
Scholar articles
D Hu, R Qian, M Jiang, X Tan, S Wen, E Ding, W Lin… - Advances in Neural Information Processing Systems, 2020