Learning Audio-Visual Correlations From Variational Cross-Modal Generation

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021, 00, pp. 4300-4304
Issue Date:
2021-06-11
Filename Description Size
Learning_Audio-Visual_Correlations_From_Variational_Cross-Modal_Generation.pdfPublished version3.07 MB
Adobe PDF
Full metadata record
People can easily imagine the potential sound while seeing an event. This natural synchronization between audio and visual signals reveals their intrinsic correlations. To this end, we propose to learn the audio-visual correlations from the perspective of cross-modal generation in a self-supervised manner, the learned correlations can be then readily applied in multiple downstream tasks such as the audio-visual cross-modal localization and retrieval. We introduce a novel Variational AutoEncoder (VAE) framework that consists of Multiple encoders and a Shared decoder (MS-VAE) with an additional Wasserstein distance constraint to tackle the problem. Extensive experiments demonstrate that the optimized latent representation of the proposed MS-VAE can effectively learn the audio-visual correlations and can be readily applied in multiple audio-visual downstream tasks to achieve competitive performance even without any given label information during training.
Please use this identifier to cite or link to this item: