Generalized Product of Experts for Learning Multimodal Representations in Noisy Environments

Abhinav Joshi, Naman Gupta, Jinang Shah, Binod Bhattarai*, Ashutosh Modi, Danail Stoyanov

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

Abstract

A real-world application or setting involves interaction between different modalities (e.g., video, speech, text). In order to process the multimodal information automatically and use it for an end application, Multimodal Representation Learning (MRL) has emerged as an active area of research in recent times. MRL involves learning reliable and robust representations of information from heterogeneous sources and fusing them. However, in practice, the data acquired from different sources are typically noisy. In some extreme cases, a noise of large magnitude can completely alter the semantics of the data leading to inconsistencies in the parallel multimodal data. In this paper, we propose a novel method for multimodal representation learning in a noisy environment via the generalized product of experts technique. In the proposed method, we train a separate network for each modality to assess the credibility of information coming from that modality, and subsequently, the contribution from each modality is dynamically varied while estimating the joint distribution. We evaluate our method on two challenging benchmarks from two diverse domains: multimodal 3D hand-pose estimation and multimodal surgical video segmentation. We attain state-of-the-art performance on both benchmarks. Our extensive quantitative and qualitative evaluations show the advantages of our method compared to previous approaches.
Original languageEnglish
Title of host publicationICMI '22: Proceedings of the 2022 International Conference on Multimodal Interaction
PublisherACM
Pages83-93
Number of pages11
ISBN (Print)9781450393904
DOIs
Publication statusPublished - 7 Nov 2022
Event2022 International Conference on Multimodal Interaction - Bangalu, India
Duration: 7 Nov 202211 Nov 2022
Conference number: 24
https://icmi.acm.org/2022/#:~:text=The%2024th%20ACM%20International%20Conference,%2C%20interfaces%2C%20and%20system%20development.

Conference

Conference2022 International Conference on Multimodal Interaction
Abbreviated titleICMI
Country/TerritoryIndia
CityBangalu
Period7/11/2211/11/22
Internet address

Bibliographical note

We would like to thank reviewers for their insightful comments. Ashutosh Modi is supported in part by SERB India (Science and
Engineering Board) (SRG/2021/000768).
Binod Bhattarai and Danail Stoyanov are funded by in whole, or in part, by the Wellcome/EPSRC Centre for Interventional and Surgical Sciences (WEISS) (203145/Z/16/Z), Engineering and Physical Sciences Research Council (EPSRC) (EP/P012841/1), the Royal Academy of Engineering Chair in Emerging Technologies scheme, and EndoMapper project by Horizon 2020 FET (GA863146).

Fingerprint

Dive into the research topics of 'Generalized Product of Experts for Learning Multimodal Representations in Noisy Environments'. Together they form a unique fingerprint.

Cite this