Back to Search Start Over

Multimodal Clustering Networks for Self-supervised Learning from Unlabeled Videos

Authors :
Chen, Brian
Rouditchenko, Andrew
Duarte, Kevin
Kuehne, Hilde
Thomas, Samuel
Boggust, Angie
Panda, Rameswar
Kingsbury, Brian
Feris, Rogerio
Harwath, David
Glass, James
Picheny, Michael
Chang, Shih-Fu
Source :
Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 8012-8021
Publication Year :
2021

Abstract

Multimodal self-supervised learning is getting more and more attention as it allows not only to train large networks without human supervision but also to search and retrieve data across various modalities. In this context, this paper proposes a self-supervised training framework that learns a common multimodal embedding space that, in addition to sharing representations across different modalities, enforces a grouping of semantically similar instances. To this end, we extend the concept of instance-level contrastive learning with a multimodal clustering step in the training pipeline to capture semantic similarities across modalities. The resulting embedding space enables retrieval of samples across all modalities, even from unseen datasets and different domains. To evaluate our approach, we train our model on the HowTo100M dataset and evaluate its zero-shot retrieval capabilities in two challenging domains, namely text-to-video retrieval, and temporal action localization, showing state-of-the-art results on four different datasets.<br />Comment: To be presented at ICCV 2021

Details

Database :
arXiv
Journal :
Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 8012-8021
Publication Type :
Report
Accession number :
edsarx.2104.12671
Document Type :
Working Paper