Back to Search Start Over

Towards Social AI: A Survey on Understanding Social Interactions

Authors :
Lee, Sangmin
Li, Minzhi
Lai, Bolin
Jia, Wenqi
Ryan, Fiona
Cao, Xu
Kara, Ozgur
Boote, Bikram
Shi, Weiyan
Yang, Diyi
Rehg, James M.
Publication Year :
2024

Abstract

Social interactions form the foundation of human societies. Artificial intelligence has made significant progress in certain areas, but enabling machines to seamlessly understand social interactions remains an open challenge. It is important to address this gap by endowing machines with social capabilities. We identify three key capabilities needed for effective social understanding: 1) understanding multimodal social cues, 2) understanding multi-party dynamics, and 3) understanding beliefs. Building upon these foundations, we classify and review existing machine learning works on social understanding from the perspectives of verbal, non-verbal, and multimodal social cues. The verbal branch focuses on understanding linguistic signals such as speaker intent, dialogue sentiment, and commonsense reasoning. The non-verbal branch addresses techniques for perceiving social meaning from visual behaviors such as body gestures, gaze patterns, and facial expressions. The multimodal branch covers approaches that integrate verbal and non-verbal multimodal cues to holistically interpret social interactions such as recognizing emotions, conversational dynamics, and social situations. By reviewing the scope and limitations of current approaches and benchmarks, we aim to clarify the development trajectory and illuminate the path towards more comprehensive intelligence for social understanding. We hope this survey will spur further research interest and insights into this area.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2409.15316
Document Type :
Working Paper