Back to Search Start Over

ProtoSound: A Personalized and Scalable Sound Recognition System for Deaf and Hard-of-Hearing Users

Authors :
Jain, Dhruv
Nguyen, Khoa Huynh Anh
Goodman, Steven
Grossman-Kahn, Rachel
Ngo, Hung
Kusupati, Aditya
Du, Ruofei
Olwal, Alex
Findlater, Leah
Froehlich, Jon E.
Publication Year :
2022

Abstract

Recent advances have enabled automatic sound recognition systems for deaf and hard of hearing (DHH) users on mobile devices. However, these tools use pre-trained, generic sound recognition models, which do not meet the diverse needs of DHH users. We introduce ProtoSound, an interactive system for customizing sound recognition models by recording a few examples, thereby enabling personalized and fine-grained categories. ProtoSound is motivated by prior work examining sound awareness needs of DHH people and by a survey we conducted with 472 DHH participants. To evaluate ProtoSound, we characterized performance on two real-world sound datasets, showing significant improvement over state-of-the-art (e.g., +9.7% accuracy on the first dataset). We then deployed ProtoSound's end-user training and real-time recognition through a mobile application and recruited 19 hearing participants who listened to the real-world sounds and rated the accuracy across 56 locations (e.g., homes, restaurants, parks). Results show that ProtoSound personalized the model on-device in real-time and accurately learned sounds across diverse acoustic contexts. We close by discussing open challenges in personalizable sound recognition, including the need for better recording interfaces and algorithmic improvements.<br />Comment: Published at the ACM CHI Conference on Human Factors in Computing Systems (CHI) 2022

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2202.11134
Document Type :
Working Paper
Full Text :
https://doi.org/10.1145/3491102.3502020