Back to Search Start Over

Temporal Relation Guided Knowledge Distillation for Continuous Sign Language Recognition

Authors :
XIAO Zheng-ye, LIN Shi-quan, WAN Xiu-an, FANGYu-chun, NI Lan
Source :
Jisuanji kexue, Vol 49, Iss 11, Pp 156-162 (2022)
Publication Year :
2022
Publisher :
Editorial office of Computer Science, 2022.

Abstract

Previous researches in continuous sign language recognition mainly focus on the RGB modality and achieve remarkable performance on real-world and laboratory datasets,but they usually require high computation intensity.On the other hand,the skeleton is a modality with small input data and fast computation speed,but poor at the real-world datasets.This paper proposes a cross-modal knowledge distillation method named temporally related knowledge distillation(TRKD) to alleviate the contradiction between RGB and skeleton modality in performance and calculation speed.TRKD utilizes the RGB modality network as a teacher to guide the skeleton modality network for fast and accurate implementation.We notice that the teacher’s understanding of sign language context is worth learning by student.It proposes to employ the graph convolutional network(GCN) to learn and align the temporally related features of teacher networks and student networks to achieve this goal.Moreover,since the supervised information from the teacher network is not available for traditional loss functions due to the learnable parameters of GCN in the teacher network,we introduce contrastive learning to provide self-supervised information.Multiple ablation experiments on the Phoenix-2014 dataset demonstrate the effectiveness of the proposed method.

Details

Language :
Chinese
ISSN :
1002137X
Volume :
49
Issue :
11
Database :
Directory of Open Access Journals
Journal :
Jisuanji kexue
Publication Type :
Academic Journal
Accession number :
edsdoj.696c8352a5cd40b79b441ab44b77c0f4
Document Type :
article
Full Text :
https://doi.org/10.11896/jsjkx.220600036