113 results on '"action units"'
Search Results
2. Characteristics of vocal cues, facial action units, and emotions that distinguish high from low self-protection participants engaged in self-protective response to self-criticizing.
- Author
-
Vráblová, Viktória and Halamová, Júlia
- Subjects
EMOTION-focused therapy ,SINGING instruction ,MULTILEVEL models ,EMOTIONS ,MENTAL health - Abstract
Introduction: Self-protection, also called protective anger or assertive anger, is a key factor in mental health. Thus, far, researchers have focused mainly on the qualitative analysis of self-protection. Methods: Therefore, we investigated facial action units, emotions, and vocal cues in low and high self-protective groups of participants in order to detect any differences. The total sample consisted of 239 participants. Using the Performance factor in the Short version of the Scale for Interpersonal Behavior (lower 15th percentile and upper 15th percentile) we selected 33 high self-protective participants (11 men, 22 women) and 25 low self-protective participants (eight men, 17 women). The self-protective dialogue was recorded using the two-chair technique script from Emotion Focused Therapy. The subsequent analysis was performed using iMotions software (for action units and emotions) and Praat software (for vocal cues of pitch and intensity). We used multilevel models in program R for the statistical analysis. Results: Compared to low self-protective participants, high self-protective participants exhibited more contempt and fear and less surprise and joy. Compared to low self-protective participants, high self-protective participants expressed the action units the following action units less often: Mouth Open (AU25), Smile (AU12), Brow Raise (AU2), Cheek Raise (AU6), Inner Brow Raise (AU1), and more often Brow Furrow (AU4), Chin Raise (AU17), Smirk (AU12), Upper Lip Raise (AU10), and Nose Wrinkle (AU9). We found no differences between the two groups in the use of vocal cues. Discussion: These findings bring us closer to understanding and diagnosing self-protection. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
3. ViTAU: Facial paralysis recognition and analysis based on vision transformer and facial action units
- Author
-
Jia GAO, Wenhao CAI, Junli ZHAO, and Fuqing DUAN
- Subjects
transformer ,action units ,multi-resolution feature maps ,generator ,heatmap regression ,Mining engineering. Metallurgy ,TN1-997 ,Environmental engineering ,TA170-171 - Abstract
Facial nerve paralysis (FNP), commonly known as Bell’s palsy or facial paralysis, significantly affects patients’ daily lives and mental well-being. Timely identification and diagnosis are crucial for early treatment and rehabilitation. With the rapid advancement of deep learning and computer vision technologies, automatic recognition of facial paralysis has become feasible, offering a more accurate and objective diagnostic approach. Current research primarily focuses on broad facial changes and often neglects finer facial details, which leads to insufficient analysis of how different areas affect recognition results. This study proposes an innovative method that combines the vision transformer (ViT) model with an action unit (AU) facial region detection network to automatically recognize and analyze facial paralysis. Initially, the ViT model utilizes its self-attention mechanism to accurately determine the presence of facial paralysis. Subsequently, we analyzed the AU data to assess the activity of facial muscles, allowing for a deeper evaluation of the affected areas. The self-attention mechanism in the transformer architecture captures the global contextual information required to recognize facial paralysis. To accurately determine the specific affected regions, we use the pixel2style2pixel (pSp) encoder and the StyleGAN2 generator to encode and decode images and extract feature maps that represent facial characteristics. These maps are then processed through a pyramid convolutional neural network interpreter to generate heatmaps. By optimizing the mean squared error between the predicted and actual heatmaps, we can effectively identify the affected paralysis areas. Our proposed method integrates ViT with facial AUs, designing a ViT-based facial paralysis recognition network that enhances the extraction of local area features through its self-attention mechanism, thereby enabling precise recognition of facial paralysis. Additionally, by incorporating facial AU data, we conducted detailed regional analyses for patients identified with facial paralysis. Experimental results demonstrate the efficacy of our approach, achieving a recognition accuracy of 99.4% for facial paralysis and 81.36% for detecting affected regions on the YouTube Facial Palsy (YFP) and extended Cohn Kanade (CK+) datasets. These results not only highlight the effectiveness of our automatic recognition method compared to the latest techniques but also validate its potential for clinical applications. Furthermore, to facilitate the observation of affected regions, we developed a visualization method that intuitively displays the impacted areas, thereby aiding patients and healthcare professionals in understanding the condition and enhancing communication regarding treatment and rehabilitation strategies. In conclusion, the proposed method illustrates the power of combining advanced deep learning techniques with a detailed analysis of facial AUs to improve the automatic recognition of facial paralysis. By addressing previous research limitations, the proposed method provides a more nuanced understanding of how specific facial areas are affected, leading to improved diagnosis, treatment, and patient care. This innovative approach not only enhances the accuracy of facial paralysis detection but also contributes to facial medical imaging.
- Published
- 2025
- Full Text
- View/download PDF
4. Employing the Artificial Intelligence Object Detection Tool YOLOv8 for Real-Time Pain Detection: A Feasibility Study
- Author
-
Cascella M, Shariff MN, Lo Bianco G, Monaco F, Gargano F, Simonini A, Ponsiglione AM, and Piazza O
- Subjects
pain ,artificial intelligence ,automatic pain assessment ,computer vision ,action units ,Medicine (General) ,R5-920 - Abstract
Marco Cascella,1 Mohammed Naveed Shariff,2 Giuliano Lo Bianco,3 Federica Monaco,4 Francesca Gargano,5 Alessandro Simonini,6 Alfonso Maria Ponsiglione,7 Ornella Piazza1 1Anesthesia and Pain Medicine, Department of Medicine, Surgery and Dentistry “scuola Medica Salernitana”, University of Salerno, Baronissi, 84081, Italy; 2Department of AI&DS, Rajalakshmi Institute of Technology, Chennai, TN, India; 3Anesthesiology and Pain Department, Fondazione Istituto G. Giglio Cefalù, Palermo, Italy; 4Anesthesia and Pain Medicine, ASL NA1, Napoli, Italy; 5Anesthesia and Intensive Care, U.O.C. Fondazione Policlinico Campus Bio-Medico, Roma, Italy; 6Pediatric Anesthesia and Intensive Care Unit, Salesi Children’s Hospital, Ancona, Italy; 7Department of Electrical Engineering and Information Technology, University of Naples “federico II”, Naples, 0125, ItalyCorrespondence: Giuliano Lo Bianco, Anesthesiology and Pain Department, Fondazione Istituto G. Giglio Cefalù, Palermo, Italy, Email giulianolobianco@gmail.comIntroduction: Effective pain management is crucial for patient care, impacting comfort, recovery, and overall well-being. Traditional subjective pain assessment methods can be challenging, particularly in specific patient populations. This research explores an alternative approach using computer vision (CV) to detect pain through facial expressions.Methods: The study implements the YOLOv8 real-time object detection model to analyze facial expressions indicative of pain. Given four pain datasets, a dataset of pain-expressing faces was compiled, and each image was carefully labeled based on the presence of pain-associated Action Units (AUs). The labeling distinguished between two classes: pain and no pain. The pain category included specific AUs (AU4, AU6, AU7, AU9, AU10, and AU43) following the Prkachin and Solomon Pain Intensity (PSPI) scoring method. Images showing these AUs with a PSPI score above 2 were labeled as expressing pain. The manual labeling process utilized an open-source tool, makesense.ai, to ensure precise annotation. The dataset was then split into training and testing subsets, each containing a mix of pain and no-pain images. The YOLOv8 model underwent iterative training over 10 epochs. The model’s performance was validated using precision, recall, and mean Average Precision (mAP) metrics, and F1 score.Results: When considering all classes collectively, our model attained a mAP of 0.893 at a threshold of 0.5. The precision for “pain” and “nopain” detection was 0.868 and 0.919, respectively. F1 scores for the classes “pain”, “nopain”, and “all classes” reached a peak value of 0.80. Finally, the model was tested on the Delaware dataset and in a real-world scenario.Discussion: Despite limitations, this study highlights the promise of using real-time computer vision models for pain detection, with potential applications in clinical settings. Future research will focus on evaluating the model’s generalizability across diverse clinical scenarios and its integration into clinical workflows to improve patient care.Keywords: pain, artificial intelligence, automatic pain assessment, computer vision, action units
- Published
- 2024
5. Demystifying Mental Health by Decoding Facial Action Unit Sequences.
- Author
-
Sharma, Deepika, Singh, Jaiteg, Sehra, Sukhjit Singh, and Sehra, Sumeet Kaur
- Subjects
EMOTION recognition ,CONVOLUTIONAL neural networks ,FACIAL expression ,MENTAL health ,EMOTIONS ,ANGER management - Abstract
Mental health is indispensable for effective daily functioning and stress management. Facial expressions may provide vital clues about the mental state of a person as they are universally consistent across cultures. This study intends to detect the emotional variances through facial micro-expressions using facial action units (AUs) to identify probable mental health issues. In addition, convolutional neural networks (CNN) were used to detect and classify the micro-expressions. Further, combinations of AUs were identified for the segmentation of micro-expressions classes using K-means square. Two benchmarked datasets CASME II and SAMM were employed for the training and evaluation of the model. The model achieved an accuracy of 95.62% on CASME II and 93.21% on the SAMM dataset, respectively. Subsequently, a case analysis was done to identify depressive patients using the proposed framework and it attained an accuracy of 92.99%. This experiment revealed the fact that emotions like disgust, sadness, anger, and surprise are the prominent emotions experienced by depressive patients during communication. The findings suggest that leveraging facial action units for micro-expression detection offers a promising approach to mental health diagnostics. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Transformer embedded spectral-based graph network for facial expression recognition.
- Author
-
Jin, Xing, Song, Xulin, Wu, Xiyin, and Yan, Wenzhu
- Abstract
Deep graph convolution networks which exploit relations among facial muscle movements for facial expression recognition (FER) have achieved great success. Due to the limited receptive field, existing graph convolution operations are difficult to model long-range muscle movement relations which plays a crucial role in FER. To alleviate this issue, we introduce the transformer encoder into graph convolution networks, in which the vision transformer enables all facial muscle movements to interact in the global receptive field and model more complex relations. Specifically, we construct facial graph data by cropping regions of interest (ROIs) which are associated with facial action units, and each ROI is represented by the representation of hidden layers from deep auto-encoder. To effectively extract features from the constructed facial graph data, we propose a novel transformer embedded spectral-based graph convolution network (TESGCN), in which the transformer encoder is exploited to interact with complex relations among facial RIOs for FER. Compared to vanilla graph convolution networks, we empirically show the superiority of the proposed model by conducting extensive experiments across four facial expression datasets. Moreover, our proposed TESGCN only has 80K parameters and 0.41MB model size, and achieves comparable results compared to existing lightweight networks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Using Diverse ConvNets to Classify Face Action Units in Dataset on Emotions Among Mexicans (DEM)
- Author
-
Marco A. Moreno-Armendariz, Alberto Espinosa-Juarez, and Esmeralda Godinez-Montero
- Subjects
Action units ,ConvNets ,CAM analysis ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
To understand how Convolutional Neural Networks (ConvNets) perceive the muscular movements of the human face, known as Action Units (AUs) in this work, we introduce a new dataset named Dataset on Emotions among Mexicans (DEM), consisting of 1557 images of Mexicans labeled with twenty-six AUs and seven emotions. As a benchmark, we used the comparison with DISFA+ labeled with 12 AUs. To address the task of detecting AUs in each image, six ConvNets were employed, and we evaluated their performance using the F1 Score. The two ConvNets with the best performance were VGG19 with 0.8180% (DEM), 0.9106 % (DISFA+), and ShuffleNetV2 with 0.7154% (DEM), 0.9440% (DISFA+). Subsequently, these ConvNets were analyzed using Grad-CAM and Grad-CAM++; this algorithms allows us to observe the areas of the face considered for prediction. In most cases, these areas consider the region of the labeled AU. Considering the F1 score and the visual study, we can conclude that using DEM as a dataset to classify AUs is promising since the experiments achieved performances similar to those of the current literature that only use ConvNets.
- Published
- 2024
- Full Text
- View/download PDF
8. Enhancing frame-level student engagement classification through knowledge transfer techniques.
- Author
-
Das, Riju and Dev, Soumyabrata
- Subjects
STUDENT engagement ,DEEP learning ,KNOWLEDGE transfer ,ABILITY grouping (Education) ,CLASSIFICATION ,LEARNING - Abstract
Assessing student engagement in educational settings is critical for monitoring and improving the learning process. Traditional methods that classify video-based student engagement datasets often assign a single engagement label to the entire video, resulting in inaccurate classification outcomes. However, student engagement varies over time, with fluctuations in concentration and interest levels. To overcome this limitation, this paper introduces a frame-level student engagement detection approach. By analyzing each frame individually, instructors gain more detailed insights into students' understanding of the course. The proposed method focuses on identifying student engagement at a granular level, enabling instructors to pinpoint specific moments of disengagement or high engagement for targeted interventions. Nevertheless, the lack of labeled frame-level engagement data presents a significant challenge. To address this challenge, we propose a novel approach for frame-level student engagement classification by leveraging the concept of knowledge transfer. Our method involves pretraining a deep learning model on a labeled image-based student engagement dataset, WACV, which serves as the base dataset for identifying frame-level engagement in our target video-based DAiSEE dataset. We then fine-tune the model on the unlabeled video dataset, utilizing the transferred knowledge to enhance engagement classification performance. Experimental results demonstrate the effectiveness of our frame-level approach, providing valuable insights for instructors to optimize instructional strategies and enhance the learning experience. This research contributes to the advancement of student engagement assessment, offering educators a more nuanced understanding of student behaviors during instructional videos. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Demystifying Mental Health by Decoding Facial Action Unit Sequences
- Author
-
Deepika Sharma, Jaiteg Singh, Sukhjit Singh Sehra, and Sumeet Kaur Sehra
- Subjects
micro-expressions ,CNN ,K-means ,emotion recognition ,action units ,Technology - Abstract
Mental health is indispensable for effective daily functioning and stress management. Facial expressions may provide vital clues about the mental state of a person as they are universally consistent across cultures. This study intends to detect the emotional variances through facial micro-expressions using facial action units (AUs) to identify probable mental health issues. In addition, convolutional neural networks (CNN) were used to detect and classify the micro-expressions. Further, combinations of AUs were identified for the segmentation of micro-expressions classes using K-means square. Two benchmarked datasets CASME II and SAMM were employed for the training and evaluation of the model. The model achieved an accuracy of 95.62% on CASME II and 93.21% on the SAMM dataset, respectively. Subsequently, a case analysis was done to identify depressive patients using the proposed framework and it attained an accuracy of 92.99%. This experiment revealed the fact that emotions like disgust, sadness, anger, and surprise are the prominent emotions experienced by depressive patients during communication. The findings suggest that leveraging facial action units for micro-expression detection offers a promising approach to mental health diagnostics.
- Published
- 2024
- Full Text
- View/download PDF
10. What Can Facial Movements Reveal? Depression Recognition and Analysis Based on Optical Flow Using Bayesian Networks
- Author
-
Yu Ma, Jian Shen, Zeguang Zhao, Huajian Liang, Yang Tan, Zhenyu Liu, Kun Qian, Minqiang Yang, and Bin Hu
- Subjects
Depression recognition ,facial expressions ,action units ,optical flow ,Bayesian networks ,Medical technology ,R855-855.5 ,Therapeutics. Pharmacology ,RM1-950 - Abstract
Recent evidence have demonstrated that facial expressions could be a valid and important aspect for depression recognition. Although various works have been achieved in automatic depression recognition, it is a challenge to explore the inherent nuances of facial expressions that might reveal the underlying differences between depressed patients and healthy subjects under different stimuli. There is a lack of an undisturbed system that monitors depressive patients’ mental states in various free-living scenarios, so this paper steps towards building a classification model where data collection, feature extraction, depression recognition and facial actions analysis are conducted to infer the differences of facial movements between depressive patients and healthy subjects. In this study, we firstly present a plan of dividing facial regions of interest to extract optical flow features of facial expressions for depression recognition. We then propose facial movements coefficients utilising discrete wavelet transformation. Specifically, Bayesian Networks equipped with construction of Pearson Correlation Coefficients based on discrete wavelet transformation is learnt, which allows for analysing movements of different facial regions. We evaluate our method on a clinically validated dataset of 30 depressed patients and 30 healthy control subjects, and experiments results obtained the accuracy and recall of 81.7%, 96.7%, respectively, outperforming other features for comparison. Most importantly, the Bayesian Networks we built on the coefficients under different stimuli may reveal some facial action patterns of depressed subjects, which have a potential to assist the automatic diagnosis of depression.
- Published
- 2023
- Full Text
- View/download PDF
11. Visualization and analysis of skin strain distribution in various human facial actions
- Author
-
Takeru MISU, Hisashi ISHIHARA, So NAGASHIMA, Yusuke DOI, and Akihiro NAKATANI
- Subjects
human face ,strain analysis ,action units ,facial expression ,artificial face design ,Mechanical engineering and machinery ,TJ1-1570 - Abstract
Human faces are mechanical systems that display emotional and linguistic information with motions of the deformable facial tissues. Since each facial action forms a different stress-strain field that characterizes the differentiated facial appearance, a detailed analysis of strain distributions for each facial action will contribute to analytical and synthetical studies on human faces. This study evaluated strain distributions of 44 facial actions of a Japanese adult male based on the three-dimensional displacements of 125 tracking markers attached to the facial surface. We investigated how much the facial skin surface is stretched and compressed in each facial region based on the evaluated area strains produced by each facial action. Then, we visualized the strain distributions and surface undulation to analyze the complexity of the deformations on a face. The results show that the positive and negative surface strains intermingled on a face even in simple facial actions, potentially reflecting the complex facial structure under the facial skin layers, where several tissues with different material properties, e.g., adipose tissues and retaining ligaments, are distributed heterogeneously. These results are beneficial for artificial face designers as a design target and evidence to consider the effective skin structure and locations of actuators for artificial faces.
- Published
- 2023
- Full Text
- View/download PDF
12. Facial expression recognition on partially occluded faces using component based ensemble stacked CNN.
- Author
-
Bellamkonda, Sivaiah, Gopalan, N. P., Mala, C., and Settipalli, Lavanya
- Abstract
Facial Expression Recognition (FER) is the basis for many applications including human-computer interaction and surveillance. While developing such applications, it is imperative to understand human emotions for better interaction with machines. Among many FER models developed so far, Ensemble Stacked Convolution Neural Networks (ES-CNN) showed an empirical impact in improving the performance of FER on static images. However, the existing ES-CNN based FER models trained with features extracted from the entire face, are unable to address the issues of ambient parameters such as pose, illumination, occlusions. To mitigate the problem of reduced performance of ES-CNN on partially occluded faces, a Component based ES-CNN (CES-CNN) is proposed. CES-CNN applies ES-CNN on action units of individual face components such as eyes, eyebrows, nose, cheek, mouth, and glabella as one subnet of the network. Max-Voting based ensemble classifier is used to ensemble the decisions of the subnets in order to obtain the optimized recognition accuracy. The proposed CES-CNN is validated by conducting experiments on benchmark datasets and the performance is compared with the state-of-the-art models. It is observed from the experimental results that the proposed model has a significant enhancement in the recognition accuracy compared to the existing models. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
13. What Can Facial Movements Reveal? Depression Recognition and Analysis Based on Optical Flow Using Bayesian Networks.
- Author
-
Ma, Yu, Shen, Jian, Zhao, Zeguang, Liang, Huajian, Tan, Yang, Liu, Zhenyu, Qian, Kun, Yang, Minqiang, and Hu, Bin
- Subjects
DISCRETE wavelet transforms ,BAYESIAN analysis ,FACIAL expression ,OPTICAL flow ,PEARSON correlation (Statistics) - Abstract
Recent evidence have demonstrated that facial expressions could be a valid and important aspect for depression recognition. Although various works have been achieved in automatic depression recognition, it is a challenge to explore the inherent nuances of facial expressions that might reveal the underlying differences between depressed patients and healthy subjects under different stimuli. There is a lack of an undisturbed system that monitors depressive patients’ mental states in various free-living scenarios, so this paper steps towards building a classification model where data collection, feature extraction, depression recognition and facial actions analysis are conducted to infer the differences of facial movements between depressive patients and healthy subjects. In this study, we firstly present a plan of dividing facial regions of interest to extract optical flow features of facial expressions for depression recognition. We then propose facial movements coefficients utilising discrete wavelet transformation. Specifically, Bayesian Networks equipped with construction of Pearson Correlation Coefficients based on discrete wavelet transformation is learnt, which allows for analysing movements of different facial regions. We evaluate our method on a clinically validated dataset of 30 depressed patients and 30 healthy control subjects, and experiments results obtained the accuracy and recall of 81.7%, 96.7%, respectively, outperforming other features for comparison. Most importantly, the Bayesian Networks we built on the coefficients under different stimuli may reveal some facial action patterns of depressed subjects, which have a potential to assist the automatic diagnosis of depression. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
14. BFFN: A novel balanced feature fusion network for fair facial expression recognition.
- Author
-
Li, Hao, Luo, Yiqin, Gu, Tianlong, and Chang, Liang
- Subjects
- *
FACIAL expression , *FAIRNESS - Abstract
Facial expression recognition (FER) technology has become increasingly mature and applicable in recent years. However, it still suffers from the bias of expression class, which can lead to unfair decisions for certain expression classes in applications. This study aims to mitigate expression class bias through both pre-processing and in-processing approaches. First, we analyze the output of existing models and demonstrate the existence of obvious class bias, particularly for underrepresented expressions. Second, we develop a class-balanced dataset constructed through data generation, mitigating unfairness at the data level. Then, we propose the Balanced Feature Fusion Network (BFFN), a class fairness-enhancing network. The BFFN mitigates the class bias by adding facial action units (AU) to enrich expression-related features and allocating weights in the AU feature fusion process to improve the extraction ability of underrepresented expression features. Finally, extensive experiments on datasets (RAF-DB and AffectNet) provide evidence that our BFFN outperforms existing FER models, improving the fairness by at least 16%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Processing Real-Life Recordings of Facial Expressions of Polish Sign Language Using Action Units.
- Author
-
Irasiak, Anna, Kozak, Jan, Piasecki, Adam, and Stęclik, Tomasz
- Subjects
- *
SIGN language , *FACIAL expression , *POLISH language , *LANGUAGE policy , *LANGUAGE & languages , *DEAF children - Abstract
Automatic translation between the national language and sign language is a complex process similar to translation between two different foreign languages. A very important aspect is the precision of not only manual gestures but also facial expressions, which are extremely important in the overall context of a sentence. In this article, we present the problem of including facial expressions in the automation of Polish-to-Polish Sign Language (PJM) translation—this is part of an ongoing project related to a comprehensive solution allowing for the animation of manual gestures, body movements and facial expressions. Our approach explores the possibility of using action unit (AU) recognition in the automatic annotation of recordings, which in the subsequent steps will be used to train machine learning models. This paper aims to evaluate entropy in real-life translation recordings and analyze the data associated with the detected action units. Our approach has been subjected to evaluation by experts related to Polish Sign Language, and the results obtained allow for the development of further work related to automatic translation into Polish Sign Language. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
16. Emotion Recognition of Down Syndrome People Based on the Evaluation of Artificial Intelligence and Statistical Analysis Methods.
- Author
-
Paredes, Nancy, Caicedo-Bravo, Eduardo F., Bacca, Bladimir, and Olmedo, Gonzalo
- Subjects
- *
EMOTION recognition , *DEEP learning , *ARTIFICIAL intelligence , *STATISTICS , *SUPERVISED learning ,PEOPLE with Down syndrome - Abstract
This article presents a study based on evaluating different techniques to automatically recognize the basic emotions of people with Down syndrome, such as anger, happiness, sadness, surprise, and neutrality, as well as the statistical analysis of the Facial Action Coding System, determine the symmetry of the Action Units present in each emotion, identify the facial features that represent this group of people. First, a dataset of images of faces of people with Down syndrome classified according to their emotions is built. Then, the characteristics of facial micro-expressions (Action Units) present in the feelings of the target group through statistical analysis are evaluated. This analysis uses the intensity values of the most representative exclusive action units to classify people's emotions. Subsequently, the collected dataset was evaluated using machine learning and deep learning techniques to recognize emotions. In the beginning, different supervised learning techniques were used, with the Support Vector Machine technique obtaining the best precision with a value of 66.20%. In the case of deep learning methods, the mini-Xception convolutional neural network was used to recognize people's emotions with typical development, obtaining an accuracy of 74.8%. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
17. Audio-Visual Stress Classification Using Cascaded RNN-LSTM Networks.
- Author
-
Gupta, Megha V., Vaikole, Shubhangi, Oza, Ankit D., Patel, Amisha, Burduhos-Nergis, Diana Petronela, and Burduhos-Nergis, Dumitru Doru
- Subjects
- *
CASCADE connections , *PERCEIVED Stress Scale , *PSYCHOLOGICAL stress , *WELL-being , *CLASSIFICATION - Abstract
The purpose of this research is to emphasize the importance of mental health and contribute to the overall well-being of humankind by detecting stress. Stress is a state of strain, whether it be mental or physical. It can result from anything that frustrates, incenses, or unnerves you in an event or thinking. Your body's response to a demand or challenge is stress. Stress affects people on a daily basis. Stress can be regarded as a hidden pandemic. Long-term (chronic) stress results in ongoing activation of the stress response, which wears down the body over time. Symptoms manifest as behavioral, emotional, and physical effects. The most common method involves administering brief self-report questionnaires such as the Perceived Stress Scale. However, self-report questionnaires frequently lack item specificity and validity, and interview-based measures can be time- and money-consuming. In this research, a novel method used to detect human mental stress by processing audio-visual data is proposed. In this paper, the focus is on understanding the use of audio-visual stress identification. Using the cascaded RNN-LSTM strategy, we achieved 91% accuracy on the RAVDESS dataset, classifying eight emotions and eventually stressed and unstressed states. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
18. Is It Me or the Robot? A Critical Evaluation of Human Affective State Recognition in a Cognitive Task.
- Author
-
Jirak, Doreen, Motonobu Aoki, Takura Yanagi, Atsushi Takamatsu, Bouet, Stephane, Tomohiro Yamamura, Sandini, Giulio, and Rea, Francesco
- Subjects
RECOGNITION (Psychology) ,AFFECTIVE computing ,HUMAN-robot interaction ,AFFECT (Psychology) ,HUMANOID robots ,SOCIAL robots - Abstract
A key goal in human-robot interaction (HRI) is to design scenarios between humanoid robots and humans such that the interaction is perceived as collaborative and natural, yet safe and comfortable for the human. Human skills like verbal and non-verbal communication are essential elements as humans tend to attribute social behaviors to robots. However, aspects like the uncanny valley and different technical affinity levels can impede the success of HRI scenarios, which has consequences on the establishment of long-term interaction qualities like trust and rapport. In the present study, we investigate the impact of a humanoid robot on human emotional responses during the performance of a cognitively demanding task. We set up three different conditions for the robot with increasing levels of social cue expressions in a between-group study design. For the analysis of emotions, we consider the eye gaze behavior, arousal-valence for affective states, and the detection of action units. Our analysis reveals that the participants display a high tendency toward positive emotions in presence of a robot with clear social skills compared to other conditions, where we show how emotions occur only at task onset. Our study also shows how different expression levels influence the analysis of the robots' role in HRI. Finally, we critically discuss the current trend of automatized emotion or affective state recognition in HRI and demonstrate issues that have direct consequences on the interpretation and, therefore, claims about human emotions in HRI studies. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
19. Is It Me or the Robot? A Critical Evaluation of Human Affective State Recognition in a Cognitive Task
- Author
-
Doreen Jirak, Motonobu Aoki, Takura Yanagi, Atsushi Takamatsu, Stephane Bouet, Tomohiro Yamamura, Giulio Sandini, and Francesco Rea
- Subjects
human-robot interaction ,social robots ,cognitive load ,affective states ,action units ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
A key goal in human-robot interaction (HRI) is to design scenarios between humanoid robots and humans such that the interaction is perceived as collaborative and natural, yet safe and comfortable for the human. Human skills like verbal and non-verbal communication are essential elements as humans tend to attribute social behaviors to robots. However, aspects like the uncanny valley and different technical affinity levels can impede the success of HRI scenarios, which has consequences on the establishment of long-term interaction qualities like trust and rapport. In the present study, we investigate the impact of a humanoid robot on human emotional responses during the performance of a cognitively demanding task. We set up three different conditions for the robot with increasing levels of social cue expressions in a between-group study design. For the analysis of emotions, we consider the eye gaze behavior, arousal-valence for affective states, and the detection of action units. Our analysis reveals that the participants display a high tendency toward positive emotions in presence of a robot with clear social skills compared to other conditions, where we show how emotions occur only at task onset. Our study also shows how different expression levels influence the analysis of the robots' role in HRI. Finally, we critically discuss the current trend of automatized emotion or affective state recognition in HRI and demonstrate issues that have direct consequences on the interpretation and, therefore, claims about human emotions in HRI studies.
- Published
- 2022
- Full Text
- View/download PDF
20. Studying Facial Activity in Parkinson's Disease Patients Using an Automated Method and Video Recording
- Author
-
Anastasia Moshkova, Andrey Samorodov, Ekaterina Ivanova, Ekaterina Fedotova, Natalia Voinova, and Alexander Volkov
- Subjects
parkinson's disease ,hypomimia ,action units ,Telecommunication ,TK5101-6720 - Abstract
Main objective of this research is studying facial activity of PD patients using an automated method for assessing facial movements based on calculating the movement signal kinematic characteristics. 16 PD patients and 16 participants in the control group have been recorded using 2D video camera while performing 4 facial tests: closing the eyes, raising the eyebrows, smiling with an effort (grin), moving the eyebrows. Each test includes a 10-fold repetition of a given facial movement with maximum effort and velocity. Intensities of the corresponding action units (AUs) in each frame have been determined by an automated method for each test. The resulting signal of the AU intensity dependence on the frame number has been marked with maximum and minimum points. From the extremum points, 11 kinematic parameters have been calculated for each AU of each patient. To assess the differences in kinematic parameters between the groups, the nonparametric Mann-Whitney test has been used. There is decrease in the facial movements frequency and velocity. Consequently, there is increase in the duration of the performance of facial tests in PD patients group compared to the control group. Only some AUs also showed a decrease in the amplitude of facial movements in PD patients group compared to the control group. Differences in the largest number of kinematic parameters have been obtained for the action units AU12 and AU14 recorded during the smile with effort test, and AU04 recorded during the eyebrow moving test.
- Published
- 2021
- Full Text
- View/download PDF
21. Facial Motion Analysis beyond Emotional Expressions.
- Author
-
Porta-Lorenzo, Manuel, Vázquez-Enríquez, Manuel, Pérez-Pérez, Ania, Alba-Castro, José Luis, and Docío-Fernández, Laura
- Subjects
- *
MOTION analysis , *SELF-expression , *FACIAL expression & emotions (Psychology) , *NONVERBAL communication , *FACIAL expression , *DEEP learning - Abstract
Facial motion analysis is a research field with many practical applications, and has been strongly developed in the last years. However, most effort has been focused on the recognition of basic facial expressions of emotion and neglects the analysis of facial motions related to non-verbal communication signals. This paper focuses on the classification of facial expressions that are of the utmost importance in sign languages (Grammatical Facial Expressions) but also present in expressive spoken language. We have collected a dataset of Spanish Sign Language sentences and extracted the intervals for three types of Grammatical Facial Expressions: negation, closed queries and open queries. A study of several deep learning models using different input features on the collected dataset (LSE_GFE) and an external dataset (BUHMAP) shows that GFEs can be learned reliably with Graph Convolutional Networks simply fed with face landmarks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
22. AU-Guided Unsupervised Domain-Adaptive Facial Expression Recognition.
- Author
-
Peng, Xiaojiang, Gu, Yuxin, and Zhang, Panpan
- Subjects
FACIAL expression ,DEEP learning - Abstract
Domain diversities, including inconsistent annotation and varied image collection conditions, inevitably exist among different facial expression recognition (FER) datasets, posing an evident challenge for adapting FER models trained on one dataset to another one. Recent works mainly focus on domain-invariant deep feature learning with adversarial learning mechanisms, ignoring the sibling facial action unit (AU) detection task, which has obtained great progress. Considering that AUs objectively determine facial expressions, this paper proposes an AU-guided unsupervised domain-adaptive FER (AdaFER) framework to relieve the annotation bias between different FER datasets. In AdaFER, we first leverage an advanced model for AU detection on both a source and a target domain. Then, we compare the AU results to perform AU-guided annotating, i.e., target faces that own the same AUs as source faces would inherit the labels from the source domain. Meanwhile, to achieve domain-invariant compact features, we utilize an AU-guided triplet training, which randomly collects anchor–positive–negative triplets on both domains with AUs. We conduct extensive experiments on several popular benchmarks and show that AdaFER achieves state-of-the-art results on all these benchmarks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. Facial Emotional Expression Assessment in Parkinson's Disease by Automated Algorithm Based on Action Units
- Author
-
Anastasia Moshkova, Andrey Samorodov, Natalia Voinova, Alexander Volkov, Ekaterina Ivanova, and Ekaterina Fedotova
- Subjects
parkinson's disease ,emotional expression ,hypomimia ,action units ,Telecommunication ,TK5101-6720 - Abstract
This work is devoted to the study of expression and interpretation of six basic emotions: anger, disgust, fear, happiness, sadness, surprise in patients with Parkinson's disease in comparison with the healthy control group of patients. The study involved 16 patients in each group. Each patients face was recorded using a 2D camera while performing 3 tasks: displaying a neutral state, displaying 6 basic emotions by researcher request, displaying 6 basic emotions depicted on the images. Action units were determined on each video frame. The percentages of emotional expressions in each video were determined, and the intensity of the recognized expressions for each task using the emotion recognition algorithm based on action units. The difference between emotional expressions and the neutral state was calculated as Euclidian distance between vectors of action units to quantify the changes in facial expression between the Parkinson's disease and healthy control groups. To analyze the differences between the groups, the non-parametric MannWhitney U-test was used. The obtained results show changes in the emotional expressions in the Parkinson's disease group in comparison with the healthy control group, Parkinson's disease patients show a decrease in the expressiveness of face and the intensity of the emotional expression.
- Published
- 2020
- Full Text
- View/download PDF
24. Faces of Pain in Dementia: Learnings From a Real-World Study Using a Technology-Enabled Pain Assessment Tool
- Author
-
Mustafa Atee, Kreshnik Hoti, Paola Chivers, and Jeffery D. Hughes
- Subjects
pain ,dementia ,facial expressions ,action units ,artificial intelligence ,PainChek® ,Neurology. Diseases of the nervous system ,RC346-429 - Abstract
Pain is common in people living with dementia (PLWD), including those with limited verbal skills. Facial expressions are key behavioral indicators of the pain experience in this group. However, there is a lack of real-world studies to report the prevalence and associations of pain-relevant facial micro-expressions in PLWD. In this observational retrospective study, pain-related facial features were studied in a sample of 3,144 PLWD [mean age 83.3 years (SD = 9.0); 59.0% female] using the Face domain of PainChek®, a point-of-care medical device application. Pain assessments were completed by 389 users from two national dementia-specific care programs and 34 Australian aged care homes. Our analysis focused on the frequency, distribution, and associations of facial action units [AU(s)] with respect to various pain intensity groups. A total of 22,194 pain assessments were completed. Of the AUs present, AU7 (eyelid tightening) was the most frequent facial expression (48.6%) detected, followed by AU43 (closing eyes; 42.9%) and AU6 (cheek raising; 42.1%) during severe pain. AU20 (horizontal mouth stretch) was the most predictive facial action of higher pain scores. Eye-related AUs (AU6, AU7, AU43) and brow-related AUs (AU4) were more common than mouth-related AUs (e.g., AU20, AU25) during higher pain intensities. No significant effect was found for age or gender. These findings offer further understanding of facial expressions during clinical pain in PLWD and confirm the usefulness of artificial intelligence (AI)-enabled real-time analysis of the face as part of the assessment of pain in aged care clinical practice.
- Published
- 2022
- Full Text
- View/download PDF
25. A Proposal for Multimodal Emotion Recognition Using Aural Transformers and Action Units on RAVDESS Dataset.
- Author
-
Luna-Jiménez, Cristina, Kleinlein, Ricardo, Griol, David, Callejas, Zoraida, Montero, Juan M., and Fernández-Martínez, Fernando
- Subjects
EMOTION recognition ,EMOTIONAL state ,EXTRACTION techniques ,EMOTIONS ,AUTONOMOUS vehicles ,EMOTICONS & emojis ,VIDEO compression - Abstract
Emotion recognition is attracting the attention of the research community due to its multiple applications in different fields, such as medicine or autonomous driving. In this paper, we proposed an automatic emotion recognizer system that consisted of a speech emotion recognizer (SER) and a facial emotion recognizer (FER). For the SER, we evaluated a pre-trained xlsr-Wav2Vec2.0 transformer using two transfer-learning techniques: embedding extraction and fine-tuning. The best accuracy results were achieved when we fine-tuned the whole model by appending a multilayer perceptron on top of it, confirming that the training was more robust when it did not start from scratch and the previous knowledge of the network was similar to the task to adapt. Regarding the facial emotion recognizer, we extracted the Action Units of the videos and compared the performance between employing static models against sequential models. Results showed that sequential models beat static models by a narrow difference. Error analysis reported that the visual systems could improve with a detector of high-emotional load frames, which opened a new line of research to discover new ways to learn from videos. Finally, combining these two modalities with a late fusion strategy, we achieved 86.70% accuracy on the RAVDESS dataset on a subject-wise 5-CV evaluation, classifying eight emotions. Results demonstrated that these modalities carried relevant information to detect users' emotional state and their combination allowed to improve the final system performance. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
26. Modelling facial action units using partial differential equations
- Author
-
Ismail, Nur Baini Binti
- Subjects
006.6 ,Geometric modelling ,3D face modelling ,Partial differential equations ,PDE method ,Action units ,Facial expressions - Abstract
This thesis discusses a novel method for modelling facial action units. It presents facial action units model based on boundary value problems for accurate representation of human facial expression in three-dimensions. In particular, a solution to a fourth order elliptic Partial Differential Equation (PDE) subject to suitable boundary conditions is utilized, where the chosen boundary curves are based on muscles movement defined by Facial Action Coding System (FACS). This study involved three stages: modelling faces, manipulating faces and application to simple facial animation. In the first stage, PDE method is used in modelling and generating a smooth 3D face. The PDE formulation using small sets of parameters contributes to the efficiency of human face representation. In the manipulation stage, a generic PDE face of neutral expression is manipulated to a face with expression using PDE descriptors that uniquely represents an action unit. A combination of the PDE descriptor results in a generic PDE face having an expression, which successfully modelled four basic expressions: happy, sad, fear and disgust. An example of application is given using simple animation technique called blendshapes. This technique uses generic PDE face in animating basic expressions.
- Published
- 2015
27. Understanding the Multidimensional and Dynamic Nature of Facial Expressions Based on Indicators for Appraisal Components as Basis for Measuring Drivers' Fear
- Author
-
Meng Zhang, Klas Ihme, Uwe Drewitz, and Meike Jipp
- Subjects
fear ,facial expression ,action units ,in-vehicle ,component process model ,Psychology ,BF1-990 - Abstract
Facial expressions are one of the commonly used implicit measurements for the in-vehicle affective computing. However, the time courses and the underlying mechanism of facial expressions so far have been barely focused on. According to the Component Process Model of emotions, facial expressions are the result of an individual's appraisals, which are supposed to happen in sequence. Therefore, a multidimensional and dynamic analysis of drivers' fear by using facial expression data could profit from a consideration of these appraisals. A driving simulator experiment with 37 participants was conducted, in which fear and relaxation were induced. It was found that the facial expression indicators of high novelty and low power appraisals were significantly activated after a fear event (high novelty: Z = 2.80, p < 0.01, rcontrast = 0.46; low power: Z = 2.43, p < 0.05, rcontrast = 0.50). Furthermore, after the fear event, the activation of high novelty occurred earlier than low power. These results suggest that multidimensional analysis of facial expression is suitable as an approach for the in-vehicle measurement of the drivers' emotions. Furthermore, a dynamic analysis of drivers' facial expressions considering of effects of appraisal components can add valuable information for the in-vehicle assessment of emotions.
- Published
- 2021
- Full Text
- View/download PDF
28. The Facial Action Coding System for Characterization of Human Affective Response to Consumer Product-Based Stimuli: A Systematic Review
- Author
-
Elizabeth A. Clark, J'Nai Kessinger, Susan E. Duncan, Martha Ann Bell, Jacob Lahne, Daniel L. Gallagher, and Sean F. O'Keefe
- Subjects
emotions ,facial action coding system ,action units ,facial analysis ,consumers ,Psychology ,BF1-990 - Abstract
To characterize human emotions, researchers have increasingly utilized Automatic Facial Expression Analysis (AFEA), which automates the Facial Action Coding System (FACS) and translates the facial muscular positioning into the basic universal emotions. There is broad interest in the application of FACS for assessing consumer expressions as an indication of emotions to consumer product-stimuli. However, the translation of FACS to characterization of emotions is elusive in the literature. The aim of this systematic review is to give an overview of how FACS has been used to investigate human emotional behavior to consumer product-based stimuli. The search was limited to studies published in English after 1978, conducted on humans, using FACS or its action units to investigate affect, where emotional response is elicited by consumer product-based stimuli evoking at least one of the five senses. The search resulted in an initial total of 1,935 records, of which 55 studies were extracted and categorized based on the outcomes of interest including (i) method of FACS implementation; (ii) purpose of study; (iii) consumer product-based stimuli used; and (iv) measures of affect validation. Most studies implemented FACS manually (73%) to develop products and/or software (20%) and used consumer product-based stimuli that had known and/or defined capacity to evoke a particular affective response, such as films and/or movie clips (20%); minimal attention was paid to consumer products with low levels of emotional competence or with unknown affective impact. The vast majority of studies (53%) did not validate FACS-determined affect and, of the validation measures that were used, most tended to be discontinuous in nature and only captured affect as it holistically related to an experience. This review illuminated some inconsistencies in how FACS is carried out as well as how emotional response is inferred from facial muscle activation. This may prompt researchers to consider measuring the total consumer experience by employing a variety of methodologies in addition to FACS and its emotion-based interpretation guide. Such strategies may better conceptualize consumers' experience with products of low, unknown, and/or undefined capacity to evoke an affective response such as product prototypes, line extensions, etc.
- Published
- 2020
- Full Text
- View/download PDF
29. Engagement detection in online learning: a review
- Author
-
M. Ali Akber Dewan, Mahbub Murshed, and Fuhua Lin
- Subjects
Engagement detection ,Affect detection ,Facial expression recognition ,Action units ,Emotion detection ,Special aspects of education ,LC8-6691 - Abstract
Abstract Online learners participate in various educational activities including reading, writing, watching video tutorials, online exams, and online meetings. During the participation in these educational activities, they show various engagement levels, such as boredom, frustration, delight, neutral, confusion, and learning gain. To provide personalized pedagogical support through interventions to online learners, it is important for online educators to detect their online learners’ engagement status precisely and efficiently. This paper presents a review of the state of the art in engagement detection in the context of online learning. We classify the existing methods into three main categories—automatic, semi-automatic and manual—considering the methods’ dependencies on learners’ participation. Methods in each category are then divided into subcategories based on the data types (e.g., audio, video, texts for learner log data etc.) they process for the engagement detection. In particular, the computer vision based methods in the automatic category that use facial expressions are examined in more details because they are found to be promising in the online learning environment. These methods are nonintrusive in nature, and the hardware and the software that these methods use to capture and analyze video data are cost-effective and easily achievable. Different techniques in the field of computer vision and machine learning are applied in these methods for the engagement detection. We then identify their challenges of engagement detection and explore available datasets and performance metrics for engagement detection, and provide recommendations for the future to advance the technology of engagement detection for online education.
- Published
- 2019
- Full Text
- View/download PDF
30. AU-Guided Unsupervised Domain-Adaptive Facial Expression Recognition
- Author
-
Xiaojiang Peng, Yuxin Gu, and Panpan Zhang
- Subjects
facial expression recognition (FER) ,action units ,unsupervised cross-domain FER ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Domain diversities, including inconsistent annotation and varied image collection conditions, inevitably exist among different facial expression recognition (FER) datasets, posing an evident challenge for adapting FER models trained on one dataset to another one. Recent works mainly focus on domain-invariant deep feature learning with adversarial learning mechanisms, ignoring the sibling facial action unit (AU) detection task, which has obtained great progress. Considering that AUs objectively determine facial expressions, this paper proposes an AU-guided unsupervised domain-adaptive FER (AdaFER) framework to relieve the annotation bias between different FER datasets. In AdaFER, we first leverage an advanced model for AU detection on both a source and a target domain. Then, we compare the AU results to perform AU-guided annotating, i.e., target faces that own the same AUs as source faces would inherit the labels from the source domain. Meanwhile, to achieve domain-invariant compact features, we utilize an AU-guided triplet training, which randomly collects anchor–positive–negative triplets on both domains with AUs. We conduct extensive experiments on several popular benchmarks and show that AdaFER achieves state-of-the-art results on all these benchmarks.
- Published
- 2022
- Full Text
- View/download PDF
31. Automated Pain Detection in Facial Videos using Transfer Learning
- Author
-
Xu, Xiaojing
- Subjects
Computer engineering ,Computer science ,action units ,deep learning ,ensemble learning ,facial video ,pain detection ,personalized pain - Abstract
Accurately determining pain levels is difficult, even for trained professionals. Facial activity provides sensitive and specific information about pain, and computer vision algorithms have been developed to automatically detect facial activities such as Facial Action Units (AUs) defined by the Facial Action Coding System (FACS). Previous work on automated pain detection from facial expressions has primarily focused on frame-level objective pain metrics, such as the Prkachin and Solomon Pain Intensity (PSPI). However, the current gold standard pain metric is the visual analog scale (VAS), which is self-reported at the video level. In this thesis, we propose machine learning models to directly evaluate VAS in video.First, we study the relationship between sequence-level metrics and frame-level metrics. Specifically, we explore an extended multitask learning model to predict VAS from human-labeled AUs with the help of other sequence-level pain measurements during training. This model consists of two parts: a multitask learning neural network model to predict multidimensional pain scores, and an ensemble learning model to linearly combine the multidimensional pain scores to best approximate VAS. Starting from human-labeled AUs, the model outperforms provided human sequence-level estimates. Secondly, we explore ways to learn sequence-level metrics based on frame-level automatically predicted AUs. We start with an AU prediction software called iMotions. We apply transfer learning by training another machine learning model to map iMotions AU codings to a subspace of manual AU codings to enable more robust pain recognition performance when only automatically coded AUs are available for the test data. We then learn our own AU prediction system which is a VGGFace neural network multitask learning model to predict AUs.Thirdly, we propose to improve our model using individual models and uncertainty estimation. For a new test video, we jointly consider which individual models generalize well generally, and which individual models are more similar/accurate to this test video, in order to choose the optimal combination of individual models and get the best performance on new test videos. Our structure achieves state-of-the-art performance on two datasets.
- Published
- 2021
32. Understanding the Multidimensional and Dynamic Nature of Facial Expressions Based on Indicators for Appraisal Components as Basis for Measuring Drivers' Fear.
- Author
-
Zhang, Meng, Ihme, Klas, Drewitz, Uwe, and Jipp, Meike
- Subjects
FACIAL expression ,AFFECTIVE computing ,IN-vehicle computing ,FEAR ,EMOTIONS ,FACIAL expression & emotions (Psychology) - Abstract
Facial expressions are one of the commonly used implicit measurements for the in-vehicle affective computing. However, the time courses and the underlying mechanism of facial expressions so far have been barely focused on. According to the Component Process Model of emotions, facial expressions are the result of an individual's appraisals, which are supposed to happen in sequence. Therefore, a multidimensional and dynamic analysis of drivers' fear by using facial expression data could profit from a consideration of these appraisals. A driving simulator experiment with 37 participants was conducted, in which fear and relaxation were induced. It was found that the facial expression indicators of high novelty and low power appraisals were significantly activated after a fear event (high novelty: Z = 2.80, p < 0.01, r
contrast = 0.46; low power: Z = 2.43, p < 0.05, rcontrast = 0.50). Furthermore, after the fear event, the activation of high novelty occurred earlier than low power. These results suggest that multidimensional analysis of facial expression is suitable as an approach for the in-vehicle measurement of the drivers' emotions. Furthermore, a dynamic analysis of drivers' facial expressions considering of effects of appraisal components can add valuable information for the in-vehicle assessment of emotions. [ABSTRACT FROM AUTHOR]- Published
- 2021
- Full Text
- View/download PDF
33. A Novel Dual CNN Architecture with LogicMax for Facial Expression Recognition.
- Author
-
AHADIT, ALAGESAN BHUVANESWARI and JATOTH, RAVI KUMAR
- Subjects
FACIAL expression ,CONVOLUTIONAL neural networks ,EMOTIONS ,SIGNAL convolution ,DEEP learning ,INTRACLASS correlation - Abstract
Facial expressions convey important features for recognizing human emotions. It is a challenging task to classify accurate facial expressions due to high intra-class correlation. Conventional methods depend on the classification of handcrafted features like scale-invariant feature transform and local binary patterns to predict the emotion. In recent years, deep learning techniques are used to boost the accuracy of FER models. Although it has improved the accuracy in standard datasets, FER models have to consider problems like face occlusion and intra-class variance. In this paper, we have used two convolutional neural networks which have vgg16 architecture as a base network using transfer learning. This paper explains the method to tackle issues on classifying high intra-class correlated facial expressions through an in-depth investigation of the Facial Action Coding System (FACS) action units. We have used a novel LogicMax layer at the end of the model to boost the accuracy of the FER model. Classification metrics like Accuracy, Precision, Recall, and Fl score are calculated for evaluating the model performance on CK+ and JAFFE datasets. The model is tested using 10-fold cross-validation and the obtained classification accuracy rate of 98.62% and 94.86% on CK+ and JAFFE datasets respectively. The experimental results also include a feature map visualization of 64 convolutional filters of the two convolutional neural networks. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
34. SliderGAN: Synthesizing Expressive Face Images by Sliding 3D Blendshape Parameters.
- Author
-
Ververas, Evangelos and Zafeiriou, Stefanos
- Subjects
- *
CONVOLUTIONAL neural networks , *THREE-dimensional imaging , *FACIAL expression , *FACE - Abstract
Image-to-image (i2i) translation is the dense regression problem of learning how to transform an input image into an output using aligned image pairs. Remarkable progress has been made in i2i translation with the advent of deep convolutional neural networks and particular using the learning paradigm of generative adversarial networks (GANs). In the absence of paired images, i2i translation is tackled with one or multiple domain transformations (i.e., CycleGAN, StarGAN etc.). In this paper, we study the problem of image-to-image translation, under a set of continuous parameters that correspond to a model describing a physical process. In particular, we propose the SliderGAN which transforms an input face image into a new one according to the continuous values of a statistical blendshape model of facial motion. We show that it is possible to edit a facial image according to expression and speech blendshapes, using sliders that control the continuous values of the blendshape model. This provides much more flexibility in various tasks, including but not limited to face editing, expression transfer and face neutralisation, comparing to models based on discrete expressions or action units. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
35. The Facial Action Coding System for Characterization of Human Affective Response to Consumer Product-Based Stimuli: A Systematic Review.
- Author
-
Clark, Elizabeth A., Kessinger, J'Nai, Duncan, Susan E., Bell, Martha Ann, Lahne, Jacob, Gallagher, Daniel L., and O'Keefe, Sean F.
- Subjects
CONSUMER behavior ,META-analysis ,HUMAN behavior ,EMOTIONS ,EMOTIONAL competence ,FACIAL expression & emotions (Psychology) - Abstract
To characterize human emotions, researchers have increasingly utilized Automatic Facial Expression Analysis (AFEA), which automates the Facial Action Coding System (FACS) and translates the facial muscular positioning into the basic universal emotions. There is broad interest in the application of FACS for assessing consumer expressions as an indication of emotions to consumer product-stimuli. However, the translation of FACS to characterization of emotions is elusive in the literature. The aim of this systematic review is to give an overview of how FACS has been used to investigate human emotional behavior to consumer product-based stimuli. The search was limited to studies published in English after 1978, conducted on humans, using FACS or its action units to investigate affect, where emotional response is elicited by consumer product-based stimuli evoking at least one of the five senses. The search resulted in an initial total of 1,935 records, of which 55 studies were extracted and categorized based on the outcomes of interest including (i) method of FACS implementation; (ii) purpose of study; (iii) consumer product-based stimuli used; and (iv) measures of affect validation. Most studies implemented FACS manually (73%) to develop products and/or software (20%) and used consumer product-based stimuli that had known and/or defined capacity to evoke a particular affective response, such as films and/or movie clips (20%); minimal attention was paid to consumer products with low levels of emotional competence or with unknown affective impact. The vast majority of studies (53%) did not validate FACS-determined affect and, of the validation measures that were used, most tended to be discontinuous in nature and only captured affect as it holistically related to an experience. This review illuminated some inconsistencies in how FACS is carried out as well as how emotional response is inferred from facial muscle activation. This may prompt researchers to consider measuring the total consumer experience by employing a variety of methodologies in addition to FACS and its emotion-based interpretation guide. Such strategies may better conceptualize consumers' experience with products of low, unknown, and/or undefined capacity to evoke an affective response such as product prototypes, line extensions, etc. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
36. Multi-modal Affect Detection Using Thermal and Optical Imaging in a Gamified Robotic Exercise
- Author
-
Mohamed, Youssef, Güneysu Özgür, Arzu, Lemaignan, Séverin, Leite, Iolanda, Mohamed, Youssef, Güneysu Özgür, Arzu, Lemaignan, Séverin, and Leite, Iolanda
- Abstract
Affect recognition, or the ability to detect and interpret emotional states, has the potential to be a valuable tool in the field of healthcare. In particular, it can be useful in gamified therapy, which involves using gaming techniques to motivate and keep the engagement of patients in therapeutic activities. This study aims to examine the accuracy of machine learning models using thermal imaging and action unit data for affect classification in a gamified robot therapy scenario. A self-report survey and three machine learning models were used to assess emotions including frustration, boredom, and enjoyment in participants during different phases of the game. The results showed that the multimodal approach with the combination of thermal imaging and action units with LSTM model had the highest accuracy of 77% for emotion classification over a 7-s sliding window, while thermal imaging had the lowest standard deviation among participants. The results suggest that thermal imaging and action units can be effective in detecting affective states and might have the potential to be used in healthcare applications, such as gamified therapy, as a promising non-intrusive method for recognizing internal states., QC 20240705
- Published
- 2023
- Full Text
- View/download PDF
37. Face image synthesis for robust facial analysis
- Author
-
Marinos, Marios (author) and Marinos, Marios (author)
- Abstract
Emotion recognition is a challenging problem in the field of computer vision. The automatic classification of emotions using facial expressions is a promising approach to understand human behavior in various applications such as marketing, health, and education. How- ever, recognizing some emotions, such as anger, jealousy, contempt, and disgust, is more challenging than others due to their subtlety and rarity in the training data. In this paper, we try to investigate if using (self)pseudo-labelled data to train an Expression Manipulator [? ] generator to generate a training set for training a classifier is a better alternative to directly using an equal amount of (self)pseudo-labelled data for training the classifier [? ]. Specifically, we focus on augmenting the Action Units (AUs) of facial expressions, which are the basic units of facial movement that correspond to specific emotions, Computer Science
- Published
- 2023
38. Facial Expression Recognition Using Computer Vision: A Systematic Review.
- Author
-
Canedo, Daniel and Neves, António J. R.
- Subjects
FACIAL expression ,HUMAN facial recognition software ,META-analysis ,EMOTION recognition ,OPTICAL flow ,MULTIMODAL user interfaces ,COMPUTER vision - Abstract
Emotion recognition has attracted major attention in numerous fields because of its relevant applications in the contemporary world: marketing, psychology, surveillance, and entertainment are some examples. It is possible to recognize an emotion through several ways; however, this paper focuses on facial expressions, presenting a systematic review on the matter. In addition, 112 papers published in ACM, IEEE, BASE and Springer between January 2006 and April 2019 regarding this topic were extensively reviewed. Their most used methods and algorithms will be firstly introduced and summarized for a better understanding, such as face detection, smoothing, Principal Component Analysis (PCA), Local Binary Patterns (LBP), Optical Flow (OF), Gabor filters, among others. This review identified a clear difficulty in translating the high facial expression recognition (FER) accuracy in controlled environments to uncontrolled and pose-variant environments. The future efforts in the FER field should be put into multimodal systems that are robust enough to face the adversities of real world scenarios. A thorough analysis on the research done on FER in Computer Vision based on the selected papers is presented. This review aims to not only become a reference for future research on emotion recognition, but also to provide an overview of the work done in this topic for potential readers. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
39. 3D Approaches and Challenges in Facial Expression Recognition Algorithms—A Literature Review.
- Author
-
Nonis, Francesca, Dagnes, Nicole, Marcolin, Federica, and Vezzetti, Enrico
- Subjects
FACIAL expression ,HUMAN facial recognition software ,FACIAL expression & emotions (Psychology) ,HUMAN-computer interaction ,SCIENTIFIC community - Abstract
In recent years, facial expression analysis and recognition (FER) have emerged as an active research topic with applications in several different areas, including the human-computer interaction domain. Solutions based on 2D models are not entirely satisfactory for real-world applications, as they present some problems of pose variations and illumination related to the nature of the data. Thanks to technological development, 3D facial data, both still images and video sequences, have become increasingly used to improve the accuracy of FER systems. Despite the advance in 3D algorithms, these solutions still have some drawbacks that make pure three-dimensional techniques convenient only for a set of specific applications; a viable solution to overcome such limitations is adopting a multimodal 2D+3D analysis. In this paper, we analyze the limits and strengths of traditional and deep-learning FER techniques, intending to provide the research community an overview of the results obtained looking to the next future. Furthermore, we describe in detail the most used databases to address the problem of facial expressions and emotions, highlighting the results obtained by the various authors. The different techniques used are compared, and some conclusions are drawn concerning the best recognition rates achieved. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
40. Human Observers and Automated Assessment of Dynamic Emotional Facial Expressions: KDEF-dyn Database Validation
- Author
-
Manuel G. Calvo, Andrés Fernández-Martín, Guillermo Recio, and Daniel Lundqvist
- Subjects
facial expression ,dynamic ,action units ,KDEF ,FACET ,Psychology ,BF1-990 - Abstract
Most experimental studies of facial expression processing have used static stimuli (photographs), yet facial expressions in daily life are generally dynamic. In its original photographic format, the Karolinska Directed Emotional Faces (KDEF) has been frequently utilized. In the current study, we validate a dynamic version of this database, the KDEF-dyn. To this end, we applied animation between neutral and emotional expressions (happy, sad, angry, fearful, disgusted, and surprised; 1,033-ms unfolding) to 40 KDEF models, with morphing software. Ninety-six human observers categorized the expressions of the resulting 240 video-clip stimuli, and automated face analysis assessed the evidence for 6 expressions and 20 facial action units (AUs) at 31 intensities. Low-level image properties (luminance, signal-to-noise ratio, etc.) and other purely perceptual factors (e.g., size, unfolding speed) were controlled. Human recognition performance (accuracy, efficiency, and confusions) patterns were consistent with prior research using static and other dynamic expressions. Automated assessment of expressions and AUs was sensitive to intensity manipulations. Significant correlations emerged between human observers’ categorization and automated classification. The KDEF-dyn database aims to provide a balance between experimental control and ecological validity for research on emotional facial expression processing. The stimuli and the validation data are available to the scientific community.
- Published
- 2018
- Full Text
- View/download PDF
41. Intelligent Sensors for Human Motion Analysis.
- Author
-
Krzeszowski, Tomasz, Calafate, Carlos Tavares, Kepski, Michal, Krzeszowski, Tomasz, and Świtoński, Adam
- Subjects
History of engineering & technology ,Technology: general issues ,3D human mesh reconstruction ,3D human pose estimation ,3D multi-person pose estimation ,Azure Kinect ,BILSTM ,Berg Balance Scale ,COVID-19 ,EMG ,F-Formation ,FFNN ,FMCW ,GRU ,Kinect v2 ,LSTM ,MFCC ,RGB-D sensors ,XGBoost ,Zed 2i ,absolute poses ,action units ,aggregation function ,anomaly detection ,artifact classification ,artifact detection ,artificial intelligence ,assessment ,balance ,biometrics ,camera-centric coordinates ,computer vision ,convolutional neural networks ,cyber-physical systems ,data augmentation ,deep learning ,deep neural network ,deep-learning ,development ,diagnosis ,elderly ,facial expression recognition ,facial landmarks ,fall risk detection ,features fusion ,features selection ,fuzzy inference ,gait analysis ,gait parameters ,gait recognition ,gap filling ,generalization ,graph convolutional networks ,grey wolf optimization ,human action recognition ,human motion analysis ,human motion modelling ,human tracking ,kinematics ,knowledge measure ,machine learning ,markerless ,markerless motion capture ,modular sensing unit ,motion capture ,movement tracking ,n/a ,neural networks ,optical sensing principle ,particle swarm optimization ,pattern recognition ,plantar pressure measurement ,pose estimation ,posture detection ,precedence indicator ,recognition ,reconstruction ,regularized discriminant analysis ,robot ,rule induction ,skeletal data ,socially occupied space ,telemedicine ,time series classification ,vital sign ,whale optimization algorithm - Abstract
Summary: The book, "Intelligent Sensors for Human Motion Analysis," contains 17 articles published in the Special Issue of the Sensors journal. These articles deal with many aspects related to the analysis of human movement. New techniques and methods for pose estimation, gait recognition, and fall detection have been proposed and verified. Some of them will trigger further research, and some may become the backbone of commercial systems.
42. Entropy in Real-World Datasets and Its Impact on Machine Learning.
- Author
-
Kozak, Jan, Juszczuk, Przemysław, and Kozak, Jan
- Subjects
Computer science ,Information technology industries ,ADF ,ARMA ,CEEMDAN ,COVID-19 ,DES ,LSTM ,Pawlak conflict analysis model ,action units ,agent-based modelling ,association rules ,automatic translation ,bike-sharing ,classification ,classification measure ,coalitions ,decision rules ,decision table ,decision tables ,decision tree ,decision trees ,differential cryptanalysis ,dispersed data ,distributed data ,dynamic stochastic general equilibrium models ,entropy measure ,entropy of real data ,fast iterative filtering ,fault diagnosis ,feature selection ,greedy heuristics ,hybrid TCN-GRU model ,hybrid model ,imbalanced data ,independent data sources ,information systems ,length ,machine learning ,memetic algorithms ,metaheuristics ,one-class classification ,optical networks ,parameter adaptive refined composite multiscale fluctuation-based dispersion entropy ,preference-driven classification ,preprocessing ,quality measure ,quality of classification ,query set ,real-world data ,reducts ,rotating machinery ,rough sets ,scenario analyses ,short-term demand prediction ,sign language ,simulated annealing ,stock index forecasting ,support ,symmetric block ciphers ,tests ,travel characteristics analysis ,vaccination ,validation of results - Abstract
Summary: The topic of the reprint is very important nowadays, because ever-evolving machine learning techniques make it possible to obtain better real-world data. Therefore, this reprint contains information related to real data in fields such as automatic sign language translation, bike-sharing travel characteristics, stock index, sports data, fake news data, and more. However, it should be noted that the reprint also contains a lot of information on new developments in machine learning, new algorithms, algorithm modifications, and a new measure of classification quality assessment that also takes into account the preferences of the decision maker.
43. Facial Expression Recognition Using Computer Vision: A Systematic Review
- Author
-
Daniel Canedo and António J. R. Neves
- Subjects
facial expression recognition ,emotion recognition ,computer vision ,machine learning ,action units ,deep learning ,facial features ,review article ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Emotion recognition has attracted major attention in numerous fields because of its relevant applications in the contemporary world: marketing, psychology, surveillance, and entertainment are some examples. It is possible to recognize an emotion through several ways; however, this paper focuses on facial expressions, presenting a systematic review on the matter. In addition, 112 papers published in ACM, IEEE, BASE and Springer between January 2006 and April 2019 regarding this topic were extensively reviewed. Their most used methods and algorithms will be firstly introduced and summarized for a better understanding, such as face detection, smoothing, Principal Component Analysis (PCA), Local Binary Patterns (LBP), Optical Flow (OF), Gabor filters, among others. This review identified a clear difficulty in translating the high facial expression recognition (FER) accuracy in controlled environments to uncontrolled and pose-variant environments. The future efforts in the FER field should be put into multimodal systems that are robust enough to face the adversities of real world scenarios. A thorough analysis on the research done on FER in Computer Vision based on the selected papers is presented. This review aims to not only become a reference for future research on emotion recognition, but also to provide an overview of the work done in this topic for potential readers.
- Published
- 2019
- Full Text
- View/download PDF
44. 3D Approaches and Challenges in Facial Expression Recognition Algorithms—A Literature Review
- Author
-
Francesca Nonis, Nicole Dagnes, Federica Marcolin, and Enrico Vezzetti
- Subjects
facial expression recognition ,3D face analysis ,deep learning-based FER ,2D/3D comparison ,facial action coding system ,action units ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
In recent years, facial expression analysis and recognition (FER) have emerged as an active research topic with applications in several different areas, including the human-computer interaction domain. Solutions based on 2D models are not entirely satisfactory for real-world applications, as they present some problems of pose variations and illumination related to the nature of the data. Thanks to technological development, 3D facial data, both still images and video sequences, have become increasingly used to improve the accuracy of FER systems. Despite the advance in 3D algorithms, these solutions still have some drawbacks that make pure three-dimensional techniques convenient only for a set of specific applications; a viable solution to overcome such limitations is adopting a multimodal 2D+3D analysis. In this paper, we analyze the limits and strengths of traditional and deep-learning FER techniques, intending to provide the research community an overview of the results obtained looking to the next future. Furthermore, we describe in detail the most used databases to address the problem of facial expressions and emotions, highlighting the results obtained by the various authors. The different techniques used are compared, and some conclusions are drawn concerning the best recognition rates achieved.
- Published
- 2019
- Full Text
- View/download PDF
45. Automatic Frustration Detection Using Thermal Imaging
- Author
-
Mohamed, Youssef, Ballardini, Giulia, Parreira, Maria Teresa, Lemaignan, Severin, Leite, Iolanda, Mohamed, Youssef, Ballardini, Giulia, Parreira, Maria Teresa, Lemaignan, Severin, and Leite, Iolanda
- Abstract
To achieve seamless interactions, robots have to be capable of reliably detecting affective states in real time. One of the possible states that humans go through while interacting with robots is frustration. Detecting frustration from RGB images can be challenging in some real-world situations; thus, we investigate in this work whether thermal imaging can be used to create a model that is capable of detecting frustration induced by cognitive load and failure. To train our model, we collected a data set from 18 participants experiencing both types of frustration induced by a robot. The model was tested using features from several modalities: thermal, RGB, Electrodermal Activity (EDA), and all three combined. When data from both frustration cases were combined and used as training input, the model reached an accuracy of 89% with just RGB features, 87% using only thermal features, 84% using EDA, and 86% when using all modalities. Furthermore, the highest accuracy for the thermal data was reached using three facial regions of interest: nose, forehead and lower lip., Part of proceedings: ISBN 978-1-6654-0731-1QC 20221216
- Published
- 2022
- Full Text
- View/download PDF
46. Faces of pain in dementia: Learnings from a real-world study using a technology-enabled pain assessment tool
- Author
-
Atee, Mustafa, Hoti, Kreshnik, Chivers, Paola, Hughes, Jeffery D., Atee, Mustafa, Hoti, Kreshnik, Chivers, Paola, and Hughes, Jeffery D.
- Abstract
Pain is common in people living with dementia (PLWD), including those with limited verbal skills. Facial expressions are key behavioral indicators of the pain experience in this group. However, there is a lack of real-world studies to report the prevalence and associations of pain-relevant facial micro-expressions in PLWD. In this observational retrospective study, pain-related facial features were studied in a sample of 3,144 PLWD [mean age 83.3 years (SD = 9.0); 59.0% female] using the Face domain of PainChek®, a point-of-care medical device application. Pain assessments were completed by 389 users from two national dementia-specific care programs and 34 Australian aged care homes. Our analysis focused on the frequency, distribution, and associations of facial action units [AU(s)] with respect to various pain intensity groups. A total of 22,194 pain assessments were completed. Of the AUs present, AU7 (eyelid tightening) was the most frequent facial expression (48.6%) detected, followed by AU43 (closing eyes; 42.9%) and AU6 (cheek raising; 42.1%) during severe pain. AU20 (horizontal mouth stretch) was the most predictive facial action of higher pain scores. Eye-related AUs (AU6, AU7, AU43) and brow-related AUs (AU4) were more common than mouth-related AUs (e.g., AU20, AU25) during higher pain intensities. No significant effect was found for age or gender. These findings offer further understanding of facial expressions during clinical pain in PLWD and confirm the usefulness of artificial intelligence (AI)-enabled real-time analysis of the face as part of the assessment of pain in aged care clinical practice. Copyright © 2022 Atee, Hoti, Chivers and Hughes.
- Published
- 2022
47. Non-rigid registration based model-free 3D facial expression recognition.
- Author
-
Savran, Arman and Sankur, Bülent
- Subjects
HUMAN facial recognition software ,THREE-dimensional imaging ,PATTERN recognition systems ,FEATURE extraction ,INFORMATION theory - Abstract
We propose a novel feature extraction approach for 3D facial expression recognition by incorporating non-rigid registration in face-model-free analysis, which in turn makes feasible data-driven, i.e., feature-model-free recognition of expressions. The resulting simplicity of feature representation is due to the fact that facial information is adapted to the input faces via shape model-free dense registration, and this provides a dynamic feature extraction mechanism. This approach eliminates the necessity of complex feature representations as required in the case of static feature extraction methods, where the complexity arises from the necessity to model the local context; higher degree of complexity persists in deep feature hierarchies enabled by end-to-end learning on large-scale datasets. Face-model-free recognition implies independence from limitations and biases due to committed face models, bypassing complications of model fitting, and avoiding the burden of manual model construction. We show via information gain maps that non-rigid registration enables extraction of highly informative features, as it provides invariance to local-shifts due to physiognomy (subject invariance) and residual pose misalignments; in addition, it allows estimation of local correspondences of expressions. To maximize the recognition rate, we use the strategy of employing a rich but computationally manageable set of local correspondence structures, and to this effect we propose a framework to optimally select multiple registration references. Our features are re-sampled surface curvature values at individual coordinates which are chosen per expression-class and per reference pair. We show the superior performance of our novel dynamic feature extraction approach on three distinct recognition problems, namely, action unit detection, basic expression recognition, and emotion dimension recognition. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
48. Emotion recognition using facial expressions.
- Author
-
Tarnowski, Paweł, Kołodziej, Marcin, Majkowski, Andrzej, and Rak, Remigiusz J.
- Subjects
EMOTION recognition ,FACIAL expression ,FEATURE selection ,ARTIFICIAL neural networks ,NUMERICAL calculations - Abstract
In the article there are presented the results of recognition of seven emotional states (neutral, joy, sadness, surprise, anger, fear, disgust) based on facial expressions. Coefficients describing elements of facial expressions, registered for six subjects, were used as features. The features have been calculated for three-dimensional face model. The classification of features were performed using k-NN classifier and MLP neural network. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
49. Studying Facial Activity in Parkinson's Disease Patients Using an Automated Method and Video Recording
- Author
-
Alexander Volkov, Ekaterina Ivanova, Andrey Samorodov, Natalia Voinova, Anastasia Moshkova, and Ekaterina Yu. Fedotova
- Subjects
Video recording ,medicine.medical_specialty ,Parkinson's disease ,business.industry ,Eyebrow ,Video camera ,Kinematics ,medicine.disease ,Intensity (physics) ,law.invention ,lcsh:Telecommunication ,Physical medicine and rehabilitation ,medicine.anatomical_structure ,law ,Facial activity ,hypomimia ,action units ,lcsh:TK5101-6720 ,parkinson's disease ,medicine ,business ,Automated method - Abstract
Main objective of this research is studying facial activity of PD patients using an automated method for assessing facial movements based on calculating the movement signal kinematic characteristics. 16 PD patients and 16 participants in the control group have been recorded using 2D video camera while performing 4 facial tests: closing the eyes, raising the eyebrows, smiling with an effort (grin), moving the eyebrows. Each test includes a 10-fold repetition of a given facial movement with maximum effort and velocity. Intensities of the corresponding action units (AUs) in each frame have been determined by an automated method for each test. The resulting signal of the AU intensity dependence on the frame number has been marked with maximum and minimum points. From the extremum points, 11 kinematic parameters have been calculated for each AU of each patient. To assess the differences in kinematic parameters between the groups, the nonparametric Mann-Whitney test has been used. There is decrease in the facial movements frequency and velocity. Consequently, there is increase in the duration of the performance of facial tests in PD patients group compared to the control group. Only some AUs also showed a decrease in the amplitude of facial movements in PD patients group compared to the control group. Differences in the largest number of kinematic parameters have been obtained for the action units AU12 and AU14 recorded during the smile with effort test, and AU04 recorded during the eyebrow moving test.
- Published
- 2021
50. Privacy-Preserving and Scalable Affect Detection in Online Synchronous Learning
- Author
-
Böttger, Felix, Cetinkaya, Ufuk, Mitri, Daniele Di, Gombert, Sebastian, Shingjergji, Krist, Iren, Deniz, Klemke, Roland, Hiliger, Isabel, Muñoz-Merino, Pedro J., De Laet, Tinne, Ortega-Arranz, Alejandro, Farrell, Tracie, RS-Research Program Educational research on activating (online) education (ERA), Department of Technology Enhanced Learning and Innovation, Department of Information Science, RS-Research Program Learning and Innovation in Resilient systems (LIRS), RS-Research Line Technology Enhanced Learning and Innovation (part of ERA program), Hilliger, Isabel, Muñoz-Merino, Pedro J., De Laet, Tinne, Ortega-Arranz, Alejandro, and Farrell, Tracie
- Subjects
Privacy ,Affect detection ,Action units ,Emotion recognition - Abstract
The recent pandemic has forced most educational institutions to shift to distance learning. Teachers can perceive various non-verbal cues in face-to-face classrooms and thus notice when students are distracted, confused, or tired. However, the students’ non-verbal cues are not observable in online classrooms. The lack of these cues poses a challenge for the teachers and hinders them in giving adequate, timely feedback in online educational settings. This can lead to learners not receiving proper guidance and may cause them to be demotivated. This paper proposes a pragmatic approach to detecting student affect in online synchronized learning classrooms. Our approach consists of a method and a privacy-preserving prototype that only collects data that is absolutely necessary to compute action units and is highly scalable by design to run on multiple devices without specialized hardware. We evaluated our prototype using a benchmark for the system performance. Our results confirm the feasibility and the applicability of the proposed approach.
- Published
- 2022
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.