222 results on '"action units"'
Search Results
2. AU-vMAE: Knowledge-Guide Action Units Detection via Video Masked Autoencoder
- Author
-
Jin, Qiaoqiao, Shi, Rui, Dou, Yishun, Ni, Bingbing, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Lin, Zhouchen, editor, Cheng, Ming-Ming, editor, He, Ran, editor, Ubul, Kurban, editor, Silamu, Wushouer, editor, Zha, Hongbin, editor, Zhou, Jie, editor, and Liu, Cheng-Lin, editor
- Published
- 2025
- Full Text
- View/download PDF
3. Characteristics of vocal cues, facial action units, and emotions that distinguish high from low self-protection participants engaged in self-protective response to self-criticizing.
- Author
-
Vráblová, Viktória and Halamová, Júlia
- Subjects
EMOTION-focused therapy ,SINGING instruction ,MULTILEVEL models ,EMOTIONS ,MENTAL health - Abstract
Introduction: Self-protection, also called protective anger or assertive anger, is a key factor in mental health. Thus, far, researchers have focused mainly on the qualitative analysis of self-protection. Methods: Therefore, we investigated facial action units, emotions, and vocal cues in low and high self-protective groups of participants in order to detect any differences. The total sample consisted of 239 participants. Using the Performance factor in the Short version of the Scale for Interpersonal Behavior (lower 15th percentile and upper 15th percentile) we selected 33 high self-protective participants (11 men, 22 women) and 25 low self-protective participants (eight men, 17 women). The self-protective dialogue was recorded using the two-chair technique script from Emotion Focused Therapy. The subsequent analysis was performed using iMotions software (for action units and emotions) and Praat software (for vocal cues of pitch and intensity). We used multilevel models in program R for the statistical analysis. Results: Compared to low self-protective participants, high self-protective participants exhibited more contempt and fear and less surprise and joy. Compared to low self-protective participants, high self-protective participants expressed the action units the following action units less often: Mouth Open (AU25), Smile (AU12), Brow Raise (AU2), Cheek Raise (AU6), Inner Brow Raise (AU1), and more often Brow Furrow (AU4), Chin Raise (AU17), Smirk (AU12), Upper Lip Raise (AU10), and Nose Wrinkle (AU9). We found no differences between the two groups in the use of vocal cues. Discussion: These findings bring us closer to understanding and diagnosing self-protection. [ABSTRACT FROM AUTHOR]
- Published
- 2025
- Full Text
- View/download PDF
4. ViTAU: Facial paralysis recognition and analysis based on vision transformer and facial action units
- Author
-
Jia GAO, Wenhao CAI, Junli ZHAO, and Fuqing DUAN
- Subjects
transformer ,action units ,multi-resolution feature maps ,generator ,heatmap regression ,Mining engineering. Metallurgy ,TN1-997 ,Environmental engineering ,TA170-171 - Abstract
Facial nerve paralysis (FNP), commonly known as Bell’s palsy or facial paralysis, significantly affects patients’ daily lives and mental well-being. Timely identification and diagnosis are crucial for early treatment and rehabilitation. With the rapid advancement of deep learning and computer vision technologies, automatic recognition of facial paralysis has become feasible, offering a more accurate and objective diagnostic approach. Current research primarily focuses on broad facial changes and often neglects finer facial details, which leads to insufficient analysis of how different areas affect recognition results. This study proposes an innovative method that combines the vision transformer (ViT) model with an action unit (AU) facial region detection network to automatically recognize and analyze facial paralysis. Initially, the ViT model utilizes its self-attention mechanism to accurately determine the presence of facial paralysis. Subsequently, we analyzed the AU data to assess the activity of facial muscles, allowing for a deeper evaluation of the affected areas. The self-attention mechanism in the transformer architecture captures the global contextual information required to recognize facial paralysis. To accurately determine the specific affected regions, we use the pixel2style2pixel (pSp) encoder and the StyleGAN2 generator to encode and decode images and extract feature maps that represent facial characteristics. These maps are then processed through a pyramid convolutional neural network interpreter to generate heatmaps. By optimizing the mean squared error between the predicted and actual heatmaps, we can effectively identify the affected paralysis areas. Our proposed method integrates ViT with facial AUs, designing a ViT-based facial paralysis recognition network that enhances the extraction of local area features through its self-attention mechanism, thereby enabling precise recognition of facial paralysis. Additionally, by incorporating facial AU data, we conducted detailed regional analyses for patients identified with facial paralysis. Experimental results demonstrate the efficacy of our approach, achieving a recognition accuracy of 99.4% for facial paralysis and 81.36% for detecting affected regions on the YouTube Facial Palsy (YFP) and extended Cohn Kanade (CK+) datasets. These results not only highlight the effectiveness of our automatic recognition method compared to the latest techniques but also validate its potential for clinical applications. Furthermore, to facilitate the observation of affected regions, we developed a visualization method that intuitively displays the impacted areas, thereby aiding patients and healthcare professionals in understanding the condition and enhancing communication regarding treatment and rehabilitation strategies. In conclusion, the proposed method illustrates the power of combining advanced deep learning techniques with a detailed analysis of facial AUs to improve the automatic recognition of facial paralysis. By addressing previous research limitations, the proposed method provides a more nuanced understanding of how specific facial areas are affected, leading to improved diagnosis, treatment, and patient care. This innovative approach not only enhances the accuracy of facial paralysis detection but also contributes to facial medical imaging.
- Published
- 2025
- Full Text
- View/download PDF
5. Employing the Artificial Intelligence Object Detection Tool YOLOv8 for Real-Time Pain Detection: A Feasibility Study
- Author
-
Cascella M, Shariff MN, Lo Bianco G, Monaco F, Gargano F, Simonini A, Ponsiglione AM, and Piazza O
- Subjects
pain ,artificial intelligence ,automatic pain assessment ,computer vision ,action units ,Medicine (General) ,R5-920 - Abstract
Marco Cascella,1 Mohammed Naveed Shariff,2 Giuliano Lo Bianco,3 Federica Monaco,4 Francesca Gargano,5 Alessandro Simonini,6 Alfonso Maria Ponsiglione,7 Ornella Piazza1 1Anesthesia and Pain Medicine, Department of Medicine, Surgery and Dentistry “scuola Medica Salernitana”, University of Salerno, Baronissi, 84081, Italy; 2Department of AI&DS, Rajalakshmi Institute of Technology, Chennai, TN, India; 3Anesthesiology and Pain Department, Fondazione Istituto G. Giglio Cefalù, Palermo, Italy; 4Anesthesia and Pain Medicine, ASL NA1, Napoli, Italy; 5Anesthesia and Intensive Care, U.O.C. Fondazione Policlinico Campus Bio-Medico, Roma, Italy; 6Pediatric Anesthesia and Intensive Care Unit, Salesi Children’s Hospital, Ancona, Italy; 7Department of Electrical Engineering and Information Technology, University of Naples “federico II”, Naples, 0125, ItalyCorrespondence: Giuliano Lo Bianco, Anesthesiology and Pain Department, Fondazione Istituto G. Giglio Cefalù, Palermo, Italy, Email giulianolobianco@gmail.comIntroduction: Effective pain management is crucial for patient care, impacting comfort, recovery, and overall well-being. Traditional subjective pain assessment methods can be challenging, particularly in specific patient populations. This research explores an alternative approach using computer vision (CV) to detect pain through facial expressions.Methods: The study implements the YOLOv8 real-time object detection model to analyze facial expressions indicative of pain. Given four pain datasets, a dataset of pain-expressing faces was compiled, and each image was carefully labeled based on the presence of pain-associated Action Units (AUs). The labeling distinguished between two classes: pain and no pain. The pain category included specific AUs (AU4, AU6, AU7, AU9, AU10, and AU43) following the Prkachin and Solomon Pain Intensity (PSPI) scoring method. Images showing these AUs with a PSPI score above 2 were labeled as expressing pain. The manual labeling process utilized an open-source tool, makesense.ai, to ensure precise annotation. The dataset was then split into training and testing subsets, each containing a mix of pain and no-pain images. The YOLOv8 model underwent iterative training over 10 epochs. The model’s performance was validated using precision, recall, and mean Average Precision (mAP) metrics, and F1 score.Results: When considering all classes collectively, our model attained a mAP of 0.893 at a threshold of 0.5. The precision for “pain” and “nopain” detection was 0.868 and 0.919, respectively. F1 scores for the classes “pain”, “nopain”, and “all classes” reached a peak value of 0.80. Finally, the model was tested on the Delaware dataset and in a real-world scenario.Discussion: Despite limitations, this study highlights the promise of using real-time computer vision models for pain detection, with potential applications in clinical settings. Future research will focus on evaluating the model’s generalizability across diverse clinical scenarios and its integration into clinical workflows to improve patient care.Keywords: pain, artificial intelligence, automatic pain assessment, computer vision, action units
- Published
- 2024
6. Advanced techniques for automated emotion recognition in dogs from video data through deep learning.
- Author
-
Franzoni, Valentina, Biondi, Giulio, and Milani, Alfredo
- Subjects
- *
AFFECTIVE computing , *EMOTIONS in animals , *EMOTION recognition , *ARTIFICIAL intelligence , *DOG behavior , *MIRROR neurons , *DEEP learning - Abstract
Inter-species emotional relationships, particularly the symbiotic interaction between humans and dogs, are complex and intriguing. Humans and dogs share fundamental mammalian neural mechanisms including mirror neurons, crucial to empathy and social behavior. Mirror neurons are activated during the execution and observation of actions, indicating inherent connections in social dynamics across species despite variations in emotional expression. This study explores the feasibility of using deep-learning Artificial Intelligence systems to accurately recognize canine emotions in general environments, to assist individuals without specialized knowledge or skills in discerning dog behavior, particularly related to aggression or friendliness. Starting with identifying key challenges in classifying pleasant and unpleasant emotions in dogs, we tested advanced deep-learning techniques and aggregated results to distinguish potentially dangerous human--dog interactions. Knowledge transfer is used to fine-tune different networks, and results are compared on original and transformed sets of frames from the Dog Clips dataset to investigate whether DogFACS action codes detailing relevant dog movements can aid the emotion recognition task. Elaborating on challenges and biases, we emphasize the need for bias mitigation to optimize performance, including different image preprocessing strategies for noise mitigation in dog recognition (i.e., face bounding boxes, segmentation of the face or body, isolating the dog on a white background, blurring the original background). Systematic experimental results demonstrate the system's capability to accurately detect emotions and effectively identify dangerous situations or signs of discomfort in the presence of humans. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Using computer vision of facial expressions to assess symptom domains and treatment response in antipsychotic‐naïve patients with first‐episode psychosis.
- Author
-
Ambrosen, Karen S., Lemvigh, Cecilie K., Nielsen, Mette Ø., Glenthøj, Birte Y., Syeda, Warda T., and Ebdrup, Bjørn H.
- Subjects
- *
FACIAL expression , *COMPUTER vision , *PATHOLOGICAL psychology , *VIDEO recording , *AMISULPRIDE - Abstract
Background Method Results Conclusion Facial expressions are a core aspect of non‐verbal communication. Reduced emotional expressiveness of the face is a common negative symptom of schizophrenia, however, quantifying negative symptoms can be clinically challenging and involves a considerable element of rater subjectivity. We used computer vision to investigate if (i) automated assessment of facial expressions captures negative as well as positive and general symptom domains, and (ii) if automated assessments are associated with treatment response in initially antipsychotic‐naïve patients with first‐episode psychosis.We included 46 patients (mean age 25.4 (6.1); 65.2% males). Psychopathology was assessed at baseline and after 6 weeks of monotherapy with amisulpride using the Positive and Negative Syndrome Scale (PANSS). Baseline interview videos were recorded. Seventeen facial action units (AUs), that is, activation of muscles, from the Facial Action Coding System were extracted using OpenFace 2.0. A correlation matrix was calculated for each patient. Facial expressions were identified using spectral clustering at group‐level. Associations between facial expressions and psychopathology were investigated using multiple linear regression.Three clusters of facial expressions were identified related to different locations of the face. Cluster 1 was associated with positive and general symptoms at baseline, Cluster 2 was associated with all symptom domains, showing the strongest association with the negative domain, and Cluster 3 was only associated with general symptoms. Cluster 1 was significantly associated with the clinically rated improvement in positive and general symptoms after treatment, and Cluster 2 was significantly associated with clinical improvement in all domains.Using automated computer vision of facial expressions during PANSS interviews did not only capture negative symptoms but also combinations of the three overall domains of psychopathology. Moreover, automated assessments of facial expressions at baseline were associated with initial antipsychotic treatment response. The findings underscore the clinical relevance of facial expressions and motivate further investigations of computer vision in clinical psychiatry. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Demystifying Mental Health by Decoding Facial Action Unit Sequences.
- Author
-
Sharma, Deepika, Singh, Jaiteg, Sehra, Sukhjit Singh, and Sehra, Sumeet Kaur
- Subjects
EMOTION recognition ,CONVOLUTIONAL neural networks ,FACIAL expression ,MENTAL health ,EMOTIONS ,ANGER management - Abstract
Mental health is indispensable for effective daily functioning and stress management. Facial expressions may provide vital clues about the mental state of a person as they are universally consistent across cultures. This study intends to detect the emotional variances through facial micro-expressions using facial action units (AUs) to identify probable mental health issues. In addition, convolutional neural networks (CNN) were used to detect and classify the micro-expressions. Further, combinations of AUs were identified for the segmentation of micro-expressions classes using K-means square. Two benchmarked datasets CASME II and SAMM were employed for the training and evaluation of the model. The model achieved an accuracy of 95.62% on CASME II and 93.21% on the SAMM dataset, respectively. Subsequently, a case analysis was done to identify depressive patients using the proposed framework and it attained an accuracy of 92.99%. This experiment revealed the fact that emotions like disgust, sadness, anger, and surprise are the prominent emotions experienced by depressive patients during communication. The findings suggest that leveraging facial action units for micro-expression detection offers a promising approach to mental health diagnostics. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Geometric Graph Representation With Learnable Graph Structure and Adaptive AU Constraint for Micro-Expression Recognition.
- Author
-
Wei, Jinsheng, Peng, Wei, Lu, Guanming, Li, Yante, Yan, Jingjie, and Zhao, Guoying
- Abstract
Micro-expression recognition (MER) holds significance in uncovering hidden emotions. Most works take image sequences as input and cannot effectively explore ME information because subtle ME-related motions are easily submerged in unrelated information. Instead, the facial landmark is a low-dimensional and compact modality, which achieves lower computational cost and potentially concentrates on ME-related movement features. However, the discriminability of facial landmarks for MER is unclear. Thus, this article investigates the contribution of facial landmarks and proposes a novel framework to efficiently recognize MEs with facial landmarks. First, a geometric two-stream graph network is constructed to aggregate the low-order and high-order geometric movement information from facial landmarks to obtain discriminative ME representation. Second, a self-learning fashion is introduced to automatically model the dynamic relationship between nodes even long-distance nodes. Furthermore, an adaptive action unit loss is proposed to reasonably build a strong correlation between landmarks, facial action units and MEs. Notably, this work provides a novel idea with much higher efficiency to promote MER, only utilizing graph-based geometric features. The experimental results demonstrate that the proposed method achieves competitive performance with a significantly reduced computational cost. Furthermore, facial landmarks significantly contribute to MER and are worth further study for high-efficient ME analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Time to retire F1-binary score for action unit detection.
- Author
-
Hinduja, Saurabh, Nourivandi, Tara, Cohn, Jeffrey F., and Canavan, Shaun
- Subjects
- *
FACIAL expression , *FACE perception , *TASK analysis , *CLASS actions - Abstract
Detecting action units is an important task in face analysis, especially in facial expression recognition. This is due, in part, to the idea that expressions can be decomposed into multiple action units. To evaluate systems that detect action units, F1-binary score is often used as the evaluation metric. In this paper, we argue that F1-binary score does not reliably evaluate these models due largely to class imbalance. Because of this, F1-binary score should be retired and a suitable replacement should be used. We justify this argument through a detailed evaluation of the negative influence of class imbalance on action unit detection. This includes an investigation into the influence of class imbalance in train and test sets and in new data (i.e., generalizability). We empirically show that F1-micro should be used as the replacement for F1-binary. • We show that AU base rates have a large influence on detection across different architectures. • We show how different evaluation metrics are impacted by AU base rates. • We argue that F1-binary should not be used, for AU detection, and that F1-micro should replace it. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Transformer embedded spectral-based graph network for facial expression recognition.
- Author
-
Jin, Xing, Song, Xulin, Wu, Xiyin, and Yan, Wenzhu
- Abstract
Deep graph convolution networks which exploit relations among facial muscle movements for facial expression recognition (FER) have achieved great success. Due to the limited receptive field, existing graph convolution operations are difficult to model long-range muscle movement relations which plays a crucial role in FER. To alleviate this issue, we introduce the transformer encoder into graph convolution networks, in which the vision transformer enables all facial muscle movements to interact in the global receptive field and model more complex relations. Specifically, we construct facial graph data by cropping regions of interest (ROIs) which are associated with facial action units, and each ROI is represented by the representation of hidden layers from deep auto-encoder. To effectively extract features from the constructed facial graph data, we propose a novel transformer embedded spectral-based graph convolution network (TESGCN), in which the transformer encoder is exploited to interact with complex relations among facial RIOs for FER. Compared to vanilla graph convolution networks, we empirically show the superiority of the proposed model by conducting extensive experiments across four facial expression datasets. Moreover, our proposed TESGCN only has 80K parameters and 0.41MB model size, and achieves comparable results compared to existing lightweight networks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. A Two-Tier GAN Architecture for Conditioned Expressions Synthesis on Categorical Emotions.
- Author
-
Lambiase, Paolo Domenico, Rossi, Alessandra, and Rossi, Silvia
- Subjects
GENERATIVE adversarial networks ,FACIAL expression & emotions (Psychology) ,ROBOTS ,EMOTIONS ,SOCIAL robots ,HUMAN-robot interaction ,FACIAL muscles - Abstract
Emotions are an effective communication mode during human–human and human–robot interactions. However, while humans can easily understand other people's emotions, and they are able to show emotions with natural facial expressions, robot-simulated emotions still represent an open challenge also due to a lack of naturalness and variety of possible expressions. In this direction, we present a two-tier Generative Adversarial Networks (GAN) architecture that generates facial expressions starting from categorical emotions (e.g. joy, sadness, etc.) to obtain a variety of synthesised expressions for each emotion. The proposed approach combines the key features of Conditional Generative Adversarial Networks (CGAN) and GANimation, overcoming their limits by allowing fine modelling of facial expressions, and generating a wide range of expressions for each class (i.e., discrete emotion). The architecture is composed of two modules for generating a synthetic Action Units (AU, i.e., a coding mechanism representing facial muscles and their activation) vector conditioned on a given emotion, and for applying an AU vector to a given image. The overall model is capable of modifying an image of a human face by modelling the facial expression to show a specific discrete emotion. Qualitative and quantitative measurements have been performed to evaluate the ability of the network to generate a variety of expressions that are consistent with the conditioned emotion. Moreover, we also collected people's responses about the quality and the legibility of the produced expressions by showing them applied to images and a social robot. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Emotion Detection Through Facial Expressions for Determining Students’ Concentration Level in E-Learning Platform
- Author
-
Hossain, Md. Noman, Long, Zalizah Awang, Seid, Norsuhaili, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Yang, Xin-She, editor, Sherratt, Simon, editor, Dey, Nilanjan, editor, and Joshi, Amit, editor
- Published
- 2024
- Full Text
- View/download PDF
14. Mapping Action Units to Valence and Arousal Space Using Machine Learning
- Author
-
Gadzhiev, Ismail M., Makarov, Alexander S., Tikhomirova, Daria V., Dolenko, Sergei A., Samsonovich, Alexei V., Kacprzyk, Janusz, Series Editor, Samsonovich, Alexei V., editor, and Liu, Tingting, editor
- Published
- 2024
- Full Text
- View/download PDF
15. ‘What Are They Looking at?’ Testing the Capacity of Action Units to Direct Attention in a 360° Recorded Virtual Reality Narrative
- Author
-
Gulyas, Napsugar, Doicaru, Miruna, Boode, Wilco, Campos, Fabio, van Gisbergen, Marnix S., tom Dieck, M. Claudia, editor, Jung, Timothy, editor, and Kim, Yen-Soon, editor
- Published
- 2024
- Full Text
- View/download PDF
16. AU-Oriented Expression Decomposition Learning for Facial Expression Recognition
- Author
-
Lin, Zehao, She, Jiahui, Shen, Qiu, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Liu, Qingshan, editor, Wang, Hanzi, editor, Ma, Zhanyu, editor, Zheng, Weishi, editor, Zha, Hongbin, editor, Chen, Xilin, editor, Wang, Liang, editor, and Ji, Rongrong, editor
- Published
- 2024
- Full Text
- View/download PDF
17. Facial action units detection using temporal context and feature reassignment.
- Author
-
Yang, Sipeng, Huang, Hongyu, Huang, Ying Sophie, and Jin, Xiaogang
- Subjects
FACIAL expression ,FACIAL muscles ,FORECASTING - Abstract
Facial action units (AUs) encode the activations of facial muscle groups, playing a crucial role in expression analysis and facial animation. However, current deep learning AU detection methods primarily focus on single‐image analysis, which limits the exploitation of rich temporal context for robust outcomes. Moreover, the scale of available datasets remains limited, leading models trained on these datasets to tend to suffer from overfitting issues. This paper proposes a novel AU detection method integrating spatial and temporal data with inter‐subject feature reassignment for accurate and robust AU predictions. Our method first extracts regional features from facial images. Then, to effectively capture both the temporal context and identity‐independent features, we introduce a temporal feature combination and feature reassignment (TC&FR) module, which transforms single‐image features into a cohesive temporal sequence and fuses features across multiple subjects. This transformation encourages the model to utilize identity‐independent features and temporal context, thus ensuring robust prediction outcomes. Experimental results demonstrate the enhancements brought by the proposed modules and the state‐of‐the‐art (SOTA) results achieved by our method. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Multi-modal Affect Detection Using Thermal and Optical Imaging in a Gamified Robotic Exercise.
- Author
-
Mohamed, Youssef, Güneysu, Arzu, Lemaignan, Séverin, and Leite, Iolanda
- Subjects
THERMOGRAPHY ,MACHINE learning ,OPTICAL images ,MULTIMODAL user interfaces ,ROBOTICS ,EMOTIONAL state ,AFFECTIVE computing ,THERMAL imaging cameras - Abstract
Affect recognition, or the ability to detect and interpret emotional states, has the potential to be a valuable tool in the field of healthcare. In particular, it can be useful in gamified therapy, which involves using gaming techniques to motivate and keep the engagement of patients in therapeutic activities. This study aims to examine the accuracy of machine learning models using thermal imaging and action unit data for affect classification in a gamified robot therapy scenario. A self-report survey and three machine learning models were used to assess emotions including frustration, boredom, and enjoyment in participants during different phases of the game. The results showed that the multimodal approach with the combination of thermal imaging and action units with LSTM model had the highest accuracy of 77% for emotion classification over a 7-s sliding window, while thermal imaging had the lowest standard deviation among participants. The results suggest that thermal imaging and action units can be effective in detecting affective states and might have the potential to be used in healthcare applications, such as gamified therapy, as a promising non-intrusive method for recognizing internal states. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Emotion-specific AUs for micro-expression recognition.
- Author
-
Leong, Shu-Min, Phan, Raphaël C.-W., and Baskaran, Vishnu Monn
- Abstract
The Facial Action Coding System (FACS) comprehensively describes facial expressions with facial action units (AUs). It is a well-used technique by researchers in emotions research to understand human emotions better. Most micro-expression datasets provide FACS-coded AU ground truths corresponding to micro-expressions classes. It is commonly accepted in computer vision-based emotions research that certain emotions are reliably revealed when specific combinations of AUs occur. However, the reliability of the ground truth AUs in the micro-expression datasets is lower than that of normal expressions, as they have lower AU intensities. Moreover, these micro-expression datasets only report the overall reliability of all AUs. It could not be identified which AUs had been accurately coded. This work aims to revisit the ground truth AUs of popular micro-expression datasets, namely CASME II, SAMM and CAS(ME) 2 , and inspect whether any AUs crucial for micro-expression recognition may need to be reconsidered. This paper also provides a detailed AU analysis which yields new AU-based RoIs for each dataset. These new RoIs improve the micro-expression recognition performances compared to the baselines considered in this work. The proposed RoIs for CASME II, SAMM and CAS(ME) 2 improve the recognition rates by 2 % , 1 % and 4 % , respectively, when compared with the existing RoIs. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. Using Diverse ConvNets to Classify Face Action Units in Dataset on Emotions Among Mexicans (DEM)
- Author
-
Marco A. Moreno-Armendariz, Alberto Espinosa-Juarez, and Esmeralda Godinez-Montero
- Subjects
Action units ,ConvNets ,CAM analysis ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
To understand how Convolutional Neural Networks (ConvNets) perceive the muscular movements of the human face, known as Action Units (AUs) in this work, we introduce a new dataset named Dataset on Emotions among Mexicans (DEM), consisting of 1557 images of Mexicans labeled with twenty-six AUs and seven emotions. As a benchmark, we used the comparison with DISFA+ labeled with 12 AUs. To address the task of detecting AUs in each image, six ConvNets were employed, and we evaluated their performance using the F1 Score. The two ConvNets with the best performance were VGG19 with 0.8180% (DEM), 0.9106 % (DISFA+), and ShuffleNetV2 with 0.7154% (DEM), 0.9440% (DISFA+). Subsequently, these ConvNets were analyzed using Grad-CAM and Grad-CAM++; this algorithms allows us to observe the areas of the face considered for prediction. In most cases, these areas consider the region of the labeled AU. Considering the F1 score and the visual study, we can conclude that using DEM as a dataset to classify AUs is promising since the experiments achieved performances similar to those of the current literature that only use ConvNets.
- Published
- 2024
- Full Text
- View/download PDF
21. Enhancing frame-level student engagement classification through knowledge transfer techniques.
- Author
-
Das, Riju and Dev, Soumyabrata
- Subjects
STUDENT engagement ,DEEP learning ,KNOWLEDGE transfer ,ABILITY grouping (Education) ,CLASSIFICATION ,LEARNING - Abstract
Assessing student engagement in educational settings is critical for monitoring and improving the learning process. Traditional methods that classify video-based student engagement datasets often assign a single engagement label to the entire video, resulting in inaccurate classification outcomes. However, student engagement varies over time, with fluctuations in concentration and interest levels. To overcome this limitation, this paper introduces a frame-level student engagement detection approach. By analyzing each frame individually, instructors gain more detailed insights into students' understanding of the course. The proposed method focuses on identifying student engagement at a granular level, enabling instructors to pinpoint specific moments of disengagement or high engagement for targeted interventions. Nevertheless, the lack of labeled frame-level engagement data presents a significant challenge. To address this challenge, we propose a novel approach for frame-level student engagement classification by leveraging the concept of knowledge transfer. Our method involves pretraining a deep learning model on a labeled image-based student engagement dataset, WACV, which serves as the base dataset for identifying frame-level engagement in our target video-based DAiSEE dataset. We then fine-tune the model on the unlabeled video dataset, utilizing the transferred knowledge to enhance engagement classification performance. Experimental results demonstrate the effectiveness of our frame-level approach, providing valuable insights for instructors to optimize instructional strategies and enhance the learning experience. This research contributes to the advancement of student engagement assessment, offering educators a more nuanced understanding of student behaviors during instructional videos. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Data Leakage and Evaluation Issues in Micro-Expression Analysis.
- Author
-
Varanka, Tuomas, Li, Yante, Peng, Wei, and Zhao, Guoying
- Abstract
Micro-expressions have drawn increasing interest lately due to various potential applications. The task is, however, difficult as it incorporates many challenges from the fields of computer vision, machine learning and emotional sciences. Due to the spontaneous and subtle characteristics of micro-expressions, the available training and testing data are limited, which make evaluation complex. We show that data leakage and fragmented evaluation protocols are issues among the micro-expression literature. We find that fixing data leaks can drastically reduce model performance, in some cases even making the models perform similarly to a random classifier. To this end, we go through common pitfalls, propose a new standardized evaluation protocol using facial action units with over 2000 micro-expression samples, and provide an open source library that implements the evaluation protocols in a standardized manner. Code is publicly available in https://github.com/tvaranka/meb. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. Learnable Eulerian Dynamics for Micro-Expression Action Unit Detection
- Author
-
Varanka, Tuomas, Peng, Wei, Zhao, Guoying, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Gade, Rikke, editor, Felsberg, Michael, editor, and Kämäräinen, Joni-Kristian, editor
- Published
- 2023
- Full Text
- View/download PDF
24. Demystifying Mental Health by Decoding Facial Action Unit Sequences
- Author
-
Deepika Sharma, Jaiteg Singh, Sukhjit Singh Sehra, and Sumeet Kaur Sehra
- Subjects
micro-expressions ,CNN ,K-means ,emotion recognition ,action units ,Technology - Abstract
Mental health is indispensable for effective daily functioning and stress management. Facial expressions may provide vital clues about the mental state of a person as they are universally consistent across cultures. This study intends to detect the emotional variances through facial micro-expressions using facial action units (AUs) to identify probable mental health issues. In addition, convolutional neural networks (CNN) were used to detect and classify the micro-expressions. Further, combinations of AUs were identified for the segmentation of micro-expressions classes using K-means square. Two benchmarked datasets CASME II and SAMM were employed for the training and evaluation of the model. The model achieved an accuracy of 95.62% on CASME II and 93.21% on the SAMM dataset, respectively. Subsequently, a case analysis was done to identify depressive patients using the proposed framework and it attained an accuracy of 92.99%. This experiment revealed the fact that emotions like disgust, sadness, anger, and surprise are the prominent emotions experienced by depressive patients during communication. The findings suggest that leveraging facial action units for micro-expression detection offers a promising approach to mental health diagnostics.
- Published
- 2024
- Full Text
- View/download PDF
25. What Can Facial Movements Reveal? Depression Recognition and Analysis Based on Optical Flow Using Bayesian Networks
- Author
-
Yu Ma, Jian Shen, Zeguang Zhao, Huajian Liang, Yang Tan, Zhenyu Liu, Kun Qian, Minqiang Yang, and Bin Hu
- Subjects
Depression recognition ,facial expressions ,action units ,optical flow ,Bayesian networks ,Medical technology ,R855-855.5 ,Therapeutics. Pharmacology ,RM1-950 - Abstract
Recent evidence have demonstrated that facial expressions could be a valid and important aspect for depression recognition. Although various works have been achieved in automatic depression recognition, it is a challenge to explore the inherent nuances of facial expressions that might reveal the underlying differences between depressed patients and healthy subjects under different stimuli. There is a lack of an undisturbed system that monitors depressive patients’ mental states in various free-living scenarios, so this paper steps towards building a classification model where data collection, feature extraction, depression recognition and facial actions analysis are conducted to infer the differences of facial movements between depressive patients and healthy subjects. In this study, we firstly present a plan of dividing facial regions of interest to extract optical flow features of facial expressions for depression recognition. We then propose facial movements coefficients utilising discrete wavelet transformation. Specifically, Bayesian Networks equipped with construction of Pearson Correlation Coefficients based on discrete wavelet transformation is learnt, which allows for analysing movements of different facial regions. We evaluate our method on a clinically validated dataset of 30 depressed patients and 30 healthy control subjects, and experiments results obtained the accuracy and recall of 81.7%, 96.7%, respectively, outperforming other features for comparison. Most importantly, the Bayesian Networks we built on the coefficients under different stimuli may reveal some facial action patterns of depressed subjects, which have a potential to assist the automatic diagnosis of depression.
- Published
- 2023
- Full Text
- View/download PDF
26. Visualization and analysis of skin strain distribution in various human facial actions
- Author
-
Takeru MISU, Hisashi ISHIHARA, So NAGASHIMA, Yusuke DOI, and Akihiro NAKATANI
- Subjects
human face ,strain analysis ,action units ,facial expression ,artificial face design ,Mechanical engineering and machinery ,TJ1-1570 - Abstract
Human faces are mechanical systems that display emotional and linguistic information with motions of the deformable facial tissues. Since each facial action forms a different stress-strain field that characterizes the differentiated facial appearance, a detailed analysis of strain distributions for each facial action will contribute to analytical and synthetical studies on human faces. This study evaluated strain distributions of 44 facial actions of a Japanese adult male based on the three-dimensional displacements of 125 tracking markers attached to the facial surface. We investigated how much the facial skin surface is stretched and compressed in each facial region based on the evaluated area strains produced by each facial action. Then, we visualized the strain distributions and surface undulation to analyze the complexity of the deformations on a face. The results show that the positive and negative surface strains intermingled on a face even in simple facial actions, potentially reflecting the complex facial structure under the facial skin layers, where several tissues with different material properties, e.g., adipose tissues and retaining ligaments, are distributed heterogeneously. These results are beneficial for artificial face designers as a design target and evidence to consider the effective skin structure and locations of actuators for artificial faces.
- Published
- 2023
- Full Text
- View/download PDF
27. Facial expression recognition on partially occluded faces using component based ensemble stacked CNN.
- Author
-
Bellamkonda, Sivaiah, Gopalan, N. P., Mala, C., and Settipalli, Lavanya
- Abstract
Facial Expression Recognition (FER) is the basis for many applications including human-computer interaction and surveillance. While developing such applications, it is imperative to understand human emotions for better interaction with machines. Among many FER models developed so far, Ensemble Stacked Convolution Neural Networks (ES-CNN) showed an empirical impact in improving the performance of FER on static images. However, the existing ES-CNN based FER models trained with features extracted from the entire face, are unable to address the issues of ambient parameters such as pose, illumination, occlusions. To mitigate the problem of reduced performance of ES-CNN on partially occluded faces, a Component based ES-CNN (CES-CNN) is proposed. CES-CNN applies ES-CNN on action units of individual face components such as eyes, eyebrows, nose, cheek, mouth, and glabella as one subnet of the network. Max-Voting based ensemble classifier is used to ensemble the decisions of the subnets in order to obtain the optimized recognition accuracy. The proposed CES-CNN is validated by conducting experiments on benchmark datasets and the performance is compared with the state-of-the-art models. It is observed from the experimental results that the proposed model has a significant enhancement in the recognition accuracy compared to the existing models. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
28. What Can Facial Movements Reveal? Depression Recognition and Analysis Based on Optical Flow Using Bayesian Networks.
- Author
-
Ma, Yu, Shen, Jian, Zhao, Zeguang, Liang, Huajian, Tan, Yang, Liu, Zhenyu, Qian, Kun, Yang, Minqiang, and Hu, Bin
- Subjects
DISCRETE wavelet transforms ,BAYESIAN analysis ,FACIAL expression ,OPTICAL flow ,PEARSON correlation (Statistics) - Abstract
Recent evidence have demonstrated that facial expressions could be a valid and important aspect for depression recognition. Although various works have been achieved in automatic depression recognition, it is a challenge to explore the inherent nuances of facial expressions that might reveal the underlying differences between depressed patients and healthy subjects under different stimuli. There is a lack of an undisturbed system that monitors depressive patients’ mental states in various free-living scenarios, so this paper steps towards building a classification model where data collection, feature extraction, depression recognition and facial actions analysis are conducted to infer the differences of facial movements between depressive patients and healthy subjects. In this study, we firstly present a plan of dividing facial regions of interest to extract optical flow features of facial expressions for depression recognition. We then propose facial movements coefficients utilising discrete wavelet transformation. Specifically, Bayesian Networks equipped with construction of Pearson Correlation Coefficients based on discrete wavelet transformation is learnt, which allows for analysing movements of different facial regions. We evaluate our method on a clinically validated dataset of 30 depressed patients and 30 healthy control subjects, and experiments results obtained the accuracy and recall of 81.7%, 96.7%, respectively, outperforming other features for comparison. Most importantly, the Bayesian Networks we built on the coefficients under different stimuli may reveal some facial action patterns of depressed subjects, which have a potential to assist the automatic diagnosis of depression. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
29. An Overview of Facial Micro-Expression Analysis: Data, Methodology and Challenge.
- Author
-
Xie, Hong-Xia, Lo, Ling, Shuai, Hong-Han, and Cheng, Wen-Huang
- Abstract
Facial micro-expressions indicate brief and subtle facial movements that appear during emotional communication. In comparison to macro-expressions, micro-expressions are more challenging to be analyzed due to the short span of time and the fine-grained changes. In recent years, micro-expression recognition (MER) has drawn much attention because it can benefit a wide range of applications, e.g., police interrogation, clinical diagnosis, depression analysis, and business negotiation. In this survey, we offer a fresh overview to discuss new research directions and challenges these days for MER tasks. For example, we review MER approaches from three novel aspects: macro-to-micro adaptation, recognition based on key apex frames, and recognition based on facial action units. Moreover, to mitigate the problem of limited and biased ME data, synthetic data generation is surveyed for the diversity enrichment of micro-expression data. Since micro-expression spotting can boost micro-expression analysis, the state-of-the-art spotting works are also introduced in this paper. At last, we discuss the challenges in MER research and provide potential solutions as well as possible directions for further investigation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
30. How to Synthesize a Large-Scale and Trainable Micro-Expression Dataset?
- Author
-
Liu, Yuchi, Wang, Zhongdao, Gedeon, Tom, Zheng, Liang, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Avidan, Shai, editor, Brostow, Gabriel, editor, Cissé, Moustapha, editor, Farinella, Giovanni Maria, editor, and Hassner, Tal, editor
- Published
- 2022
- Full Text
- View/download PDF
31. Privacy-Preserving and Scalable Affect Detection in Online Synchronous Learning
- Author
-
Böttger, Felix, Cetinkaya, Ufuk, Di Mitri, Daniele, Gombert, Sebastian, Shingjergji, Krist, Iren, Deniz, Klemke, Roland, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Hilliger, Isabel, editor, Muñoz-Merino, Pedro J., editor, De Laet, Tinne, editor, Ortega-Arranz, Alejandro, editor, and Farrell, Tracie, editor
- Published
- 2022
- Full Text
- View/download PDF
32. FMeAR: FACS Driven Ensemble Model for Micro-Expression Action Unit Recognition
- Author
-
Chauhan, Anjaly and Jain, Shikha
- Published
- 2024
- Full Text
- View/download PDF
33. Statistical analysis of the units of action in facial expressions for emotion recognition.
- Author
-
Olmedo, Gonzalo and Paredes, Nancy
- Subjects
EMOTION recognition ,ARTIFICIAL intelligence ,QUANTITATIVE research ,PROBABILITY density function ,STATISTICAL sampling ,FACIAL expression - Abstract
In this article, the statistical behavior of the activation intensities of the action units that represent the micro-expressions of the facial expressions for four main emotions, happiness, anger, sadness, and surprise, is analyzed. Based on the results obtained, the distribution of each unit of action is modeled through probability density functions, which will allow the creation of an infinity of random samples, which contribute to emotion evaluation processes and especially to artificial intelligence techniques that require a high number of samples for their training processes. [ABSTRACT FROM AUTHOR]
- Published
- 2023
34. Talking Face Generation via Facial Anatomy.
- Author
-
SHIGUANG LIU and HUIXIN WANG
- Subjects
EYEBROWS ,ANATOMY ,CHEEK ,ANNOTATIONS - Abstract
To generate the corresponding talking face from a speech audio and a face image, it is essential to match the variations in the facial appearance with the speech audio in subtle movements of different face regions. Nevertheless, the facial movements generated by the existing methods lack detail and vividness, or the methods are only oriented toward a specific person. In this article, we propose a novel two-stage network to generate talking faces for any target identity through annotations of the action units (AUs). In the first stage, the relationship between the audio and the AUs in the audio-to-AU network is learned. The audio-to-AU network needs to produce the consistent AU group for the input audio. In the second stage, the AU group in the first stage and a face image are fed into the generation network to output the resulting talking face image. Various results confirm that, compared to state-of-the-art methods, our approach is able to produce more realistic and vivid talking faces for arbitrary targets with richer details of facial movements, such as the cheek motion and eyebrow motion. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
35. MERASTC : Micro-Expression Recognition Using Effective Feature Encodings and 2D Convolutional Neural Network.
- Author
-
Gupta, Puneet
- Abstract
Facial micro-expression (ME) can disclose genuine and concealed human feelings. It makes MEs extensively useful in real-world applications pertaining to affective computing and psychology. Unfortunately, they are induced by subtle facial movements for a short duration of time, which makes the ME recognition, a highly challenging problem even for human beings. In automatic ME recognition, the well-known features encode either incomplete or redundant information, and there is a lack of sufficient training data. The proposed method, Micro-Expression Recognition by Analysing Spatial and Temporal Characteristics, $MERASTC$ M E R A S T C mitigates these issues for improving the ME recognition. It compactly encodes the subtle deformations using action units (AUs), landmarks, gaze, and appearance features of all the video frames while preserving most of the relevant ME information. Furthermore, it improves the efficacy by introducing a novel neutral face normalization for ME and initiating the utilization of gaze features in deep learning-based ME recognition. The features are provided to the 2D convolutional neural network that jointly analyses the spatial and temporal behavior for correct ME classification. Experimental results1 on publicly available datasets indicate that the proposed method exhibits better performance than the well-known methods. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
36. BFFN: A novel balanced feature fusion network for fair facial expression recognition.
- Author
-
Li, Hao, Luo, Yiqin, Gu, Tianlong, and Chang, Liang
- Subjects
- *
FACIAL expression , *FAIRNESS - Abstract
Facial expression recognition (FER) technology has become increasingly mature and applicable in recent years. However, it still suffers from the bias of expression class, which can lead to unfair decisions for certain expression classes in applications. This study aims to mitigate expression class bias through both pre-processing and in-processing approaches. First, we analyze the output of existing models and demonstrate the existence of obvious class bias, particularly for underrepresented expressions. Second, we develop a class-balanced dataset constructed through data generation, mitigating unfairness at the data level. Then, we propose the Balanced Feature Fusion Network (BFFN), a class fairness-enhancing network. The BFFN mitigates the class bias by adding facial action units (AU) to enrich expression-related features and allocating weights in the AU feature fusion process to improve the extraction ability of underrepresented expression features. Finally, extensive experiments on datasets (RAF-DB and AffectNet) provide evidence that our BFFN outperforms existing FER models, improving the fairness by at least 16%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Processing Real-Life Recordings of Facial Expressions of Polish Sign Language Using Action Units.
- Author
-
Irasiak, Anna, Kozak, Jan, Piasecki, Adam, and Stęclik, Tomasz
- Subjects
- *
SIGN language , *FACIAL expression , *POLISH language , *LANGUAGE policy , *LANGUAGE & languages , *DEAF children - Abstract
Automatic translation between the national language and sign language is a complex process similar to translation between two different foreign languages. A very important aspect is the precision of not only manual gestures but also facial expressions, which are extremely important in the overall context of a sentence. In this article, we present the problem of including facial expressions in the automation of Polish-to-Polish Sign Language (PJM) translation—this is part of an ongoing project related to a comprehensive solution allowing for the animation of manual gestures, body movements and facial expressions. Our approach explores the possibility of using action unit (AU) recognition in the automatic annotation of recordings, which in the subsequent steps will be used to train machine learning models. This paper aims to evaluate entropy in real-life translation recordings and analyze the data associated with the detected action units. Our approach has been subjected to evaluation by experts related to Polish Sign Language, and the results obtained allow for the development of further work related to automatic translation into Polish Sign Language. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. Deep Learning Pipeline for Spotting Macro- and Micro-expressions in Long Video Sequences Based on Action Units and Optical Flow.
- Author
-
Yang, Bo, Wu, Jianming, Ikeda, Kazushi, Hattori, Gen, Sugano, Masaru, Iwasawa, Yusuke, and Matsuo, Yutaka
- Subjects
- *
DEEP learning , *OPTICAL flow , *FACIAL expression - Abstract
• For the first time, AUs and optical flow features are combined to spot either macro- or micro- expression intervals. • The proposal can eliminate the influence of facial image change caused by noises, such as body or head movement. • The proposed Concat-CNN model can learn both the inner features of a single frame and the correlation between frames. • The re-labeling method considers the overall change process of a specific expression and improves the detection performance. • The proposal shows remarkable improvement in the F1 scores on datasets, such as the CAS(ME) 2 -cropped and the SAMM-LV. [Display omitted] This paper is an extension of our previously published ACM Multimedia 2022 paper, which was ranked 3rd in the macro-expressions (MaEs) and micro-expressions (MEs) spotting task of the FME challenge 2021. In our earlier work, a deep learning framework based on facial action units (AUs) was proposed to emphasize both local and global features to deal with the MaEs and MEs spotting tasks. In this paper, an advanced Concat-CNN model is proposed to not only utilize facial action units (AU) features, which our previous work proved were more effective in detecting MaEs, but also to fuse the optical flow features to improve the detection performance of MEs. The advanced Concat-CNN proposed in this paper not only considers the intra-features correlation of a single frame but also the inter-features correlation between frames. Further, we devise a new adaptive re-labeling method by labeling the emotional frames with distinctive scores. This method takes into account the dynamic changes in expressions to further improve the overall detection performance. Compared with our earlier work and several existing works, the newly proposed deep learning pipeline is able to achieve a better performance in terms of the overall F1-scores: 0.2623 on CAS(ME) 2 , 0.2839 on CAS(ME) 2 -cropped, and 0.3241 on SAMM-LV, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. Emotion Recognition of Down Syndrome People Based on the Evaluation of Artificial Intelligence and Statistical Analysis Methods.
- Author
-
Paredes, Nancy, Caicedo-Bravo, Eduardo F., Bacca, Bladimir, and Olmedo, Gonzalo
- Subjects
- *
EMOTION recognition , *DEEP learning , *ARTIFICIAL intelligence , *STATISTICS , *SUPERVISED learning ,PEOPLE with Down syndrome - Abstract
This article presents a study based on evaluating different techniques to automatically recognize the basic emotions of people with Down syndrome, such as anger, happiness, sadness, surprise, and neutrality, as well as the statistical analysis of the Facial Action Coding System, determine the symmetry of the Action Units present in each emotion, identify the facial features that represent this group of people. First, a dataset of images of faces of people with Down syndrome classified according to their emotions is built. Then, the characteristics of facial micro-expressions (Action Units) present in the feelings of the target group through statistical analysis are evaluated. This analysis uses the intensity values of the most representative exclusive action units to classify people's emotions. Subsequently, the collected dataset was evaluated using machine learning and deep learning techniques to recognize emotions. In the beginning, different supervised learning techniques were used, with the Support Vector Machine technique obtaining the best precision with a value of 66.20%. In the case of deep learning methods, the mini-Xception convolutional neural network was used to recognize people's emotions with typical development, obtaining an accuracy of 74.8%. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
40. The Loop of Nonverbal Communication Between Human and Virtual Actor: Mapping Between Spaces
- Author
-
Shirokiy, Vladimir R., Tikhomirova, Daria V., Vladimirov, Roman D., Dolenko, Sergei A., Samsonovich, Alexei V., Kacprzyk, Janusz, Series Editor, Pal, Nikhil R., Advisory Editor, Bello Perez, Rafael, Advisory Editor, Corchado, Emilio S., Advisory Editor, Hagras, Hani, Advisory Editor, Kóczy, László T., Advisory Editor, Kreinovich, Vladik, Advisory Editor, Lin, Chin-Teng, Advisory Editor, Lu, Jie, Advisory Editor, Melin, Patricia, Advisory Editor, Nedjah, Nadia, Advisory Editor, Nguyen, Ngoc Thanh, Advisory Editor, Wang, Jun, Advisory Editor, Samsonovich, Alexei V., editor, Gudwin, Ricardo R., editor, and Simões, Alexandre da Silva, editor
- Published
- 2021
- Full Text
- View/download PDF
41. Automatic Recognition of Learning-Centered Emotions
- Author
-
González-Meneses, Yesenia N., Guerrero-García, Josefina, Reyes-García, Carlos Alberto, Zatarain-Cabada, Ramón, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Roman-Rangel, Edgar, editor, Kuri-Morales, Ángel Fernando, editor, Martínez-Trinidad, José Francisco, editor, Carrasco-Ochoa, Jesús Ariel, editor, and Olvera-López, José Arturo, editor
- Published
- 2021
- Full Text
- View/download PDF
42. Automatic Frustration Detection Using Thermal Imaging.
- Author
-
Mohamed, Youssef, Ballardini, Giulia, Parreira, Maria Teresa, Lemaignan, Séverin, and Leite, Iolanda
- Subjects
THERMOGRAPHY ,FRUSTRATION ,COGNITIVE load ,HUMAN-robot interaction ,THERMAL imaging cameras ,AFFECT (Psychology) - Abstract
To achieve seamless interactions, robots have to be capable of reliably detecting affective states in real time. One of the possible states that humans go through while interacting with robots is frustration. Detecting frustration from RGB images can be challenging in some real-world situations; thus, we investigate in this work whether thermal imaging can be used to create a model that is capable of detecting frustration induced by cognitive load and failure. To train our model, we collected a data set from 18 participants experiencing both types of frustration induced by a robot. The model was tested using features from several modalities: thermal, RGB, Electrodermal Activity (EDA), and all three combined. When data from both frustration cases were combined and used as training input, the model reached an accuracy of 89% with just RGB features, 87% using only thermal features, 84% using EDA, and 86% when using all modalities. Furthermore, the highest accuracy for the thermal data was reached using three facial regions of interest: nose, forehead and lower lip. [ABSTRACT FROM AUTHOR]
- Published
- 2022
43. Audio-Visual Stress Classification Using Cascaded RNN-LSTM Networks.
- Author
-
Gupta, Megha V., Vaikole, Shubhangi, Oza, Ankit D., Patel, Amisha, Burduhos-Nergis, Diana Petronela, and Burduhos-Nergis, Dumitru Doru
- Subjects
- *
CASCADE connections , *PERCEIVED Stress Scale , *PSYCHOLOGICAL stress , *WELL-being , *CLASSIFICATION - Abstract
The purpose of this research is to emphasize the importance of mental health and contribute to the overall well-being of humankind by detecting stress. Stress is a state of strain, whether it be mental or physical. It can result from anything that frustrates, incenses, or unnerves you in an event or thinking. Your body's response to a demand or challenge is stress. Stress affects people on a daily basis. Stress can be regarded as a hidden pandemic. Long-term (chronic) stress results in ongoing activation of the stress response, which wears down the body over time. Symptoms manifest as behavioral, emotional, and physical effects. The most common method involves administering brief self-report questionnaires such as the Perceived Stress Scale. However, self-report questionnaires frequently lack item specificity and validity, and interview-based measures can be time- and money-consuming. In this research, a novel method used to detect human mental stress by processing audio-visual data is proposed. In this paper, the focus is on understanding the use of audio-visual stress identification. Using the cascaded RNN-LSTM strategy, we achieved 91% accuracy on the RAVDESS dataset, classifying eight emotions and eventually stressed and unstressed states. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
44. Basic Facial Expressions Analysis on a 3D Model: Based on Action Units and the Nose Tip
- Author
-
Hailemariam, Meareg A., Akan, Ozgur, Editorial Board Member, Bellavista, Paolo, Editorial Board Member, Cao, Jiannong, Editorial Board Member, Coulson, Geoffrey, Editorial Board Member, Dressler, Falko, Editorial Board Member, Ferrari, Domenico, Editorial Board Member, Gerla, Mario, Editorial Board Member, Kobayashi, Hisashi, Editorial Board Member, Palazzo, Sergio, Editorial Board Member, Sahni, Sartaj, Editorial Board Member, Shen, Xuemin (Sherman), Editorial Board Member, Stan, Mircea, Editorial Board Member, Jia, Xiaohua, Editorial Board Member, Zomaya, Albert Y., Editorial Board Member, Habtu, Nigus Gabbiye, editor, Ayele, Delele Worku, editor, Fanta, Solomon Workneh, editor, Admasu, Bimrew Tamrat, editor, and Bitew, Mekuanint Agegnehu, editor
- Published
- 2020
- Full Text
- View/download PDF
45. Facial Expressions as Indicator for Discomfort in Automated Driving
- Author
-
Beggiato, Matthias, Rauh, Nadine, Krems, Josef, Kacprzyk, Janusz, Series Editor, Pal, Nikhil R., Advisory Editor, Bello Perez, Rafael, Advisory Editor, Corchado, Emilio S., Advisory Editor, Hagras, Hani, Advisory Editor, Kóczy, László T., Advisory Editor, Kreinovich, Vladik, Advisory Editor, Lin, Chin-Teng, Advisory Editor, Lu, Jie, Advisory Editor, Melin, Patricia, Advisory Editor, Nedjah, Nadia, Advisory Editor, Nguyen, Ngoc Thanh, Advisory Editor, Wang, Jun, Advisory Editor, Ahram, Tareq, editor, Karwowski, Waldemar, editor, Vergnano, Alberto, editor, Leali, Francesco, editor, and Taiar, Redha, editor
- Published
- 2020
- Full Text
- View/download PDF
46. Is It Me or the Robot? A Critical Evaluation of Human Affective State Recognition in a Cognitive Task.
- Author
-
Jirak, Doreen, Motonobu Aoki, Takura Yanagi, Atsushi Takamatsu, Bouet, Stephane, Tomohiro Yamamura, Sandini, Giulio, and Rea, Francesco
- Subjects
RECOGNITION (Psychology) ,AFFECTIVE computing ,HUMAN-robot interaction ,AFFECT (Psychology) ,HUMANOID robots ,SOCIAL robots - Abstract
A key goal in human-robot interaction (HRI) is to design scenarios between humanoid robots and humans such that the interaction is perceived as collaborative and natural, yet safe and comfortable for the human. Human skills like verbal and non-verbal communication are essential elements as humans tend to attribute social behaviors to robots. However, aspects like the uncanny valley and different technical affinity levels can impede the success of HRI scenarios, which has consequences on the establishment of long-term interaction qualities like trust and rapport. In the present study, we investigate the impact of a humanoid robot on human emotional responses during the performance of a cognitively demanding task. We set up three different conditions for the robot with increasing levels of social cue expressions in a between-group study design. For the analysis of emotions, we consider the eye gaze behavior, arousal-valence for affective states, and the detection of action units. Our analysis reveals that the participants display a high tendency toward positive emotions in presence of a robot with clear social skills compared to other conditions, where we show how emotions occur only at task onset. Our study also shows how different expression levels influence the analysis of the robots' role in HRI. Finally, we critically discuss the current trend of automatized emotion or affective state recognition in HRI and demonstrate issues that have direct consequences on the interpretation and, therefore, claims about human emotions in HRI studies. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
47. Facial identity protection using deep learning technologies: an application in affective computing
- Author
-
Mase, Jimiama M., Leesakul, Natalie, Figueredo, Grazziela P., and Torres, Mercedes Torres
- Published
- 2023
- Full Text
- View/download PDF
48. Is It Me or the Robot? A Critical Evaluation of Human Affective State Recognition in a Cognitive Task
- Author
-
Doreen Jirak, Motonobu Aoki, Takura Yanagi, Atsushi Takamatsu, Stephane Bouet, Tomohiro Yamamura, Giulio Sandini, and Francesco Rea
- Subjects
human-robot interaction ,social robots ,cognitive load ,affective states ,action units ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
A key goal in human-robot interaction (HRI) is to design scenarios between humanoid robots and humans such that the interaction is perceived as collaborative and natural, yet safe and comfortable for the human. Human skills like verbal and non-verbal communication are essential elements as humans tend to attribute social behaviors to robots. However, aspects like the uncanny valley and different technical affinity levels can impede the success of HRI scenarios, which has consequences on the establishment of long-term interaction qualities like trust and rapport. In the present study, we investigate the impact of a humanoid robot on human emotional responses during the performance of a cognitively demanding task. We set up three different conditions for the robot with increasing levels of social cue expressions in a between-group study design. For the analysis of emotions, we consider the eye gaze behavior, arousal-valence for affective states, and the detection of action units. Our analysis reveals that the participants display a high tendency toward positive emotions in presence of a robot with clear social skills compared to other conditions, where we show how emotions occur only at task onset. Our study also shows how different expression levels influence the analysis of the robots' role in HRI. Finally, we critically discuss the current trend of automatized emotion or affective state recognition in HRI and demonstrate issues that have direct consequences on the interpretation and, therefore, claims about human emotions in HRI studies.
- Published
- 2022
- Full Text
- View/download PDF
49. Studying Facial Activity in Parkinson's Disease Patients Using an Automated Method and Video Recording
- Author
-
Anastasia Moshkova, Andrey Samorodov, Ekaterina Ivanova, Ekaterina Fedotova, Natalia Voinova, and Alexander Volkov
- Subjects
parkinson's disease ,hypomimia ,action units ,Telecommunication ,TK5101-6720 - Abstract
Main objective of this research is studying facial activity of PD patients using an automated method for assessing facial movements based on calculating the movement signal kinematic characteristics. 16 PD patients and 16 participants in the control group have been recorded using 2D video camera while performing 4 facial tests: closing the eyes, raising the eyebrows, smiling with an effort (grin), moving the eyebrows. Each test includes a 10-fold repetition of a given facial movement with maximum effort and velocity. Intensities of the corresponding action units (AUs) in each frame have been determined by an automated method for each test. The resulting signal of the AU intensity dependence on the frame number has been marked with maximum and minimum points. From the extremum points, 11 kinematic parameters have been calculated for each AU of each patient. To assess the differences in kinematic parameters between the groups, the nonparametric Mann-Whitney test has been used. There is decrease in the facial movements frequency and velocity. Consequently, there is increase in the duration of the performance of facial tests in PD patients group compared to the control group. Only some AUs also showed a decrease in the amplitude of facial movements in PD patients group compared to the control group. Differences in the largest number of kinematic parameters have been obtained for the action units AU12 and AU14 recorded during the smile with effort test, and AU04 recorded during the eyebrow moving test.
- Published
- 2021
- Full Text
- View/download PDF
50. Facial Motion Analysis beyond Emotional Expressions.
- Author
-
Porta-Lorenzo, Manuel, Vázquez-Enríquez, Manuel, Pérez-Pérez, Ania, Alba-Castro, José Luis, and Docío-Fernández, Laura
- Subjects
- *
MOTION analysis , *SELF-expression , *FACIAL expression & emotions (Psychology) , *NONVERBAL communication , *FACIAL expression , *DEEP learning - Abstract
Facial motion analysis is a research field with many practical applications, and has been strongly developed in the last years. However, most effort has been focused on the recognition of basic facial expressions of emotion and neglects the analysis of facial motions related to non-verbal communication signals. This paper focuses on the classification of facial expressions that are of the utmost importance in sign languages (Grammatical Facial Expressions) but also present in expressive spoken language. We have collected a dataset of Spanish Sign Language sentences and extracted the intervals for three types of Grammatical Facial Expressions: negation, closed queries and open queries. A study of several deep learning models using different input features on the collected dataset (LSE_GFE) and an external dataset (BUHMAP) shows that GFEs can be learned reliably with Graph Convolutional Networks simply fed with face landmarks. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.