重庆理工大学学报(自然科学) ›› 2023, Vol. 37 ›› Issue (9): 217-226.

• 信息·计算机 • 上一篇    下一篇

新型融合注意力机制的遮挡面部表情识别框架

张本文,高瑞玮,乔少杰   

  1. (1.四川民族学院 理工学院,四川 康定 626001; 2.成都信息工程大学 软件工程学院,成都 610225)
  • 出版日期:2023-10-17 发布日期:2023-10-17
  • 作者简介:张本文,男,副教授,主要从事机器学习、数据挖掘、粗糙集理论与代价敏感研究,Email:25926940@qq.com;通 信作者 乔少杰,男,博士,教授,主要从事数据库、人工智能、机器学习研究,Email:sjqiao@cuit.edu.cn。

A novel framework for occluded facial expression recognition by integrating attention mechanism

  • Online:2023-10-17 Published:2023-10-17

摘要: 现有基于深度学习的面部表情识别模型不能有效地应对面部遮挡部分的干扰,无 法准确捕捉面部未遮挡部分的特征,会导致识别准确率降低。为此,提出一种新型融合注意力 机制的遮挡面部表情识别框架 FERAM(facialexpressionrecognitionframeworkbasedonattention mechanism),应用局部特征网络提取面部表情的局部关键特征,设计全局特征网络学习整个面 部表情中的互补信息,采用注意力机制处理面部遮挡部分如眼镜、口罩和围巾等。在 RAFDB、 AffectNet、CK+(CohnKanade)及 FEDRO数据集进行大量实验,结果表明:FERAM的 7种表 情分类性能均优于基于深度学习的代表性人脸面部表情识别模型,识别准确率达到 88.1%。

关键词: 遮挡面部表情识别, 特征提取, 特征分类, 深度学习, 注意力机制

Abstract: Traditional facial expression recognition technologies rely heavily on manually formulated feature extraction rules, while deep learning-based facial expression recognition technologies can automatically perform the operations of feature extraction, feature selection and feature classification. However, for the faces with occluded parts, the existing facial expression recognition models based on deep learning cannot effectively deal with the interference of the occluded part of the face, and cannot accurately capture the features of facial unobstructed parts, thus leading to the degradation of recognition accuracy. To solve the aforementioned problems, a novel occluded facial expression recognition framework by integrating attention mechanism called FER-AM(facial expression recognition framework based on attention mechanism) is proposed, the local feature network is used to extract the local key features of facial expressions, and the global feature network is designed to learn the complementary information in the whole face, and the attention mechanism can effectively deal with facial occluded parts, such as glasses, masks and scarves. A large number of experiments are conducted on RAF-DB, AffectNet, CK+(Cohn Kanade) and FED-RO data sets, and the results show that the seven expression classification performance of FER-AM is better than the representative facial expression recognition models based on deep learning, and the recognition accuracy can reach 88.1%.

中图分类号: 

  • TP311