直近一年間の累計
アクセス数 : ?
ダウンロード数 : ?
ID 116332
著者
Wang, Xiaohua Hefei University of Technology
Gong, Jianqiao Hefei University of Technology
Hu, Min Hefei University of Technology
Gu, Yu Hefei University of Technology
任, 福継 Hefei University of Technology|University of Tokushima 徳島大学 教育研究者総覧 KAKEN研究者をさがす
キーワード
Facial expression recognition
data enhancement
generative adversarial networks
self-attention
資料タイプ
学術雑誌論文
抄録
In the field of facial expression recognition, deep learning is extensively used. However, insufficient and unbalanced facial training data in available public databases is a major challenge for improving the expression recognition rate. Generative Adversarial Networks (GANs) can produce more one-to-one faces with different expressions, which can be used to enhance databases. StarGAN can perform one-to-many translations for multiple expressions. Compared with original GANs, StarGAN can increase the efficiency of sample generation. Nevertheless, there are some defects in essential areas of the generated face, such as the mouth and the fuzzy side face image generation. To address these limitations, we improved StarGAN to alleviate the defects of images generation by modifying the reconstruction loss and adding the Contextual loss. Meanwhile, we added the Attention U-Net to StarGAN's generator, replacing StarGAN's original generator. Therefore, we proposed the Contextual loss and Attention U-Net (LAUN) improved StarGAN. The U-shape structure and skip connection in Attention U-Net can effectively integrate the details and semantic features of images. The network's attention structure can pay attention to the essential areas of the human face. The experimental results demonstrate that the improved model can alleviate some flaws in the face generated by the original StarGAN. Therefore, it can generate person images with better quality with different poses and expressions. The experiments were conducted on the Karolinska Directed Emotional Faces database, and the accuracy of facial expression recognition is 95.97%, 2.19% higher than that by using StarGAN. Meanwhile, the experiments were carried out on the MMI Facial Expression Database, and the accuracy of expression is 98.30%, 1.21% higher than that by using StarGAN. Moreover, experiment results have better performance based on the LAUN improved StarGAN enhanced databases than those without enhancement.
掲載誌名
IEEE Access
ISSN
21693536
出版者
IEEE
8
開始ページ
161509
終了ページ
161518
発行日
2020-09-03
権利情報
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4.0/
EDB ID
出版社版DOI
出版社版URL
フルテキストファイル
言語
eng
著者版フラグ
出版社版
部局
理工学系