Facial action unit detection using attention and relation learning

Attention mechanism has recently attracted increasing attentions in the field of facial action unit (AU) detection. By finding the region of interest of each AU with the attention mechanism, AU-related local features can be captured. Most of the existing attention based AU detection works use prior...

Full description

Saved in:
Bibliographic Details
Main Authors: Shao, Zhiwen, Liu, Zhilei, Cai, Jianfei, Wu, Yunsheng, Ma, Lizhuang
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/163146
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-163146
record_format dspace
spelling sg-ntu-dr.10356-1631462022-11-25T02:30:35Z Facial action unit detection using attention and relation learning Shao, Zhiwen Liu, Zhilei Cai, Jianfei Wu, Yunsheng Ma, Lizhuang School of Computer Science and Engineering Engineering::Computer science and engineering Pixel-Level Relation Learning Facial AU Detection Attention mechanism has recently attracted increasing attentions in the field of facial action unit (AU) detection. By finding the region of interest of each AU with the attention mechanism, AU-related local features can be captured. Most of the existing attention based AU detection works use prior knowledge to predefine fixed attentions or refine the predefined attentions within a small range, which limits their capacity to model various AUs. In this paper, we propose an end-to-end deep learning based attention and relation learning framework for AU detection with only AU labels, which has not been explored before. In particular, multi-scale features shared by each AU are learned firstly, and then both channel-wise and spatial attentions are adaptively learned to select and extract AU-related local features. Moreover, pixel-level relations for AUs are further captured to refine spatial attentions so as to extract more relevant local features. Without changing the network architecture, our framework can be easily extended for AU intensity estimation. Extensive experiments show that our framework (i) soundly outperforms the state-of-the-art methods for both AU detection and AU intensity estimation on the challenging BP4D, DISFA, FERA 2015 and BP4D+ benchmarks, (ii) can adaptively capture the correlated regions of each AU, and (iii) also works well under severe occlusions and large poses. Nanyang Technological University This work was supported by the National Natural Science Foundation of China (No. 61503277 and No. 61472245), the National Social Science Foundation of China (No. 18ZD22), and the Science and Technology Commission of Shanghai Municipality Program (No. 18D1205903). It was also partially supported by Data Science & Artificial Intelligence Research Centre@NTU (DSAIR) and SINGTEL-NTU Cognitive & Artificial Intelligence Joint Lab (SCALE@NTU), and the joint project of Tencent YouTu and Shanghai Jiao Tong University. 2022-11-25T02:30:35Z 2022-11-25T02:30:35Z 2019 Journal Article Shao, Z., Liu, Z., Cai, J., Wu, Y. & Ma, L. (2019). Facial action unit detection using attention and relation learning. IEEE Transactions On Affective Computing, 13(3), 1274-1289. https://dx.doi.org/10.1109/TAFFC.2019.2948635 1949-3045 https://hdl.handle.net/10356/163146 10.1109/TAFFC.2019.2948635 2-s2.0-85139183863 3 13 1274 1289 en IEEE Transactions on Affective Computing © 2019 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Pixel-Level Relation Learning
Facial AU Detection
spellingShingle Engineering::Computer science and engineering
Pixel-Level Relation Learning
Facial AU Detection
Shao, Zhiwen
Liu, Zhilei
Cai, Jianfei
Wu, Yunsheng
Ma, Lizhuang
Facial action unit detection using attention and relation learning
description Attention mechanism has recently attracted increasing attentions in the field of facial action unit (AU) detection. By finding the region of interest of each AU with the attention mechanism, AU-related local features can be captured. Most of the existing attention based AU detection works use prior knowledge to predefine fixed attentions or refine the predefined attentions within a small range, which limits their capacity to model various AUs. In this paper, we propose an end-to-end deep learning based attention and relation learning framework for AU detection with only AU labels, which has not been explored before. In particular, multi-scale features shared by each AU are learned firstly, and then both channel-wise and spatial attentions are adaptively learned to select and extract AU-related local features. Moreover, pixel-level relations for AUs are further captured to refine spatial attentions so as to extract more relevant local features. Without changing the network architecture, our framework can be easily extended for AU intensity estimation. Extensive experiments show that our framework (i) soundly outperforms the state-of-the-art methods for both AU detection and AU intensity estimation on the challenging BP4D, DISFA, FERA 2015 and BP4D+ benchmarks, (ii) can adaptively capture the correlated regions of each AU, and (iii) also works well under severe occlusions and large poses.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Shao, Zhiwen
Liu, Zhilei
Cai, Jianfei
Wu, Yunsheng
Ma, Lizhuang
format Article
author Shao, Zhiwen
Liu, Zhilei
Cai, Jianfei
Wu, Yunsheng
Ma, Lizhuang
author_sort Shao, Zhiwen
title Facial action unit detection using attention and relation learning
title_short Facial action unit detection using attention and relation learning
title_full Facial action unit detection using attention and relation learning
title_fullStr Facial action unit detection using attention and relation learning
title_full_unstemmed Facial action unit detection using attention and relation learning
title_sort facial action unit detection using attention and relation learning
publishDate 2022
url https://hdl.handle.net/10356/163146
_version_ 1751548584668430336