Arrow Research search
Back to AAAI

AAAI 2018

Residual Encoder Decoder Network and Adaptive Prior for Face Parsing

Conference Paper AAAI Technical Track: Vision Artificial Intelligence

Abstract

Face parsing assigns every pixel in a facial image with a semantic label, which could be applied in various applications including face expression recognition, facial beautification, affective computing and animation. While lots of progress have been made in this field, current state-of-the-art methods still fail to extract real effective feature and restore accurate score map, especially for those facial parts which have large variations of deformation and fairly similar appearance, e. g. mouth, eyes and thin eyebrows. In this paper, we propose a novel pixel-wise face parsing method called Residual Encoder Decoder Network (RED-Net), which combines a feature-rich encoder-decoder framework with adaptive prior mechanism. Our encoder-decoder framework extracts feature with ResNet and decodes the feature by elaborately fusing the residual architectures into deconvolution. This framework learns more effective feature comparing to that learnt by decoding with interpolation or classic deconvolution operations. To overcome the appearance ambiguity between facial parts, an adaptive prior mechanism is proposed in term of the decoder prediction confidence, allowing refining the final result. The experimental results on two public databases demonstrate that our method outperforms the state-of-thearts significantly, achieving improvements of F-measure from 0. 854 to 0. 905 on the Helen dataset, and pixel accuracy from 95. 12% to 97. 59% on the LFW dataset. In particular, convincing qualitative examples show that our method parses eye, eyebrow and lip regions more accurately.

Authors

Keywords

No keywords are indexed for this paper.

Context

Venue
AAAI Conference on Artificial Intelligence
Archive span
1980-2026
Indexed papers
28718
Paper id
851248041107824911