Publication: Spontaneous smile intensity estimation by fusing saliency maps and convolutional neural networks
dc.contributor.coauthor | Wei, Qinglan | |
dc.contributor.coauthor | Morency, Louis-Philippe | |
dc.contributor.coauthor | Sun, Bo | |
dc.contributor.department | N/A | |
dc.contributor.kuauthor | Bozkurt, Elif | |
dc.contributor.kuprofile | PhD Student | |
dc.contributor.schoolcollegeinstitute | Graduate School of Sciences and Engineering | |
dc.contributor.yokid | N/A | |
dc.date.accessioned | 2024-11-09T23:19:46Z | |
dc.date.issued | 2019 | |
dc.description.abstract | Smile intensity estimation plays important roles in applications such as affective disorder prediction, life satisfaction prediction, camera technique improvement, etc. In recent studies, many researchers applied only traditional features, such as local binary pattern and local phase quantization (LPQ) to represent smile intensity. To improve the performance of spontaneous smile intensity estimation, we introduce a feature set that combines the saliency map (SM)-based handcrafted feature and non-low-level convolutional neural network (CNN) features. We took advantage of the opponent-color characteristic of SMs and the multiple convolutional level features, which were assumed to be mutually complementary. Experiments were made on the Binghamton-Pittsburgh 4D (BP4D) database and Denver Intensity of Spontaneous Facial Action (DISFA) database. We set the local binary patterns on three orthogonal planes (LBPTOP) method as a baseline, and the experimental results show that the CNN features can better estimate smile intensity. Finally, through the proposed SM-LBPTOP feature fusion with the median- and high-level CNN features, we obtained the best result (52.08% on BP4D, 70.55% on DISFA), demonstrating our hypothesis is reasonable: the SM-based handcrafted feature is a good supplement to CNNs in spontaneous smile intensity estimation. (C) 2019 SPIE and IS&T | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.issue | 2 | |
dc.description.openaccess | NO | |
dc.description.publisherscope | International | |
dc.description.sponsorship | Beijing Natural Science Foundation [4182031] This work was supported by the Beijing Natural Science Foundation (Grant No. 4182031) on students affect recognition research based on deep spatial filter network and multitask learning. | |
dc.description.volume | 28 | |
dc.identifier.doi | 10.1117/1.JEI.28.2.023031 | |
dc.identifier.eissn | 1560-229X | |
dc.identifier.issn | 1017-9909 | |
dc.identifier.quartile | Q4 | |
dc.identifier.scopus | 2-s2.0-85065475503 | |
dc.identifier.uri | http://dx.doi.org/10.1117/1.JEI.28.2.023031 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/10600 | |
dc.identifier.wos | 473731200045 | |
dc.keywords | Smile intensity | |
dc.keywords | Saliency maps | |
dc.keywords | Convolutional neural network | |
dc.language | English | |
dc.publisher | Spie-Soc Photo-Optical Instrumentation Engineers | |
dc.source | Journal of Electronic Imaging | |
dc.subject | Engineering | |
dc.subject | Electrical electronic engineering | |
dc.subject | Optics | |
dc.subject | Imaging science | |
dc.subject | Photographic technology | |
dc.title | Spontaneous smile intensity estimation by fusing saliency maps and convolutional neural networks | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.authorid | N/A | |
local.contributor.kuauthor | Bozkurt, Elif |