Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos
Facial Micro-Expressions (MEs) reveal a person's hidden emotions in high stake situations within a fraction of a second and at a low intensity. The broad range of potential real-world applications that can be applied has drawn considerable attention from researchers in recent years. However, bo...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Published: |
Elsevier
2023
|
Subjects: | |
Online Access: | http://eprints.um.edu.my/39363/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my.um.eprints.39363 |
---|---|
record_format |
eprints |
spelling |
my.um.eprints.393632023-11-28T08:17:43Z http://eprints.um.edu.my/39363/ Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos Liong, Gen-Bing See, John Chan, Chee Seng QA75 Electronic computers. Computer science Facial Micro-Expressions (MEs) reveal a person's hidden emotions in high stake situations within a fraction of a second and at a low intensity. The broad range of potential real-world applications that can be applied has drawn considerable attention from researchers in recent years. However, both spotting and recognition tasks are often treated separately. In this paper, we present Micro-Expression Analysis Network (MEAN), a shallow multi-stream multi-output network architecture comprising of task-specific (spotting and recognition) networks that is designed to effectively learn a meaningful representation from both ME class labels and location-wise pseudo-labels. Notably, this is the first known work that addresses ME analysis on long videos using a deep learning approach, whereby ME spotting and recognition are performed sequentially in a two-step procedure: first spotting the ME intervals using the spotting network, and proceeding to predict their emotion classes using the recognition network. We report extensive benchmark results on the ME analysis task on both short video datasets (CASME II, SMIC-E-HS, SMIC-E-VIS, and SMIC-E-NIR), and long video datasets (CAS(ME)2 and SAMMLV); the latter in particular demonstrates the capability of the proposed approach under unconstrained settings. Besides the standard measures, we promote the usage of fairer metrics in evaluating the performance of a complete ME analysis system. We also provide visual explanations of where the network is ``looking''and showcasing the effectiveness of inductive transfer applied during network training. An analysis is performed on the in-the-wild dataset (MEVIEW) to open up further research into real-world scenarios. Elsevier 2023-01 Article PeerReviewed Liong, Gen-Bing and See, John and Chan, Chee Seng (2023) Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos. Signal Processing-Image Communication, 110. ISSN 0923-5965, DOI https://doi.org/10.1016/j.image.2022.116875 <https://doi.org/10.1016/j.image.2022.116875>. 10.1016/j.image.2022.116875 |
institution |
Universiti Malaya |
building |
UM Library |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Malaya |
content_source |
UM Research Repository |
url_provider |
http://eprints.um.edu.my/ |
topic |
QA75 Electronic computers. Computer science |
spellingShingle |
QA75 Electronic computers. Computer science Liong, Gen-Bing See, John Chan, Chee Seng Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
description |
Facial Micro-Expressions (MEs) reveal a person's hidden emotions in high stake situations within a fraction of a second and at a low intensity. The broad range of potential real-world applications that can be applied has drawn considerable attention from researchers in recent years. However, both spotting and recognition tasks are often treated separately. In this paper, we present Micro-Expression Analysis Network (MEAN), a shallow multi-stream multi-output network architecture comprising of task-specific (spotting and recognition) networks that is designed to effectively learn a meaningful representation from both ME class labels and location-wise pseudo-labels. Notably, this is the first known work that addresses ME analysis on long videos using a deep learning approach, whereby ME spotting and recognition are performed sequentially in a two-step procedure: first spotting the ME intervals using the spotting network, and proceeding to predict their emotion classes using the recognition network. We report extensive benchmark results on the ME analysis task on both short video datasets (CASME II, SMIC-E-HS, SMIC-E-VIS, and SMIC-E-NIR), and long video datasets (CAS(ME)2 and SAMMLV); the latter in particular demonstrates the capability of the proposed approach under unconstrained settings. Besides the standard measures, we promote the usage of fairer metrics in evaluating the performance of a complete ME analysis system. We also provide visual explanations of where the network is ``looking''and showcasing the effectiveness of inductive transfer applied during network training. An analysis is performed on the in-the-wild dataset (MEVIEW) to open up further research into real-world scenarios. |
format |
Article |
author |
Liong, Gen-Bing See, John Chan, Chee Seng |
author_facet |
Liong, Gen-Bing See, John Chan, Chee Seng |
author_sort |
Liong, Gen-Bing |
title |
Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
title_short |
Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
title_full |
Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
title_fullStr |
Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
title_full_unstemmed |
Spot-then-recognize: A micro-expression analysis network for seamless evaluation of long videos |
title_sort |
spot-then-recognize: a micro-expression analysis network for seamless evaluation of long videos |
publisher |
Elsevier |
publishDate |
2023 |
url |
http://eprints.um.edu.my/39363/ |
_version_ |
1783876687002337280 |
score |
13.211869 |