Literature DB >> 24475068

CASME II: an improved spontaneous micro-expression database and the baseline evaluation.

Wen-Jing Yan1, Xiaobai Li2, Su-Jing Wang3, Guoying Zhao2, Yong-Jin Liu4, Yu-Hsin Chen1, Xiaolan Fu3.   

Abstract

A robust automatic micro-expression recognition system would have broad applications in national safety, police interrogation, and clinical diagnosis. Developing such a system requires high quality databases with sufficient training samples which are currently not available. We reviewed the previously developed micro-expression databases and built an improved one (CASME II), with higher temporal resolution (200 fps) and spatial resolution (about 280×340 pixels on facial area). We elicited participants' facial expressions in a well-controlled laboratory environment and proper illumination (such as removing light flickering). Among nearly 3000 facial movements, 247 micro-expressions were selected for the database with action units (AUs) and emotions labeled. For baseline evaluation, LBP-TOP and SVM were employed respectively for feature extraction and classifier with the leave-one-subject-out cross-validation method. The best performance is 63.41% for 5-class classification.

Entities:  

Mesh:

Year:  2014        PMID: 24475068      PMCID: PMC3903513          DOI: 10.1371/journal.pone.0086041

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

A micro-expression is a brief facial movement which reveals a genuine emotion that a person tries to conceal [1]–[3]. In addition, micro-expressions might be unaware and/or uncontrollable to the actor, thus may provide effective clues for detecting lies. Therefore, micro-expression recognition has many potential applications such as clinical diagnosis and interrogation. In the clinical field, micro-expressions may be used for understanding genuine emotions of the patients and promoting better therapies. For instance, Ekman [4] analyzed an interview video of a patient stricken with depression and that patient displayed a desperate micro-expression which could predict a commit suicide. During an interrogation or interview, micro-expression discloses what the interviewee really feels and thus helps further investigation. For instance, a micro-expression of scorn from Kato Kaolin's testimony in the O.J. Simpson trial betrayed his genuine feelings [5]. A micro-expression is featured by its short duration. The generally accepted upper limit of the duration is 1/2 s [3], [6]. These fleeting facial expressions, as observed, are usually low in intensity – it might be so brief for the facial muscles to become fully-stretched with suppression. Because of the short duration and low intensity, it is usually imperceptible or neglected by the naked eyes [1]. To better analyze micro-expressions and to help reveal people's feelings, an automatic micro-expression recognition system is in great need. Automatic facial expression recognition is booming. Researchers had developed many algorithms and the accuracy has reached over 90% for the posed six basic facial expressions (anger, disgust, fear, happiness, sadness and surprise). On the contrary, studies on automatic micro-expression recognition have just started recently and only several pieces of work were available. Shreve et al. [7] divided the face into sub-regions (mouth, cheeks, forehead, and eyes) and calculated the facial strain of each sub-region. The strain pattern in each sub-region is then analyzed to detect micro-expression from the video clips. Pfister et al. [8] proposed a framework using temporal interpolation model and multiple kernel learning to recognize micro-expressions. Polikovsky et al. [9] used a 3D-gradient descriptor for micro-expression recognition. Wu et al. [10] developed an automatic micro-expression recognition system by employing Gabor features and using GentleSVM as the classifier. Wang et al. [11] treated a gray-scale micro-expression video clip as a 3rd-order tensor and utilized Discriminant Tensor Subspace Analysis (DTSA) and Extreme Learning Machine (ELM) to recognize micro-expressions. Ruiz-Hernandez and Pietikäinen [12] proposed to encode the Local Binary Patterns (LBP) using a re-parametrization of the second local order Gaussian jet to generate more robust and reliable histograms for micro-expression representation. The success in conventional facial expression recognition largely rely on sufficient facial expression databases, such as the popularly used CK+[13], MUG [14], MMI [15], JAFFE [16], Multi-PIE [17], and also several 3-D facial expression databases [18], [19]. In contrast, there are very few well-developed micro-expression databases, which hindered the development of micro-expression recognition research. Eliciting micro-expression is difficult because it only presents at certain situations when an individual tries to suppress felt emotions but fails. According to Ekman, when a person tries to conceal his or her feelings, the true emotions would leak quickly and may be manifested as micro-expressions (Ekman & Friesen, 1969). In the early years, researchers constructed high-stakes situations to elicit micro-expressions, such as by asking people to lie about what they saw in video episodes [20], and by constructing crime scenarios (mock theft) and opinion scenarios (lie about own opinions) [21]. Micro-expressions elicited in these studies were commonly confounded by other facial movements (such as conversational facial movements), which are irrelevant to emotions and may “contaminate” the performance of the currently-not-robust micro-expression analysis algorithms. Some researchers developed small databases of micro-expressions by asking participants to pose facial expressions quickly [7], [9]. However, these posed “micro-expressions” are different from the spontaneous ones [2], [3]. Therefore, spontaneous micro-expression databases are necessary for academic research and practical applications. It was found that watching emotional video episodes while neutralizing faces is an effective method to elicit spontaneous micro-expressions without many irrelevant facial movements [3], [22]. Until now, there are only four micro-expression databases and only two of them contain spontaneous micro-expressions, as briefly summarized in Table 1.
Table 1

The current micro-expression databases.

USF-HD [7] Polikovsky's database [9] SMIC database [22] CASME database [23]
subjects /1016 valid* 19 valid
samples 10042164195
Frames per sec. 3020010060
Posed/spontaneous PosedPosedspontaneousspontaneous
AU labels NoNonoyes
Emotion class 6637
remarks Their criteria set for micro- expressions (2/3 s) is longer than most accepted durations/Emotions are classified as positive, negative and surpriseTense and repression in addition to the basic emotions

Since not all of the participants showed micro-expressions, the recruited were more than the valid.

Since not all of the participants showed micro-expressions, the recruited were more than the valid. The first reason for developing a new database is that current spontaneous micro-expression databases are too small to provide sufficient samples. A large number of samples is needed for robust automatic micro-expression recognition research and related applications. We aim to provide more samples for training and test. The second reason is that the video quality of current micro-expression databases cannot satisfy the need for micro-expression analysis. Micro-expression is brief in duration and low in intensity, which makes it difficult to detect and recognize. Higher spatial and temporal resolution may help to train the algorithms (which remains unclear). Compared with the 100 fps in SMIC and 60 fps in CASME, the sampling rate of CASME II is 200 fps, aiming to provide more detailed information on the facial muscle movements. In addition, compared with the face size of about 190 pixels×230 pixels in SMIC and about 150 pixels×190 pixels in CASME, the samples in CASME II have a larger face size about 280 pixels×340 pixels. Third, we tuned the instruction in elicitation to increase the variety of the micro-expressions. Ekman (1969) mentioned two types of micro displays which may leak individual's genuine feelings: “The time-reduced full affect micro displays (i.e. micro-expressions) may well be those which the ego is not aware of, while the squelched micro displays may be those which the ego senses and interrupts in mid performance.” The elicitation paradigm used by Li et al. [22] and Yan et al. [23] only focused on the former, i.e. asking participants to fully suppress facial movements. The other type of micro-expression elicitation paradigm, i.e. asking participants to suppress facial movements when they are self-aware, is not used for collecting micro-expressions. It is possible that these two types of micro-expressions have different dynamic features. We aim to give participants different instructions to elicit both types of micro-expressions in order to increase the variety of our data. This paper consists of two parts. In the first part, we introduce the procedure of developing CASME II, including methods of eliciting micro-expressions and how we select samples for the database. In the second part, we evaluate this database with Local Binary Patterns from Three Orthogonal Planes (LBP-TOP) for feature extraction and Support Vector Machine (SVM) for classification.

Methods: Elicitation and analysis

The method of micro-expression elicitation

Ethics statement

The experimental procedure was approved by the IRB of the Institute of Psychology, Chinese Academy of Sciences. The participants signed informed consent and had the right to quit the experiment at any time. The subject of the photograph has given written informed consent, as outlined in the PLOS consent form, to publication of their photograph.

Participants

Thirty-five participants were recruited, with a mean age of 22.03 years (Standard Deviation (SD) = 1.60) in the study. They all signed informed consent and had the right to quit the experiment at any time.

Apparatus

Four strictly selected LED lamps (to avoid possible flickering light caused by alternating current, 50 Hz) under umbrella reflectors (to focus the light on the participant's face) are placed to provide steady and high-intensity illumination. We designed such a setup because flickering light usually appear in high speed recordings (such as those in SMIC) and the video recorded usually turn out to be dark and noisy. We used Point Grey GRAS-03K2C camera to capture participants' faces, with the resolution at 640×480 pixels. “Raw 8” mode is used to reach 200 fps. We saved recordings as MJPEG format without any inter-frame compression.

Materials

Videos episodes with high emotional valence proved to be effective materials for eliciting micro-expressions [3], [22], [23]. Video episodes have a relatively high degree of ecological validity (Gross & Robert, 1995) and are usually better than pictures in term of emotional valence. Furthermore, video episodes are lasting and dynamic emotional stimuli, making inhibition more difficult (Yan et al., 2013). We used video clips similar to those for eliciting micro-expressions in CASME, but several invalid episodes were removed and several new ones were added. 20 participants rated the video episodes by choosing one or two emotion key words from a list and rating the intensity on a 7-point Likert scale (0 as the lowest and 6 as the highest). If a certain basic expression (e.g. happiness) was chosen by one third of the participants or more for a video episode, that emotion would be assumed as the principal emotion(s) of the video episode. There are individual differences of the subjective feelings at the same video episodes. Meanwhile, one episode may elicit several emotions, which is clearly exemplified by episode #12 and #17. Details of selected video episodes and their corresponding rating scores are listed in Table 2.
Table 2

Video episodes used for eliciting micro-expressions and participants' rating scores.

Episode NO.ContentDurationMain elicited emotion(s)Rate of selectionMean score (intensity)
1Ad of 7 Up2′52″happiness0.693.27
2Sports fault1′18″happiness0.713.60
3Jokes on names51′happiness0.703.14
4Larva (animation)1′32″happiness0.644.43
5Larva 2 (animation)1′32″happiness0.943.93
6Jokes on stool1′28″disgust0.814.15
7Tooth extraction1′7″disgust0.694.18
8Eating worm1′35″disgust0.784.00
9A brazen man1′34″disgust0.813.23
10Operation for Near-sighted eye1′56″fear0.632.90
11Ring(Horror film)2′4″fear0.672.83
12Meat grinder2′25″disgust (fear)0.60(0.33)3.78(3.6)
13Final destination(movie)1′43″fear0.53.72
14Roots and Branches (Movie)2′26″sadness1.004.27
15Derek Redmond in 1992 Olympics game1′31″sadness0.714.08
16A girl killed by cars1′57″sadness1.005.00
17Train accident and officer's dereliction1′57″anger (sadness)0.69 (0.61)4.33(4.88)
18Torturing dog1′37″anger0.754.67
19Beating a pregnant women2′25″anger0.944.93

Procedure

To elicit such facial expressions, we induced participants to experience a high arousal and promoted a motivation to disguise. Similar procedures were conducted as for building SMIC and CASME (see Figure 1). The difference is on the instructions given to the participants: 18 participants were asked to keep neutralized faces when watching video clips while 17 participants only tried to suppress the facial movements when they realized there was a facial expression. This design was for eliciting two types of micro-expression [1] as explained above in the Introduction. The participants were asked to watch the video clips in front of a screen and avoid any body movement. The experimenter monitored participants' face on-line from the other monitor. This setup helped the experimenter to pre-define some habitual movements which will be verified by the participant after each video play finished.
Figure 1

Acquisition setup for elicitation and recording of micro-expressions.

Sample selection and category labeling

Two coders were involved in the analysis of the micro-expressions. We processed the raw video recordings in following steps to select micro-expressions:

Step 1

Remove the irrelevant facial movements. There is a large quantity of facial movements in the recordings but many are obviously irrelevant, such as head movements, pressing the lips when swallowing saliva or other habitual movements (such as blowing the nose). All the unemotional movements were ruled out.

Step 2

Pre-select micro-expression candidates by restricting the to-be-analyzed samples. The coders browsed and spotted the onset and offset frames of each emotional movement clip and only those whose duration last less than one second were saved as micro-expression candidates for a further coding. Expressions that are too subtle to be precisely coded are also excluded.

Step 3

Candidate micro-expression clips were converted into frame sequences to facilitate precise labeling of onset and offset frames. By employing a frame-by-frame approach, the precise frame numbers of onset and offset of each sequence were decided, and only those met the selection criteria (total duration less than 500 ms or onset duration less than 250 ms) were chosen as the final micro-expression samples. After the micro-expressions were selected, AU(s) were marked and emotion labels were given for each micro-expression by two coders. The reliability is 0.846, which is calculated by:where #AU(C is the number of AUs on which Coder 1 and Coder 2 agreed and #All_AU is the total number of AUs in a micro-expression scored by the two coders. Afterwards, the two coders discussed and arbitrated the disagreements. FACS [10] is an objective method for labeling facial movements in terms of component actions. Different groups may have different micro-expression classes but could have the same AU coding system. Unlike posed facial expressions, for which people were asked to generate several preset facial movements, the spontaneous micro-expressions in this database were elicited by strong emotional stimuli. Therefore, it is inappropriate to forcefully classify these micro-expressions into the same six categories as for ordinary facial expressions. For example, AU4 (frown) may indicate disgust, anger, attention or tense [3]. In CASME II database, the micro-expressions are labeled based on the AUs, participants' self-report, and the contents of the video episodes. We provided five main categories (as listed in Table 3).
Table 3

Criteria for labeling the emotions and the frequency in the database*.

EmotionCriteriaN
Happiness either AU6 or AU1233
Disgust one of AU9, AU10 or AU4+AU760
Surprise AU1+2, AU25 or AU225
Repression AU15 or AU17 alone or in combination27
Others Other emotion-related facial movements102

The emotion labeling are just partly based on the AUs because micro-expressions are usually partial and in low intensity.

Therefore, we also take account of participants' self-report and the content of the video episodes.

The emotion labeling are just partly based on the AUs because micro-expressions are usually partial and in low intensity. Therefore, we also take account of participants' self-report and the content of the video episodes.

Profile of the database

The developed database called CASME II contains 247 micro-expression samples from 26 participants. They are selected from nearly 3,000 elicited facial movements. These samples are coded with the onset and offset frames, with action units (AUs) marked and emotions labeled and five main categories are provided, as listed in Table 3. The criteria employed to label emotion categories are also shown in Table 3, and one example of micro-expression clip is illustrated in Figure 2. Very subtle micro-expressions were removed because they are almost unable to code the onset and the offset (difficult to detect the turning points due to the almost imperceptible changes). The CASME II database has the following characteristics:
Figure 2

A demonstration of the frame sequence in a micro-expression.

The apex frame presents at about 110-expression is 4+9 (with AU 17 kept almost unchanged), which indicates disgust. The three rectangles above the images show the right inner brow (AU 4) in “zoom in” mode. The movement is more obvious in video play than picture sequence.

A demonstration of the frame sequence in a micro-expression.

The apex frame presents at about 110-expression is 4+9 (with AU 17 kept almost unchanged), which indicates disgust. The three rectangles above the images show the right inner brow (AU 4) in “zoom in” mode. The movement is more obvious in video play than picture sequence. The samples are spontaneous and dynamic micro-expressions. Baseline (usually neutral) frames are kept before and after each micro-expression, making it possible to evaluate different detection algorithms. The recordings have high temporal resolution (200 fps) and relatively higher face resolution (280×340 pixels). Micro-expression labeling is based on FACS investigator's guide [24] and Yan et al.'s findings [3]. The labeling criteria are different from the traditional 6 categories on ordinary facial expressions. The recordings have proper illumination without flickering light and with reduced highlight regions on the face. Some types of facial expressions are difficult to elicit in laboratory situations, thus the samples in different categories distributed unequally, e.g., there are 60 disgust samples but only 7 sadness samples. In CASME II, we provide 5 classes of micro-expressions.

Database evaluation

Preprocessing

Before running any formal experiment on this dataset, three steps of preprocessing were carried out on the raw sample clips that have been labeled. Let be the set of micro-expression clips. The ith sample , is the frame number of sequence . First, a frontal face image M with neutral expression was selected as the model face. 68 facial landmark points of the model face were detected using the Active Shape Model [25]. Second, the first frame of each micro-expression clip was normalized to the model face using a Local Weighted Mean (LWM) [26] transformation, and the transform matrix T is written as:Where is the coordinates of 68 landmark points of the first frame of the micro-expression sample . Then all frames of were normalized using the same matrix . There are two reasons why we detected ASM landmark points only on the first frame but not on all frames. The first reason is that, since the time duration of a micro-expression is very short the rigid head movement within the duration can be neglected. The second reason is that the landmark points detected by ASM may be not accurate enough; if applied on a sequence of frames, there could be big deviation of locations of the same point even when the face did not move at all. The normalized image was computed as a 2D transformation of the original image: is the jth frame of the normalized micro-expression sequence . Third, the eye coordinates of the first frame of each normalized micro-expression sequence were located and then the face of each frame of was cropped out using a rectangle determined by eye positions .

Experiments

We carried out a micro-expression recognition experiment on the samples preprocessed by the steps described in the previous section. LBP-TOP features were extracted to describe micro-expressions from a spatiotemporal point of view, and SVM was employed as the classifier to provide baseline performance for the future evaluation of our database (Figure 3). The details are described below.
Figure 3

Block diagram of the micro-expression classification.

Methods

Local Binary Patterns from Three Orthogonal Planes (LBP-TOP)

The conventional LBP method has been demonstrated to be effective for describing 2D textures of static images [27]. Considering one pixel C in the image, LBP operator describes its local texture pattern by comparing and thresholding the gray values of its neighboring pixels. As illustrated in Figure 4(a), for the center pixel C with its P neighboring pixels with the radius R, the LBP value is calculated as:Where are the coordinates of the center pixel C and is its gray value, and is the gray value of its pth neighboring pixel on the radius R. is the weight corresponding to the neighboring pixels locations, which is used to transform a binary pattern string to a decimal index for this pattern. The function is a sign function defined as:
Figure 4

The LBP operator and its histogram.

Given an input image of size M by N, the LBP value b of each pixel could be calculated using the above two equations. Then we can form a feature vector to represent the input image by calculating the histogram distribution of all of its LBPs:Where n denotes the total number of interested patterns, and: One example of LBP histogram is illustrated in Figure 4(b). More details about LBP please refer to [27]. As an extension of the basic LBP idea, Local Binary Pattern on Three Orthogonal Planes (LBP-TOP) was proposed by Zhao et al [28] for dynamic texture analysis in the spatial-temporal domain. Given a video sequence of time length T, usually it can be thought as a stack of XY planes along the time axis T, but it can also be thought as a stack of XT planes on axis Y, or a stack of YT planes on axis X. The XT and YT planes provide information about the space-time transitions. Taking micro-expression as an example, the XT and YT planes contain information of how the gray values of a row or a column of pixels change along the time dimension; if the input data volume is of a mouth corner, they describe the mouth movement as dynamic texture in the XT or YT plane accordingly (see Figure 5(a)).
Figure 5

The texture of three planes and the corresponding histograms. (a) XY, XT and YT planes of a micro-expression sample, (b) concatenated LBP-TOP feature.

The basic idea of LBP-TOP is the similar as LBP, which is to use an index code to represent the local pattern of each pixel by comparing (and thresholding) its gray value to its neighbors' gray values. To include information from 3D spatial-temporal domain, LBP code is extracted from the XY, XT and YT planes of each pixel, respectively. Then one histogram of each plane is formed by using equation (3) and (4) and then concatenated into a single histogram as the final LBP-TOP feature vector. Note that we can set different values for the radius of X, Y and T axis as , and , so finally we can have elliptical sampling than circular sampling. The procedure of extracting LBP-TOP feature histogram is demonstrated in Figure 5(b).

Results

The radii in axes X and Y vary from 1 to 4. The radius of T varies from 2 to 4 (we don't consider T = 1 because the samples are in 200 fps where the change of two neighboring frames is little). The number P of the neighboring points in the XY, XT and YT planes was set as 4. SVM was used as the classifier. For this baseline evaluation, we selected 5 classes of the micro-expressions–happiness, surprise, disgust, repression and others–for training and test. Considering samples unequally distributed in the five selected classes, leave-one-subject-out cross validation was applied. The LBP-TOP features were extracted in 5×5 blocks. The performance is shown in Table 4. The best performance is 63.41% where the radii are 1, 1 and 4 for XY, YT and XT planes respectively.
Table 4

Performance on recognizing 5-class micro-expressions with LBP-TOP features extraction and leave-one-out cross-validation.

RX RY RT 5×5 blocks (%)
11263.01
11362.60
1 1 4 63.41
22261.38
22361.79
22462.20
33258.54
33361.79
33458.55
44258.94
44361.38
44460.57

Discussions and Limitations

There are still some limitations regarding the elicited micro-expressions and this database: First, the labeling for these spontaneous micro-expressions is not very satisfactory. The materials for eliciting micro-expressions are video episodes which are complex stimuli and can have different meanings to different people. For example, the scene of chewing worms is not always disgusting—some participants actually reported “funny” or “interesting”. Categorizing the micro-expressions simply based on AUs (especially for AU combinations that haven't been defined) requires further in-depth studies. Participants are required to control and suppress their facial expressions, which often makes spontaneous micro-expressions appear partial and sometimes with only a single AU (also see Porter, 2008; Yan et al., 2013). Even though we considered the AU combinations, the content of the video episodes and the participants' report, the emotion labels for these micro-expressions remain debatable. Second, the micro-expressions in this database are elicited under one specific lab situation and may not cover the micro-expressions elicited under other situations. Micro-expressions that occur when people are under an police interrogation or lying to their boss, may be different from those we observe in a laboratory situation. Up to now, very few studies have investigated the micro-expressions elicited in different situations. Future studies should be conducted to explore the variations of micro-expressions in different situations. Third, it's unclear which method is most suitable for classifying micro-expressions. The micro-expressions have the following characteristics which are different from the ordinary facial expressions in the following ways: (1) low-intensity facial movements and (2) partial facial expressions (fragments). Therefore, previous methods that were suitable for classifying ordinary facial expressions may not work well for micro-expressions. We provided this database and a baseline evaluation, so future research may compare with our initial results. Fourth, the participants came from a limited range of age. Those who participated in the elicitation section are young people. Most of them are university students. It would be better if the database contain samples from models at various ages.

Conclusions

This paper reviews the previously developed micro-expression databases and presents a new and improved one. This new spontaneous micro-expression database, CASME II, includes 247 micro-expression samples which is larger than previous micro-expression databases. The micro-expressions were elicited in a well-controlled laboratory environment and they were recorded by a high-speed camera (200 fps) so the temporal resolution is much higher than previous spontaneous micro-expression databases. For the spatial resolution, the size of the faces is larger than previous spontaneous micro-expression samples; as larger scale of images helps to reveal tiny changes, which would benefit feature extraction and futher better classification. The recordings have proper illumination without flicking light and with reduced highlight regions on the face. These samples were coded with onset and offset frames as well as AUs and emotion labels. We also reported the baseline results of evaluation using a combination of LBP-TOP and SVM. Micro-expression analysis is a new research area. Starting from a controlled laboratory environment can make the database well focused on the subtle expressions and rule out irrelevant factors. We provide the CASME II database as a benchmark for the future exploration of potential micro-expression analysis algorithms that are suitable for detecting subtle facial muscle movements for emotional states recognition. In the future, we will extend our research to a more natural environment for expression and micro-expression analysis in natural conversation and interaction. The CASME II database is online public available now for testing (see http://fu.psych.ac.cn/CASME/casme2-en.php for details).
  5 in total

1.  Dynamic texture recognition using local binary patterns with an application to facial expressions.

Authors:  Guoying Zhao; Matti Pietikäinen
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2007-06       Impact factor: 6.226

2.  The ability to detect deceit generalizes across different types of high-stake lies.

Authors:  M G Frank; P Ekman
Journal:  J Pers Soc Psychol       Date:  1997-06

3.  Nonverbal leakage and clues to deception.

Authors:  P Ekman; W V Friesen
Journal:  Psychiatry       Date:  1969-02       Impact factor: 2.458

4.  Reading between the lies: identifying concealed and falsified emotions in universal facial expressions.

Authors:  Stephen Porter; Leanne ten Brinke
Journal:  Psychol Sci       Date:  2008-05

5.  Multi-PIE.

Authors:  Ralph Gross; Iain Matthews; Jeff Cohn; Takeo Kanade; Simon Baker
Journal:  Proc Int Conf Autom Face Gesture Recognit       Date:  2010-05-01
  5 in total
  24 in total

1.  Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning.

Authors:  Yanju Liu; Yange Li; Xinhai Yi; Zuojin Hu; Huiyu Zhang; Yanzhong Liu
Journal:  Front Neurorobot       Date:  2022-06-30       Impact factor: 3.493

2.  Micro-Expression-Based Emotion Recognition Using Waterfall Atrous Spatial Pyramid Pooling Networks.

Authors:  Marzuraikah Mohd Stofa; Mohd Asyraf Zulkifley; Muhammad Ammirrul Atiqi Mohd Zainuri
Journal:  Sensors (Basel)       Date:  2022-06-19       Impact factor: 3.847

3.  Using Facial Micro-Expressions in Combination With EEG and Physiological Signals for Emotion Recognition.

Authors:  Nastaran Saffaryazdi; Syed Talal Wasim; Kuldeep Dileep; Alireza Farrokhi Nia; Suranga Nanayakkara; Elizabeth Broadbent; Mark Billinghurst
Journal:  Front Psychol       Date:  2022-06-28

4.  Micro-expression recognition model based on TV-L1 optical flow method and improved ShuffleNet.

Authors:  Yanju Liu; Yange Li; Xinhan Yi; Zuojin Hu; Huiyu Zhang; Yanzhong Liu
Journal:  Sci Rep       Date:  2022-10-20       Impact factor: 4.996

5.  Efficient spatio-temporal local binary patterns for spontaneous facial micro-expression recognition.

Authors:  Yandan Wang; John See; Raphael C-W Phan; Yee-Hui Oh
Journal:  PLoS One       Date:  2015-05-19       Impact factor: 3.240

6.  Dual Temporal Scale Convolutional Neural Network for Micro-Expression Recognition.

Authors:  Min Peng; Chongyang Wang; Tong Chen; Guangyuan Liu; Xiaolan Fu
Journal:  Front Psychol       Date:  2017-10-13

7.  Emotion recognition based on customized smart bracelet with built-in accelerometer.

Authors:  Zhan Zhang; Yufei Song; Liqing Cui; Xiaoqian Liu; Tingshao Zhu
Journal:  PeerJ       Date:  2016-07-26       Impact factor: 2.984

8.  Emotion recognition using Kinect motion capture data of human gaits.

Authors:  Shun Li; Liqing Cui; Changye Zhu; Baobin Li; Nan Zhao; Tingshao Zhu
Journal:  PeerJ       Date:  2016-09-15       Impact factor: 2.984

9.  High-Speed Video System for Micro-Expression Detection and Recognition.

Authors:  Diana Borza; Radu Danescu; Razvan Itu; Adrian Darabant
Journal:  Sensors (Basel)       Date:  2017-12-14       Impact factor: 3.576

10.  A Brief Review of Facial Emotion Recognition Based on Visual Information.

Authors:  Byoung Chul Ko
Journal:  Sensors (Basel)       Date:  2018-01-30       Impact factor: 3.576

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.